

Senior Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
π - Country
United Kingdom
π± - Currency
Β£ GBP
-
π° - Day rate
680
-
ποΈ - Date discovered
September 16, 2025
π - Project duration
Unknown
-
ποΈ - Location type
Hybrid
-
π - Contract type
Inside IR35
-
π - Security clearance
Unknown
-
π - Location detailed
London Area, United Kingdom
-
π§ - Skills detailed
#S3 (Amazon Simple Storage Service) #Deployment #Security #Terraform #SQL (Structured Query Language) #API (Application Programming Interface) #MLflow #TensorFlow #Big Data #Python #Spark (Apache Spark) #Data Pipeline #Delta Lake #Cloud #Apache Spark #Data Science #IAM (Identity and Access Management) #Automation #Infrastructure as Code (IaC) #Data Quality #Computer Science #AI (Artificial Intelligence) #ML (Machine Learning) #DevOps #Data Governance #Databricks #Datasets #Kafka (Apache Kafka) #Lambda (AWS Lambda) #Observability #EC2 #Monitoring #Scala #AWS (Amazon Web Services) #Data Engineering #Compliance #Data Warehouse #Code Reviews #PySpark
Role description
Senior Data Engineer - Β£650pd (Inside IR35) - Hybrid/London
We are seeking a highly skilled Senior Data Engineer to help shape and scale a cutting-edge global analytics platform that supports front office trading across multiple asset classes. This role offers the opportunity to work at the intersection of data engineering, cloud technologies, and advanced analytics, building solutions that power decision-making in a fast-moving, data-driven environment.
Key Responsibilities
β’ Design, build, and maintain scalable data pipelines and Delta Lake architectures in Databricks on AWS.
β’ Develop and enhance enterprise-level data warehouses to ensure performance, reliability, and data quality for trading analytics.
β’ Collaborate with data scientists, quants, and ML engineers to prepare ML-ready datasets and deliver production-grade ML/AI pipelines.
β’ Implement CI/CD pipelines, testing frameworks, and observability tools for robust data workflows.
β’ Contribute to MLOps practices, including model tracking, deployment, and monitoring with Databricks and MLflow.
β’ Participate in solution design, code reviews, and cross-functional collaboration.
β’ Ensure compliance with data governance, security, and performance standards.
β’ Stay current with Databricks and cloud data ecosystem developments to drive innovation and best practices.
Skills & Experience
β’ BS/MS in Computer Science, Software Engineering, or related technical field.
β’ 8+ years of hands-on experience designing and implementing large-scale distributed data pipelines.
β’ Deep expertise in Apache Spark, PySpark, and Databricks (Delta Lake, Unity Catalog, MLflow, Databricks Workflows).
β’ Strong proficiency in Python and SQL, with experience building modular, testable, reusable pipeline components.
β’ Proven track record with AWS services (S3, Lambda, Glue, API Gateway, IAM, EC2, EKS) and their integration with Databricks.
β’ Advanced skills in Infrastructure as Code (Terraform) for provisioning and managing data infrastructure.
β’ Experience in MLOps pipelines and modern ML frameworks (e.g., scikit-learn, XGBoost, TensorFlow).
β’ Familiarity with real-time streaming architectures (Kafka, Structured Streaming) is a plus.
β’ Strong DevOps practices in data engineering: versioning, testing, deployment automation, monitoring.
β’ Understanding of data governance, access control, and compliance in regulated industries.
β’ Excellent communication, collaboration, and problem-solving skills with a strong sense of ownership.
β’ Experience in commodity trading markets (e.g., power, gas, crude, freight) is advantageous.
β’ Relevant certifications in Databricks, AWS, or big data technologies are preferred.
Senior Data Engineer - Β£650pd (Inside IR35) - Hybrid/London
We are seeking a highly skilled Senior Data Engineer to help shape and scale a cutting-edge global analytics platform that supports front office trading across multiple asset classes. This role offers the opportunity to work at the intersection of data engineering, cloud technologies, and advanced analytics, building solutions that power decision-making in a fast-moving, data-driven environment.
Key Responsibilities
β’ Design, build, and maintain scalable data pipelines and Delta Lake architectures in Databricks on AWS.
β’ Develop and enhance enterprise-level data warehouses to ensure performance, reliability, and data quality for trading analytics.
β’ Collaborate with data scientists, quants, and ML engineers to prepare ML-ready datasets and deliver production-grade ML/AI pipelines.
β’ Implement CI/CD pipelines, testing frameworks, and observability tools for robust data workflows.
β’ Contribute to MLOps practices, including model tracking, deployment, and monitoring with Databricks and MLflow.
β’ Participate in solution design, code reviews, and cross-functional collaboration.
β’ Ensure compliance with data governance, security, and performance standards.
β’ Stay current with Databricks and cloud data ecosystem developments to drive innovation and best practices.
Skills & Experience
β’ BS/MS in Computer Science, Software Engineering, or related technical field.
β’ 8+ years of hands-on experience designing and implementing large-scale distributed data pipelines.
β’ Deep expertise in Apache Spark, PySpark, and Databricks (Delta Lake, Unity Catalog, MLflow, Databricks Workflows).
β’ Strong proficiency in Python and SQL, with experience building modular, testable, reusable pipeline components.
β’ Proven track record with AWS services (S3, Lambda, Glue, API Gateway, IAM, EC2, EKS) and their integration with Databricks.
β’ Advanced skills in Infrastructure as Code (Terraform) for provisioning and managing data infrastructure.
β’ Experience in MLOps pipelines and modern ML frameworks (e.g., scikit-learn, XGBoost, TensorFlow).
β’ Familiarity with real-time streaming architectures (Kafka, Structured Streaming) is a plus.
β’ Strong DevOps practices in data engineering: versioning, testing, deployment automation, monitoring.
β’ Understanding of data governance, access control, and compliance in regulated industries.
β’ Excellent communication, collaboration, and problem-solving skills with a strong sense of ownership.
β’ Experience in commodity trading markets (e.g., power, gas, crude, freight) is advantageous.
β’ Relevant certifications in Databricks, AWS, or big data technologies are preferred.