

Haystack
Lead PySpark Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead PySpark Engineer with a contract length of long-term, offering £281 - £292 per day. Key skills include expert PySpark, AWS Data stack experience, and financial services background. Remote work is available.
🌎 - Country
United Kingdom
💱 - Currency
£ GBP
-
💰 - Day rate
292
-
🗓️ - Date
March 27, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Remote
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
London, England, United Kingdom
-
🧠 - Skills detailed
#Data Accuracy #Macros #Distributed Computing #Scala #Migration #Datasets #Jenkins #AWS (Amazon Web Services) #Leadership #Cloud #Lambda (AWS Lambda) #Data Reconciliation #PySpark #Infrastructure as Code (IaC) #Spark (Apache Spark) #S3 (Amazon Simple Storage Service) #Athena #Data Mart #GitLab #SAS #Terraform #Unit Testing #GIT #DevOps #Data Engineering #Python #AWS EMR (Amazon Elastic MapReduce) #"ETL (Extract #Transform #Load)" #Base
Role description
Lead PySpark Engineer | £281 - £292
We're working with a Tier-1 global financial services powerhouse on this exciting opportunity.
Step into a high-impact leadership role driving a massive cloud-native transformation. You will move beyond a simple "lift and shift" to architect and refactor legacy SAS analytics into high-performance, distributed PySpark pipelines on AWS for one of the world's most respected financial institutions.
The Role
• Lead the technical migration of complex SAS Base and Macro logic into modern, modular Python code using PySpark and SAS2PY.
• Design and develop robust ETL/ELT pipelines and scalable Data Marts using AWS EMR, Glue, and S3.
• Tackle performance tuning at scale by optimizing Spark execution, including partitioning, shuffling, and caching strategies for massive datasets.
• Architect modern data models utilizing SCD Type 2, Fact/Dimension tables, and Star Schemas within a production-ready cloud environment.
• Drive quality through technical leadership, implementing rigorous CI/CD workflows, unit testing, and automated data reconciliation frameworks.
What You'll Need
• Expert-level proficiency in PySpark and Python, with a deep understanding of Clean Code and SOLID principles.
• Strong hands-on experience across the AWS Data stack, specifically EMR, Glue, S3, Athena, and Lambda.
• Proven ability to read, debug, and translate legacy SAS (Base, Macros, or DI Studio) into modern distributed computing architectures.
• Experience with DevOps tooling and modern workflows, including Git, Jenkins, GitLab CI, and Infrastructure as Code via Terraform.
• A background in Financial Services or a similar highly regulated environment where "penny-perfect" data accuracy is non-negotiable.
What's On Offer
• Competitive daily rate of £281 - £292 in a high-stakes, Tier-1 banking environment.
• Remote-friendly working model providing excellent work-life balance.
• Long-term contract potential within a massive digital transformation program.
• Opportunity to work with the latest AWS data engineering tools at an enterprise scale.
Apply via Haystack today!
Lead PySpark Engineer | £281 - £292
We're working with a Tier-1 global financial services powerhouse on this exciting opportunity.
Step into a high-impact leadership role driving a massive cloud-native transformation. You will move beyond a simple "lift and shift" to architect and refactor legacy SAS analytics into high-performance, distributed PySpark pipelines on AWS for one of the world's most respected financial institutions.
The Role
• Lead the technical migration of complex SAS Base and Macro logic into modern, modular Python code using PySpark and SAS2PY.
• Design and develop robust ETL/ELT pipelines and scalable Data Marts using AWS EMR, Glue, and S3.
• Tackle performance tuning at scale by optimizing Spark execution, including partitioning, shuffling, and caching strategies for massive datasets.
• Architect modern data models utilizing SCD Type 2, Fact/Dimension tables, and Star Schemas within a production-ready cloud environment.
• Drive quality through technical leadership, implementing rigorous CI/CD workflows, unit testing, and automated data reconciliation frameworks.
What You'll Need
• Expert-level proficiency in PySpark and Python, with a deep understanding of Clean Code and SOLID principles.
• Strong hands-on experience across the AWS Data stack, specifically EMR, Glue, S3, Athena, and Lambda.
• Proven ability to read, debug, and translate legacy SAS (Base, Macros, or DI Studio) into modern distributed computing architectures.
• Experience with DevOps tooling and modern workflows, including Git, Jenkins, GitLab CI, and Infrastructure as Code via Terraform.
• A background in Financial Services or a similar highly regulated environment where "penny-perfect" data accuracy is non-negotiable.
What's On Offer
• Competitive daily rate of £281 - £292 in a high-stakes, Tier-1 banking environment.
• Remote-friendly working model providing excellent work-life balance.
• Long-term contract potential within a massive digital transformation program.
• Opportunity to work with the latest AWS data engineering tools at an enterprise scale.
Apply via Haystack today!



