

Pyspark Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a PySpark Data Engineer with a 6-month contract, paying up to £450/day, remote with occasional London travel. Key skills include PySpark, Python, Databricks, Delta Lake, and active SC clearance is essential.
🌎 - Country
United Kingdom
💱 - Currency
£ GBP
-
💰 - Day rate
450
-
🗓️ - Date discovered
June 21, 2025
🕒 - Project duration
Unknown
-
🏝️ - Location type
Remote
-
📄 - Contract type
Inside IR35
-
🔒 - Security clearance
Yes
-
📍 - Location detailed
England, United Kingdom
-
🧠 - Skills detailed
#PySpark #Databricks #Spark (Apache Spark) #Python #Data Processing #YAML (YAML Ain't Markup Language) #Azure DevOps #Azure #Data Lake #Delta Lake #DevOps #Scala #Data Engineering #Data Pipeline #Agile #GIT
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
PySpark Data Engineer | up to £450/day Inside | Remote with occasional London travel
We are seeking a PySpark Data Engineer to support the development of a modern, scalable data lake for a new strategic programme. This is a greenfield initiative to replace fragmented legacy reporting solutions, offering the opportunity to shape a long-term, high-impact platform from the ground up.
Key Responsibilities:
• Design, build, and maintain scalable data pipelines using PySpark 3/4 and Python 3.
• Contribute to the creation of a unified data lake following medallion architecture principles.
• Leverage Databricks and Delta Lake (Parquet format) for efficient, reliable data processing.
• Apply BDD testing practices using Python Behave and ensure code quality with Python Coverage.
• Collaborate with cross-functional teams and participate in Agile delivery workflows.
• Manage configurations and workflows using YAML, Git, and Azure DevOps.
Required Skills & Experience:
• Proven expertise in PySpark 3/4 and Python 3 for large-scale data engineering.
• Hands-on experience with Databricks, Delta Lake, and medallion architecture.
• Familiarity with Python Behave for Behaviour Driven Development.
• Strong understanding of YAML, code quality tools (e.g. Python Coverage), and CI/CD pipelines.
• Knowledge of Azure DevOps and Git best practices.
• Active SC clearance is essential - applicants without this cannot be considered.
Contract Details:
• 6-month initial contract with long-term extension potential (multi-year programme).
• Inside IR35.
This is an excellent opportunity to join a high-profile programme at its inception and help build a critical data platform from the ground up. If you are a mission-driven engineer with a passion for scalable data solutions and secure environments, we'd love to hear from you.