

Hays
Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with a contract length of "unknown," offering a pay rate of "unknown," and is remote. Key skills include Databricks, PySpark, Delta Lake, and Azure services. Experience in data governance and performance optimization is required.
🌎 - Country
United Kingdom
💱 - Currency
£ GBP
-
💰 - Day rate
Unknown
-
🗓️ - Date
December 1, 2025
🕒 - Duration
Unknown
-
🏝️ - Location
Unknown
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
England, United Kingdom
-
🧠 - Skills detailed
#Azure #Batch #Data Engineering #ADLS (Azure Data Lake Storage) #Compliance #Data Lake #DevOps #Databricks #Azure cloud #Automation #Data Governance #"ETL (Extract #Transform #Load)" #PySpark #Data Lineage #Cloud #Delta Lake #Deployment #Azure DevOps #Spark (Apache Spark) #Data Pipeline #Storage #Monitoring #ML (Machine Learning) #Data Management #Azure ADLS (Azure Data Lake Storage) #Microsoft Power BI #Vault #Data Processing #Data Quality #Documentation #"ACID (Atomicity #Consistency #Isolation #Durability)" #BI (Business Intelligence) #Data Analysis #Synapse
Role description
We are looking for a Databricks Data Engineer with strong expertise in developing and optimizing large-scale data engineering solutions within the Databricks Data Intelligence Platform. The ideal candidate will have practical experience in workflow orchestration, performance optimization, and data governance, alongside broad proficiency in PySpark, Delta Lake, and Azure services.
Key Responsibilities:
• Design, build, and maintain robust data pipelines using Databricks notebooks, Jobs, and Workflows for batch and streaming data processing.
• Optimize Spark and Delta Lake performance on Databricks clusters through efficient cluster configuration, adaptive query execution, and caching strategies.
• Conduct performance testing and cluster tuning to ensure cost-efficient and high-performing workloads.
• Implement data quality, lineage tracking, and access control policies aligned with Databricks Unity Catalog and data governance best practices.
• Develop PySpark applications for ETL, data transformation, and analytical use cases, adhering to modular and reusable design principles.
• Create and manage Delta Lake tables with a focus on ACID compliance, schema evolution, and time travel for versioned data management.
• Integrate Databricks solutions with Azure services including Azure Data Lake Storage, Key Vault, and Azure Functions.
• Collaborate with cloud architects and data analysts to design end-to-end workflows supporting analytics, machine learning, and reporting use cases.
• Support CI/CD deployment of Databricks assets using Azure DevOps or similar automation frameworks.
• Maintain detailed technical documentation on architecture, performance benchmarks, and governance configurations.
Required Skills and Experience:
• In-depth knowledge of Databricks Data Intelligence Platform and multi-cloud ecosystem integration.
• Experience configuring, scheduling, and monitoring Databricks Jobs and Workflows.
• Strong proficiency in PySpark, including advanced data transformation, schema management, and optimization techniques.
• Solid understanding of Delta Lake architecture, transactional processing, and incremental data pipeline design.
• Proven ability to conduct Spark performance tuning and cluster optimization based on workload profiles.
• Experience implementing fine-grained data governance with Unity Catalog, access policies, and data lineage tracking.
• Hands-on experience with Azure Cloud components such as Data Lake Storage (Gen2), Key Vault, and Azure Functions.
• Familiarity with CI/CD frameworks for Databricks asset deployment and environment automation.
• Strong analytical and troubleshooting skills in distributed data environments.
Preferred Qualifications:
• Experience supporting enterprise-scale Databricks environments with multiple workspaces and governed catalogs.
• Knowledge of Azure Synapse, Power BI, or related analytics services.
• Understanding of cost optimization strategies for data compute on Databricks clusters.
Excellent problem-solving skills, technical communication, and cross-functional collaboration abilitie
We are looking for a Databricks Data Engineer with strong expertise in developing and optimizing large-scale data engineering solutions within the Databricks Data Intelligence Platform. The ideal candidate will have practical experience in workflow orchestration, performance optimization, and data governance, alongside broad proficiency in PySpark, Delta Lake, and Azure services.
Key Responsibilities:
• Design, build, and maintain robust data pipelines using Databricks notebooks, Jobs, and Workflows for batch and streaming data processing.
• Optimize Spark and Delta Lake performance on Databricks clusters through efficient cluster configuration, adaptive query execution, and caching strategies.
• Conduct performance testing and cluster tuning to ensure cost-efficient and high-performing workloads.
• Implement data quality, lineage tracking, and access control policies aligned with Databricks Unity Catalog and data governance best practices.
• Develop PySpark applications for ETL, data transformation, and analytical use cases, adhering to modular and reusable design principles.
• Create and manage Delta Lake tables with a focus on ACID compliance, schema evolution, and time travel for versioned data management.
• Integrate Databricks solutions with Azure services including Azure Data Lake Storage, Key Vault, and Azure Functions.
• Collaborate with cloud architects and data analysts to design end-to-end workflows supporting analytics, machine learning, and reporting use cases.
• Support CI/CD deployment of Databricks assets using Azure DevOps or similar automation frameworks.
• Maintain detailed technical documentation on architecture, performance benchmarks, and governance configurations.
Required Skills and Experience:
• In-depth knowledge of Databricks Data Intelligence Platform and multi-cloud ecosystem integration.
• Experience configuring, scheduling, and monitoring Databricks Jobs and Workflows.
• Strong proficiency in PySpark, including advanced data transformation, schema management, and optimization techniques.
• Solid understanding of Delta Lake architecture, transactional processing, and incremental data pipeline design.
• Proven ability to conduct Spark performance tuning and cluster optimization based on workload profiles.
• Experience implementing fine-grained data governance with Unity Catalog, access policies, and data lineage tracking.
• Hands-on experience with Azure Cloud components such as Data Lake Storage (Gen2), Key Vault, and Azure Functions.
• Familiarity with CI/CD frameworks for Databricks asset deployment and environment automation.
• Strong analytical and troubleshooting skills in distributed data environments.
Preferred Qualifications:
• Experience supporting enterprise-scale Databricks environments with multiple workspaces and governed catalogs.
• Knowledge of Azure Synapse, Power BI, or related analytics services.
• Understanding of cost optimization strategies for data compute on Databricks clusters.
Excellent problem-solving skills, technical communication, and cross-functional collaboration abilitie






