

JRD Systems
Lead Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead Data Engineer with 8+ years of experience, focusing on Databricks, PySpark, and SQL. Contract length is unspecified, with a competitive pay rate. Must possess strong data engineering skills and experience in cloud environments.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
January 29, 2026
π - Duration
Unknown
-
ποΈ - Location
Unknown
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Michigan, United States
-
π§ - Skills detailed
#Kafka (Apache Kafka) #Databricks #Observability #ML (Machine Learning) #SQL (Structured Query Language) #Azure #BI (Business Intelligence) #Cloud #Data Engineering #Data Quality #PySpark #Airflow #Data Pipeline #Batch #Data Processing #GCP (Google Cloud Platform) #Delta Lake #Strategy #Data Modeling #Data Science #"ETL (Extract #Transform #Load)" #Datasets #Data Lake #Scala #Data Strategy #Spark (Apache Spark) #AWS (Amazon Web Services) #ADF (Azure Data Factory)
Role description
About the Role
Weβre looking for a Lead Data Engineer to design, build, and scale modern data platforms using Databricks, PySpark, and SQL. In this role, youβll take ownership of complex data pipelines, guide architectural decisions, and mentor a team of data engineers while partnering closely with analytics, product, and business stakeholders.
Key Responsibilities
β’ Lead the design and development of scalable, high-performance data pipelines using Databricks
β’ Build and optimize ETL/ELT workflows using PySpark and SQL
β’ Architect data solutions for batch and (if applicable) streaming workloads
β’ Ensure data quality, reliability, performance, and governance best practices
β’ Collaborate with data scientists, analysts, and business teams to enable analytics and ML use cases
β’ Optimize Databricks jobs for cost, performance, and scalability
β’ Review code, enforce engineering standards, and mentor junior engineers
β’ Participate in technical design discussions and contribute to long-term data strategy
Required Qualifications
β’ 8+ years of experience in Data Engineering roles, with recent experience in a Lead or Senior capacity
β’ Strong hands-on experience with Databricks (workspaces, notebooks, jobs, Delta Lake)
β’ Advanced proficiency in PySpark for large-scale data processing
β’ Strong SQL skills for data transformation, performance tuning, and analytics
β’ Experience working with large datasets in cloud or distributed environments
β’ Solid understanding of data modeling, partitioning, and performance optimization
β’ Experience leading technical initiatives or mentoring engineers
Nice to Have
β’ Experience with cloud platforms (AWS, Azure, or GCP)
β’ Familiarity with Delta Lake, data lakes, or lakehouse architectures
β’ Exposure to streaming technologies (Kafka, Structured Streaming)
β’ Experience supporting analytics, BI, or ML workloads
β’ Knowledge of CI/CD, data observability, or orchestration tools (Airflow, ADF, etc.)
About the Role
Weβre looking for a Lead Data Engineer to design, build, and scale modern data platforms using Databricks, PySpark, and SQL. In this role, youβll take ownership of complex data pipelines, guide architectural decisions, and mentor a team of data engineers while partnering closely with analytics, product, and business stakeholders.
Key Responsibilities
β’ Lead the design and development of scalable, high-performance data pipelines using Databricks
β’ Build and optimize ETL/ELT workflows using PySpark and SQL
β’ Architect data solutions for batch and (if applicable) streaming workloads
β’ Ensure data quality, reliability, performance, and governance best practices
β’ Collaborate with data scientists, analysts, and business teams to enable analytics and ML use cases
β’ Optimize Databricks jobs for cost, performance, and scalability
β’ Review code, enforce engineering standards, and mentor junior engineers
β’ Participate in technical design discussions and contribute to long-term data strategy
Required Qualifications
β’ 8+ years of experience in Data Engineering roles, with recent experience in a Lead or Senior capacity
β’ Strong hands-on experience with Databricks (workspaces, notebooks, jobs, Delta Lake)
β’ Advanced proficiency in PySpark for large-scale data processing
β’ Strong SQL skills for data transformation, performance tuning, and analytics
β’ Experience working with large datasets in cloud or distributed environments
β’ Solid understanding of data modeling, partitioning, and performance optimization
β’ Experience leading technical initiatives or mentoring engineers
Nice to Have
β’ Experience with cloud platforms (AWS, Azure, or GCP)
β’ Familiarity with Delta Lake, data lakes, or lakehouse architectures
β’ Exposure to streaming technologies (Kafka, Structured Streaming)
β’ Experience supporting analytics, BI, or ML workloads
β’ Knowledge of CI/CD, data observability, or orchestration tools (Airflow, ADF, etc.)






