

Staffing Technologies
Lead Data Engineer (Databricks | Python | PySpark)
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead Data Engineer with a contract length of "unknown," offering a pay rate of "$/hour." Required skills include Databricks, Python, PySpark, and SQL. Candidates should have 15+ years in data engineering and 3+ years in a lead role.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
January 16, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Unknown
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Atlanta Metropolitan Area
-
🧠 - Skills detailed
#Data Quality #Code Reviews #Databricks #Scrum #Data Engineering #Data Warehouse #"ETL (Extract #Transform #Load)" #Deployment #Infrastructure as Code (IaC) #Python #Scala #Agile #Data Processing #PySpark #Terraform #Data Ingestion #Data Lineage #SQL (Structured Query Language) #GIT #Spark (Apache Spark) #Leadership #Model Optimization #DevOps #Monitoring #Cloud #Metadata
Role description
Lead Data Engineer (Databricks | Python | PySpark)
Overview
We are seeking a Lead Data Engineer to design, build, and lead modern cloud-based data platforms with a strong focus on Databricks, Python, and PySpark. This role combines hands-on engineering with technical leadership, owning architecture decisions, delivery standards, and scalable data solutions.
Key Responsibilities
• Lead the design and delivery of cloud-native data platforms using Databricks
• Architect and implement Lakehouse and Data Warehouse patterns
• Build and optimize ETL/ELT pipelines using Python and PySpark
• Establish engineering standards, reusable frameworks, and metadata-driven orchestration
• Review designs, vet solutions with the team, and lead demos and retros prior to deployment
• Enforce data quality, lineage, monitoring, and alerting across pipelines
• Mentor engineers and provide hands-on technical leadership
• Partner with analytics and business teams to align solutions with data and reporting needs
Required Experience & Skills
Core Experience
• ~15 years of total experience in data or software engineering
• 3+ years in a technical lead role
• 5+ years building cloud-based data platforms
• Proven delivery of production-grade, scalable data systems
• Excellent Communication Skills are critical here.
Databricks (Strong Focus)
• Hands-on experience with Databricks Notebooks, Jobs, and workload optimization
• Building pipelines using Lakeflow / Declarative Pipelines
• Data ingestion via Databricks connectors
• Implementing data lineage, quality checks, monitoring, and alerting
• Table, compute, and performance optimization within Databricks
Python, PySpark & Spark
• Advanced Python with strong packaging and dependency management
• Expert PySpark for distributed data processing
• Clear understanding of Spark vs single-node execution
• Spark performance tuning and troubleshooting
SQL
• Strong SQL for mid-to-complex transformations
• Query and data model optimization to reduce compute and improve performance
Engineering & DevOps Practices
• Strong adherence to SOLID and DRY principles
• Experience building parameterized, reusable frameworks
• Agile/SCRUM delivery experience
• Git-based development workflows and code reviews
• Testing strategies: unit, integration, and end-to-end
• CI/CD pipelines and Infrastructure as Code (Terraform)
Lead Data Engineer (Databricks | Python | PySpark)
Overview
We are seeking a Lead Data Engineer to design, build, and lead modern cloud-based data platforms with a strong focus on Databricks, Python, and PySpark. This role combines hands-on engineering with technical leadership, owning architecture decisions, delivery standards, and scalable data solutions.
Key Responsibilities
• Lead the design and delivery of cloud-native data platforms using Databricks
• Architect and implement Lakehouse and Data Warehouse patterns
• Build and optimize ETL/ELT pipelines using Python and PySpark
• Establish engineering standards, reusable frameworks, and metadata-driven orchestration
• Review designs, vet solutions with the team, and lead demos and retros prior to deployment
• Enforce data quality, lineage, monitoring, and alerting across pipelines
• Mentor engineers and provide hands-on technical leadership
• Partner with analytics and business teams to align solutions with data and reporting needs
Required Experience & Skills
Core Experience
• ~15 years of total experience in data or software engineering
• 3+ years in a technical lead role
• 5+ years building cloud-based data platforms
• Proven delivery of production-grade, scalable data systems
• Excellent Communication Skills are critical here.
Databricks (Strong Focus)
• Hands-on experience with Databricks Notebooks, Jobs, and workload optimization
• Building pipelines using Lakeflow / Declarative Pipelines
• Data ingestion via Databricks connectors
• Implementing data lineage, quality checks, monitoring, and alerting
• Table, compute, and performance optimization within Databricks
Python, PySpark & Spark
• Advanced Python with strong packaging and dependency management
• Expert PySpark for distributed data processing
• Clear understanding of Spark vs single-node execution
• Spark performance tuning and troubleshooting
SQL
• Strong SQL for mid-to-complex transformations
• Query and data model optimization to reduce compute and improve performance
Engineering & DevOps Practices
• Strong adherence to SOLID and DRY principles
• Experience building parameterized, reusable frameworks
• Agile/SCRUM delivery experience
• Git-based development workflows and code reviews
• Testing strategies: unit, integration, and end-to-end
• CI/CD pipelines and Infrastructure as Code (Terraform)






