

Motion Recruitment
Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with 5+ years of experience, focusing on PySpark, Databricks, and Azure data platforms. It is a 6+ month hybrid contract in Chicago, IL, offering a pay rate of "$XX per hour."
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
640
-
ποΈ - Date
March 25, 2026
π - Duration
More than 6 months
-
ποΈ - Location
Hybrid
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Chicago, IL
-
π§ - Skills detailed
#ADLS (Azure Data Lake Storage) #Java #Databricks #PySpark #.Net #Logging #ML (Machine Learning) #TypeScript #AI (Artificial Intelligence) #Azure SQL #Triggers #Integration Testing #Monitoring #Data Transformations #Code Reviews #Azure cloud #Python #ADF (Azure Data Factory) #Spring Boot #Angular #Infrastructure as Code (IaC) #Data Pipeline #Cloud #Azure Data Platforms #Documentation #R #"ETL (Extract #Transform #Load)" #Data Engineering #Datasets #Computer Science #Delta Lake #Spark (Apache Spark) #SQL (Structured Query Language) #Azure Data Factory #Azure #Azure Portal #Storage
Role description
About the Company
We are looking for a strong core Data Engineer with hands-on experience in PySpark, Databricks, and Azure data platforms to design, build, and support end-to-end data pipelines. The ideal candidate will develop and optimize data transformations, build production-grade Python components, and maintain cloud-native Azure environments while collaborating with application teams and ensuring high-quality, reliable data delivery. This role offers the opportunity to work with large-scale datasets, implement ETL/ELT best practices, optimize Databricks clusters, and leverage modern cloud technologies to support AI/ML initiatives.
About the Role
Location: 3 days Hybrid in Chicago, IL
Duration: 6+ Month Contract
Interview: 2 video interview and final onsite
Responsibilities
β’ Design, build, and support end-to-end data pipelines, including ingestion, transformation, validation, and publishing.
β’ Develop and optimize SQL and PySpark/Databricks transformations for large datasets.
β’ Build production-grade Python modules with logging, error handling, testing, and integration with APIs/files.
β’ Create, maintain, and operate Azure Data Factory (ADF) pipelines, including triggers, parameterization, monitoring, and failure handling.
β’ Work within Azure environments: ADLS Gen2 (Blob Storage), Azure SQL, Azure App Service, and resource groups.
β’ Provision and maintain Azure components using Pulumi (Infrastructure as Code).
β’ Optimize Databricks clusters, workflows, and jobs for performance and reliability.
β’ Participate in code reviews, documentation, and operational support, including triage and root cause analysis.
β’ Collaborate with application teams for integration, troubleshooting, and operational coordination.
Qualifications
β’ Education: Bachelorβs degree in Computer Science, Engineering, or a related technical field (or equivalent experience).
Required Skills
β’ Experience: 5+ years as a Data Engineer; 3+ years in ETL/ELT concepts, PySpark, and SQL.
β’ SQL: Advanced querying, CTEs, views, joins, complex transformations, and performance tuning.
β’ Python: 2+ years building production-quality modules, unit/integration testing, logging, and CI/CD integration.
β’ Databricks: 1+ year working with notebooks, jobs, workflows, external/managed tables, Delta Lake, and basic cluster configuration.
β’ Azure Data Factory (ADF): 1+ year creating and maintaining pipelines, including triggers, parameterization, monitoring, and error handling.
β’ Azure Cloud: Hands-on with ADLS Gen2, Azure SQL, Azure App Service, and general Azure portal/resource group operations.
β’ Infrastructure as Code: Experience provisioning Azure resources with Pulumi.
β’ ETL/ELT Concepts: Strong understanding of pipeline patterns, incremental loads, data validation, and troubleshooting.
Preferred Skills
β’ Additional Skills (nice-to-have): R for data validation, TypeScript for Pulumi pipelines, Java/.NET for integration, Angular/Spring Boot for minor troubleshooting.
About the Company
We are looking for a strong core Data Engineer with hands-on experience in PySpark, Databricks, and Azure data platforms to design, build, and support end-to-end data pipelines. The ideal candidate will develop and optimize data transformations, build production-grade Python components, and maintain cloud-native Azure environments while collaborating with application teams and ensuring high-quality, reliable data delivery. This role offers the opportunity to work with large-scale datasets, implement ETL/ELT best practices, optimize Databricks clusters, and leverage modern cloud technologies to support AI/ML initiatives.
About the Role
Location: 3 days Hybrid in Chicago, IL
Duration: 6+ Month Contract
Interview: 2 video interview and final onsite
Responsibilities
β’ Design, build, and support end-to-end data pipelines, including ingestion, transformation, validation, and publishing.
β’ Develop and optimize SQL and PySpark/Databricks transformations for large datasets.
β’ Build production-grade Python modules with logging, error handling, testing, and integration with APIs/files.
β’ Create, maintain, and operate Azure Data Factory (ADF) pipelines, including triggers, parameterization, monitoring, and failure handling.
β’ Work within Azure environments: ADLS Gen2 (Blob Storage), Azure SQL, Azure App Service, and resource groups.
β’ Provision and maintain Azure components using Pulumi (Infrastructure as Code).
β’ Optimize Databricks clusters, workflows, and jobs for performance and reliability.
β’ Participate in code reviews, documentation, and operational support, including triage and root cause analysis.
β’ Collaborate with application teams for integration, troubleshooting, and operational coordination.
Qualifications
β’ Education: Bachelorβs degree in Computer Science, Engineering, or a related technical field (or equivalent experience).
Required Skills
β’ Experience: 5+ years as a Data Engineer; 3+ years in ETL/ELT concepts, PySpark, and SQL.
β’ SQL: Advanced querying, CTEs, views, joins, complex transformations, and performance tuning.
β’ Python: 2+ years building production-quality modules, unit/integration testing, logging, and CI/CD integration.
β’ Databricks: 1+ year working with notebooks, jobs, workflows, external/managed tables, Delta Lake, and basic cluster configuration.
β’ Azure Data Factory (ADF): 1+ year creating and maintaining pipelines, including triggers, parameterization, monitoring, and error handling.
β’ Azure Cloud: Hands-on with ADLS Gen2, Azure SQL, Azure App Service, and general Azure portal/resource group operations.
β’ Infrastructure as Code: Experience provisioning Azure resources with Pulumi.
β’ ETL/ELT Concepts: Strong understanding of pipeline patterns, incremental loads, data validation, and troubleshooting.
Preferred Skills
β’ Additional Skills (nice-to-have): R for data validation, TypeScript for Pulumi pipelines, Java/.NET for integration, Angular/Spring Boot for minor troubleshooting.






