Motion Recruitment

Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with 5+ years of experience, focusing on PySpark, Databricks, and Azure data platforms. It is a 6+ month hybrid contract in Chicago, IL, offering a pay rate of "$XX per hour."
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
640
-
πŸ—“οΈ - Date
March 25, 2026
πŸ•’ - Duration
More than 6 months
-
🏝️ - Location
Hybrid
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
Chicago, IL
-
🧠 - Skills detailed
#ADLS (Azure Data Lake Storage) #Java #Databricks #PySpark #.Net #Logging #ML (Machine Learning) #TypeScript #AI (Artificial Intelligence) #Azure SQL #Triggers #Integration Testing #Monitoring #Data Transformations #Code Reviews #Azure cloud #Python #ADF (Azure Data Factory) #Spring Boot #Angular #Infrastructure as Code (IaC) #Data Pipeline #Cloud #Azure Data Platforms #Documentation #R #"ETL (Extract #Transform #Load)" #Data Engineering #Datasets #Computer Science #Delta Lake #Spark (Apache Spark) #SQL (Structured Query Language) #Azure Data Factory #Azure #Azure Portal #Storage
Role description
About the Company We are looking for a strong core Data Engineer with hands-on experience in PySpark, Databricks, and Azure data platforms to design, build, and support end-to-end data pipelines. The ideal candidate will develop and optimize data transformations, build production-grade Python components, and maintain cloud-native Azure environments while collaborating with application teams and ensuring high-quality, reliable data delivery. This role offers the opportunity to work with large-scale datasets, implement ETL/ELT best practices, optimize Databricks clusters, and leverage modern cloud technologies to support AI/ML initiatives. About the Role Location: 3 days Hybrid in Chicago, IL Duration: 6+ Month Contract Interview: 2 video interview and final onsite Responsibilities β€’ Design, build, and support end-to-end data pipelines, including ingestion, transformation, validation, and publishing. β€’ Develop and optimize SQL and PySpark/Databricks transformations for large datasets. β€’ Build production-grade Python modules with logging, error handling, testing, and integration with APIs/files. β€’ Create, maintain, and operate Azure Data Factory (ADF) pipelines, including triggers, parameterization, monitoring, and failure handling. β€’ Work within Azure environments: ADLS Gen2 (Blob Storage), Azure SQL, Azure App Service, and resource groups. β€’ Provision and maintain Azure components using Pulumi (Infrastructure as Code). β€’ Optimize Databricks clusters, workflows, and jobs for performance and reliability. β€’ Participate in code reviews, documentation, and operational support, including triage and root cause analysis. β€’ Collaborate with application teams for integration, troubleshooting, and operational coordination. Qualifications β€’ Education: Bachelor’s degree in Computer Science, Engineering, or a related technical field (or equivalent experience). Required Skills β€’ Experience: 5+ years as a Data Engineer; 3+ years in ETL/ELT concepts, PySpark, and SQL. β€’ SQL: Advanced querying, CTEs, views, joins, complex transformations, and performance tuning. β€’ Python: 2+ years building production-quality modules, unit/integration testing, logging, and CI/CD integration. β€’ Databricks: 1+ year working with notebooks, jobs, workflows, external/managed tables, Delta Lake, and basic cluster configuration. β€’ Azure Data Factory (ADF): 1+ year creating and maintaining pipelines, including triggers, parameterization, monitoring, and error handling. β€’ Azure Cloud: Hands-on with ADLS Gen2, Azure SQL, Azure App Service, and general Azure portal/resource group operations. β€’ Infrastructure as Code: Experience provisioning Azure resources with Pulumi. β€’ ETL/ELT Concepts: Strong understanding of pipeline patterns, incremental loads, data validation, and troubleshooting. Preferred Skills β€’ Additional Skills (nice-to-have): R for data validation, TypeScript for Pulumi pipelines, Java/.NET for integration, Angular/Spring Boot for minor troubleshooting.