Addison Group

Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with expertise in Azure Databricks, requiring hands-on experience in PySpark, ETL/ELT processes, and data governance. Contract length is unspecified, with a competitive pay rate. Remote work is allowed.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
560
-
πŸ—“οΈ - Date
January 6, 2026
πŸ•’ - Duration
Unknown
-
🏝️ - Location
Unknown
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
Houston, TX
-
🧠 - Skills detailed
#Azure ADLS (Azure Data Lake Storage) #Spark (Apache Spark) #Python #Data Architecture #DevOps #Deployment #Scala #Data Modeling #Data Lake #Data Pipeline #Storage #ADLS (Azure Data Lake Storage) #Databricks #SaaS (Software as a Service) #Azure Databricks #Security #PySpark #"ETL (Extract #Transform #Load)" #SQL (Structured Query Language) #Data Governance #Azure DevOps #Batch #Delta Lake #Data Quality #Compliance #Data Engineering #Debugging #Azure #Schema Design #Monitoring #GIT
Role description
We’re looking for a highly skilled Data Engineer with deep expertise in Azure Databricks and modern data engineering practices. This role is ideal for someone who thrives in fast‑moving environments, collaborates well with scientists and engineers, and can independently translate business needs into scalable data solutions. Responsibilities β€’ Design, build, and optimize data pipelines using Azure Databricks, PySpark, Delta Lake, and Workflows β€’ Develop robust ETL/ELT processes for both batch and streaming workloads β€’ Ingest and integrate data from RESTful SaaS APIs, including sources with and without CDC capabilities β€’ Implement strong governance and security practices using Unity Catalog, role-based access, data quality checks, and lineage tracking β€’ Collaborate with scientists, engineers, and business stakeholders to understand workflows and translate them into scalable data models β€’ Ensure reliability and performance of production pipelines, including monitoring, alerting, and SLA adherence β€’ Contribute to CI/CD processes using Azure DevOps, Git, and automated deployment patterns β€’ Support schema design and data modeling for scientific and laboratory data environments Required Qualifications (Technical Must‑Haves) β€’ Hands-on experience with Databricks on Azure: Spark (PySpark), Delta Lake, Workflows β€’ Strong skills in optimization, tuning, and debugging distributed data workloads β€’ Expertise in ETL/ELT design, batch/stream processing, and scalable data architecture β€’ Experience ingesting data from RESTful APIs β€’ Solid understanding of data governance, security, and compliance best practices β€’ Proficiency in Python, SQL, Git, and CI/CD pipelines (Azure DevOps) β€’ Experience with Azure Data Lake Storage Gen2 and production monitoring/alerting β€’ Proven ability to productionize pipelines with performance tuning and SLA management