Centraprise

Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with strong experience in Databricks, Apache Spark, and Azure cloud services, offering a 12+ month contract in Plano, TX. Key skills include Python, SQL, ETL/ELT design, and data modeling.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
October 10, 2025
🕒 - Duration
More than 6 months
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Dallas, TX
-
🧠 - Skills detailed
#Data Science #Azure #Data Governance #Jenkins #BI (Business Intelligence) #Spark (Apache Spark) #Spark SQL #Terraform #Debugging #Scala #GIT #Agile #Azure ADLS (Azure Data Lake Storage) #Python #Compliance #Cloud #ADF (Azure Data Factory) #Azure DevOps #"ETL (Extract #Transform #Load)" #Synapse #Data Architecture #DevOps #ADLS (Azure Data Lake Storage) #Databricks #GCP (Google Cloud Platform) #Data Quality #Data Lake #PySpark #Airflow #Monitoring #Version Control #AWS (Amazon Web Services) #Azure Data Factory #Storage #MLflow #Data Ingestion #Logging #Data Engineering #Data Pipeline #Datasets #Delta Lake #Security #Apache Spark #Infrastructure as Code (IaC) #SQL (Structured Query Language) #Data Lakehouse #Data Modeling
Role description
Databricks Engineer Plano, TX (Day 1 onsite) 12+ months contract Job Description: Required Skills & Qualifications: • Strong hands-on experience with Databricks, Apache Spark, PySpark, and Spark SQL. • Proficiency in Python, SQL, and data modeling concepts. • Experience working with Azure Data Lake Storage (ADLS), Azure Synapse, Azure Data Factory, or equivalent cloud services (AWS/GCP). • Solid understanding of ETL/ELT design patterns, data warehousing, and data lakehouse architectures. • Experience with Delta Lake, Unity Catalog, and MLflow is a plus. • Knowledge of CI/CD pipelines and infrastructure-as-code (IaC) tools such as Terraform or ARM templates. • Familiarity with Git-based version control and Agile delivery methodologies. • Excellent problem-solving, debugging, and performance tuning skills. Key Responsibilities: • Design, build, and maintain data pipelines and ETL/ELT workflows using Databricks (Spark, PySpark, SQL, Delta Lake). • Develop scalable data ingestion, cleansing, and transformation processes from various structured and unstructured data sources. • Implement and optimize Delta Lake for data versioning, reliability, and performance. • Work closely with data architects, data scientists, and BI teams to deliver high-quality, production-grade datasets. • Tune and optimize Spark clusters and jobs for performance and cost efficiency. • Integrate Databricks with Azure Data Lake Storage (ADLS), Azure Synapse, Event Hubs, and other Azure services. • Manage job scheduling, orchestration, and monitoring (e.g., using Azure Data Factory, Airflow, or Databricks Workflows). • Implement data quality checks, logging, and error handling mechanisms. • Follow DevOps and CI/CD practices for data engineering projects (e.g., Git, Azure DevOps, Jenkins). • Ensure compliance with data governance, security, and privacy standards.