Mindlance

Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for an Azure Data Engineer in Washington, DC, on a 3-month contract. Key skills include Azure Data Factory, DBT, Databricks, and Python. Experience with structured/unstructured data and data governance is required. Hybrid work model, 4 days onsite.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
520
-
πŸ—“οΈ - Date
March 17, 2026
πŸ•’ - Duration
3 to 6 months
-
🏝️ - Location
Hybrid
-
πŸ“„ - Contract
1099 Contractor
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
Washington DC-Baltimore Area
-
🧠 - Skills detailed
#Scala #Data Modeling #Datasets #Data Architecture #Cloud #Data Science #"ETL (Extract #Transform #Load)" #ML (Machine Learning) #Azure Data Factory #Data Lake #Azure #dbt (data build tool) #Data Quality #Programming #Data Pipeline #Data Governance #ADF (Azure Data Factory) #Data Processing #AI (Artificial Intelligence) #Python #Data Engineering #Databricks
Role description
Position: Azure Data Engineer Location: Washington, DC 20433 (Hybrid – 4 days onsite per week from Day 1; expected to transition to 100% onsite) Duration: 3 Months (Contract) Job Description: We are seeking an experienced Azure Data Engineer to support the K360 (Knowledge 360) initiative in building a strong and scalable data foundation. This role will focus on designing and implementing modern data pipelines that support both structured and unstructured data, enabling advanced analytics and AI-driven capabilities across the organization. The ideal candidate will have hands-on experience with Azure-based data platforms, modern data transformation tools, and large-scale data processing frameworks. This position will play a key role in establishing the core data infrastructure that powers downstream AI and analytics initiatives. Key Responsibilities β€’ Design, develop, and maintain scalable data pipelines using Azure Data Factory (ADF). β€’ Build and manage data transformation workflows using DBT (Data Build Tool). β€’ Develop and optimize data processing solutions using Databricks and Python. β€’ Integrate and manage both structured and unstructured data sources within the data platform. β€’ Collaborate with data scientists, analysts, and business teams to deliver reliable and high-quality datasets. β€’ Implement best practices for data modeling, data quality, and data governance. β€’ Optimize data workflows for performance, scalability, and cost efficiency. β€’ Support the development of a modern data architecture that enables advanced analytics and AI acceleration. β€’ Troubleshoot and resolve data pipeline issues while ensuring high availability and reliability. Required Skills & Experience β€’ Strong experience in Azure Data Engineering and modern cloud data platforms. β€’ Hands-on experience with Azure Data Factory (ADF) for building and orchestrating data pipelines. β€’ Experience with DBT (Data Build Tool) for data transformation and modeling. β€’ Proficiency in Databricks for large-scale data processing. β€’ Strong programming experience with Python for data engineering tasks. β€’ Experience working with both structured and unstructured data. β€’ Knowledge of data warehousing, ETL/ELT processes, and data modeling. β€’ Familiarity with building scalable data platforms that support analytics and AI workloads. Preferred Qualifications β€’ Experience working on modern data lake or lakehouse architectures. β€’ Familiarity with AI/ML data pipelines and data preparation for machine learning. β€’ Experience working in large enterprise or global data environments. β€’ Knowledge of data governance and data quality frameworks. β€œMindlance is an Equal Opportunity Employer and does not discriminate in employment on the basis of – Minority/Gender/Disability/Religion/LGBTQI/Age/Veterans.”