New York Technology Partners

Lead Data Engineer (Only W2 or Selfcorp)

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead Data Engineer in Silver Spring, MD, on a contract basis, focusing on Databricks, Azure, and CI/CD automation. Requires 7+ years of experience in cloud data solutions and a Bachelor's degree in a related field.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
November 22, 2025
πŸ•’ - Duration
Unknown
-
🏝️ - Location
On-site
-
πŸ“„ - Contract
W2 Contractor
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
Silver Spring, MD
-
🧠 - Skills detailed
#Indexing #Data Lake #Airflow #Strategy #Scripting #Data Analysis #Data Processing #Data Warehouse #Model Deployment #Databricks #Monitoring #Metadata #Azure #AI (Artificial Intelligence) #Data Access #Vault #ADF (Azure Data Factory) #Compliance #Data Architecture #Data Ingestion #Scala #SQL Server #PostgreSQL #Programming #Kubernetes #Azure cloud #Data Engineering #Apache Spark #Data Storage #Kafka (Apache Kafka) #DevOps #Data Governance #Deployment #Data Management #Big Data #Automation #BI (Business Intelligence) #Data Pipeline #Data Strategy #MLflow #SQL (Structured Query Language) #Security #GitHub #GDPR (General Data Protection Regulation) #JSON (JavaScript Object Notation) #NoSQL #Data Quality #Infrastructure as Code (IaC) #Computer Science #MongoDB #PySpark #Debugging #Azure Data Factory #Docker #ML (Machine Learning) #Python #Data Science #Storage #Synapse #Cloud #Batch #Azure DevOps #Spark (Apache Spark) #"ETL (Extract #Transform #Load)" #Terraform #Business Analysis #Databases
Role description
Job Title: Lead Data Engineer (Only W2 or Selfcorp) Location: Silver Spring, MD (Onsite usually every Wednesday) Position Type: Contract Top Skills β€’ Databricks β€’ Azure β€’ CI/CD automation experience β€’ Data Lake β€’ SQL/Python β€’ Someone who can mentor and lead. (will not be managing anyone) β€’ Association background – Nice to have Job Description: What You Will Do: β€’ Design, build, and optimize scalable, high-performance data pipelines to support analytics, AI/ML, and business intelligence workloads. β€’ Troubleshoot data pipeline failures, latency issues, and infrastructure bottlenecks, implementing proactive monitoring and alerting. β€’ Lead the adoption of DevOps and automation practices, including CI/CD pipelines, Infrastructure as Code (IaC), and workflow orchestration. β€’ Implement ETL/ELT solutions for efficient data ingestion, transformation, and integration from multiple sources. β€’ Build and manage RESTful APIs to integrate data from various data sources with strong proficiency in handling OAuth authentication, JSON data parsing, and building scalable data pipelines. β€’ Translate complex business requirements into technical solutions, ensuring alignment with enterprise data strategies. β€’ Manage and optimize Azure cloud data storage solutions for cost and performance efficiency. β€’ Ensure data quality, integrity, and consistency by implementing validation, monitoring and error-handling frameworks. β€’ Evaluate and recommend emerging data technologies and architectures to enhance scalability, efficiency, and innovation. β€’ Define and enforce data governance, security, and compliance policies, ensuring alignment with industry standards. β€’ Optimize query performance, indexing, and partitioning strategies to enhance data accessibility and speed. β€’ Collaborate with data architects, data analysts and business stakeholders to deliver data-driven solutions that support decision-making. β€’ Mentor and provide technical guidance to junior data engineers, fostering a culture of best practices and continuous improvement. β€’ Takes authority, responsibility, and accountability for exploiting the value of enterprise information assets and of the analytics used to render insights for decision making automated decisions and augmentation of human performance. Education β€’ Bachelor’s degree in computer science, engineering, or related field. Related Work Experience β€’ Seven (7+) plus years building enterprise-level data solutions on cloud platforms. Skills: β€’ Primary expertise in Databricks, including Databricks Workflows for orchestrating scalable, production-grade data pipelines and ML workflows. β€’ Advanced programming skills in Python and SQL for developing robust data engineering, analytics, and machine learning solutions. β€’ Proficient in Apache Spark (PySpark) for large-scale distributed data processing and real-time analytics on big data platforms. β€’ Experience in integrating data from social media platforms using RESTful APIs, with strong proficiency in handling OAuth authentication, JSON data parsing, and building scalable data pipelines. β€’ Strong DevOps capabilities, including CI/CD pipeline design, automated deployments, and monitoring for data solutions using tools like Shelling Scripting, Azure DevOps, GitHub Actions, and Terraform. β€’ Deep experience with data pipeline design (ETL/ELT), including structured streaming and batch ingestion, transformation, and orchestration using tools like Databricks, Airflow, and Kafka. β€’ Proven expertise in working with relational (PostgreSQL, SQL Server) and NoSQL (MongoDB, Cassandra) databases, as well as data warehouse platforms such as Azure Synapse. β€’ Experience in data governance frameworks, including data quality, lineage, cataloging, access control, and compliance with security and regulatory standards (e.g., GDPR, CCPA). β€’ Experience implementing data quality controls, metadata management, and lineage tracking using Unity Catalog, Great Expectations, and custom rule-based validation frameworks. β€’ Proficient in cloud-native data architectures, especially on Azure, leveraging services such as Azure Data Factory, Azure Key Vault, and Azure Monitor for end-to-end solution management. β€’ Proven experience designing and managing the full machine learning lifecycle using tools such as MLflow, encompassing model deployment, performance tracking, and scalable retraining processes. β€’ Familiarity with containerization and orchestration using Docker and Kubernetes for deploying scalable and portable data services is a plus. β€’ Excellent collaborator with cross-functional teams including data scientists, business analysts, and IT, translating complex technical requirements into actionable, value-driven data solutions. β€’ Strong analytical, debugging, and optimization skills for improving pipeline performance, data quality, and reliability. β€’ Adept at aligning technical data solutions with business goals by understanding data sources, analytical objectives, and enterprise data strategy. If you believe you are qualified for this position and are currently in the job market or interested in making a change, please email me the resume along with contact details at roshni@nytpcorp.com