

New York Technology Partners
Azure Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for an Azure Data Engineer focused on building scalable data pipelines using Azure Data Factory and Databricks, requiring expertise in PySpark and SQL. Ideal candidates should have Oil & Gas experience and a strong background in data ingestion and real-time analytics. Contract length and pay rate are unspecified.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
January 8, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Hybrid
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Houston, TX
-
🧠 - Skills detailed
#Azure #Spark (Apache Spark) #Data Ingestion #PySpark #"ETL (Extract #Transform #Load)" #Storage #Cloud #Data Management #Data Lake #Python #SQL (Structured Query Language) #Data Engineering #BI (Business Intelligence) #Data Pipeline #Databricks #ADLS (Azure Data Lake Storage) #SQL Queries #Azure ADLS (Azure Data Lake Storage) #AI (Artificial Intelligence) #Data Integrity #Scala #Datasets #ADF (Azure Data Factory) #Azure Data Factory #Spark SQL
Role description
Azure Data Engineer (Databricks & PySpark)
• Job Title: Azure Data Engineer
• Location: Hybrid – Houston, TX (Remote Dallas option with travel)
Job Summary
We are seeking an Azure Data Engineer to join a high-impact project modernizing business data into a cloud-native architecture. This role focuses on moving digitalized business data to a modern Azure structure to deliver BI insights, advanced analytics, and AI capabilities. You will be responsible for building complex, scalable end-to-end pipelines using Azure Data Factory (ADF) and Databricks. The ideal candidate has deep experience in data ingestion, curation using PySpark, and a background in Oil & Gas (O&G), specifically working with time-series data and real-time use cases.
Key Responsibilities & Required Skills
Data Pipeline Engineering & Orchestration
• ADF Implementation: Design and develop end-to-end scalable pipelines in Azure Data Factory for seamless data ingestion into Azure Data Lake Storage (ADLS).
• Scalable ETL: Build and implement robust ETL processes to move data across environments while ensuring data integrity and high availability.
• Real-time Use Cases: Optimize pipelines to handle time-series data and support real-time analytics requirements.
Databricks & PySpark Development
• Data Curation: Build curated datasets within Databricks using PySpark and SQL, applying complex pivot logic and aggregations.
• Advanced Analytics: Enhance data management by extracting data from multiple disparate sources and transforming it for downstream AI and BI consumption.
• Coding Mastery: Write complex SQL queries from scratch and develop reusable Python/PySpark scripts for data transformation.
Technical Environment & Standards
• Cloud Modernization: Support the transition of legacy digital data to a modern cloud structure within the complete Azure ecosystem.
• Collaboration: Work closely with BI and AI teams to ensure curated data meets the requirements for advanced analytics and insights.
• Emerging Tech: Leverage or integrate with Microsoft Fabric (Nice to Have) to further streamline the data engineering workspace.
Mandatory Technical Skills
1. Azure Data Factory: Experience with complex, scalable ingestion pipelines.
1. Databricks: Expert-level PySpark, SQL, and curated dataset building.
1. SQL: Proven ability to write high-performance queries from scratch.
Azure Data Engineer (Databricks & PySpark)
• Job Title: Azure Data Engineer
• Location: Hybrid – Houston, TX (Remote Dallas option with travel)
Job Summary
We are seeking an Azure Data Engineer to join a high-impact project modernizing business data into a cloud-native architecture. This role focuses on moving digitalized business data to a modern Azure structure to deliver BI insights, advanced analytics, and AI capabilities. You will be responsible for building complex, scalable end-to-end pipelines using Azure Data Factory (ADF) and Databricks. The ideal candidate has deep experience in data ingestion, curation using PySpark, and a background in Oil & Gas (O&G), specifically working with time-series data and real-time use cases.
Key Responsibilities & Required Skills
Data Pipeline Engineering & Orchestration
• ADF Implementation: Design and develop end-to-end scalable pipelines in Azure Data Factory for seamless data ingestion into Azure Data Lake Storage (ADLS).
• Scalable ETL: Build and implement robust ETL processes to move data across environments while ensuring data integrity and high availability.
• Real-time Use Cases: Optimize pipelines to handle time-series data and support real-time analytics requirements.
Databricks & PySpark Development
• Data Curation: Build curated datasets within Databricks using PySpark and SQL, applying complex pivot logic and aggregations.
• Advanced Analytics: Enhance data management by extracting data from multiple disparate sources and transforming it for downstream AI and BI consumption.
• Coding Mastery: Write complex SQL queries from scratch and develop reusable Python/PySpark scripts for data transformation.
Technical Environment & Standards
• Cloud Modernization: Support the transition of legacy digital data to a modern cloud structure within the complete Azure ecosystem.
• Collaboration: Work closely with BI and AI teams to ensure curated data meets the requirements for advanced analytics and insights.
• Emerging Tech: Leverage or integrate with Microsoft Fabric (Nice to Have) to further streamline the data engineering workspace.
Mandatory Technical Skills
1. Azure Data Factory: Experience with complex, scalable ingestion pipelines.
1. Databricks: Expert-level PySpark, SQL, and curated dataset building.
1. SQL: Proven ability to write high-performance queries from scratch.





