

MM International, LLC
Data Engineer / Databricks Developer (Onsite Interview)
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer / Databricks Developer in New York, requiring over 6 months of full-time commitment. Key skills include Azure Databricks, Apache Spark, and SQL, with mandatory experience in Investment Banking. Onsite work is required 3 days/week.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
April 14, 2026
🕒 - Duration
More than 6 months
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
New York, United States
-
🧠 - Skills detailed
#Data Lineage #"ETL (Extract #Transform #Load)" #Azure Databricks #Security #Data Quality #Datasets #Azure #Data Architecture #Data Engineering #PySpark #Cloud #Spark (Apache Spark) #ADF (Azure Data Factory) #Data Access #Azure Data Factory #SQL (Structured Query Language) #MDM (Master Data Management) #Data Processing #Data Pipeline #Compliance #Databricks #Data Governance #Data Management #Data Integrity #Apache Spark #Scala
Role description
Job Title: Data Engineer / Databricks Developer
Location: New York (Onsite – 3 days/week)
Employment Type: Full-time
Note: Only local candidates will be considered. In-person interview is mandatory.
Job Overview
We are seeking an experienced Data Engineer / Databricks Developer with a strong background in the Investment Banking domain to join our team in New York. The ideal candidate will be responsible for designing, developing, and optimizing scalable data solutions using Azure-based technologies, with a focus on Databricks and modern data architecture.
Key Responsibilities
• Design, develop, and implement scalable data pipelines using Azure Databricks and Azure Data Factory.
• Build and maintain robust ETL/ELT workflows using Apache Spark and PySpark DataFrames.
• Develop high-performance data transformation processes using Spark and SQL.
• Optimize data pipelines for efficient ingestion, processing, and transformation of large datasets.
• Implement and maintain data governance frameworks, including data access, security, compliance, and lifecycle management.
• Ensure data quality, consistency, and lineage across enterprise data platforms.
• Design and support modern Lakehouse architecture and distributed data processing systems.
• Apply Master Data Management (MDM) principles to maintain data integrity and standardization.
• Collaborate with cross-functional teams to deliver scalable, reliable, enterprise-grade data solutions.
Required Skills & Qualifications
• Strong hands-on experience with Azure Databricks
• Proficiency in Apache Spark / PySpark
• Experience with Azure Data Factory
• Advanced knowledge of SQL
• Expertise in building data pipelines and ETL/ELT processes
• Experience with data governance, data lineage, and data quality frameworks
• Strong understanding of Lakehouse architecture
• Experience working in Investment Banking or Financial Services domain (Mandatory)
Preferred Qualifications
• Experience with large-scale distributed data systems
• Familiarity with cloud-based data platforms and enterprise data architecture
• Strong problem-solving and performance optimization skills
Job Title: Data Engineer / Databricks Developer
Location: New York (Onsite – 3 days/week)
Employment Type: Full-time
Note: Only local candidates will be considered. In-person interview is mandatory.
Job Overview
We are seeking an experienced Data Engineer / Databricks Developer with a strong background in the Investment Banking domain to join our team in New York. The ideal candidate will be responsible for designing, developing, and optimizing scalable data solutions using Azure-based technologies, with a focus on Databricks and modern data architecture.
Key Responsibilities
• Design, develop, and implement scalable data pipelines using Azure Databricks and Azure Data Factory.
• Build and maintain robust ETL/ELT workflows using Apache Spark and PySpark DataFrames.
• Develop high-performance data transformation processes using Spark and SQL.
• Optimize data pipelines for efficient ingestion, processing, and transformation of large datasets.
• Implement and maintain data governance frameworks, including data access, security, compliance, and lifecycle management.
• Ensure data quality, consistency, and lineage across enterprise data platforms.
• Design and support modern Lakehouse architecture and distributed data processing systems.
• Apply Master Data Management (MDM) principles to maintain data integrity and standardization.
• Collaborate with cross-functional teams to deliver scalable, reliable, enterprise-grade data solutions.
Required Skills & Qualifications
• Strong hands-on experience with Azure Databricks
• Proficiency in Apache Spark / PySpark
• Experience with Azure Data Factory
• Advanced knowledge of SQL
• Expertise in building data pipelines and ETL/ELT processes
• Experience with data governance, data lineage, and data quality frameworks
• Strong understanding of Lakehouse architecture
• Experience working in Investment Banking or Financial Services domain (Mandatory)
Preferred Qualifications
• Experience with large-scale distributed data systems
• Familiarity with cloud-based data platforms and enterprise data architecture
• Strong problem-solving and performance optimization skills






