

MokshaaLLC
Cloud Data Engineer - Boston, MA (Onsite)
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Cloud Data Engineer in Boston, MA, offering a 6-month+ contract at $55-$60/hr. Requires 5-10 years of data engineering experience, with 3+ years in Databricks and PySpark, and proficiency in SQL, Python, and Apache Airflow.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
480
-
ποΈ - Date
October 25, 2025
π - Duration
More than 6 months
-
ποΈ - Location
On-site
-
π - Contract
W2 Contractor
-
π - Security
Unknown
-
π - Location detailed
Boston, MA
-
π§ - Skills detailed
#Apache Airflow #Data Storage #Monitoring #Scala #Security #Compliance #Storage #Data Management #Data Transformations #Data Integration #S3 (Amazon Simple Storage Service) #Cloud #Data Catalog #Data Pipeline #MDM (Master Data Management) #Automation #Delta Lake #Data Quality #Spark (Apache Spark) #Data Engineering #Data Processing #Databricks #Airflow #SQL (Structured Query Language) #"ETL (Extract #Transform #Load)" #Collibra #Data Modeling #Complex Queries #Lambda (AWS Lambda) #IAM (Identity and Access Management) #PySpark #EC2 #AWS (Amazon Web Services) #Python #Data Governance #Data Science
Role description
Cloud Data Engineer with Databricks and PySpark Expertise
Location: Austin, TX (Onsite)
Duration: 6 months+ Contract
Contract - C2C/W2/1099
Rate: $55/hr to $60/hr (depending on experience and engagement type)
Job Summary:
We are seeking a highly skilled Data Engineer with strong expertise in Databricks and PySpark to design, develop, and optimize large-scale data integration and transformation solutions. The ideal candidate will have proven experience in architecting Lakehouse-based solutions, integrating data from multiple enterprise systems, and implementing best practices in ETL/ELT pipelines, data modeling, and orchestration frameworks.
Key Responsibilities:
β’ Architect, design, and implement scalable data processing solutions using Databricks, PySpark, and Delta Lake.
β’ Develop and manage ETL/ELT pipelines integrating data from various sources including Salesforce, MDM systems, and other enterprise applications.
β’ Design, configure, and optimize Databricks clusters, jobs, and workflows for performance and cost efficiency.
β’ Integrate Databricks with data governance tools (e.g., Collibra) and ensure compliance with data management policies.
β’ Implement and maintain Apache Airflow DAGs for orchestration, monitoring, and automation of data pipelines.
β’ Work with AWS services including S3, EC2, Lambda, and IAM for data storage, compute, and access management.
β’ Perform complex data transformations and ensure data quality using SQL and Python.
β’ Design and maintain Lakehouse architectures following data warehousing best practices.
β’ Collaborate with cross-functional teams (data scientists, analysts, business users) to deliver data-driven insights.
β’ Ensure system reliability, security, and performance tuning for data platforms.
Required Skills & Experience:
β’ 5β10 years of hands-on experience in data engineering, with at least 3+ years in Databricks and PySpark.
β’ Strong understanding of Delta Lake, Lakehouse architecture, and data modeling principles.
β’ Experience integrating Databricks with Salesforce, MDM platforms, and other enterprise applications.
β’ Proficiency in SQL (complex queries, optimization) and Python for data processing and automation.
β’ Strong experience with Apache Airflow (DAG creation, scheduling, and monitoring).
β’ Hands-on experience with AWS services (S3, EC2, Lambda, IAM).
β’ Familiarity with data catalog and governance tools such as Collibra.
β’ Excellent problem-solving skills, attention to detail, and ability to work in a collaborative environment.
Cloud Data Engineer with Databricks and PySpark Expertise
Location: Austin, TX (Onsite)
Duration: 6 months+ Contract
Contract - C2C/W2/1099
Rate: $55/hr to $60/hr (depending on experience and engagement type)
Job Summary:
We are seeking a highly skilled Data Engineer with strong expertise in Databricks and PySpark to design, develop, and optimize large-scale data integration and transformation solutions. The ideal candidate will have proven experience in architecting Lakehouse-based solutions, integrating data from multiple enterprise systems, and implementing best practices in ETL/ELT pipelines, data modeling, and orchestration frameworks.
Key Responsibilities:
β’ Architect, design, and implement scalable data processing solutions using Databricks, PySpark, and Delta Lake.
β’ Develop and manage ETL/ELT pipelines integrating data from various sources including Salesforce, MDM systems, and other enterprise applications.
β’ Design, configure, and optimize Databricks clusters, jobs, and workflows for performance and cost efficiency.
β’ Integrate Databricks with data governance tools (e.g., Collibra) and ensure compliance with data management policies.
β’ Implement and maintain Apache Airflow DAGs for orchestration, monitoring, and automation of data pipelines.
β’ Work with AWS services including S3, EC2, Lambda, and IAM for data storage, compute, and access management.
β’ Perform complex data transformations and ensure data quality using SQL and Python.
β’ Design and maintain Lakehouse architectures following data warehousing best practices.
β’ Collaborate with cross-functional teams (data scientists, analysts, business users) to deliver data-driven insights.
β’ Ensure system reliability, security, and performance tuning for data platforms.
Required Skills & Experience:
β’ 5β10 years of hands-on experience in data engineering, with at least 3+ years in Databricks and PySpark.
β’ Strong understanding of Delta Lake, Lakehouse architecture, and data modeling principles.
β’ Experience integrating Databricks with Salesforce, MDM platforms, and other enterprise applications.
β’ Proficiency in SQL (complex queries, optimization) and Python for data processing and automation.
β’ Strong experience with Apache Airflow (DAG creation, scheduling, and monitoring).
β’ Hands-on experience with AWS services (S3, EC2, Lambda, IAM).
β’ Familiarity with data catalog and governance tools such as Collibra.
β’ Excellent problem-solving skills, attention to detail, and ability to work in a collaborative environment.






