MokshaaLLC

Cloud Data Engineer - Boston, MA (Onsite)

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Cloud Data Engineer in Boston, MA, offering a 6-month+ contract at $55-$60/hr. Requires 5-10 years of data engineering experience, with 3+ years in Databricks and PySpark, and proficiency in SQL, Python, and Apache Airflow.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
480
-
🗓️ - Date
October 25, 2025
🕒 - Duration
More than 6 months
-
🏝️ - Location
On-site
-
📄 - Contract
W2 Contractor
-
🔒 - Security
Unknown
-
📍 - Location detailed
Boston, MA
-
🧠 - Skills detailed
#Apache Airflow #Data Storage #Monitoring #Scala #Security #Compliance #Storage #Data Management #Data Transformations #Data Integration #S3 (Amazon Simple Storage Service) #Cloud #Data Catalog #Data Pipeline #MDM (Master Data Management) #Automation #Delta Lake #Data Quality #Spark (Apache Spark) #Data Engineering #Data Processing #Databricks #Airflow #SQL (Structured Query Language) #"ETL (Extract #Transform #Load)" #Collibra #Data Modeling #Complex Queries #Lambda (AWS Lambda) #IAM (Identity and Access Management) #PySpark #EC2 #AWS (Amazon Web Services) #Python #Data Governance #Data Science
Role description
Cloud Data Engineer with Databricks and PySpark Expertise Location: Austin, TX (Onsite) Duration: 6 months+ Contract Contract - C2C/W2/1099 Rate: $55/hr to $60/hr (depending on experience and engagement type) Job Summary: We are seeking a highly skilled Data Engineer with strong expertise in Databricks and PySpark to design, develop, and optimize large-scale data integration and transformation solutions. The ideal candidate will have proven experience in architecting Lakehouse-based solutions, integrating data from multiple enterprise systems, and implementing best practices in ETL/ELT pipelines, data modeling, and orchestration frameworks. Key Responsibilities: • Architect, design, and implement scalable data processing solutions using Databricks, PySpark, and Delta Lake. • Develop and manage ETL/ELT pipelines integrating data from various sources including Salesforce, MDM systems, and other enterprise applications. • Design, configure, and optimize Databricks clusters, jobs, and workflows for performance and cost efficiency. • Integrate Databricks with data governance tools (e.g., Collibra) and ensure compliance with data management policies. • Implement and maintain Apache Airflow DAGs for orchestration, monitoring, and automation of data pipelines. • Work with AWS services including S3, EC2, Lambda, and IAM for data storage, compute, and access management. • Perform complex data transformations and ensure data quality using SQL and Python. • Design and maintain Lakehouse architectures following data warehousing best practices. • Collaborate with cross-functional teams (data scientists, analysts, business users) to deliver data-driven insights. • Ensure system reliability, security, and performance tuning for data platforms. Required Skills & Experience: • 5–10 years of hands-on experience in data engineering, with at least 3+ years in Databricks and PySpark. • Strong understanding of Delta Lake, Lakehouse architecture, and data modeling principles. • Experience integrating Databricks with Salesforce, MDM platforms, and other enterprise applications. • Proficiency in SQL (complex queries, optimization) and Python for data processing and automation. • Strong experience with Apache Airflow (DAG creation, scheduling, and monitoring). • Hands-on experience with AWS services (S3, EC2, Lambda, IAM). • Familiarity with data catalog and governance tools such as Collibra. • Excellent problem-solving skills, attention to detail, and ability to work in a collaborative environment.