Nova Web Technologies LLC

Data Engineer – AWS Databricks

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is a Data Engineer – AWS Databricks position in Santa Clara, CA, for a contract longer than 6 months, offering a competitive pay rate. Requires 10+ years of experience, expertise in Databricks, AWS, PySpark, SQL, and batch data pipeline development.
🌎 - Country
United States
💱 - Currency
Unknown
-
💰 - Day rate
Unknown
-
🗓️ - Date
February 14, 2026
🕒 - Duration
More than 6 months
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Santa Clara, CA
-
🧠 - Skills detailed
#Spark SQL #Data Pipeline #Batch #"ETL (Extract #Transform #Load)" #GitLab #Scala #GIT #Data Governance #Databricks #SQL (Structured Query Language) #Databases #BI (Business Intelligence) #Delta Lake #Data Architecture #Data Transformations #AWS (Amazon Web Services) #Data Processing #Triggers #Spark (Apache Spark) #Datasets #Data Engineering #PySpark #IAM (Identity and Access Management) #Cloud #S3 (Amazon Simple Storage Service)
Role description
Location: Santa Clara, CA Experience: 10+ Years Employment Type: Full-Time / Contract Position Overview: We are seeking a highly experienced AWS Databricks Data Engineer to join our data engineering team in Santa Clara. The ideal candidate will have deep expertise in Databricks, AWS, PySpark, SQL, and large-scale data pipeline development. This role focuses on designing and optimizing modern cloud-based data platforms that support analytics, BI, and enterprise reporting use cases. You will collaborate with cross-functional teams and business stakeholders to deliver scalable, secure, and high-performance data solutions built on a Lakehouse architecture. Key Responsibilities: Design and maintain scalable ETL/ELT pipelines using Databricks on AWS Develop high-performance data transformations using PySpark and SQL Implement and optimize Lakehouse (Medallion) architecture for batch data processing Integrate data from S3, databases, and AWS-native services Optimize Spark workloads for performance, cost, and scalability Implement data governance and access controls using Unity Catalog Deploy and manage jobs using Databricks Workflows and CI/CD pipelines Collaborate with business and analytics teams to deliver reliable, production-ready datasets Required Technical Skills: Strong expertise in Databricks: Delta Lake Unity Catalog Lakehouse Architecture Workflows Delta Live Pipelines Table Triggers Databricks Runtime Advanced proficiency in PySpark and SQL Experience designing and rebuilding batch-heavy data pipelines Strong knowledge of Medallion Architecture Expertise in performance tuning and Spark optimization Experience with Databricks Workflows & orchestration Familiarity with Genie enablement concepts (working understanding required) Experience with CI/CD and Git-based development Strong AWS fundamentals: IAM Networking basics S3 Glue Catalog Preferred Qualifications: Experience with Spark Structured Streaming Knowledge of real-time or near real-time data solutions Advanced Databricks Runtime configurations Experience with GitLab CI/CD pipelines Exposure to scalable enterprise data architectures Certifications (Optional) Databricks Certified Data Engineer (Associate/Professional) AWS Data Engineer or AWS Solutions Architect Certification