

Ampstek
Data Engineer (Contract W2)
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer (Contract W2) in Texas, requiring 6+ years of experience with Databricks, Spark, Hive, AWS QuickSight, Python, and Django. The contract length and pay rate are unspecified.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
March 19, 2026
π - Duration
Unknown
-
ποΈ - Location
On-site
-
π - Contract
W2 Contractor
-
π - Security
Unknown
-
π - Location detailed
Texas, United States
-
π§ - Skills detailed
#Automated Testing #Azure #ADF (Azure Data Factory) #AWS EMR (Amazon Elastic MapReduce) #Spark (Apache Spark) #Scala #Big Data #Apache Spark #Datasets #Django #Spark SQL #Batch #"ETL (Extract #Transform #Load)" #PySpark #Documentation #Databricks #Data Processing #AWS (Amazon Web Services) #dbt (data build tool) #Data Engineering #Delta Lake #Python #SQL (Structured Query Language) #Azure Data Factory #Data Pipeline #Data Documentation
Role description
Role : Data Engineering
Location : Texas (Onsite)
Must Have Skills:
β¦ Databricks
β¦ Spark
β¦ Hive
β¦ AWS QuickSight
β¦ Python
β¦ Django
Minimum Years of Experience:
β¦ 6+ years
Nice to Have Skills:
Detailed Job Description:
Data Engineering & Big Data Development
β¦ Design and develop scalable, highβperformance data pipelines using:
1. Databricks (PySpark/SQL)
1. Apache Spark (batch & streaming)
1. Hive (query optimization, partitioning, bucketing)
1. AWS EMR (PySpark jobs for large-scale data processing)
1. Azure Data Factory (ADF) for ingestion and pipeline orchestration.
β¦ Build data processing frameworks to handle structured, semiβstructured, and unstructured datasets.
β¦ Develop highly optimized ETL/ELT workflows using Spark, SQL, Python.
β¦ Create curated data models (Bronze/Silver/Gold) using Databricks Delta Lake.
β¦ Optimize Spark transformations through:
1. Caching, checkpointing
1. Partition pruning
1. Adaptive query execution (AQE)
β¦ Build DBT models for:
1. SQL-based transformations
1. Automated testing
1. Lineage graphs
1. Data documentation to provide transparency across pipelines.
Thanks
Sandra
Sandra@ampstek.com
Role : Data Engineering
Location : Texas (Onsite)
Must Have Skills:
β¦ Databricks
β¦ Spark
β¦ Hive
β¦ AWS QuickSight
β¦ Python
β¦ Django
Minimum Years of Experience:
β¦ 6+ years
Nice to Have Skills:
Detailed Job Description:
Data Engineering & Big Data Development
β¦ Design and develop scalable, highβperformance data pipelines using:
1. Databricks (PySpark/SQL)
1. Apache Spark (batch & streaming)
1. Hive (query optimization, partitioning, bucketing)
1. AWS EMR (PySpark jobs for large-scale data processing)
1. Azure Data Factory (ADF) for ingestion and pipeline orchestration.
β¦ Build data processing frameworks to handle structured, semiβstructured, and unstructured datasets.
β¦ Develop highly optimized ETL/ELT workflows using Spark, SQL, Python.
β¦ Create curated data models (Bronze/Silver/Gold) using Databricks Delta Lake.
β¦ Optimize Spark transformations through:
1. Caching, checkpointing
1. Partition pruning
1. Adaptive query execution (AQE)
β¦ Build DBT models for:
1. SQL-based transformations
1. Automated testing
1. Lineage graphs
1. Data documentation to provide transparency across pipelines.
Thanks
Sandra
Sandra@ampstek.com






