

Purple Austyn technologies
Senior Big Data Engineer - AWS
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is a Senior Big Data Engineer - AWS, offering a contract for an unspecified length at $50.00 - $65.00 per hour. Key skills required include AWS, Kubernetes, SQL, Python, and CI/CD. Candidates must reside in the MVD area.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
520
-
🗓️ - Date
March 18, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Rockville, MD 20853
-
🧠 - Skills detailed
#Datasets #AWS (Amazon Web Services) #Big Data #Scripting #Scala #AI (Artificial Intelligence) #Presto #S3 (Amazon Simple Storage Service) #Cloud #Trino #Athena #Jenkins #AWS Glue #HDFS (Hadoop Distributed File System) #Data Lake #Scrum #Data Catalog #Python #Grafana #SQL (Structured Query Language) #AWS S3 (Amazon Simple Storage Service) #Spark (Apache Spark) #ML (Machine Learning) #Data Access #Data Engineering #Kubernetes #Shell Scripting #Metadata #Hadoop #Data Pipeline
Role description
Responsibilities
Design and build scalable data lakes and pipelines on AWS using cloud-native and automated solutions.
Enable fast, federated analytics using Amazon Athena and Trino, with performance tuning for large-scale queries.
Manage metadata, schemas, and discovery using AWS Glue Data Catalog.
Implement fine-grained data access and governance using AWS Lake Formation, KMS encryption, and SSL.
Build and operate data services on EKS (Kubernetes).
Work with the Hadoop ecosystem (Spark, Hive, HDFS) using partitioning, bucketing, and columnar formats (Parquet, ORC).
Troubleshoot and resolve complex big data issues across pipelines, clusters, and queries.
Design, implement, and maintain CI/CD pipelines using Jenkins or similar tools.
Monitor and observe pipelines and clusters using CloudWatch and Grafana.
Prepare high-quality datasets for AI/ML use cases.
Build, configure, and operate an MCP server for AI/ML integration.
Collaborate in Scrum teams; proactively identify gaps and propose out-of-the-box, scalable solutions. Qualifications
Qualifications
Required:
8+ years of experience in Big Data / Data Engineering.
Strong hands-on experience with AWS (S3, Glue Data Catalog, Lake Formation, Athena, EMR, EKS).
Proven experience with Trino (or Presto) and optimizing query performance.
Working knowledge of Kubernetes / EKS for data workloads.
Strong SQL, Python, and shell scripting skills.
Experience with CI/CD pipelines and Jenkins.
Experience building and configuring MCP server for AI/ML integration.
Ownership mindset — problem solver, not a task executor.
Preferred:
AI/ML data pipeline exposure
Cloud-native data modernization experience
Candidates are required to be living in the MVD area at the time of the interview.
Job Type: Contract
Pay: $50.00 - $65.00 per hour
Work Location: In person
Responsibilities
Design and build scalable data lakes and pipelines on AWS using cloud-native and automated solutions.
Enable fast, federated analytics using Amazon Athena and Trino, with performance tuning for large-scale queries.
Manage metadata, schemas, and discovery using AWS Glue Data Catalog.
Implement fine-grained data access and governance using AWS Lake Formation, KMS encryption, and SSL.
Build and operate data services on EKS (Kubernetes).
Work with the Hadoop ecosystem (Spark, Hive, HDFS) using partitioning, bucketing, and columnar formats (Parquet, ORC).
Troubleshoot and resolve complex big data issues across pipelines, clusters, and queries.
Design, implement, and maintain CI/CD pipelines using Jenkins or similar tools.
Monitor and observe pipelines and clusters using CloudWatch and Grafana.
Prepare high-quality datasets for AI/ML use cases.
Build, configure, and operate an MCP server for AI/ML integration.
Collaborate in Scrum teams; proactively identify gaps and propose out-of-the-box, scalable solutions. Qualifications
Qualifications
Required:
8+ years of experience in Big Data / Data Engineering.
Strong hands-on experience with AWS (S3, Glue Data Catalog, Lake Formation, Athena, EMR, EKS).
Proven experience with Trino (or Presto) and optimizing query performance.
Working knowledge of Kubernetes / EKS for data workloads.
Strong SQL, Python, and shell scripting skills.
Experience with CI/CD pipelines and Jenkins.
Experience building and configuring MCP server for AI/ML integration.
Ownership mindset — problem solver, not a task executor.
Preferred:
AI/ML data pipeline exposure
Cloud-native data modernization experience
Candidates are required to be living in the MVD area at the time of the interview.
Job Type: Contract
Pay: $50.00 - $65.00 per hour
Work Location: In person






