

Aaura Softwares
Lead AWS Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead AWS Data Engineer in Houston, TX, on a long-term contract. Requires 12+ years of data engineering experience, 5+ years in a lead role, AWS expertise, Python/PySpark proficiency, and strong communication skills. AWS certifications preferred.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
November 13, 2025
π - Duration
Unknown
-
ποΈ - Location
On-site
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Houston, TX
-
π§ - Skills detailed
#Automated Testing #Data Governance #Cloud #Metadata #Apache Iceberg #Compliance #IAM (Identity and Access Management) #ML (Machine Learning) #Data Warehouse #S3 (Amazon Simple Storage Service) #Computer Science #Data Engineering #PySpark #Data Modeling #Scala #Workday #Data Storage #Storage #Documentation #Data Lake #Data Lakehouse #Airflow #Data Management #SQL (Structured Query Language) #AWS (Amazon Web Services) #DevOps #Athena #Oracle #"ETL (Extract #Transform #Load)" #Terraform #Python #Redshift #Datasets #SaaS (Software as a Service) #Data Pipeline #Security #Spark (Apache Spark) #Lambda (AWS Lambda) #Data Catalog #AWS Glue
Role description
Role: Lead AWS Data Engineer
Location: Houston, TX-Onsite
Duration: Long Term Contract
Minimum Requirements:
β’ Bachelorβs or Masterβs degree in Computer Science, Information Systems, Engineering, or a related discipline plus at least 12 years of hands-on data engineering experience, or demonstrated equivalency of experience and/or education
β’ 5+ years in a technical-lead or team-lead capacity delivering enterprise-grade solutions.
β’ Deep expertise in AWS data and analytics services: e.g.; S3, Glue, Redshift, Athena, EMR/Spark, Lambda, IAM, and Lake Formation.
β’ Proficiency in Python/PySpark or Scala for data engineering, along with advanced SQL for warehousing and analytics workloads.
β’ Demonstrated success designing and operating large-scale ELT/ETL pipelines, data lakes, and dimensional/columnar data warehouses.
β’ Experience with workflow orchestration (e.g.; Airflow, Step Functions) and modern DevOps practicesβCI/CD, automated testing, and infrastructure-as-code (e.g.; Terraform or CloudFormation).
β’ Experience with data lakehouse architecture and frameworks (e.g.; Apache Iceberg).
β’ Experience in integrating with enterprise (onprem, SaaS) systems (Oracle e-business, Salesforce, Workday)
β’ Strong communication, stakeholder-management, and documentation skills; aptitude for translating business needs into technical roadmaps.
Preferred Qualifications:
β’ Solid understanding of data modeling, data governance, security best practices (encryption, key management), and compliance requirements.
β’ Experience working within similarly large, complex organizations
β’ Experience building integrations for enterprise back-office applications
β’ AWS Certified Data Analytics β Specialty or AWS Solutions Architect certification (or equivalent) preferred; experience with other cloud platforms is a plus.
β’ Proficiency in modern data storage formats and table management systems, with a strong understanding of Apache Iceberg for managing large-scale datasets and Parquet for efficient, columnar data storage.
β’ In-depth knowledge of data cataloging, metadata management, and lineage tools (AWS Glue Data Catalog, Apache Atlas, Amundsen) to bolster data discovery and governance.
β’ Knowledge of how machine learning models are developed, trained, and deployed, as well as the ability to design data pipelines that support these processes.
β’ Experience migrating on-prem data sources onto AWS.
β’ Experience building high quality Data Products.
Role: Lead AWS Data Engineer
Location: Houston, TX-Onsite
Duration: Long Term Contract
Minimum Requirements:
β’ Bachelorβs or Masterβs degree in Computer Science, Information Systems, Engineering, or a related discipline plus at least 12 years of hands-on data engineering experience, or demonstrated equivalency of experience and/or education
β’ 5+ years in a technical-lead or team-lead capacity delivering enterprise-grade solutions.
β’ Deep expertise in AWS data and analytics services: e.g.; S3, Glue, Redshift, Athena, EMR/Spark, Lambda, IAM, and Lake Formation.
β’ Proficiency in Python/PySpark or Scala for data engineering, along with advanced SQL for warehousing and analytics workloads.
β’ Demonstrated success designing and operating large-scale ELT/ETL pipelines, data lakes, and dimensional/columnar data warehouses.
β’ Experience with workflow orchestration (e.g.; Airflow, Step Functions) and modern DevOps practicesβCI/CD, automated testing, and infrastructure-as-code (e.g.; Terraform or CloudFormation).
β’ Experience with data lakehouse architecture and frameworks (e.g.; Apache Iceberg).
β’ Experience in integrating with enterprise (onprem, SaaS) systems (Oracle e-business, Salesforce, Workday)
β’ Strong communication, stakeholder-management, and documentation skills; aptitude for translating business needs into technical roadmaps.
Preferred Qualifications:
β’ Solid understanding of data modeling, data governance, security best practices (encryption, key management), and compliance requirements.
β’ Experience working within similarly large, complex organizations
β’ Experience building integrations for enterprise back-office applications
β’ AWS Certified Data Analytics β Specialty or AWS Solutions Architect certification (or equivalent) preferred; experience with other cloud platforms is a plus.
β’ Proficiency in modern data storage formats and table management systems, with a strong understanding of Apache Iceberg for managing large-scale datasets and Parquet for efficient, columnar data storage.
β’ In-depth knowledge of data cataloging, metadata management, and lineage tools (AWS Glue Data Catalog, Apache Atlas, Amundsen) to bolster data discovery and governance.
β’ Knowledge of how machine learning models are developed, trained, and deployed, as well as the ability to design data pipelines that support these processes.
β’ Experience migrating on-prem data sources onto AWS.
β’ Experience building high quality Data Products.






