Aaura Softwares

Lead AWS Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead AWS Data Engineer in Houston, TX, on a long-term contract. Requires 12+ years of data engineering experience, 5+ years in a lead role, AWS expertise, Python/PySpark proficiency, and strong communication skills. AWS certifications preferred.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
November 13, 2025
πŸ•’ - Duration
Unknown
-
🏝️ - Location
On-site
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
Houston, TX
-
🧠 - Skills detailed
#Automated Testing #Data Governance #Cloud #Metadata #Apache Iceberg #Compliance #IAM (Identity and Access Management) #ML (Machine Learning) #Data Warehouse #S3 (Amazon Simple Storage Service) #Computer Science #Data Engineering #PySpark #Data Modeling #Scala #Workday #Data Storage #Storage #Documentation #Data Lake #Data Lakehouse #Airflow #Data Management #SQL (Structured Query Language) #AWS (Amazon Web Services) #DevOps #Athena #Oracle #"ETL (Extract #Transform #Load)" #Terraform #Python #Redshift #Datasets #SaaS (Software as a Service) #Data Pipeline #Security #Spark (Apache Spark) #Lambda (AWS Lambda) #Data Catalog #AWS Glue
Role description
Role: Lead AWS Data Engineer Location: Houston, TX-Onsite Duration: Long Term Contract Minimum Requirements: β€’ Bachelor’s or Master’s degree in Computer Science, Information Systems, Engineering, or a related discipline plus at least 12 years of hands-on data engineering experience, or demonstrated equivalency of experience and/or education β€’ 5+ years in a technical-lead or team-lead capacity delivering enterprise-grade solutions. β€’ Deep expertise in AWS data and analytics services: e.g.; S3, Glue, Redshift, Athena, EMR/Spark, Lambda, IAM, and Lake Formation. β€’ Proficiency in Python/PySpark or Scala for data engineering, along with advanced SQL for warehousing and analytics workloads. β€’ Demonstrated success designing and operating large-scale ELT/ETL pipelines, data lakes, and dimensional/columnar data warehouses. β€’ Experience with workflow orchestration (e.g.; Airflow, Step Functions) and modern DevOps practicesβ€”CI/CD, automated testing, and infrastructure-as-code (e.g.; Terraform or CloudFormation). β€’ Experience with data lakehouse architecture and frameworks (e.g.; Apache Iceberg). β€’ Experience in integrating with enterprise (onprem, SaaS) systems (Oracle e-business, Salesforce, Workday) β€’ Strong communication, stakeholder-management, and documentation skills; aptitude for translating business needs into technical roadmaps. Preferred Qualifications: β€’ Solid understanding of data modeling, data governance, security best practices (encryption, key management), and compliance requirements. β€’ Experience working within similarly large, complex organizations β€’ Experience building integrations for enterprise back-office applications β€’ AWS Certified Data Analytics – Specialty or AWS Solutions Architect certification (or equivalent) preferred; experience with other cloud platforms is a plus. β€’ Proficiency in modern data storage formats and table management systems, with a strong understanding of Apache Iceberg for managing large-scale datasets and Parquet for efficient, columnar data storage. β€’ In-depth knowledge of data cataloging, metadata management, and lineage tools (AWS Glue Data Catalog, Apache Atlas, Amundsen) to bolster data discovery and governance. β€’ Knowledge of how machine learning models are developed, trained, and deployed, as well as the ability to design data pipelines that support these processes. β€’ Experience migrating on-prem data sources onto AWS. β€’ Experience building high quality Data Products.