

Data Lake AWS Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Lakehouse Engineer (AWS + Python + DevOps) on a long-term contract in Dallas, TX, requiring independent visa holders. Key skills include AWS Lake Formation, Python, PySpark, and DevOps tools.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
July 12, 2025
π - Project duration
Unknown
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Dallas, TX
-
π§ - Skills detailed
#Data Storage #Spark (Apache Spark) #DynamoDB #Python #Version Control #Data Governance #AWS Glue #Terraform #Batch #Scala #AWS IAM (AWS Identity and Access Management) #Programming #Infrastructure as Code (IaC) #Data Lake #PySpark #Data Engineering #AWS Lambda #Data Lakehouse #DevOps #S3 (Amazon Simple Storage Service) #Cloud #IAM (Identity and Access Management) #GitLab #Security #AWS (Amazon Web Services) #"ETL (Extract #Transform #Load)" #Lambda (AWS Lambda) #Storage #Data Processing
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Job Title: Data Lakehouse Engineer (AWS + Python + DevOps)
Location: Dallas, TX - 1 week onsite every quarter
Type: Long Term Contract
Note: Looking for independent visa holders
Job Summary:
We are seeking a highly skilled Data Lakehouse Engineer to join our team and help solve critical business and technology challenges. You will play a key role in building and maintaining a scalable Data Lakehouse solution to ensure the right data reaches the right users at the right time β empowering both business and technical teams with trusted, governed data.
Key Responsibilities
β’ Design and implement a modern Data Lakehouse architecture using AWS native services.
β’ Develop and maintain ETL pipelines using AWS Glue, Lambda, Step Functions, and Python.
β’ Configure and manage AWS Lake Formation for secure, governed access to data in S3.
β’ Integrate and optimize data workflows using PySpark and serverless technologies.
β’ Build and deploy infrastructure as code using CloudFormation, Terraform, Stacker, and the Serverless Framework.
β’ Collaborate with DevOps teams using GitLab for version control and CI/CD pipelines.
β’ Work with IAM for access control and DynamoDB for data storage requirements.
Technical Skills Required
Cloud & Data Engineering (AWS):
β’ AWS Lake Formation
β’ S3
β’ AWS Glue (Crawler, Catalog, Registry, Glue Jobs)
β’ AWS Step Functions
β’ AWS Lambda
β’ AWS IAM
β’ DynamoDB
Programming & Data Processing:
β’ Python (AWS-specific development)
β’ PySpark
DevOps & Infrastructure as Code:
β’ GitLab (CI/CD)
β’ Serverless Framework
β’ Stacker
β’ CloudFormation
β’ Terraform
Preferred Qualifications
β’ Proven experience in building data lakehouse or lake-based data platforms on AWS.
β’ Strong knowledge of data governance and access management using AWS Lake Formation.
β’ Hands-on experience with real-time and batch data processing.
β’ Familiarity with best practices in cloud security, DevOps, and CI/CD pipelines.
β’ Excellent problem-solving, communication, and collaboration skills.