Python Developer - 100% Onsite in Houston

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Python Developer/Data Engineer on a 12-month contract, 100% onsite in Houston. Requires 5+ years of Python, DBT, data pipeline architecture, and Data Lakehouse technologies. Experience in midstream oil and gas is preferred.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
-
πŸ—“οΈ - Date discovered
August 28, 2025
πŸ•’ - Project duration
More than 6 months
-
🏝️ - Location type
On-site
-
πŸ“„ - Contract type
Fixed Term
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
Houston, TX
-
🧠 - Skills detailed
#Data Lakehouse #Delta Lake #Version Control #Dremio #Kubernetes #Data Quality #Apache Iceberg #Visualization #Airflow #Data Lake #Pytest #Data Vault #SQL (Structured Query Language) #Physical Data Model #Storage #S3 (Amazon Simple Storage Service) #dbt (data build tool) #Data Integration #Programming #Snowflake #Cloud #Data Engineering #Vault #AWS (Amazon Web Services) #NumPy #Apache Airflow #Data Pipeline #Python #Datasets #Pandas
Role description
We're looking for a Python Developer/Data Engineer to join a 12-month contract. 100% onsite in downtown Houston. The Python Developer/Data Engineer will work closely with business domain experts to create an Enterprise Data Lakehouse to support data analytic use cases for the midstream oil and gas operations, engineering, and measurements business units. Must Haves: β€’ Software development/software engineering experience β€’ 5+ years Python β€’ Data Build Tool (DBT) β€’ Knowledge of Data Lakehouse technologies, Apache Iceberg, or Delta Lake β€’ Working with S3 object storage Nice to Haves: β€’ Python UI development, Dash β€’ Dremio β€’ Kubernetes/AWS EKS β€’ AWS Cloud Responsibilities include: β€’ Design and implement reliable data pipelines to integrate disparate data sources into a single Data Lakehouse β€’ Design and implement data quality pipelines to ensure data correctness and building trusted datasets β€’ Design and implement a Data Lakehouse solution which accurately reflects business operations β€’ Assist with data platform performance tuning and physical data model support including partitioning and compaction β€’ Provide guidance in data visualizations and reporting efforts to ensure solutions are aligned to business objectives. Qualifications include: β€’ 5+ years Data Engineer designing and maintaining data pipeline architectures β€’ 5+ years of programming experience in Python and ANSI SQL β€’ 2+ years of development experience with DBT β€’ Various data modelling methods such as Star Schema, Snowflake, Data Vault design β€’ Implementing Data Lakehouse using a Medallion Architecture with Apache Iceberg on S3 Object Storage β€’ Various data integration patters including ELT, Pub/Sub, and Change Data Capture β€’ Common Python Data Engineering packages (Pandas, Numpy, Pyarrow, Pytest, Scikit-Learn, Boto3) β€’ Excellent communication and experience presenting complex concepts to technical and non-technical stakeholders β€’ Software development practices such as Design Principles and Patters, Testing, Refactoring, CI/CD, and version control. β€’ Dremio, Apache Airflow, and Airbyte preferred