

Oremda Infotech Inc
Senior Data Engineer
โญ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer in Houston, Texas, on a long-term contract. Requires 7+ years of experience, including 3+ years with Snowflake. Key skills include Snowflake SQL, Python, dbt, and Airflow. Local candidates only.
๐ - Country
United States
๐ฑ - Currency
$ USD
-
๐ฐ - Day rate
512
-
๐๏ธ - Date
November 8, 2025
๐ - Duration
Unknown
-
๐๏ธ - Location
Hybrid
-
๐ - Contract
W2 Contractor
-
๐ - Security
Unknown
-
๐ - Location detailed
Houston, TX 77001
-
๐ง - Skills detailed
#S3 (Amazon Simple Storage Service) #Palantir Foundry #Spark (Apache Spark) #SQL (Structured Query Language) #Streamlit #Observability #UAT (User Acceptance Testing) #Datasets #Data Extraction #Cloudera #SnowPipe #Trino #Data Quality #Python #dbt (data build tool) #Microsoft Power BI #Batch #Clustering #Documentation #Storage #Cloud #Airflow #Version Control #GitLab #Snowpark #ADLS (Azure Data Lake Storage) #"ETL (Extract #Transform #Load)" #BI (Business Intelligence) #Security #Snowflake #Data Engineering #GitHub #Macros #Kubernetes
Role description
Job Title:- Senior Data Engineer
Location:- Houston, Texas (On-Site/Hybrid)
Job Type:- Long Term Contract
Must be Local to Houston, TX
Only GC, USC, H4EAD Candidates on W2 or 1099 will be entertained
Overview:
Delivers the Palantir Foundry exit on a modern Snowflake stack by building reliable, performant, and testable ELT pipelines; recreates Foundry transformations and rule-based event logic; and ensures historical data extraction, reconciliation, and cutover readiness.
Years of Experience:
7+ years overall; 3+ years hands-on with Snowflake.
Key Responsibilities:Extract historical datasets from Palantir (dataset export, parquet) to S3/ADLS and load into Snowflake; implement checksum and reconciliation controls.
Rebuild Foundry transformations as dbt models and/or Snowflake SQL; implement curated schemas and incremental patterns using Streams and Tasks.
Implement the batch event/rules engine that evaluates time-series plus reference data on a schedule (e.g., 30รขโฌโ60 minutes) and produces auditable event tables.
Configure orchestration in Airflow running on AKS and, where appropriate, Snowflake Tasks; monitor, alert, and document operational runbooks.
Optimize warehouses, queries, clustering, and caching; manage cost with Resource Monitors and usage telemetry.
Author automated tests (dbt tests, Great Expectations or equivalent), validate parity versus legacy outputs, and support UAT and cutover.
Collaborate with BI/analytics teams (Sigma, Power BI) on dataset contracts, performance, and security requirements.
Required Qualifications:Strong Snowflake SQL and Python for ELT, utilities, and data validation.
Production experience with dbt (models, tests, macros, documentation, lineage).
Orchestration with Airflow (preferably on AKS/Kubernetes) and use of Snowflake Tasks/Streams for incrementals.
Proficiency with cloud object storage (S3/ADLS), file formats (Parquet/CSV), and bulk/incremental load patterns (Snowpipe, External Tables).
Version control and CI/CD with GitHub/GitLab; environment promotion and release hygiene.
Data quality and reconciliation fundamentals, including checksums, row/aggregate parity, and schema integrity tests.
Performance and cost tuning using query profiles, micro-partitioning behavior, and warehouse sizing policies.
Preferred Qualifications:Experience migrating from legacy platforms (Palantir Foundry, Cloudera/Hive/Spark) and familiarity with Trino/Starburst federation patterns.
Time-series data handling and rules/pattern detection; exposure to Snowpark or UDFs for complex transforms.
Familiarity with consumption patterns in Sigma and Power BI (Import, DirectQuery, composite models, RLS/OLS considerations).
Security and governance in Snowflake (RBAC, masking, row/column policies), tagging, and cost allocation.
Exposure to containerized workloads on AKS, lightweight apps for surfacing data (e.g., Streamlit), and basic observability practices.
Job Title:- Senior Data Engineer
Location:- Houston, Texas (On-Site/Hybrid)
Job Type:- Long Term Contract
Must be Local to Houston, TX
Only GC, USC, H4EAD Candidates on W2 or 1099 will be entertained
Overview:
Delivers the Palantir Foundry exit on a modern Snowflake stack by building reliable, performant, and testable ELT pipelines; recreates Foundry transformations and rule-based event logic; and ensures historical data extraction, reconciliation, and cutover readiness.
Years of Experience:
7+ years overall; 3+ years hands-on with Snowflake.
Key Responsibilities:Extract historical datasets from Palantir (dataset export, parquet) to S3/ADLS and load into Snowflake; implement checksum and reconciliation controls.
Rebuild Foundry transformations as dbt models and/or Snowflake SQL; implement curated schemas and incremental patterns using Streams and Tasks.
Implement the batch event/rules engine that evaluates time-series plus reference data on a schedule (e.g., 30รขโฌโ60 minutes) and produces auditable event tables.
Configure orchestration in Airflow running on AKS and, where appropriate, Snowflake Tasks; monitor, alert, and document operational runbooks.
Optimize warehouses, queries, clustering, and caching; manage cost with Resource Monitors and usage telemetry.
Author automated tests (dbt tests, Great Expectations or equivalent), validate parity versus legacy outputs, and support UAT and cutover.
Collaborate with BI/analytics teams (Sigma, Power BI) on dataset contracts, performance, and security requirements.
Required Qualifications:Strong Snowflake SQL and Python for ELT, utilities, and data validation.
Production experience with dbt (models, tests, macros, documentation, lineage).
Orchestration with Airflow (preferably on AKS/Kubernetes) and use of Snowflake Tasks/Streams for incrementals.
Proficiency with cloud object storage (S3/ADLS), file formats (Parquet/CSV), and bulk/incremental load patterns (Snowpipe, External Tables).
Version control and CI/CD with GitHub/GitLab; environment promotion and release hygiene.
Data quality and reconciliation fundamentals, including checksums, row/aggregate parity, and schema integrity tests.
Performance and cost tuning using query profiles, micro-partitioning behavior, and warehouse sizing policies.
Preferred Qualifications:Experience migrating from legacy platforms (Palantir Foundry, Cloudera/Hive/Spark) and familiarity with Trino/Starburst federation patterns.
Time-series data handling and rules/pattern detection; exposure to Snowpark or UDFs for complex transforms.
Familiarity with consumption patterns in Sigma and Power BI (Import, DirectQuery, composite models, RLS/OLS considerations).
Security and governance in Snowflake (RBAC, masking, row/column policies), tagging, and cost allocation.
Exposure to containerized workloads on AKS, lightweight apps for surfacing data (e.g., Streamlit), and basic observability practices.






