Sibitalent Corp

Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer specializing in GCP, located onsite in Dallas, TX. The long-term contract requires expertise in BigQuery, Airflow, Python, and data lake engineering, along with experience in ETL/ELT pipelines and data ingestion.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
February 25, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Texas, United States
-
🧠 - Skills detailed
#Data Pipeline #"ETL (Extract #Transform #Load)" #Datasets #Data Lake #Spark (Apache Spark) #Airflow #Data Engineering #Batch #Python #BigQuery #Dataflow #Clustering #Data Processing #GCP (Google Cloud Platform) #Cloud #Data Ingestion #Storage #Hadoop
Role description
Job Title: Data Engineer - GCP Location: Dallas, TX (Onsite) Duration: Long Term Contract Job description:- Key Responsibilities 1. Data Lake Engineering & Storage • Develop and maintain multi-layered data lake structures (Bronze/Silver/Gold) • Design GCS buckets, lifecycle policies, naming conventions, and access configurations • Work with columnar formats such as Parquet, Avro, ORC • Implement partitioning, clustering, and optimized data organization • Build analytics-friendly data models and curated datasets 1. Data Ingestion & Orchestration • Build batch and streaming pipelines using Dataflow, Pub/Sub, Dataproc, BigQuery • Implement CDC, incremental loads, and deduplication logic • Set up Airflow/Cloud Composer pipelines for orchestration • Build robust error-handling, replay, and backfill mechanisms 1. Data Processing & Transformation • Develop ETL/ELT data pipelines using Dataflow (Beam) or Spark • Write optimized BigQuery SOL (partitioning, clustering, cost controls) • Manage schema evolution with minimal downstream disruption • Write clean, modular Python code with appropriate test coverage • Utilize Hadoop ecosystem tools when required 1. Analytics & Data Serving • Optimize BigQuery tables for cost and performance • Build semantic layers and standardized metric definitions • Expose data via views, curated datasets, or APIs • Partner with Bl teams to support dashboard and reporting needs