

Sibitalent Corp
Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer specializing in GCP, located onsite in Dallas, TX. The long-term contract requires expertise in BigQuery, Airflow, Python, and data lake engineering, along with experience in ETL/ELT pipelines and data ingestion.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
February 25, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Texas, United States
-
🧠 - Skills detailed
#Data Pipeline #"ETL (Extract #Transform #Load)" #Datasets #Data Lake #Spark (Apache Spark) #Airflow #Data Engineering #Batch #Python #BigQuery #Dataflow #Clustering #Data Processing #GCP (Google Cloud Platform) #Cloud #Data Ingestion #Storage #Hadoop
Role description
Job Title: Data Engineer - GCP
Location: Dallas, TX (Onsite)
Duration: Long Term Contract
Job description:-
Key Responsibilities
1. Data Lake Engineering & Storage
• Develop and maintain multi-layered data lake structures (Bronze/Silver/Gold)
• Design GCS buckets, lifecycle policies, naming conventions, and access configurations
• Work with columnar formats such as Parquet, Avro, ORC
• Implement partitioning, clustering, and optimized data organization
• Build analytics-friendly data models and curated datasets
1. Data Ingestion & Orchestration
• Build batch and streaming pipelines using Dataflow, Pub/Sub, Dataproc, BigQuery
• Implement CDC, incremental loads, and deduplication logic
• Set up Airflow/Cloud Composer pipelines for orchestration
• Build robust error-handling, replay, and backfill mechanisms
1. Data Processing & Transformation
• Develop ETL/ELT data pipelines using Dataflow (Beam) or Spark
• Write optimized BigQuery SOL (partitioning, clustering, cost controls)
• Manage schema evolution with minimal downstream disruption
• Write clean, modular Python code with appropriate test coverage
• Utilize Hadoop ecosystem tools when required
1. Analytics & Data Serving
• Optimize BigQuery tables for cost and performance
• Build semantic layers and standardized metric definitions
• Expose data via views, curated datasets, or APIs
• Partner with Bl teams to support dashboard and reporting needs
Job Title: Data Engineer - GCP
Location: Dallas, TX (Onsite)
Duration: Long Term Contract
Job description:-
Key Responsibilities
1. Data Lake Engineering & Storage
• Develop and maintain multi-layered data lake structures (Bronze/Silver/Gold)
• Design GCS buckets, lifecycle policies, naming conventions, and access configurations
• Work with columnar formats such as Parquet, Avro, ORC
• Implement partitioning, clustering, and optimized data organization
• Build analytics-friendly data models and curated datasets
1. Data Ingestion & Orchestration
• Build batch and streaming pipelines using Dataflow, Pub/Sub, Dataproc, BigQuery
• Implement CDC, incremental loads, and deduplication logic
• Set up Airflow/Cloud Composer pipelines for orchestration
• Build robust error-handling, replay, and backfill mechanisms
1. Data Processing & Transformation
• Develop ETL/ELT data pipelines using Dataflow (Beam) or Spark
• Write optimized BigQuery SOL (partitioning, clustering, cost controls)
• Manage schema evolution with minimal downstream disruption
• Write clean, modular Python code with appropriate test coverage
• Utilize Hadoop ecosystem tools when required
1. Analytics & Data Serving
• Optimize BigQuery tables for cost and performance
• Build semantic layers and standardized metric definitions
• Expose data via views, curated datasets, or APIs
• Partner with Bl teams to support dashboard and reporting needs





