

Cozen Technology Solutions Inc
GCP Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a GCP Data Engineer in Phoenix, AZ, with a contract length of "unknown." The pay rate is "unknown." Key skills include GCP, Dataflow, Pub/Sub, BigQuery, and Cloud Composer. Experience in marketing technology is preferred.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
March 18, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Phoenix, AZ
-
🧠 - Skills detailed
#BigQuery #Data Security #GCP (Google Cloud Platform) #Datasets #Scala #Batch #Programming #Security #Cloud #Apache Beam #SQL Queries #Airflow #Compliance #Dataflow #Python #"ETL (Extract #Transform #Load)" #Customer Segmentation #SQL (Structured Query Language) #Apache Airflow #Data Engineering #Data Processing #Data Pipeline
Role description
Job Title: GCP Data Engineer
Location: Phoenix, AZ, USA
Job Description:
We are seeking a Senior Data Engineer with strong hands-on experience in GCP to design, develop, and operate batch and real-time data pipelines for marketing technology use cases. The ideal candidate will have deep practical experience building Dataflow pipelines, working with Pub/Sub for real-time streaming, and orchestrating workflows using Cloud Composer (Apache Airflow).
This role requires a strong engineering mindset, real-world pipeline design experience, and the ability to work across projects and teams to deliver scalable, reliable data solutions.
Key Responsibilities
Data Pipeline Development
• Design, develop, and maintain Dataflow (Apache Beam) pipelines for both batch and real-time data processing.
• Implement Pub/Sub-based streaming architectures, including listening to source topics, transforming messages, and publishing to downstream topics.
• Build cross-project data pipelines in GCP, ensuring secure and reliable data exchange between projects.
Data Transformation & Processing
• Perform complex transformations using BigQuery, including:
• Advanced SQL queries
• Joins across large datasets
• Window functions and aggregations
• Implement encryption and decryption mechanisms as part of data processing workflows.
• Enrich streaming data by querying BigQuery or reference datasets within Dataflow pipelines.
Workflow Orchestration
• Develop and manage Airflow DAGs using Cloud Composer.
• Design DAGs to orchestrate batch and streaming pipelines.
• Implement DAG-to-DAG triggering patterns and dependency management.
• Monitor, troubleshoot, and optimize pipeline execution and scheduling.
Real-Time Processing
• Build and enhance real-time data processing components using Pub/Sub and Dataflow.
• Handle late-arriving data, retries, error handling, and idempotency in streaming pipelines.
• Ensure low-latency and high-throughput data processing for marketing use cases.
Platform & Integration
• Integrate GCP data pipelines with Adobe Experience Platform (AEP) and related Adobe services.
• Support marketing use cases such as customer segmentation, decisioning, and personalization.
• Collaborate with analytics, marketing, and platform teams to align data solutions with business needs.
Required Technical Skills
Google Cloud Platform (GCP)
• Strong hands-on experience with:
• Cloud Dataflow (Apache Beam) – streaming and batch pipelines
• Cloud Pub/Sub – message publishing, subscriptions, and streaming patterns
• BigQuery – advanced SQL, performance optimization, partitioning
• Cloud Composer (Apache Airflow) – DAG development and orchestration
Programming & Querying
• Python (advanced):
• Writing and maintaining Airflow DAGs
• Data processing logic and pipeline components
• SQL (advanced):
• Complex joins
• Window functions
• Analytical queries on large datasets
Streaming & Pipeline Design
• Practical experience designing real-time data processing architectures.
• Understanding of:
• Event-driven architectures
• Exactly-once / at-least-once processing semantics
• Error handling and replay strategies
Preferred / Nice-to-Have Skills
• Experience with Adobe Experience Platform (AEP), Adobe Segmentation, or Adobe Decisioning.
• Familiarity with marketing data models (customer profiles, events, attributes).
• Knowledge of data security and compliance, including encryption and access controls.
• Experience working in multi-project GCP environments.
• Exposure to CI/CD practices for data pipelines.
Typical Interview Focus Areas (Scenario-Based)
Candidates should be able to clearly explain and demonstrate real-world experience with scenarios such as:
• Designing a Dataflow pipeline that listens to messages from one Pub/Sub topic and publishes transformed messages to another topic.
• Orchestrating batch and streaming pipelines using Airflow DAGs.
• Triggering one Airflow DAG from another and managing dependencies.
• Handling real-time data processing challenges, such as late data, failures, retries, and scaling.
• Designing secure and scalable cross-project GCP pipelines.
Ideal Candidate Profile
• Strong hands-on implementation experience with GCP data services.
• Proven track record of building production-grade batch and real-time pipelines.
• Comfortable discussing architecture, trade-offs, and design decisions.
• Ability to translate business and marketing requirements into robust data solutions.
Priority will be given to candidates with demonstrated, real-world experience implementing Dataflow and Pub/Sub-based streaming pipelines in GCP, especially within marketing or customer data platforms.
Job Title: GCP Data Engineer
Location: Phoenix, AZ, USA
Job Description:
We are seeking a Senior Data Engineer with strong hands-on experience in GCP to design, develop, and operate batch and real-time data pipelines for marketing technology use cases. The ideal candidate will have deep practical experience building Dataflow pipelines, working with Pub/Sub for real-time streaming, and orchestrating workflows using Cloud Composer (Apache Airflow).
This role requires a strong engineering mindset, real-world pipeline design experience, and the ability to work across projects and teams to deliver scalable, reliable data solutions.
Key Responsibilities
Data Pipeline Development
• Design, develop, and maintain Dataflow (Apache Beam) pipelines for both batch and real-time data processing.
• Implement Pub/Sub-based streaming architectures, including listening to source topics, transforming messages, and publishing to downstream topics.
• Build cross-project data pipelines in GCP, ensuring secure and reliable data exchange between projects.
Data Transformation & Processing
• Perform complex transformations using BigQuery, including:
• Advanced SQL queries
• Joins across large datasets
• Window functions and aggregations
• Implement encryption and decryption mechanisms as part of data processing workflows.
• Enrich streaming data by querying BigQuery or reference datasets within Dataflow pipelines.
Workflow Orchestration
• Develop and manage Airflow DAGs using Cloud Composer.
• Design DAGs to orchestrate batch and streaming pipelines.
• Implement DAG-to-DAG triggering patterns and dependency management.
• Monitor, troubleshoot, and optimize pipeline execution and scheduling.
Real-Time Processing
• Build and enhance real-time data processing components using Pub/Sub and Dataflow.
• Handle late-arriving data, retries, error handling, and idempotency in streaming pipelines.
• Ensure low-latency and high-throughput data processing for marketing use cases.
Platform & Integration
• Integrate GCP data pipelines with Adobe Experience Platform (AEP) and related Adobe services.
• Support marketing use cases such as customer segmentation, decisioning, and personalization.
• Collaborate with analytics, marketing, and platform teams to align data solutions with business needs.
Required Technical Skills
Google Cloud Platform (GCP)
• Strong hands-on experience with:
• Cloud Dataflow (Apache Beam) – streaming and batch pipelines
• Cloud Pub/Sub – message publishing, subscriptions, and streaming patterns
• BigQuery – advanced SQL, performance optimization, partitioning
• Cloud Composer (Apache Airflow) – DAG development and orchestration
Programming & Querying
• Python (advanced):
• Writing and maintaining Airflow DAGs
• Data processing logic and pipeline components
• SQL (advanced):
• Complex joins
• Window functions
• Analytical queries on large datasets
Streaming & Pipeline Design
• Practical experience designing real-time data processing architectures.
• Understanding of:
• Event-driven architectures
• Exactly-once / at-least-once processing semantics
• Error handling and replay strategies
Preferred / Nice-to-Have Skills
• Experience with Adobe Experience Platform (AEP), Adobe Segmentation, or Adobe Decisioning.
• Familiarity with marketing data models (customer profiles, events, attributes).
• Knowledge of data security and compliance, including encryption and access controls.
• Experience working in multi-project GCP environments.
• Exposure to CI/CD practices for data pipelines.
Typical Interview Focus Areas (Scenario-Based)
Candidates should be able to clearly explain and demonstrate real-world experience with scenarios such as:
• Designing a Dataflow pipeline that listens to messages from one Pub/Sub topic and publishes transformed messages to another topic.
• Orchestrating batch and streaming pipelines using Airflow DAGs.
• Triggering one Airflow DAG from another and managing dependencies.
• Handling real-time data processing challenges, such as late data, failures, retries, and scaling.
• Designing secure and scalable cross-project GCP pipelines.
Ideal Candidate Profile
• Strong hands-on implementation experience with GCP data services.
• Proven track record of building production-grade batch and real-time pipelines.
• Comfortable discussing architecture, trade-offs, and design decisions.
• Ability to translate business and marketing requirements into robust data solutions.
Priority will be given to candidates with demonstrated, real-world experience implementing Dataflow and Pub/Sub-based streaming pipelines in GCP, especially within marketing or customer data platforms.






