

Extend Information Systems Inc.
Big Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Big Data Engineer in Phoenix, AZ, on a long-term contract. Requires 7+ years in Big Data, strong expertise in Hadoop, Spark, GCP, and proficiency in Python/Scala/Java. Hybrid work arrangement.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
480
-
🗓️ - Date
November 20, 2025
🕒 - Duration
Unknown
-
🏝️ - Location
Hybrid
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Phoenix, AZ
-
🧠 - Skills detailed
#Docker #GCP (Google Cloud Platform) #Deployment #Scala #"ETL (Extract #Transform #Load)" #Dataflow #Apache Spark #Kubernetes #Python #Kafka (Apache Kafka) #Terraform #Storage #Data Science #Data Processing #Security #Data Engineering #Data Modeling #Data Governance #Pig #Data Quality #SQL (Structured Query Language) #Airflow #BigQuery #Spark (Apache Spark) #Hadoop #Big Data #GIT #Sqoop (Apache Sqoop) #Cloud #Data Pipeline #Java #HDFS (Hadoop Distributed File System) #Data Cleansing
Role description
Job Title: Big Data Engineer
Location: Phoenix, AZ (Hybrid Onsite)
Duration: Long Term Contract
Job Description:
We are seeking an experienced Big Data Engineer with strong expertise in Hadoop, Spark, and Google Cloud Platform (GCP). The ideal candidate will design, develop, and optimize large-scale data processing pipelines and analytical solutions on the cloud.
Responsibilities:
• Design and implement data pipelines and ETL processes using Spark, Hadoop, and GCP services (BigQuery, Dataflow, Dataproc, Pub/Sub, Cloud Storage).
• Work with structured and unstructured data from multiple sources and perform data cleansing, transformation, and aggregation.
• Collaborate with data scientists, analysts, and application teams to deliver scalable data solutions.
• Optimize data performance and ensure reliability, availability, and scalability of data systems.
• Implement data governance, quality, and security best practices.
• Troubleshoot performance and data quality issues in distributed systems.
Required Skills:
• 7+ years of experience in Big Data technologies.
• Strong hands-on experience with Hadoop ecosystem (HDFS, Hive, Pig, Sqoop, Oozie).
• Expertise in Apache Spark (Core, SQL, Streaming).
• Strong experience with GCP data services – BigQuery, Dataflow, Dataproc, Composer, Cloud Storage, Pub/Sub.
• Proficiency in Python/Scala/Java for data processing.
• Good knowledge of SQL and data modeling concepts.
• Familiarity with CI/CD, Git, and Cloud Deployment tools.
Nice to Have:
• Experience with Airflow, Terraform, or Dataform.
• Knowledge of Kafka or real-time streaming.
• Familiarity with Docker/Kubernetes.
Job Title: Big Data Engineer
Location: Phoenix, AZ (Hybrid Onsite)
Duration: Long Term Contract
Job Description:
We are seeking an experienced Big Data Engineer with strong expertise in Hadoop, Spark, and Google Cloud Platform (GCP). The ideal candidate will design, develop, and optimize large-scale data processing pipelines and analytical solutions on the cloud.
Responsibilities:
• Design and implement data pipelines and ETL processes using Spark, Hadoop, and GCP services (BigQuery, Dataflow, Dataproc, Pub/Sub, Cloud Storage).
• Work with structured and unstructured data from multiple sources and perform data cleansing, transformation, and aggregation.
• Collaborate with data scientists, analysts, and application teams to deliver scalable data solutions.
• Optimize data performance and ensure reliability, availability, and scalability of data systems.
• Implement data governance, quality, and security best practices.
• Troubleshoot performance and data quality issues in distributed systems.
Required Skills:
• 7+ years of experience in Big Data technologies.
• Strong hands-on experience with Hadoop ecosystem (HDFS, Hive, Pig, Sqoop, Oozie).
• Expertise in Apache Spark (Core, SQL, Streaming).
• Strong experience with GCP data services – BigQuery, Dataflow, Dataproc, Composer, Cloud Storage, Pub/Sub.
• Proficiency in Python/Scala/Java for data processing.
• Good knowledge of SQL and data modeling concepts.
• Familiarity with CI/CD, Git, and Cloud Deployment tools.
Nice to Have:
• Experience with Airflow, Terraform, or Dataform.
• Knowledge of Kafka or real-time streaming.
• Familiarity with Docker/Kubernetes.






