Extend Information Systems Inc.

Big Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Big Data Engineer in Phoenix, AZ, on a long-term contract. Requires 7+ years in Big Data, strong expertise in Hadoop, Spark, GCP, and proficiency in Python/Scala/Java. Hybrid work arrangement.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
480
-
🗓️ - Date
November 20, 2025
🕒 - Duration
Unknown
-
🏝️ - Location
Hybrid
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Phoenix, AZ
-
🧠 - Skills detailed
#Docker #GCP (Google Cloud Platform) #Deployment #Scala #"ETL (Extract #Transform #Load)" #Dataflow #Apache Spark #Kubernetes #Python #Kafka (Apache Kafka) #Terraform #Storage #Data Science #Data Processing #Security #Data Engineering #Data Modeling #Data Governance #Pig #Data Quality #SQL (Structured Query Language) #Airflow #BigQuery #Spark (Apache Spark) #Hadoop #Big Data #GIT #Sqoop (Apache Sqoop) #Cloud #Data Pipeline #Java #HDFS (Hadoop Distributed File System) #Data Cleansing
Role description
Job Title: Big Data Engineer Location: Phoenix, AZ (Hybrid Onsite) Duration: Long Term Contract Job Description: We are seeking an experienced Big Data Engineer with strong expertise in Hadoop, Spark, and Google Cloud Platform (GCP). The ideal candidate will design, develop, and optimize large-scale data processing pipelines and analytical solutions on the cloud. Responsibilities: • Design and implement data pipelines and ETL processes using Spark, Hadoop, and GCP services (BigQuery, Dataflow, Dataproc, Pub/Sub, Cloud Storage). • Work with structured and unstructured data from multiple sources and perform data cleansing, transformation, and aggregation. • Collaborate with data scientists, analysts, and application teams to deliver scalable data solutions. • Optimize data performance and ensure reliability, availability, and scalability of data systems. • Implement data governance, quality, and security best practices. • Troubleshoot performance and data quality issues in distributed systems. Required Skills: • 7+ years of experience in Big Data technologies. • Strong hands-on experience with Hadoop ecosystem (HDFS, Hive, Pig, Sqoop, Oozie). • Expertise in Apache Spark (Core, SQL, Streaming). • Strong experience with GCP data services – BigQuery, Dataflow, Dataproc, Composer, Cloud Storage, Pub/Sub. • Proficiency in Python/Scala/Java for data processing. • Good knowledge of SQL and data modeling concepts. • Familiarity with CI/CD, Git, and Cloud Deployment tools. Nice to Have: • Experience with Airflow, Terraform, or Dataform. • Knowledge of Kafka or real-time streaming. • Familiarity with Docker/Kubernetes.