United Software Group Inc

Hadoop Cloudera Engineer

โญ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Hadoop Cloudera Engineer in Irving, TX (hybrid, 3 days onsite). It is a full-time position with a duration of over 6 months, requiring expertise in Hadoop, Spark, ETL development, and certifications like CCA Data Analyst or CCP Data Engineer.
๐ŸŒŽ - Country
United States
๐Ÿ’ฑ - Currency
$ USD
-
๐Ÿ’ฐ - Day rate
Unknown
-
๐Ÿ—“๏ธ - Date
May 13, 2026
๐Ÿ•’ - Duration
More than 6 months
-
๐Ÿ๏ธ - Location
Hybrid
-
๐Ÿ“„ - Contract
Unknown
-
๐Ÿ”’ - Security
Unknown
-
๐Ÿ“ - Location detailed
Irving, TX
-
๐Ÿง  - Skills detailed
#Version Control #Data Analysis #GIT #HDFS (Hadoop Distributed File System) #Impala #Python #HBase #ML (Machine Learning) #Security #Storage #Big Data #JSON (JavaScript Object Notation) #Docker #Spark (Apache Spark) #Scala #Kafka (Apache Kafka) #Programming #"ETL (Extract #Transform #Load)" #Kubernetes #Sqoop (Apache Sqoop) #Data Processing #DevOps #GCP (Google Cloud Platform) #Data Governance #Cloud #Data Storage #AWS (Amazon Web Services) #Data Security #Hadoop #Azure #SQL (Structured Query Language) #YARN (Yet Another Resource Negotiator) #Data Ingestion #Datasets #AI (Artificial Intelligence) #PySpark #Cloudera #Data Engineering
Role description
Job Title: Hadoop Cloudera Engineer Location : Irving, TX (Hybrid Onsite Role-3 Days Onsite Per Week) Duration : Fulltime Teams Meeting Interview Job Description: Skills : โ€ข Hadoop & Cloudera , Ecosystem (CDP/CDH, HDFS, YARN, MapReduce) โ€ข Spark & Data Processing(Spark Core, SQL, PySpark) โ€ข Version Control (Git) โ€ข ETL Pipeline Development &Optimization โ€ข Cloud & DevOps (CI/CD, Docker, Kubernetes, CDP (Cloud) โ€ข Big Data Tools (Hive,Impala, HBase) โ€ข Programming (Python / Scala / Shell) โ€ข Data Storage & File Formats (Parquet, Avro,ORC, JSON) โ€ข Cluster Management & Performance Tuning โ€ข Data Ingestion Tools (Sqoop, Flume, Kafka) โ€ข Experience with real-time processing (Spark Streaming / Kafka) โ€ข Knowledge of data governance and security frameworks โ€ข Exposure to cloud platforms (Azure / AWS / GCP) ยท Basic understanding of data warehousing concepts โ€ข Develop and maintain big data solutions using Cloudera Hadoop platform โ€ข Design and implement scalable ETL pipelines for large datasets โ€ข Work on data ingestion, processing, and transformation using Spark and Hadoop tools ยท Optimize data workflows, query performance, and storage strategies โ€ข Manage and maintain Hadoop clusters (Cloudera distribution) โ€ข Work with structured and unstructured data across multiple sources โ€ข Collaborate with Data Engineers, Analysts, and Business teams โ€ข Implement data security, governance, and access controls โ€ข Ensure system performance, reliability, and scalabilityCollaborate with cross-functional teams (Data Engineers, DevOps, Business teams) โ€ข Stay updated with the latest advancements in AI/ML and GenAI ecosystem Certification โ€ข Cloudera Certified Associate (CCA) Data Analyst/ Cloudera Certified Professional (CCP) Data Engineer (highly preferred) Cloudera Certified Administrator (CCA Admin)