

Collabera
Hadoop Python Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Hadoop Python Data Engineer, located in Chicago, Denver, or Washington, DC, for an 18+ month contract at $63-$64 per hour. Requires expertise in Python, Hadoop, data pipeline engineering, and banking industry experience.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
520
-
ποΈ - Date
February 18, 2026
π - Duration
More than 6 months
-
ποΈ - Location
On-site
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Washington, DC
-
π§ - Skills detailed
#Datasets #Programming #Kafka (Apache Kafka) #"ETL (Extract #Transform #Load)" #YARN (Yet Another Resource Negotiator) #Python #Scala #Data Engineering #Data Processing #HDFS (Hadoop Distributed File System) #PHP #Spark (Apache Spark) #Base #Hadoop #Batch #Data Pipeline #Data Ingestion #Data Integration
Role description
Job Title: Hadoop Data Engineer
Location: Chicago, IL/Denver, CO/Washington, DC
Work Arrangement: Fully Onsite
Client Industry: Enterprise BFSI client
Duration: 18+ Months Contract
Schedule: Monday to Friday, Standard working hours
About the Role:
Seeking a highly skilled Data Engineer (hadoop plus python) to join theΒ clientΒ in Chicago or Denver or Washington. This is a 100% onsite role supporting critical data ingestion, pipeline development, and performance optimization initiatives in a secure banking environment.
Top Three skills:
β’ Python-Based Data Pipeline Engineering
β’ Building production-grade pipelines
β’ Performance tuning, memory management, error handling
β’ Batch & near-real-time processing
β’ Hadoop Ecosystem Expertise
β’ Strong understanding of HDFS, YARN, MapReduce
β’ Distributed processing with Spark/Hive
β’ Scalable and reliable data workflows
β’ Custom Data Ingestion & Integration
β’ Designing ingestion from APIs, queues, DBs, files
β’ Kafka & message-driven architectures
β’ PHP-based data services & integration endpoints
What Weβre Looking For:
Β· Design, develop, and maintain data pipelines primarily using Python, writing production-grade code to ingest, process, and move large-scale datasets across distributed environments.
Β· Build and enhance data processing workflows on Hadoop-based platforms, leveraging a strong understanding of Hadoop architecture (HDFS, YARN, MapReduce) to ensure scalability, reliability, and performance.
Β· Develop custom data ingestion and transformation logic, rather than relying solely on prebuilt tools, to support batch and near-real-time processing requirements.
Β· Write, maintain, and optimize Python code for distributed data processing, including performance tuning, memory management, and error handling in large data environments.
Β· Contribute hands-on code in PHP to support data services, ingestion endpoints, or integration layers that interact with data pipelines or upstream/downstream system
Compensation:
Hourly Rate: $63 - $64 per hour
This range reflects base compensation and may vary based on location, market conditions, experience, and candidate qualifications.
Benefits:
The Company offers the following benefits for this position, subject to applicable eligibility requirements: medical insurance, dental insurance, vision insurance, 401(k) retirement plan, life insurance, long-term disability insurance, short-term disability insurance, paid parking/public transportation, (paid time , paid sick and safe time , hours of paid vacation time, weeks of paid parental leave, paid holidays annually - AS Applicable)
About Us:
At Collabera, we donβt just offer jobs—we build careers. As a global leader in talent solutions, we provide opportunities to work with top organizations, cutting-edge technologies, and dynamic teams. Our culture thrives on innovation, collaboration, and a commitment to excellence. With continuous learning, career growth, and a people-first approach, we empower you to achieve your full potential. Join us and be part of a company that values passion, integrity, and making an impact.
Ready to Apply?
Apply now on shivani.soni@collabera.com/973-841-2468
Note: we are looking for python programming with data integration rather than ETL Tooling.
Job Title: Hadoop Data Engineer
Location: Chicago, IL/Denver, CO/Washington, DC
Work Arrangement: Fully Onsite
Client Industry: Enterprise BFSI client
Duration: 18+ Months Contract
Schedule: Monday to Friday, Standard working hours
About the Role:
Seeking a highly skilled Data Engineer (hadoop plus python) to join theΒ clientΒ in Chicago or Denver or Washington. This is a 100% onsite role supporting critical data ingestion, pipeline development, and performance optimization initiatives in a secure banking environment.
Top Three skills:
β’ Python-Based Data Pipeline Engineering
β’ Building production-grade pipelines
β’ Performance tuning, memory management, error handling
β’ Batch & near-real-time processing
β’ Hadoop Ecosystem Expertise
β’ Strong understanding of HDFS, YARN, MapReduce
β’ Distributed processing with Spark/Hive
β’ Scalable and reliable data workflows
β’ Custom Data Ingestion & Integration
β’ Designing ingestion from APIs, queues, DBs, files
β’ Kafka & message-driven architectures
β’ PHP-based data services & integration endpoints
What Weβre Looking For:
Β· Design, develop, and maintain data pipelines primarily using Python, writing production-grade code to ingest, process, and move large-scale datasets across distributed environments.
Β· Build and enhance data processing workflows on Hadoop-based platforms, leveraging a strong understanding of Hadoop architecture (HDFS, YARN, MapReduce) to ensure scalability, reliability, and performance.
Β· Develop custom data ingestion and transformation logic, rather than relying solely on prebuilt tools, to support batch and near-real-time processing requirements.
Β· Write, maintain, and optimize Python code for distributed data processing, including performance tuning, memory management, and error handling in large data environments.
Β· Contribute hands-on code in PHP to support data services, ingestion endpoints, or integration layers that interact with data pipelines or upstream/downstream system
Compensation:
Hourly Rate: $63 - $64 per hour
This range reflects base compensation and may vary based on location, market conditions, experience, and candidate qualifications.
Benefits:
The Company offers the following benefits for this position, subject to applicable eligibility requirements: medical insurance, dental insurance, vision insurance, 401(k) retirement plan, life insurance, long-term disability insurance, short-term disability insurance, paid parking/public transportation, (paid time , paid sick and safe time , hours of paid vacation time, weeks of paid parental leave, paid holidays annually - AS Applicable)
About Us:
At Collabera, we donβt just offer jobs—we build careers. As a global leader in talent solutions, we provide opportunities to work with top organizations, cutting-edge technologies, and dynamic teams. Our culture thrives on innovation, collaboration, and a commitment to excellence. With continuous learning, career growth, and a people-first approach, we empower you to achieve your full potential. Join us and be part of a company that values passion, integrity, and making an impact.
Ready to Apply?
Apply now on shivani.soni@collabera.com/973-841-2468
Note: we are looking for python programming with data integration rather than ETL Tooling.






