TechDoQuest

Senior Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer with a contract length of "unknown" and a pay rate of "$$$". The position requires 14+ years of experience in data projects, expertise in Hadoop, Spark, Kafka, and IBM DataStage, and banking domain experience. Only 1099 candidates accepted.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
May 15, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Unknown
-
📄 - Contract
1099 Contractor
-
🔒 - Security
Unknown
-
📍 - Location detailed
United States
-
🧠 - Skills detailed
#Monitoring #Airflow #Shell Scripting #DataStage #SQL (Structured Query Language) #Linux #Hadoop #Distributed Computing #Java #Data Engineering #Automation #Data Processing #Data Quality #Code Reviews #Scripting #Scrum #SQL Queries #Cloud #"ETL (Extract #Transform #Load)" #Big Data #Scala #Agile #Spark (Apache Spark) #HBase #Kafka (Apache Kafka) #Programming #Data Pipeline #Debugging #Batch #Python #Data Integration #Logging #Data Ingestion #Unix
Role description
Skills : Hadoop, Spark, Kafka, DataStage Domain : Banking is Must Only 1099 is accepted . Key Responsibilities: • Build and support real-time and batch-based data pipelines using Big Data distributed systems and streaming technologies. • Design, develop, test, and maintain scalable ETL and ELT pipelines for processing large volumes of structured and unstructured data. • Develop data ingestion, transformation, and orchestration workflows using ETL tools such as IBM DataStage and modern scheduling/orchestration platforms. • Write complex SQL queries, stored procedures, and optimization logic for high-performance data processing. • Work extensively in UNIX/Linux environments for scripting, job automation, file handling, monitoring, and troubleshooting. • Develop and maintain applications using programming languages such as Python and Java for automation, data processing, and integration tasks. • Monitor production data pipelines, troubleshoot failures, perform root cause analysis, and provide production support within SLA timelines. • Implement best practices for data quality, validation, reconciliation, logging, monitoring, and operational support. • Work with orchestration and scheduling tools such as Airflow, Control-M, Autosys, or equivalent workflow automation platforms. • Support cloud and Big Data initiatives involving Hadoop, Spark, Kafka, distributed processing systems, and real-time streaming frameworks. • Participate in code reviews, technical design discussions, and Agile/Scrum ceremonies. Required Qualifications: • Must have 14+ years work experience in successful delivery of complex data related projects end to end. • Experience with Big Data technologies such as Hadoop, Spark, Kafka, Hive, or distributed computing systems. • Strong hands-on experience in ETL development and enterprise data integration projects. • Expertise in IBM DataStage and data warehousing concepts. • Strong SQL development and query optimization experience. • Proficiency in UNIX/Linux shell scripting and system-level operations. • Strong programming experience in Python and/or Java. • Knowledge of real-time data pipeline development and streaming architectures. • Experience with scheduling and orchestration tools such as Airflow, Autosys, Control-M, or equivalent. • Experience supporting production environments including incident management, debugging, monitoring, and performance tuning. • Strong analytical, troubleshooting, and communication skills.