Ehub Global Inc

Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer in New York City, NY, with a contract length of unspecified duration. The position requires expertise in Hadoop, Snowflake, Kafka, and Spark, along with 12+ years of experience in data engineering within the banking/finance domain.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
January 7, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
New York City Metropolitan Area
-
🧠 - Skills detailed
#Data Quality #Cloud #Snowflake #GCP (Google Cloud Platform) #Python #Data Lake #Java #Scripting #NiFi (Apache NiFi) #Airflow #Pig #Visualization #Kubernetes #Programming #Hadoop #SQL (Structured Query Language) #Storage #BigQuery #Data Processing #AWS (Amazon Web Services) #Data Storage #"ETL (Extract #Transform #Load)" #Tableau #Data Analysis #Scala #HDFS (Hadoop Distributed File System) #Spark (Apache Spark) #Docker #Automation #DevOps #Redshift #Security #Data Engineering #Microsoft Power BI #BI (Business Intelligence) #Azure #Datasets #Data Pipeline #Big Data #Batch #Kafka (Apache Kafka)
Role description
Role: Data Engineer Location: New York City, NY (Onsite) Domain: Banking/ Finance Overview We are seeking a highly skilled Data Engineer with deep expertise in Big Data technologies, data lakes, and modern analytics platforms. The ideal candidate will design, build, and optimize scalable data pipelines that support advanced analytics and business intelligence. This role requires strong hands-on experience with Hadoop ecosystems, Snowflake, Kafka, Spark, and other distributed data platforms. Responsibilities • Design, develop, and maintain data pipelines for ingesting, transforming, and delivering large-scale datasets. • Manage and optimize data lake architectures to ensure scalability, reliability, and performance. • Implement and support Hadoop-based solutions for distributed data processing. • Integrate and manage Snowflake for cloud-based data warehousing and analytics. • Build and maintain real-time streaming solutions using Kafka. • Develop and optimize Spark applications for batch and streaming workloads. • Collaborate with data analysts, scientists, and business stakeholders to deliver actionable insights. • Ensure data quality, governance, and security across all platforms. • Monitor and troubleshoot data pipelines to maintain high availability and performance. Skills & Qualifications Core Skills • Big Data Ecosystem: Hadoop (HDFS, Hive, Pig, MapReduce), Spark, Kafka. • Cloud Data Warehousing: Snowflake (preferred), Redshift, BigQuery. • Data Lake Management: Experience with large-scale data storage and retrieval. • Data Pipelines: ETL/ELT design, orchestration tools (Airflow, NiFi, etc.). • Programming & Scripting: Python, Scala, Java, SQL. • Data Analysis: Strong ability to query, analyze, and interpret large datasets. • Distributed Systems: Understanding of scalability, fault tolerance, and performance optimization. • DevOps & Automation: CI/CD pipelines, containerization (Docker, Kubernetes). • Visualization & BI Tools: Familiarity with Tableau, Power BI, or similar. Preferred Qualifications • 12+ years of experience in data engineering or big data roles. • Experience with cloud platforms (AWS, Azure, GCP). • Strong problem-solving and analytical mindset. • Excellent communication and collaboration skills.