

The Judge Group
Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer in Jersey City, requiring 2+ years of experience with Python, PySpark, and AWS. Contract length is unspecified, with a pay rate of "unknown". Local candidates only; no C2C or third parties.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
December 6, 2025
π - Duration
Unknown
-
ποΈ - Location
On-site
-
π - Contract
W2 Contractor
-
π - Security
Unknown
-
π - Location detailed
Jersey City, NJ
-
π§ - Skills detailed
#Python #Security #PySpark #S3 (Amazon Simple Storage Service) #Data Pipeline #Batch #SQL (Structured Query Language) #Data Vault #Data Analysis #Data Science #Code Reviews #Airflow #Cloud #Databricks #Snowflake #JSON (JavaScript Object Notation) #AI (Artificial Intelligence) #Redshift #MongoDB #Agile #Compliance #Scala #Data Governance #NoSQL #Data Lifecycle #Vault #Data Processing #Data Lakehouse #Storage #Scripting #Data Storage #Complex Queries #React #AWS (Amazon Web Services) #Lambda (AWS Lambda) #Spark (Apache Spark) #Data Warehouse #Hadoop #Physical Data Model #Data Access #Unix #Data Quality #Data Engineering #Data Lake #Oracle #Databases #Java
Role description
Location: Jersey City/ONSITE/ NO C2C/NO C2H/NO THIRD PARTIES/ONLY W2/
ONLY LOCALS TO NJ/NY - NO RELOCATION CANDIDATES
Skillset: Data Engineer
Must Haves: Python, PySpark, AWS β ECS, Glue, Lambda, S3
Nice to Haves: Java, Spark, React Js
Interview Process: Interview Process: 2 rounds, 2nd will be on site
Youβre ready to gain the skills and experience needed to grow within your role and advance your career β and we have the perfect software engineering opportunity for you.
As a Data Engineer III - Python / Spark / Data Lake at JPMorgan Chase within the Consumer and Community Bank , you will be a seasoned member of an agile team, tasked with designing and delivering reliable data collection, storage, access, and analytics solutions that are secure, stable, and scalable. Your responsibilities will include developing, testing, and maintaining essential data pipelines and architectures across diverse technical areas, supporting various business functions to achieve the firm's business objectives.
Job responsibilities:
β’ Supports review of controls to ensure sufficient protection of enterprise data.
β’ Advises and makes custom configuration changes in one to two tools to generate a product at the business or customer request.
β’ Updates logical or physical data models based on new use cases.
β’ Frequently uses SQL and understands NoSQL databases and their niche in the marketplace.
β’ Adds to team culture of diversity, opportunity, inclusion, and respect.
β’ Develop enterprise data models, Design/ develop/ maintain large-scale data processing pipelines (and infrastructure), Lead code reviews and provide mentoring thru the process, Drive data quality, Ensure data accessibility (to analysts and data scientists), Ensure compliance with data governance requirements, and Ensure business alignment (ensure data engineering practices align with business goals).
β’ Supports review of controls to ensure sufficient protection of enterprise data
Required qualifications, capabilities, and skills
β’ Formal training or certification on data engineering concepts and 2+ years applied experience
β’ Experience across the data lifecycle, advanced experience with SQL (e.g., joins and aggregations), and working understanding of NoSQL databases
β’ Experience with statistical data analysis and ability to determine appropriate tools and data patterns to perform analysis
β’ Extensive experience in AWS, design, implementation, and maintenance of data pipelines using Python and PySpark.
β’ Proficient in Python and PySpark, able to write and execute complex queries to perform curation and build views required by end users (single and multi-dimensional).
β’ Proven experience in performance and tuning to ensure jobs are running at optimal levels and no performance bottleneck.
β’ Advanced proficiency in leveraging Gen AI models from Anthropic (or OpenAI, or Google) using APIs/SDKs
β’ Advanced proficiency in cloud data lakehouse platform such as AWS data lake services, Databricks or Hadoop, relational data store such as Postgres, Oracle or similar, and at least one NOSQL data store such as Cassandra, Dynamo, MongoDB or similar
β’ Advanced proficiency in Cloud Data Warehouse Snowflake, AWS Redshift
β’ Advanced proficiency in at least one scheduling/orchestration tool such as Airflow, AWS Step Functions or similar
β’ Proficiency in Unix scripting, data structures, data serialization formats such as JSON, AVRO, Protobuf, or similar, big-data storage formats such as Parquet, Iceberg, or similar, data processing methodologies such as batch, micro-batching, or stream, one or more data modelling techniques such as Dimensional, Data Vault, Kimball, Inmon, etc., Agile methodology, TDD or BDD and CI/CD tools.
Preferred qualifications, capabilities, and skills
β’ Knowledge of data governance and security best practices.
β’ Experience in carrying out data analysis to support business insights.
β’ Strong Python and Spark
Location: Jersey City/ONSITE/ NO C2C/NO C2H/NO THIRD PARTIES/ONLY W2/
ONLY LOCALS TO NJ/NY - NO RELOCATION CANDIDATES
Skillset: Data Engineer
Must Haves: Python, PySpark, AWS β ECS, Glue, Lambda, S3
Nice to Haves: Java, Spark, React Js
Interview Process: Interview Process: 2 rounds, 2nd will be on site
Youβre ready to gain the skills and experience needed to grow within your role and advance your career β and we have the perfect software engineering opportunity for you.
As a Data Engineer III - Python / Spark / Data Lake at JPMorgan Chase within the Consumer and Community Bank , you will be a seasoned member of an agile team, tasked with designing and delivering reliable data collection, storage, access, and analytics solutions that are secure, stable, and scalable. Your responsibilities will include developing, testing, and maintaining essential data pipelines and architectures across diverse technical areas, supporting various business functions to achieve the firm's business objectives.
Job responsibilities:
β’ Supports review of controls to ensure sufficient protection of enterprise data.
β’ Advises and makes custom configuration changes in one to two tools to generate a product at the business or customer request.
β’ Updates logical or physical data models based on new use cases.
β’ Frequently uses SQL and understands NoSQL databases and their niche in the marketplace.
β’ Adds to team culture of diversity, opportunity, inclusion, and respect.
β’ Develop enterprise data models, Design/ develop/ maintain large-scale data processing pipelines (and infrastructure), Lead code reviews and provide mentoring thru the process, Drive data quality, Ensure data accessibility (to analysts and data scientists), Ensure compliance with data governance requirements, and Ensure business alignment (ensure data engineering practices align with business goals).
β’ Supports review of controls to ensure sufficient protection of enterprise data
Required qualifications, capabilities, and skills
β’ Formal training or certification on data engineering concepts and 2+ years applied experience
β’ Experience across the data lifecycle, advanced experience with SQL (e.g., joins and aggregations), and working understanding of NoSQL databases
β’ Experience with statistical data analysis and ability to determine appropriate tools and data patterns to perform analysis
β’ Extensive experience in AWS, design, implementation, and maintenance of data pipelines using Python and PySpark.
β’ Proficient in Python and PySpark, able to write and execute complex queries to perform curation and build views required by end users (single and multi-dimensional).
β’ Proven experience in performance and tuning to ensure jobs are running at optimal levels and no performance bottleneck.
β’ Advanced proficiency in leveraging Gen AI models from Anthropic (or OpenAI, or Google) using APIs/SDKs
β’ Advanced proficiency in cloud data lakehouse platform such as AWS data lake services, Databricks or Hadoop, relational data store such as Postgres, Oracle or similar, and at least one NOSQL data store such as Cassandra, Dynamo, MongoDB or similar
β’ Advanced proficiency in Cloud Data Warehouse Snowflake, AWS Redshift
β’ Advanced proficiency in at least one scheduling/orchestration tool such as Airflow, AWS Step Functions or similar
β’ Proficiency in Unix scripting, data structures, data serialization formats such as JSON, AVRO, Protobuf, or similar, big-data storage formats such as Parquet, Iceberg, or similar, data processing methodologies such as batch, micro-batching, or stream, one or more data modelling techniques such as Dimensional, Data Vault, Kimball, Inmon, etc., Agile methodology, TDD or BDD and CI/CD tools.
Preferred qualifications, capabilities, and skills
β’ Knowledge of data governance and security best practices.
β’ Experience in carrying out data analysis to support business insights.
β’ Strong Python and Spark






