

Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with a contract length of "unknown" in Jersey City, NJ. Pay rate is "unknown." Key skills include AWS, Databricks, PySpark, and Splunk. Experience with ETL processes and big data environments is required.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
July 26, 2025
π - Project duration
Unknown
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Jersey City, NJ
-
π§ - Skills detailed
#Data Engineering #Data Pipeline #Automation #AWS (Amazon Web Services) #Agile #Delta Lake #Data Quality #Compliance #Data Modeling #Big Data #S3 (Amazon Simple Storage Service) #Deployment #Docker #"ETL (Extract #Transform #Load)" #SQL (Structured Query Language) #Monitoring #Observability #Cloud #Debugging #Python #Spark (Apache Spark) #Scripting #Scala #Terraform #Data Analysis #Data Warehouse #Lambda (AWS Lambda) #IAM (Identity and Access Management) #PySpark #Data Processing #Data Lake #Databricks #DevOps #Splunk #Airflow #Snowflake
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Job Title: Data Engineer β AWS, Databricks, PySpark, Splunk
Location: Jersey City, NJ (onsite)
Role Type: Contract
Job Overview:
We are looking for an experienced Data Engineer to join our data team and contribute to building scalable, high-performance data pipelines and analytics solutions. The ideal candidate will have hands-on expertise in AWS, Databricks, PySpark, and experience with Splunk for monitoring and log analytics.
Key Responsibilities:
β’ Design, develop, and optimize ETL/ELT data pipelines using Databricks and PySpark.
β’ Work with large-scale structured and unstructured data across various cloud and data lake environments.
β’ Implement data processing workflows on AWS (e.g., S3, Glue, EMR, Lambda).
β’ Monitor pipeline performance and troubleshoot issues using Splunk and other observability tools.
β’ Ensure data quality, integrity, and compliance through robust validation and testing processes.
β’ Collaborate with data analysts, architects, and DevOps teams to align solutions with business needs and enterprise architecture.
β’ Automate workflows and deployments using CI/CD pipelines and Infrastructure-as-Code principles.
Required Skills
β’ Strong hands-on experience with AWS services, especially S3, Glue, EMR, Lambda, IAM, and CloudWatch.
β’ Expertise in Databricks, including notebooks, jobs, Delta Lake, and workspace management.
β’ Proficiency in PySpark and distributed data processing.
β’ Solid understanding of data modeling, partitioning, and performance tuning in big data environments.
β’ Experience using Splunk for log monitoring, dashboarding, and root cause analysis.
β’ Strong SQL skills and knowledge of performance optimization techniques.
Nice to Have:
β’ Experience with Airflow or other orchestration tools.
β’ Familiarity with Snowflake or other cloud data warehouses.
β’ Knowledge of CI/CD, Docker, or Terraform for deploying data applications.
β’ Python scripting for automation and data engineering tasks.
Soft Skills:
β’ Strong problem-solving and debugging capabilities.
β’ Excellent communication and collaboration in cross-functional teams.
β’ Ability to work independently in a fast-paced, agile environment.