

Data Engineer (Scala/Spark/AWS EMR)
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer (Scala/Spark/AWS EMR) on a 6-month contract in Durham, NC, requiring 10+ years of experience, strong skills in Scala, Apache Spark, and AWS EMR, and a background in secure financial data aggregation.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
July 9, 2025
π - Project duration
More than 6 months
-
ποΈ - Location type
Hybrid
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Raleigh-Durham-Chapel Hill Area
-
π§ - Skills detailed
#Computer Science #Cloud #Spark (Apache Spark) #Databases #Java #AWS (Amazon Web Services) #Scala #Data Pipeline #Data Processing #Data Aggregation #Data Engineering #"ETL (Extract #Transform #Load)" #AWS EMR (Amazon Elastic MapReduce) #REST (Representational State Transfer) #Amazon EMR (Amazon Elastic MapReduce) #NoSQL #Big Data #Datasets #Apache Spark #Snowflake #Batch #Kafka (Apache Kafka)
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Job Title: Data Engineer β Leave of Absence Coverage (6-Month Contract)
Location: Durham, NC β Hybrid (2 Weeks Onsite per Month)
Weβre looking for an experienced Data Engineer to provide 6 months of coverage for a team focused on secure financial data aggregationβpowering customer access to their financial data across 3rd party platforms like Mint and Morningstar.
Youβll work with batch systems that collect data from multiple business lines (brokerage, 401(k), stock plans, etc.) and deliver it via secure APIs based on consumer consent. The environment is on-premises, but the role involves modernizing with cloud and big data technologies.
Must-Have Skills
β’ Strong Data Engineering background
β’ Expert-level Scala
β’ Solid experience with Apache Spark and Amazon EMR
β’ Experience with ETL pipelines and Control-M job scheduling
β’ Familiarity with Cassandra or other distributed NoSQL databases
β’ Some Java knowledge
Nice to Have
β’ Experience with streaming platforms (especially Kafka)
Responsibilities
β’ Design and develop scalable data pipelines for aggregating customer financial data
β’ Build and maintain batch Spark and Spring Batch ETL jobs
β’ Monitor and troubleshoot ETL workflows using Control-M
β’ Consume and integrate REST/SOAP APIs
β’ Work on data processing using Apache Spark on AWS EMR
β’ Store large datasets efficiently using Parquet, Cassandra, or similar tools
β’ Collaborate across teams and advocate for best practices in software engineering
Required Experience
β’ Bachelorβs degree in Computer Science or related field
β’ 10+ years of hands-on development experience with Spark or Spring Batch
β’ Experience with cloud data tools, such as AWS and Snowflake.
β’ Proven experience working with big data solutions and distributed systems