

SBS Creatix
Data Engineer (Databricks/ Scala/ Spark)
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer (Databricks/Scala/Spark) on a contract-to-hire basis, fully remote, requiring 3+ years of experience, strong skills in Scala, Spark, and Databricks, and familiarity with cloud platforms (AWS, Azure, GCP).
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
640
-
ποΈ - Date
January 17, 2026
π - Duration
Unknown
-
ποΈ - Location
Remote
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
United States
-
π§ - Skills detailed
#Programming #Big Data #Cloud #PyTorch #AI (Artificial Intelligence) #AWS (Amazon Web Services) #Data Modeling #Databricks #Spark (Apache Spark) #Consulting #Kafka (Apache Kafka) #Data Science #Delta Lake #Scala #TensorFlow #Migration #Azure #"ETL (Extract #Transform #Load)" #ML (Machine Learning) #DataOps #Data Pipeline #Batch #GCP (Google Cloud Platform) #Data Engineering #Apache Spark #StreamSets
Role description
General Info:
Must be a US Citizen or GC Holder
Fully Remote
Contract to Hire
Summary:
The Data Engineer will design, build, and optimize scalable data platforms to support analytics and business decision-making. This role focuses on Databricks and cloud-based data engineering, integrating structured and unstructured data using Scala, Apache Spark, and both batch and streaming architectures. You will work closely with data scientists, architects, and client stakeholders in a consulting environment.
Key Responsibilities:
β’ Design and deliver end-to-end big data and AI-enabled analytics platforms
β’ Build and optimize batch and streaming data pipelines
β’ Apply Databricks, Spark, and cloud best practices
β’ Serve as technical lead for complex solutions; document and support delivery
β’ Support estimation, technical risk management, and knowledge transfer
Required Experience:
β’ 3+ years as a Data Engineer / Big Data Engineer
β’ Hands-on Databricks experience in production environments
β’ Strong Scala and Apache Spark expertise
β’ Experience with batch and streaming pipelines (Kafka, Structured Streaming)
β’ Experience with DataOps, CI/CD, and cloud migrations
β’ Experience with at least two cloud platforms: AWS, Azure, or GCP
β’ Bachelorβs degree or equivalent experience
Core Technical Skills:
β’ Scala & Spark: Functional programming, immutability, performance tuning
β’ Streaming: Kafka, watermarking, windowed aggregations
β’ Data Engineering: ETL/ELT, data modeling, orchestration
β’ Databricks: Notebooks, Jobs, Delta Lake, CI/CD, performance optimization
Nice to Have:
β’ ML frameworks (Spark MLlib, TensorFlow, PyTorch)
β’ Experience with StreamSets, MapReduce
β’ Databricks, Azure, AWS, or GCP certifications
General Info:
Must be a US Citizen or GC Holder
Fully Remote
Contract to Hire
Summary:
The Data Engineer will design, build, and optimize scalable data platforms to support analytics and business decision-making. This role focuses on Databricks and cloud-based data engineering, integrating structured and unstructured data using Scala, Apache Spark, and both batch and streaming architectures. You will work closely with data scientists, architects, and client stakeholders in a consulting environment.
Key Responsibilities:
β’ Design and deliver end-to-end big data and AI-enabled analytics platforms
β’ Build and optimize batch and streaming data pipelines
β’ Apply Databricks, Spark, and cloud best practices
β’ Serve as technical lead for complex solutions; document and support delivery
β’ Support estimation, technical risk management, and knowledge transfer
Required Experience:
β’ 3+ years as a Data Engineer / Big Data Engineer
β’ Hands-on Databricks experience in production environments
β’ Strong Scala and Apache Spark expertise
β’ Experience with batch and streaming pipelines (Kafka, Structured Streaming)
β’ Experience with DataOps, CI/CD, and cloud migrations
β’ Experience with at least two cloud platforms: AWS, Azure, or GCP
β’ Bachelorβs degree or equivalent experience
Core Technical Skills:
β’ Scala & Spark: Functional programming, immutability, performance tuning
β’ Streaming: Kafka, watermarking, windowed aggregations
β’ Data Engineering: ETL/ELT, data modeling, orchestration
β’ Databricks: Notebooks, Jobs, Delta Lake, CI/CD, performance optimization
Nice to Have:
β’ ML frameworks (Spark MLlib, TensorFlow, PyTorch)
β’ Experience with StreamSets, MapReduce
β’ Databricks, Azure, AWS, or GCP certifications






