SBS Creatix

Data Engineer (Databricks/ Scala/ Spark)

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer (Databricks/Scala/Spark) on a contract-to-hire basis, fully remote, requiring 3+ years of experience, strong skills in Scala, Spark, and Databricks, and familiarity with cloud platforms (AWS, Azure, GCP).
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
640
-
πŸ—“οΈ - Date
January 17, 2026
πŸ•’ - Duration
Unknown
-
🏝️ - Location
Remote
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
United States
-
🧠 - Skills detailed
#Programming #Big Data #Cloud #PyTorch #AI (Artificial Intelligence) #AWS (Amazon Web Services) #Data Modeling #Databricks #Spark (Apache Spark) #Consulting #Kafka (Apache Kafka) #Data Science #Delta Lake #Scala #TensorFlow #Migration #Azure #"ETL (Extract #Transform #Load)" #ML (Machine Learning) #DataOps #Data Pipeline #Batch #GCP (Google Cloud Platform) #Data Engineering #Apache Spark #StreamSets
Role description
General Info: Must be a US Citizen or GC Holder Fully Remote Contract to Hire Summary: The Data Engineer will design, build, and optimize scalable data platforms to support analytics and business decision-making. This role focuses on Databricks and cloud-based data engineering, integrating structured and unstructured data using Scala, Apache Spark, and both batch and streaming architectures. You will work closely with data scientists, architects, and client stakeholders in a consulting environment. Key Responsibilities: β€’ Design and deliver end-to-end big data and AI-enabled analytics platforms β€’ Build and optimize batch and streaming data pipelines β€’ Apply Databricks, Spark, and cloud best practices β€’ Serve as technical lead for complex solutions; document and support delivery β€’ Support estimation, technical risk management, and knowledge transfer Required Experience: β€’ 3+ years as a Data Engineer / Big Data Engineer β€’ Hands-on Databricks experience in production environments β€’ Strong Scala and Apache Spark expertise β€’ Experience with batch and streaming pipelines (Kafka, Structured Streaming) β€’ Experience with DataOps, CI/CD, and cloud migrations β€’ Experience with at least two cloud platforms: AWS, Azure, or GCP β€’ Bachelor’s degree or equivalent experience Core Technical Skills: β€’ Scala & Spark: Functional programming, immutability, performance tuning β€’ Streaming: Kafka, watermarking, windowed aggregations β€’ Data Engineering: ETL/ELT, data modeling, orchestration β€’ Databricks: Notebooks, Jobs, Delta Lake, CI/CD, performance optimization Nice to Have: β€’ ML frameworks (Spark MLlib, TensorFlow, PyTorch) β€’ Experience with StreamSets, MapReduce β€’ Databricks, Azure, AWS, or GCP certifications