

Galaxy i technologies Inc
Senior Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer with 8+ years of experience, specializing in Scala and Apache Spark, on a contract basis in Sacramento, CA. Requires expertise in data pipelines, cloud platforms, and CI/CD practices. US citizens and GC holders only.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
October 22, 2025
π - Duration
Unknown
-
ποΈ - Location
On-site
-
π - Contract
W2 Contractor
-
π - Security
Unknown
-
π - Location detailed
Sacramento, CA
-
π§ - Skills detailed
#Data Lake #ML (Machine Learning) #GDPR (General Data Protection Regulation) #Redshift #Cloud #Apache Spark #Schema Design #Spark SQL #Data Warehouse #Computer Science #Programming #Kafka (Apache Kafka) #Security #Data Processing #Monitoring #Data Quality #Airflow #Kubernetes #Data Pipeline #Azure #Databricks #GIT #SQL (Structured Query Language) #Scala #Spark (Apache Spark) #API (Application Programming Interface) #Python #Data Science #Java #GCP (Google Cloud Platform) #Data Governance #Delta Lake #BI (Business Intelligence) #DevOps #Snowflake #Docker #Data Engineering #AWS (Amazon Web Services) #Batch #BigQuery #NoSQL #"ETL (Extract #Transform #Load)"
Role description
Job Title: Data Engineer (Scala & Spark)
Location: Sacramento, CA - Onsite
Employment Type: Contract W2
Visa type: Only US citizen and GC holders
Experience Level: 8+ years
About the Role
We are seeking a highly skilled Data Engineer experienced in Scala and Apache Spark to design, build, and maintain scalable data pipelines and processing systems. You will collaborate with data scientists, analysts, and other engineers to ensure efficient data flow across our platforms and support business intelligence, analytics, and machine learning initiatives.
Key Responsibilities
Design, develop, and optimize data pipelines using Apache Spark (batch and streaming).
Write efficient, scalable, and maintainable Scala code for ETL/ELT processes.
Integrate data from multiple structured and unstructured sources.
Work with data lake and data warehouse technologies (e.g., Delta Lake, Snowflake, Redshift, BigQuery).
Collaborate with cross-functional teams to ensure data quality, consistency, and governance.
Implement data validation, monitoring, and alerting frameworks.
Contribute to architecture design for large-scale data systems and distributed processing.
Optimize data processing performance and troubleshoot pipeline issues.
Follow best practices for code versioning, CI/CD, and DevOps in a data engineering context.
Required Skills & Qualifications
Bachelorβs or Masterβs degree in Computer Science, Engineering, or related field.
6+ years of experience in data engineering or backend software development.
Strong hands-on experience with Apache Spark (DataFrame API, Spark SQL, Structured Streaming).
Proficiency in Scala (functional programming, performance optimization, testing).
Solid understanding of distributed systems, parallel processing, and data partitioning concepts.
Experience with data lakes, data warehouses, and ETL workflows.
Familiarity with cloud platforms (AWS, Azure, or GCP) and orchestration tools (Airflow, Databricks Jobs, etc.).
Working knowledge of SQL, NoSQL, and schema design.
Experience with CI/CD pipelines, Git, and containerization (Docker/Kubernetes).
Preferred / Nice-to-Have
Experience with Databricks, Delta Lake, or Glue.
Knowledge of Python or Java in data engineering contexts.
Exposure to Kafka, Kinesis, or other streaming frameworks.
Understanding of data governance, security, and privacy regulations (GDPR, CCPA).
Familiarity with machine learning pipelines or feature engineering.
Soft Skills
Strong problem-solving and analytical skills.
Excellent communication and teamwork abilities.
Proactive, detail-oriented, and able to work independently.
Passion for data-driven decision-making and scalable design.
Job Title: Data Engineer (Scala & Spark)
Location: Sacramento, CA - Onsite
Employment Type: Contract W2
Visa type: Only US citizen and GC holders
Experience Level: 8+ years
About the Role
We are seeking a highly skilled Data Engineer experienced in Scala and Apache Spark to design, build, and maintain scalable data pipelines and processing systems. You will collaborate with data scientists, analysts, and other engineers to ensure efficient data flow across our platforms and support business intelligence, analytics, and machine learning initiatives.
Key Responsibilities
Design, develop, and optimize data pipelines using Apache Spark (batch and streaming).
Write efficient, scalable, and maintainable Scala code for ETL/ELT processes.
Integrate data from multiple structured and unstructured sources.
Work with data lake and data warehouse technologies (e.g., Delta Lake, Snowflake, Redshift, BigQuery).
Collaborate with cross-functional teams to ensure data quality, consistency, and governance.
Implement data validation, monitoring, and alerting frameworks.
Contribute to architecture design for large-scale data systems and distributed processing.
Optimize data processing performance and troubleshoot pipeline issues.
Follow best practices for code versioning, CI/CD, and DevOps in a data engineering context.
Required Skills & Qualifications
Bachelorβs or Masterβs degree in Computer Science, Engineering, or related field.
6+ years of experience in data engineering or backend software development.
Strong hands-on experience with Apache Spark (DataFrame API, Spark SQL, Structured Streaming).
Proficiency in Scala (functional programming, performance optimization, testing).
Solid understanding of distributed systems, parallel processing, and data partitioning concepts.
Experience with data lakes, data warehouses, and ETL workflows.
Familiarity with cloud platforms (AWS, Azure, or GCP) and orchestration tools (Airflow, Databricks Jobs, etc.).
Working knowledge of SQL, NoSQL, and schema design.
Experience with CI/CD pipelines, Git, and containerization (Docker/Kubernetes).
Preferred / Nice-to-Have
Experience with Databricks, Delta Lake, or Glue.
Knowledge of Python or Java in data engineering contexts.
Exposure to Kafka, Kinesis, or other streaming frameworks.
Understanding of data governance, security, and privacy regulations (GDPR, CCPA).
Familiarity with machine learning pipelines or feature engineering.
Soft Skills
Strong problem-solving and analytical skills.
Excellent communication and teamwork abilities.
Proactive, detail-oriented, and able to work independently.
Passion for data-driven decision-making and scalable design.