

Sr. Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Sr. Data Engineer, offering a contract length of "unknown" with a pay rate of $55-$67/hr. Key skills include experience with Spark, Kafka, SQL, and cloud platforms like AWS. A minimum of 5 years in data pipeline development is required.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
536
-
ποΈ - Date discovered
July 4, 2025
π - Project duration
Unknown
-
ποΈ - Location type
Unknown
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Seattle, WA
-
π§ - Skills detailed
#Datasets #Metadata #AWS (Amazon Web Services) #Schema Design #Spark (Apache Spark) #Code Reviews #"ETL (Extract #Transform #Load)" #Java #Python #Data Lake #Data Integrity #SQL (Structured Query Language) #Batch #Kafka (Apache Kafka) #Observability #ML (Machine Learning) #Data Warehouse #Data Engineering #MLflow #Monitoring #Cloud #Data Pipeline #Data Processing #Data Quality #Scala
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
We are seeking a Senior Data Engineer with a deep understanding of scalable data systems, streaming pipelines, and ML-focused data workflows to help design and implement robust data infrastructure supporting real-time machine learning. In this role, you will work closely with ML engineers, platform engineers, and product teams to build end-to-end data pipelines, feature stores, and monitoring systems that power production-grade recommendation models.
Responsibilities and Duties of the Role:
-Focus on major areas of work, typically 20% or more of role % of the time.
Β· Design, build, and maintain scalable data pipelines (batch and streaming) for ingesting, transforming, and serving user interaction data, content metadata, and model features.
Β· Develop and operate real-time and offline feature stores to support low-latency ML inference and model training workflows.
Β· Partner with ML engineers to define and implement data schemas, data validation processes, and high-quality datasets for model development.
Β· Implement and maintain data quality monitoring, observability, and alerting to ensure reliable production systems.
Β· Optimize data processing workflows for performance, scalability, and cost efficiency across large-scale datasets.
Β· Collaborate with product managers and ML teams to translate personalization requirements into data infrastructure solutions.
Β· Contribute to the selection and implementation of modern data tooling and cloud-native services.
Β· Participate in code reviews, design discussions, and team knowledge sharing to raise the bar for data engineering excellence.
Basic Qualifications:
Strong experience with modern data processing frameworks such as Spark, Flink, Beam, Kafka Streams, or equivalent.
Β· Experience designing and implementing real-time streaming data pipelines.
Β· Proficiency with SQL and schema design for large-scale analytical datasets.
Β· Familiarity with cloud data platforms (e.g., AWS) and modern data infrastructure components (e.g.,
data lakes, data warehouses, feature stores).
Β· Experience supporting ML workflows (model training pipelines, feature engineering, data validation).
Β· Solid software engineering skills with experience in Python, Java, Scala, or similar languages.
Β· Strong problem-solving skills and ability to work independently in a fast-paced environment.
Preferred Qualifications:
Prior experience building data infrastructure for personalization, recommendation systems, or other ML-powered products.
Β· Familiarity with ML lifecycle tools (MLflow, TFX, Kubeflow) and MLOps best practices.
Β· Experience implementing data validation, monitoring, and lineage tools to ensure high data integrity for ML models.
Β· Knowledge of real-time ML serving architectures and online feature generation. Β· Experience optimizing large-scale data workflows for latency-sensitive applications.
Β· Prior experience operating in product development or startup environments. Β· Strong collaboration skills working closely with machine learning engineers and product teams.
Experience with: 5+ years of experience building and maintaining production-grade data pipelines and distributed data processing systems.
Compensation:
The compensation range for this position is $55/hr. - $67/hr. (approx. $110,000 - $135,000 annualized). Final compensation for this role will be determined by various factors such as a candidateβs relevant work experience, skills, certifications, and geographic location.
Benefits:
Medical, Dental, Vision, Hearing, Employee Assistance Program, Life, AD&D, and Long-term Disability Insurance, Retirement program with a 3% company match (no vesting limitation), interactive employee reward & recognition program, and equipment provision.
- EEO Employer
- E-Verify Employer