

Hope Tech
Contract Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Contract Data Engineer lasting 6 weeks, part-time (15–20 hours/week), with a pay rate of "unknown." Remote work is required. Key skills include AWS, data pipelines (Python, SQL), and machine learning support experience.
🌎 - Country
United States
💱 - Currency
Unknown
-
💰 - Day rate
Unknown
-
🗓️ - Date
January 22, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Remote
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Remote
-
🧠 - Skills detailed
#Documentation #Batch #Airflow #Observability #Storage #Normalization #OpenSearch #dbt (data build tool) #Data Pipeline #Predictive Modeling #AWS (Amazon Web Services) #Spark (Apache Spark) #S3 (Amazon Simple Storage Service) #Datasets #ML (Machine Learning) #Python #Monitoring #SQL (Structured Query Language) #IAM (Identity and Access Management) #Data Quality #"ETL (Extract #Transform #Load)" #SNS (Simple Notification Service) #Data Science #Data Engineering #Databases #Scala #Security #Lambda (AWS Lambda) #Cloud
Role description
Contract Duration: 6 weeks Hours: Part-time, 15–20 hours/week Start Date: ASAP Reports To: Chief Data Scientist, Backstroke.com
About the Role
Backstroke.com is seeking a part-time contract Data Engineer to support critical data engineering work powering our predictive modeling efforts. In this 6-week engagement, you'll help bring raw data into production-grade pipelines, improve data reliability and observability, and help maintain a large-scale dataset used for machine learning and embedding-based predictive models.
This role is hands-on and execution-focused, working closely with the Chief Data Scientist to accelerate modeling throughput and strengthen the stability and usability of our data foundation.
Key Responsibilities Ingest raw data into production data pipelines used for data science modeling (batch and/or near real-time as needed)
Build and enhance AWS-based data workflows, leveraging best practices for scalability and security Set up alerts and notifications in AWS to monitor pipeline health, failures, latency, and data quality issues Create and manage a database layer that stores transformed data, including embeddings used for predictive models Support management of a large-scale dataset, including movement, cleaning, normalization, and maintaining consistency for modeling use
Required Qualifications Strong experience as a Data Engineer supporting machine learning or data science teams
Deep working knowledge of AWS services, such as (or similar): S3, IAM, Lambda, CloudWatch, SNS, EventBridge
Glue, ECS/EKS, Step Functions (nice to have)
Experience building data pipelines (e.g., Python, SQL, Spark, dbt, Airflow, Dagster, Prefect, or similar tools) Experience designing and maintaining databases for ML workflows, including embedding stores and feature-like datasets Comfort working with large datasets and ensuring performance, reliability, and correctness Ability to work independently, communicate clearly, and deliver quickly in a contractor environment
Preferred / Nice-to-Have Familiarity with vector databases and embedding storage patterns (e.g., pgvector, OpenSearch, Pinecone, FAISS, etc.)
Exposure to MLOps concepts (feature pipelines, training dataset versioning, model monitoring) Experience with data quality tooling (e.g., Great Expectations, Monte Carlo, custom checks)
Deliverables & Outcomes (6-Week Goals) Reliable ingestion of raw data into modeling pipelines
Monitoring and alerting for critical pipeline workflows in AWS Operational database/storage system for embedding-ready transformed data Improved processes for handling and cleaning a large dataset used in predictive models Clear documentation of pipeline architecture and handoff notes for the internal team
Working Style
You'll collaborate directly with the Chief Data Scientist and contribute to a fast-moving, data-driven team. We value pragmatic engineering, clear documentation, and systems that are reliable and easy to operate.
Contract Duration: 6 weeks Hours: Part-time, 15–20 hours/week Start Date: ASAP Reports To: Chief Data Scientist, Backstroke.com
About the Role
Backstroke.com is seeking a part-time contract Data Engineer to support critical data engineering work powering our predictive modeling efforts. In this 6-week engagement, you'll help bring raw data into production-grade pipelines, improve data reliability and observability, and help maintain a large-scale dataset used for machine learning and embedding-based predictive models.
This role is hands-on and execution-focused, working closely with the Chief Data Scientist to accelerate modeling throughput and strengthen the stability and usability of our data foundation.
Key Responsibilities Ingest raw data into production data pipelines used for data science modeling (batch and/or near real-time as needed)
Build and enhance AWS-based data workflows, leveraging best practices for scalability and security Set up alerts and notifications in AWS to monitor pipeline health, failures, latency, and data quality issues Create and manage a database layer that stores transformed data, including embeddings used for predictive models Support management of a large-scale dataset, including movement, cleaning, normalization, and maintaining consistency for modeling use
Required Qualifications Strong experience as a Data Engineer supporting machine learning or data science teams
Deep working knowledge of AWS services, such as (or similar): S3, IAM, Lambda, CloudWatch, SNS, EventBridge
Glue, ECS/EKS, Step Functions (nice to have)
Experience building data pipelines (e.g., Python, SQL, Spark, dbt, Airflow, Dagster, Prefect, or similar tools) Experience designing and maintaining databases for ML workflows, including embedding stores and feature-like datasets Comfort working with large datasets and ensuring performance, reliability, and correctness Ability to work independently, communicate clearly, and deliver quickly in a contractor environment
Preferred / Nice-to-Have Familiarity with vector databases and embedding storage patterns (e.g., pgvector, OpenSearch, Pinecone, FAISS, etc.)
Exposure to MLOps concepts (feature pipelines, training dataset versioning, model monitoring) Experience with data quality tooling (e.g., Great Expectations, Monte Carlo, custom checks)
Deliverables & Outcomes (6-Week Goals) Reliable ingestion of raw data into modeling pipelines
Monitoring and alerting for critical pipeline workflows in AWS Operational database/storage system for embedding-ready transformed data Improved processes for handling and cleaning a large dataset used in predictive models Clear documentation of pipeline architecture and handoff notes for the internal team
Working Style
You'll collaborate directly with the Chief Data Scientist and contribute to a fast-moving, data-driven team. We value pragmatic engineering, clear documentation, and systems that are reliable and easy to operate.



