

Net2Source Inc.
MLOps Lead Engineer (Dataiku and AWS SageMaker)
β - Featured Role | Apply direct with Data Freelance Hub
This role is for an "MLOps Lead Engineer" with a long-term contract in Reading, Pennsylvania, offering competitive pay. Candidates must have expertise in Dataiku and AWS SageMaker, along with experience in building agentic AI systems and RAG pipelines.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
March 7, 2026
π - Duration
Unknown
-
ποΈ - Location
On-site
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Reading, PA
-
π§ - Skills detailed
#Grafana #S3 (Amazon Simple Storage Service) #Scala #Lambda (AWS Lambda) #Infrastructure as Code (IaC) #AWS Lambda #ML (Machine Learning) #Computer Science #Monitoring #AWS SageMaker #SageMaker #IAM (Identity and Access Management) #AWS (Amazon Web Services) #Indexing #Deployment #OpenSearch #Model Evaluation #Dataiku #Docker #Data Science #DevOps #Databases #Observability #Kubernetes #A/B Testing #API (Application Programming Interface) #Automation #Cloud #DynamoDB #Data Governance #GIT #AI (Artificial Intelligence) #Data Engineering
Role description
Job Title: MLOps Engineer (Dataiku and AWS SageMaker)
Location: Reading, Pennsylvania (Onsite β 5 Days/Week at Client Location)
Employment Type: Contract / Long-Term
Role Overview
We are seeking a hands-on MLOps Engineer with strong experience in Dataiku and AWS SageMaker to design, deploy, and operate scalable machine learning and generative AI solutions. The ideal candidate will have experience building agentic AI systems, RAG pipelines, and production-grade ML infrastructure on AWS while ensuring reliability, governance, and performance at scale.
This role requires deep expertise in LLMOps, CI/CD automation, containerization, cloud infrastructure, and observability frameworks to support enterprise AI workloads.
Key Responsibilities
Agentic AI System Design
β’ Design and implement multi-agent architectures including planner, researcher, retriever, executor, and reviewer agents.
β’ Define agent collaboration policies, memory strategies (short/long-term), and tool orchestration frameworks.
β’ Implement supervisor policies and guardrails to ensure safe agent collaboration.
Retrieval-Augmented Generation (RAG) Development
β’ Build high-quality RAG pipelines including ingestion, chunking, embeddings, indexing, and retrieval workflows.
β’ Implement evaluation frameworks for precision, recall, groundedness, and hallucination detection.
β’ Ensure proper citation mechanisms and guardrails for enterprise-grade AI applications.
AWS-Based AI/ML Production Deployment
β’ Deploy and manage AI solutions using AWS services including:
β’ Amazon Bedrock (Agents, Knowledge Bases, Flows)
β’ AWS Lambda
β’ API Gateway
β’ S3
β’ DynamoDB
β’ OpenSearch / Vector Databases
β’ Step Functions
β’ CloudWatch
β’ Enable scalable, secure, and fault-tolerant AI systems in production environments.
MLOps / LLMOps Implementation
β’ Build automated CI/CD pipelines using GitOps practices.
β’ Implement containerization using Docker and Kubernetes.
β’ Manage Infrastructure as Code (IaC) and deployment pipelines.
β’ Implement secure secrets management, IAM policies, blue-green deployments, and rollback mechanisms.
Observability and Model Evaluation
β’ Instrument telemetry including traces, token usage, cost tracking, and latency monitoring.
β’ Build dashboards using Grafana or CloudWatch for operational visibility.
β’ Implement human-in-the-loop review systems, A/B testing, and continuous evaluation pipelines.
Reliability and Scalability
β’ Implement caching strategies, queue management, rate limiting, and retry/backoff mechanisms.
β’ Ensure system reliability through idempotency patterns and drift detection mechanisms.
β’ Monitor and optimize system performance under scale.
Collaboration and Communication
β’ Work closely with DevOps, Data Engineering, Infrastructure, and Architecture teams.
β’ Document system architectures, SLIs/SLOs, and operational runbooks.
β’ Communicate technical updates and insights to both technical and non-technical stakeholders.
Required Qualifications
β’ Bachelorβs degree in Computer Science, Data Science, Engineering, or related field (or equivalent experience).
β’ Proven experience building production-grade MLOps pipelines and AI systems.
β’ Hands-on experience with Dataiku and AWS SageMaker.
β’ Experience designing and deploying RAG pipelines and agent-based AI architectures.
β’ Strong expertise in cloud platforms for AI/ML workloads (AWS preferred).
β’ Solid experience with CI/CD pipelines, Git, Docker, and Kubernetes.
β’ Understanding of model governance, data governance, and AI lifecycle management.
β’ Excellent communication, problem-solving, and collaboration skills.
Preferred / Nice to Have Skills
β’ Experience with AWS Bedrock (Agents, Knowledge Bases, Flows).
β’ Experience with OpenSearch or other vector databases.
β’ Familiarity with LangGraph, CrewAI, Semantic Kernel, or AutoGen frameworks.
β’ Experience with Step Functions, Lambda, API Gateway, DynamoDB, and S3.
β’ Knowledge of evaluation frameworks for LLMs including groundedness and hallucination detection.
β’ Dataiku platform expertise including governance, approvals, artifacts, and MLOps deployment flows.
Certifications (Nice to Have)
β’ Dataiku ML Practitioner
β’ Dataiku Advanced Designer
β’ Dataiku MLOps Practitioner
Job Title: MLOps Engineer (Dataiku and AWS SageMaker)
Location: Reading, Pennsylvania (Onsite β 5 Days/Week at Client Location)
Employment Type: Contract / Long-Term
Role Overview
We are seeking a hands-on MLOps Engineer with strong experience in Dataiku and AWS SageMaker to design, deploy, and operate scalable machine learning and generative AI solutions. The ideal candidate will have experience building agentic AI systems, RAG pipelines, and production-grade ML infrastructure on AWS while ensuring reliability, governance, and performance at scale.
This role requires deep expertise in LLMOps, CI/CD automation, containerization, cloud infrastructure, and observability frameworks to support enterprise AI workloads.
Key Responsibilities
Agentic AI System Design
β’ Design and implement multi-agent architectures including planner, researcher, retriever, executor, and reviewer agents.
β’ Define agent collaboration policies, memory strategies (short/long-term), and tool orchestration frameworks.
β’ Implement supervisor policies and guardrails to ensure safe agent collaboration.
Retrieval-Augmented Generation (RAG) Development
β’ Build high-quality RAG pipelines including ingestion, chunking, embeddings, indexing, and retrieval workflows.
β’ Implement evaluation frameworks for precision, recall, groundedness, and hallucination detection.
β’ Ensure proper citation mechanisms and guardrails for enterprise-grade AI applications.
AWS-Based AI/ML Production Deployment
β’ Deploy and manage AI solutions using AWS services including:
β’ Amazon Bedrock (Agents, Knowledge Bases, Flows)
β’ AWS Lambda
β’ API Gateway
β’ S3
β’ DynamoDB
β’ OpenSearch / Vector Databases
β’ Step Functions
β’ CloudWatch
β’ Enable scalable, secure, and fault-tolerant AI systems in production environments.
MLOps / LLMOps Implementation
β’ Build automated CI/CD pipelines using GitOps practices.
β’ Implement containerization using Docker and Kubernetes.
β’ Manage Infrastructure as Code (IaC) and deployment pipelines.
β’ Implement secure secrets management, IAM policies, blue-green deployments, and rollback mechanisms.
Observability and Model Evaluation
β’ Instrument telemetry including traces, token usage, cost tracking, and latency monitoring.
β’ Build dashboards using Grafana or CloudWatch for operational visibility.
β’ Implement human-in-the-loop review systems, A/B testing, and continuous evaluation pipelines.
Reliability and Scalability
β’ Implement caching strategies, queue management, rate limiting, and retry/backoff mechanisms.
β’ Ensure system reliability through idempotency patterns and drift detection mechanisms.
β’ Monitor and optimize system performance under scale.
Collaboration and Communication
β’ Work closely with DevOps, Data Engineering, Infrastructure, and Architecture teams.
β’ Document system architectures, SLIs/SLOs, and operational runbooks.
β’ Communicate technical updates and insights to both technical and non-technical stakeholders.
Required Qualifications
β’ Bachelorβs degree in Computer Science, Data Science, Engineering, or related field (or equivalent experience).
β’ Proven experience building production-grade MLOps pipelines and AI systems.
β’ Hands-on experience with Dataiku and AWS SageMaker.
β’ Experience designing and deploying RAG pipelines and agent-based AI architectures.
β’ Strong expertise in cloud platforms for AI/ML workloads (AWS preferred).
β’ Solid experience with CI/CD pipelines, Git, Docker, and Kubernetes.
β’ Understanding of model governance, data governance, and AI lifecycle management.
β’ Excellent communication, problem-solving, and collaboration skills.
Preferred / Nice to Have Skills
β’ Experience with AWS Bedrock (Agents, Knowledge Bases, Flows).
β’ Experience with OpenSearch or other vector databases.
β’ Familiarity with LangGraph, CrewAI, Semantic Kernel, or AutoGen frameworks.
β’ Experience with Step Functions, Lambda, API Gateway, DynamoDB, and S3.
β’ Knowledge of evaluation frameworks for LLMs including groundedness and hallucination detection.
β’ Dataiku platform expertise including governance, approvals, artifacts, and MLOps deployment flows.
Certifications (Nice to Have)
β’ Dataiku ML Practitioner
β’ Dataiku Advanced Designer
β’ Dataiku MLOps Practitioner






