Openkyber

Docker MLOps Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Docker MLOps Engineer with 15+ years of experience, offering a long-term contract at a hybrid location in Warren, NJ. Key skills include Python, LLM/SLM solutions, data pipelines, and MLOps practices.
🌎 - Country
United States
💱 - Currency
Unknown
-
💰 - Day rate
Unknown
-
🗓️ - Date
February 13, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Hybrid
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
New Jersey
-
🧠 - Skills detailed
#Terraform #Scala #Docker #Kubernetes #Datasets #Regression #Python #TypeScript #Infrastructure as Code (IaC) #Observability #Security #DevOps #Containers #Azure DevOps #Data Pipeline #AI (Artificial Intelligence) #Cloud #Prometheus #AWS (Amazon Web Services) #Data Quality #GitHub #Langchain #Compliance #Databases #GCP (Google Cloud Platform) #Model Deployment #Batch #MLflow #Deployment #ML (Machine Learning) #Azure
Role description
Job Title: Gen AI Developer - Architect (Need 15+ Experience) Location: Hybrid role in Warren, NJ Duration: Long Term Contract About the Role: We are looking for a GenAI / Small Language Model (SLM) Engineer to design, deploy, and maintain agentic AI solutions that are safe, scalable, and business ready. You will own end to end delivery-from prompt and agent design to data pipelines, model deployment, observability, and rigorous validation-partnering with product, architecture, security, and QA to ship AI features that perform reliably in production.hat You'll Do SLM Design and Fine Tuning: Collect, clean, and preprocess domain-specific datasets for SLM training and fine-tuning. Ensure data quality, diversity, and compliance with privacy and security standards. Fine-tune small language models on curated datasets using techniques like LoRA, adapters, or parameter-efficient tuning. Optimize hyperparameters for performance, latency, and resource efficiency. Agent & Prompt Implementation: Help design and implement agent orchestration (single and multi-agent) and function/tool use strategies. Craft, version, and optimize prompts and system instructions for accuracy, coherence, and domain alignment. Integrate external tools/APIs and establish content safety guardrails (e.g., policy enforcement, PII redaction, jailbreak prevention). Implementation, Testing & Maintenance: Build resilient agent workflows and services; harden reliability with retries, fallbacks, circuit breakers. Develop automated tests for prompts, tools, and agent behaviors; maintain regression suites and golden datasets. Operate AI services in production: performance tuning, cost optimization, incident response, and iterative improvement. Data & MLOps: Design and manage data pipelines for fine tuning and retrieval (RAG), including cleansing, labeling, and governance. Monitor drift, quality, latency, and safety signals; implement model/agent observability and alerting. Run structured evaluations of agent outputs (functional, coherence, safety, bias); track precision/recall and hallucination rates. Perform risk assessments for agent behaviors and tool actions; document mitigations and approval workflows. Collaborate with security/compliance to meet regulatory, privacy, and usage policy requirements. Minimum Qualifications: 4 8+ years in software/ML engineering, with 2+ years building LLM/SLM/GenAI solutions in production. Proficiency in Python (and/or TypeScript) and modern AI orchestration frameworks (e.g., Microsoft Agent Framework, Google Agent Development Kit, LangChain, Semantic Kernel). Hands on with retrieval augmented generation (RAG), function calling, prompt optimization, and agent design patterns. Experience building data pipelines (batch/stream), and managing datasets for training/fine tuning and evaluation. Practical understanding of AI guardrails: content filtering, safety policies, redaction, rate limiting, and misuse prevention. Strong willingness to learn advanced agent orchestration and MLOps practices. Preferred Qualifications: MLOps fluency: model packaging, CI/CD, experiment tracking (e.g., MLflow), deployment on cloud/container platforms. IaC (e.g., Terraform/Bicep) and DevOps tooling (e.g., GitHub Actions/Azure DevOps); strong grasp of observability. Experience with multi agent systems, toolformer patterns, and complex orchestration graphs. Knowledge of vector databases and retrieval systems; evaluation frameworks (e.g., Ragas, DeepEval) and custom metrics. Familiarity with privacy, compliance, and model risk management practices for AI. Background in tuning open source and hosted models; comfort with hybrid cloud environments. Tools & Technologies: Python; TypeScript; MAF/Google ADK/LangChain/Semantic Kernel; Vector DBs and frameworks (e.g., Qdrant/FAISS/Pinecone); CI/CD (GitHub Actions/Azure DevOps); IaC (Terraform/Bicep); Observability (OpenTelemetry/Prometheus); Experiment tracking (MLflow); Cloud AI services (e.g., Azure AI FGoundry, Azure OpenAI, Google Cloud Platform Vertex AI, AWS Bedrock); Containers (Docker/Kubernetes). Working Model: Partner with Product, Architecture, Security, and QA to plan, design, and ship safe AI features. Contribute to internal prompt standards, evaluation datasets, and reuseable components. Document designs, decisions, and risks; mentor peers and champion responsible AI practices. For applications and inquiries, contact: hirings@openkyber.com