Salve.Lab

Senior MLOps Platform Architect

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior MLOps Platform Architect on a long-term B2B contract, offering competitive pay. It requires extensive experience in AWS, Kubernetes, Terraform, CI/CD, and deploying ML models. Remote work is available across Europe.
🌎 - Country
United Kingdom
πŸ’± - Currency
Β£ GBP
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
December 6, 2025
πŸ•’ - Duration
Unknown
-
🏝️ - Location
Remote
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
United Kingdom
-
🧠 - Skills detailed
#Deployment #Kubernetes #Model Deployment #Automation #Python #Security #Leadership #Grafana #NLU (Natural Language Understanding) #Apache Airflow #Data Science #Airflow #Cloud #Jenkins #Docker #Observability #Programming #SageMaker #AI (Artificial Intelligence) #ML (Machine Learning) #Prometheus #Scala #Infrastructure as Code (IaC) #Storage #AWS (Amazon Web Services) #"ETL (Extract #Transform #Load)" #MLflow #GitLab #Terraform #Monitoring #DevOps #Automatic Speech Recognition (ASR)
Role description
Remote | B2B Contract | Europe Role Overview We are hiring a senior MLOps who can build an entire AI platform infrastructure end-to-end. This is not a research role and not a standard ML Engineer role. If you haven’t designed production-grade MLOps infrastructure, haven’t built CI/CD for ML, or haven’t deployed ML workloads on Kubernetes at scale, this role is not a fit. You will design, build, and own the AWS-based infrastructure, Kubernetes platform, CI/CD pipelines, and observability stack that supports our AI models (Agentic AI, NLU, ASR, Voice Biometrics, TTS). You will be the technical owner of MLOps infrastructure decisions, patterns, and standards. Key Responsibilities: MLOps Platform Architecture (from scratch) β€’ Design and build AWS-based AI/ML infrastructure using Terraform (required). β€’ Define standards for security, automation, cost efficiency, and governance. β€’ Architect infrastructure for ML workloads, GPU/accelerators, scaling, and high availability. Kubernetes & Model Deployment β€’ Architect, build, and operate production Kubernetes clusters. β€’ Containerize and productize ML models (Docker, Helm). β€’ Deploy latency-sensitive and high-throughput models (ASR/TTS/NLU/Agentic AI). β€’ Ensure GPU and accelerator nodes are properly integrated and optimized. CI/CD for Machine Learning β€’ Build automated training, validation, and deployment pipelines (GitLab/Jenkins). β€’ Implement canary, blue-green, and automated rollback strategies. β€’ Integrate MLOps lifecycle tools (MLflow, Kubeflow, SageMaker Model Registry, etc.). Observability & Reliability β€’ Implement full observability (Prometheus + Grafana). β€’ Own uptime, performance, and reliability for ML production services. β€’ Establish monitoring for latency, drift, model health, and infrastructure health. Collaboration & Technical Leadership β€’ Work closely with ML engineers, researchers, and data scientists. β€’ Translate experimental models into production-ready deployments. β€’ Define best practices for MLOps across the company. Requirements: We’re looking for a senior engineer with a strong DevOps/SRE background who has worked extensively with ML systems in production. The ideal candidate brings a combination of infrastructure, automation, and hands-on MLOps experience. β€’ 5+ years in a Senior DevOps, SRE, or MLOps Engineering role supporting production environments. β€’ Strong experience designing, building, and maintaining Kubernetes clusters in production. β€’ Hands-on expertise with Terraform (or similar IaC tools) to manage cloud infrastructure. β€’ Solid programming skills in Python or Go for building automation, tooling, and ML workflows. β€’ Proven experience creating and maintaining CI/CD pipelines (GitLab or Jenkins). β€’ Practical experience deploying and supporting ML models in production (e.g., ASR, TTS, NLU, LLM/Agentic AI). β€’ Familiarity with ML workflow orchestration tools such as Kubeflow, Apache Airflow, or similar. β€’ Experience with experiment tracking and model registry tools (e.g., MLflow, SageMaker Model Registry). β€’ Exposure to deploying models on GPU or specialized hardware (e.g., Inferentia, Trainium). β€’ Solid understanding of cloud infrastructure on AWS, including networking, scaling, storage, and security best practices. β€’ Experience with deployment tooling (Docker, Helm) and observability stacks (Prometheus, Grafana). Ways to Know You’ll Succeed β€’ You enjoy building platforms from the ground up and owning technical decisions. β€’ You’re comfortable collaborating with ML engineers, researchers, and software teams to turn research into stable production systems. β€’ You like solving performance, automation, and reliability challenges in distributed systems. β€’ You bring a structured, pragmatic, and scalable approach to infrastructure design. β€’ Energetic and proactive individual, with a natural drive to take initiative and move things forward. β€’ Enjoys working closely with people - researchers, ML engineers, cloud architects, product teams. β€’ Comfortable sharing ideas openly, challenging assumptions, and contributing to technical discussions. β€’ Collaborative mindset: you like to build together, not work in isolation. β€’ Strong ownership mentality - you enjoy taking responsibility for systems end-to-end. β€’ Curious, hands-on, and motivated by solving complex technical challenges. β€’ Clear communicator who can translate technical work into practical recommendations. β€’ Thrives in a fast-paced environment where you can experiment, improve, and shape how things are done. What's on Offer: β€’ Competitive fixed compensation based on experience and expertise. β€’ Work on cutting-edge AI systems used globall. β€’ Dynamic, multi-disciplinary teams engaged in digital transformation. β€’ Remote-first work model β€’ Long-term B2B contract β€’ 20+ days paid time off β€’ Apple gear β€’ Training & development budget Diversity and Inclusion Commitment We are dedicated to creating and sustaining an inclusive, respectful workplace for all -regardless of gender, ethnicity, or background. We actively encourage applicants from all identities and experience levels to apply and bring your authentic self to our fast-paced, supportive team.