

PTR Global
DevOps MLOps Engineer (Python), Legal Operations
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a DevOps MLOps Engineer (Python) in Legal Operations, offering a long-term hybrid contract (3 days onsite in Cupertino, CA or Austin, TX). Requires 3+ years in MLOps, strong Python skills, and experience with CI/CD pipelines. Pay rate: $70-$75/hr.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
600
-
ποΈ - Date
April 14, 2026
π - Duration
More than 6 months
-
ποΈ - Location
Hybrid
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Cupertino, CA
-
π§ - Skills detailed
#API (Application Programming Interface) #Deployment #REST API #Security #Scala #Documentation #AI (Artificial Intelligence) #Logging #Database Management #Docker #Data Warehouse #Data Access #REST (Representational State Transfer) #Observability #Data Science #Automation #DevOps #Python #Snowflake #Compliance #dbt (data build tool) #GitHub #Monitoring #EDW (Enterprise Data Warehouse) #Kubernetes
Role description
DevOps MLOps Engineer (Python), Legal Operations
Day 1 onsite Cupertino, CA / Austin, TX (Prefers local folks)
Hybrid β 3 days onsite and 2 days remote
Long term contract
Direct client opportunity
No mid layer / No Implementation partners are Involved
For this role, ideally someone has previous working exp with same client, because this role is specific to how we deploy solutions/ products/ code in Client environment.
β’ The Applied Data Science team within Legal Operations builds AI-powered tools, analytics capabilities, and data infrastructure that enable Clientβs legal organization to work smarter and faster.
β’ The team uses Clientβs internal AI platform to build and deploy production-grade AI applications β from document analysis and legal research to spend analytics and conversational intelligence.
β’ As the volume and complexity of AI deployments grows, the team needs a dedicated engineering discipline to own the path from prototype to production.
β’ The MLOps Engineer is the bridge between the AI applications the Applied Data Science team builds and the production environment where stakeholders rely on them.
β’ You will own deployment pipelines, integration infrastructure, access governance, and scalability for every AI-powered tool the team ships.
β’ This is not a model training or data science role β it is an engineering role focused on making AI applications reliable, governed, and scalable in an enterprise environment.
Key Responsibilities
Deployment & Delivery
β’ Own CI/CD pipelines for AI-powered applications β from development environments through production release
β’ Build and maintain reusable deployment templates that enable the team to ship AI tools faster and more consistently
β’ Manage environment promotion β dev β staging β production β with appropriate testing and validation gates at each stage
β’ Coordinate with infrastructure and platform teams on deployment standards and security requirements
Integration & Data Connectivity
β’ Own API integrations connecting AI applications to live Legal data sources
β’ Manage live data refresh pipelines, ensuring AI tools reflect current data without manual intervention
β’ Version and manage API contracts β handling changes in upstream data sources without breaking downstream AI applications
β’ Troubleshoot and resolve integration failures with minimal impact to end users
Governance & Security
β’ Implement role-based access control (RBAC) for all deployed AI applications β ensuring the right stakeholders have access to the right tools
β’ Build and maintain audit logging β capturing usage, queries, and responses for compliance and accountability
β’ Ensure all deployments meet Client's enterprise security standards β secrets management, authentication, data handling policies
β’ Partner with governance and data teams to enforce data access policies at the application layer
Observability & Reliability
β’ Instrument deployed AI applications with monitoring for latency, error rates, token usage, and response quality
β’ Build alerting for production failures and performance degradation
β’ Establish SLAs for AI application uptime and response time; own resolution when they are breached
β’ Maintain deployment documentation and runbooks for every production application
Enablement
β’ Reduce the deployment burden on data scientists and AI engineers β abstract complexity so builders can focus on building
β’ Build and maintain developer tooling that makes the path from prototype to production faster and more self-service over time
Minimum Qualifications
β’ 3+ years of experience in MLOps, DevOps, or platform engineering roles
β’ Strong Python skills β this is the primary language for tooling and automation
β’ Hands-on experience deploying and operating LLM-powered applications in production
β’ Experience building CI/CD pipelines for AI or software applications (GitHub Actions or equivalent)
β’ Experience with REST API development and integration β connecting applications to live data sources
β’ Working knowledge of containerization β Docker; Kubernetes basics
β’ Experience implementing access control, authentication, and audit logging in enterprise environments
Preferred Qualifications
β’ Exp with LLM observability tooling β monitoring latency, token usage, response quality in production
β’ Familiarity with dbt, Snowflake, or enterprise data warehouse environments
β’ Exp with vector database management β Chroma, Pinecone, Weaviate, or equivalent
β’ Exp in a regulated or compliance-sensitive environment where auditability and data access governance are non-negotiable
β’ Familiarity with RAG (Retrieval-Augmented Generation) application architecture β not to build them, but to deploy and operate them reliably.
Pay Range: $70/hr - $75/hr
The specific compensation for this position will be determined by a number of factors, including the scope, complexity and location of the role as well as the cost of labor in the market; the skills, education, training, credentials and experience of the candidate; and other conditions of employment. Our full-time consultants have access to benefits including medical, dental, and vision as well as 401K contributions.
DevOps MLOps Engineer (Python), Legal Operations
Day 1 onsite Cupertino, CA / Austin, TX (Prefers local folks)
Hybrid β 3 days onsite and 2 days remote
Long term contract
Direct client opportunity
No mid layer / No Implementation partners are Involved
For this role, ideally someone has previous working exp with same client, because this role is specific to how we deploy solutions/ products/ code in Client environment.
β’ The Applied Data Science team within Legal Operations builds AI-powered tools, analytics capabilities, and data infrastructure that enable Clientβs legal organization to work smarter and faster.
β’ The team uses Clientβs internal AI platform to build and deploy production-grade AI applications β from document analysis and legal research to spend analytics and conversational intelligence.
β’ As the volume and complexity of AI deployments grows, the team needs a dedicated engineering discipline to own the path from prototype to production.
β’ The MLOps Engineer is the bridge between the AI applications the Applied Data Science team builds and the production environment where stakeholders rely on them.
β’ You will own deployment pipelines, integration infrastructure, access governance, and scalability for every AI-powered tool the team ships.
β’ This is not a model training or data science role β it is an engineering role focused on making AI applications reliable, governed, and scalable in an enterprise environment.
Key Responsibilities
Deployment & Delivery
β’ Own CI/CD pipelines for AI-powered applications β from development environments through production release
β’ Build and maintain reusable deployment templates that enable the team to ship AI tools faster and more consistently
β’ Manage environment promotion β dev β staging β production β with appropriate testing and validation gates at each stage
β’ Coordinate with infrastructure and platform teams on deployment standards and security requirements
Integration & Data Connectivity
β’ Own API integrations connecting AI applications to live Legal data sources
β’ Manage live data refresh pipelines, ensuring AI tools reflect current data without manual intervention
β’ Version and manage API contracts β handling changes in upstream data sources without breaking downstream AI applications
β’ Troubleshoot and resolve integration failures with minimal impact to end users
Governance & Security
β’ Implement role-based access control (RBAC) for all deployed AI applications β ensuring the right stakeholders have access to the right tools
β’ Build and maintain audit logging β capturing usage, queries, and responses for compliance and accountability
β’ Ensure all deployments meet Client's enterprise security standards β secrets management, authentication, data handling policies
β’ Partner with governance and data teams to enforce data access policies at the application layer
Observability & Reliability
β’ Instrument deployed AI applications with monitoring for latency, error rates, token usage, and response quality
β’ Build alerting for production failures and performance degradation
β’ Establish SLAs for AI application uptime and response time; own resolution when they are breached
β’ Maintain deployment documentation and runbooks for every production application
Enablement
β’ Reduce the deployment burden on data scientists and AI engineers β abstract complexity so builders can focus on building
β’ Build and maintain developer tooling that makes the path from prototype to production faster and more self-service over time
Minimum Qualifications
β’ 3+ years of experience in MLOps, DevOps, or platform engineering roles
β’ Strong Python skills β this is the primary language for tooling and automation
β’ Hands-on experience deploying and operating LLM-powered applications in production
β’ Experience building CI/CD pipelines for AI or software applications (GitHub Actions or equivalent)
β’ Experience with REST API development and integration β connecting applications to live data sources
β’ Working knowledge of containerization β Docker; Kubernetes basics
β’ Experience implementing access control, authentication, and audit logging in enterprise environments
Preferred Qualifications
β’ Exp with LLM observability tooling β monitoring latency, token usage, response quality in production
β’ Familiarity with dbt, Snowflake, or enterprise data warehouse environments
β’ Exp with vector database management β Chroma, Pinecone, Weaviate, or equivalent
β’ Exp in a regulated or compliance-sensitive environment where auditability and data access governance are non-negotiable
β’ Familiarity with RAG (Retrieval-Augmented Generation) application architecture β not to build them, but to deploy and operate them reliably.
Pay Range: $70/hr - $75/hr
The specific compensation for this position will be determined by a number of factors, including the scope, complexity and location of the role as well as the cost of labor in the market; the skills, education, training, credentials and experience of the candidate; and other conditions of employment. Our full-time consultants have access to benefits including medical, dental, and vision as well as 401K contributions.






