

On-Device AI Runtime Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for an On-Device AI Runtime Engineer on a 3-month contract, with a pay rate of $78.00 - $93.00 (DOE), located in San Diego or Sunnyvale, CA. Key skills include Swift, Metal Performance Shaders, and ML model optimization.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
744
-
ποΈ - Date discovered
August 30, 2025
π - Project duration
3 to 6 months
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Cupertino, CA
-
π§ - Skills detailed
#TensorFlow #Data Science #Data Analysis #ML (Machine Learning) #PyTorch #AI (Artificial Intelligence) #Model Optimization #Scala #Consulting
Role description
A globally leading technology company is looking for an On-Device AI Runtime Engineer to join its cutting-edge AI team. In this position, you will be a key contributor in building high-performance machine learning inference systems, developing optimized AI drivers for edge devices, creating scalable model lifecycle management solutions, and delivering efficient on-device runtime drivers for AI inference across a wide range of hardware platforms.
This role will focus on optimizing and deploying machine learning models on edge and mobile devices. Please note that this is not a Data Science or Data Analyst role.
Job Responsibilities:
β’ Design and implement robust Core ML model optimization pipelines for deploying large-scale ML models on resource-constrained devices.
β’ Support product engineering teams by consulting on AI model performance, iterating on inference solutions to solve real-world mobile/edge AI problems, and developing/delivering custom on-device AI frameworks.
β’ Interface with hardware and platform teams to ensure optimal utilization of neural processing units (NPUs), GPUs, and specialized AI accelerators across the device ecosystem.
Minimum Qualifications:
β’ Strong proficiency in Swift/Objective-C and Metal Performance Shaders.
β’ Familiar with various ML model formats such as Core ML, ONNX, TensorFlow Lite, and PyTorch Mobile.
β’ Strong critical thinking, performance optimization, and low-level system design skills.
β’ Experience with model quantization, pruning, and hardware-aware neural architecture optimization.
β’ Experience with real-time inference pipelines and latency-critical AI applications.
β’ Understanding of mobile device thermal management, power consumption patterns, and compute resource allocation for AI workloads.
Type: Contract
Duration: 3 months (with a possibility for extension)
Work Location: San Diego or Sunnyvale, CA (On-site)
Pay rate: $78.00 - $93.00 (DOE)
A globally leading technology company is looking for an On-Device AI Runtime Engineer to join its cutting-edge AI team. In this position, you will be a key contributor in building high-performance machine learning inference systems, developing optimized AI drivers for edge devices, creating scalable model lifecycle management solutions, and delivering efficient on-device runtime drivers for AI inference across a wide range of hardware platforms.
This role will focus on optimizing and deploying machine learning models on edge and mobile devices. Please note that this is not a Data Science or Data Analyst role.
Job Responsibilities:
β’ Design and implement robust Core ML model optimization pipelines for deploying large-scale ML models on resource-constrained devices.
β’ Support product engineering teams by consulting on AI model performance, iterating on inference solutions to solve real-world mobile/edge AI problems, and developing/delivering custom on-device AI frameworks.
β’ Interface with hardware and platform teams to ensure optimal utilization of neural processing units (NPUs), GPUs, and specialized AI accelerators across the device ecosystem.
Minimum Qualifications:
β’ Strong proficiency in Swift/Objective-C and Metal Performance Shaders.
β’ Familiar with various ML model formats such as Core ML, ONNX, TensorFlow Lite, and PyTorch Mobile.
β’ Strong critical thinking, performance optimization, and low-level system design skills.
β’ Experience with model quantization, pruning, and hardware-aware neural architecture optimization.
β’ Experience with real-time inference pipelines and latency-critical AI applications.
β’ Understanding of mobile device thermal management, power consumption patterns, and compute resource allocation for AI workloads.
Type: Contract
Duration: 3 months (with a possibility for extension)
Work Location: San Diego or Sunnyvale, CA (On-site)
Pay rate: $78.00 - $93.00 (DOE)