

Hope Tech
Machine Vision & Perception Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Machine Vision & Perception Engineer with 5+ years of experience in real-time perception systems. It offers a 12–24 month contract, remote work, and a pay rate of $60,000 - $150,000. Key skills include Python, C++, object detection, and sensor fusion.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
681
-
🗓️ - Date
February 21, 2026
🕒 - Duration
More than 6 months
-
🏝️ - Location
Remote
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Remote
-
🧠 - Skills detailed
#R #ML (Machine Learning) #Deployment #Object Detection #Visual Odometry #PyTorch #Python #Cloud #C++ #OpenCV (Open Source Computer Vision Library) #Computer Science
Role description
Familiarity with:
We are seeking a highly capable Machine Vision & Perception Engineer to design, implement, and deploy real-time scene understanding systems for complex, high-stakes operational environments.
This role focuses on multimodal perception under extreme conditions, including low light, occlusion, motion blur, environmental variability, and limited compute resources. You will build production-grade vision systems that interpret human actions, detect and track individuals, understand context, and operate reliably on edge devices without cloud connectivity.
This is a fixed-term, full-time position (12–24 months) with strong potential for renewal based on performance and program continuation.
Key Responsibilities
- Design and deploy real-time computer vision pipelines for:
Multi-person detection, registration, and tracking
Human pose estimation and skeletal tracking
Body-part segmentation under occlusion
Action and scene interpretation
Object detection and spatial association
- Develop robust perception systems that operate in:
Low-light and high dynamic range environments
Outdoor conditions (sunlight, weather variability)
Cluttered and partially occluded scenes
- Integrate multimodal sensors including:
Stereo RGB / depth cameras
Thermal imaging systems
IMUs and localization sensors
- Fuse RGB, depth, and thermal data for improved robustness
- Optimize models for edge deployment (Jetson-class or similar)
- Implement latency-aware, real-time inference pipelines (≤250 ms target)
- Develop dynamic ID tracking across field-of-view changes
- Architect and maintain production-level ML pipelines:
Dataset versioning
Training/validation workflows
Model benchmarking under stress conditions
Quantization / pruning / TensorRT optimization
- Collaborate on hardware–software integration
- Conduct structured field testing and performance evaluation
Required Qualifications
- B.S., M.S., or Ph.D. in Computer Vision, Robotics, Electrical Engineering, Computer Science, or related field
- 5+ years of industry experience (or Ph.D. + 3+ years industry) in real-time perception systems
- Strong experience in:
Object detection (YOLO, Faster R-CNN, etc.)
Human pose estimation (HRNet, OpenPose, MediaPipe, etc.)
Multi-object tracking (DeepSORT, ByteTrack, etc.)
Semantic/instance segmentation
- Experience with low-light or adverse visual condition modeling
Depth cameras (stereo or structured light)
Thermal imaging systems
Sensor fusion
- Strong Python and C++ proficiency
- Experience deploying models on edge devices (NVIDIA Jetson, embedded GPU systems)
TensorRT, ONNX, model quantization
ROS2 or similar robotics middleware
OpenCV, PyTorch
- Experience building production-grade ML systems (not just research prototypes)
Preferred Qualifications
Experience interpreting human actions or complex scenes
Experience with ego-centric (first-person) vision
Experience in robotics perception
Experience with real-time constraints and power-limited systems
Familiarity with localization techniques (visual odometry, IMU fusion)
Experience developing for rugged or field-deployed systems
Background in safety-critical or regulated environments
Published research in computer vision or robotics
Experience bridging academic research into production systems
What Success Looks Like
Reliable multi-person tracking in dynamic scenes
Robust performance across lighting extremes
Edge-deployable models meeting real-time constraints
Clean, maintainable, production-ready perception stack
Demonstrated improvement in robustness via sensor fusion
Compensation & Benefits
We offer a competitive compensation package including:
Competitive salary based on experience
Health, dental, and vision insurance
Paid time off and holidays
Flexible work arrangements (role dependent)
Opportunity to work on high-impact, technically challenging systems
Renewal potential and long-term growth opportunity based on performance and funding continuity
Employment Term
This is a full-time fixed-term role (12–24 months) aligned with a funded development program. Continuation or conversion to long-term employment may be available based on performance, program needs, and future funding.
Ideal Candidate Profile
You are:
Technically rigorous
Comfortable operating with ambiguity
Capable of balancing research innovation with engineering discipline
Excited by real-world constraints (compute, power, lighting, weather)
Able to move from model training to deployed system integration
Pay: $60,000.00 - $150,000.00 per year
Benefits:
Dental insurance
Health insurance
Paid time off
Stock options
Vision insurance
Work Location: Remote
Familiarity with:
We are seeking a highly capable Machine Vision & Perception Engineer to design, implement, and deploy real-time scene understanding systems for complex, high-stakes operational environments.
This role focuses on multimodal perception under extreme conditions, including low light, occlusion, motion blur, environmental variability, and limited compute resources. You will build production-grade vision systems that interpret human actions, detect and track individuals, understand context, and operate reliably on edge devices without cloud connectivity.
This is a fixed-term, full-time position (12–24 months) with strong potential for renewal based on performance and program continuation.
Key Responsibilities
- Design and deploy real-time computer vision pipelines for:
Multi-person detection, registration, and tracking
Human pose estimation and skeletal tracking
Body-part segmentation under occlusion
Action and scene interpretation
Object detection and spatial association
- Develop robust perception systems that operate in:
Low-light and high dynamic range environments
Outdoor conditions (sunlight, weather variability)
Cluttered and partially occluded scenes
- Integrate multimodal sensors including:
Stereo RGB / depth cameras
Thermal imaging systems
IMUs and localization sensors
- Fuse RGB, depth, and thermal data for improved robustness
- Optimize models for edge deployment (Jetson-class or similar)
- Implement latency-aware, real-time inference pipelines (≤250 ms target)
- Develop dynamic ID tracking across field-of-view changes
- Architect and maintain production-level ML pipelines:
Dataset versioning
Training/validation workflows
Model benchmarking under stress conditions
Quantization / pruning / TensorRT optimization
- Collaborate on hardware–software integration
- Conduct structured field testing and performance evaluation
Required Qualifications
- B.S., M.S., or Ph.D. in Computer Vision, Robotics, Electrical Engineering, Computer Science, or related field
- 5+ years of industry experience (or Ph.D. + 3+ years industry) in real-time perception systems
- Strong experience in:
Object detection (YOLO, Faster R-CNN, etc.)
Human pose estimation (HRNet, OpenPose, MediaPipe, etc.)
Multi-object tracking (DeepSORT, ByteTrack, etc.)
Semantic/instance segmentation
- Experience with low-light or adverse visual condition modeling
Depth cameras (stereo or structured light)
Thermal imaging systems
Sensor fusion
- Strong Python and C++ proficiency
- Experience deploying models on edge devices (NVIDIA Jetson, embedded GPU systems)
TensorRT, ONNX, model quantization
ROS2 or similar robotics middleware
OpenCV, PyTorch
- Experience building production-grade ML systems (not just research prototypes)
Preferred Qualifications
Experience interpreting human actions or complex scenes
Experience with ego-centric (first-person) vision
Experience in robotics perception
Experience with real-time constraints and power-limited systems
Familiarity with localization techniques (visual odometry, IMU fusion)
Experience developing for rugged or field-deployed systems
Background in safety-critical or regulated environments
Published research in computer vision or robotics
Experience bridging academic research into production systems
What Success Looks Like
Reliable multi-person tracking in dynamic scenes
Robust performance across lighting extremes
Edge-deployable models meeting real-time constraints
Clean, maintainable, production-ready perception stack
Demonstrated improvement in robustness via sensor fusion
Compensation & Benefits
We offer a competitive compensation package including:
Competitive salary based on experience
Health, dental, and vision insurance
Paid time off and holidays
Flexible work arrangements (role dependent)
Opportunity to work on high-impact, technically challenging systems
Renewal potential and long-term growth opportunity based on performance and funding continuity
Employment Term
This is a full-time fixed-term role (12–24 months) aligned with a funded development program. Continuation or conversion to long-term employment may be available based on performance, program needs, and future funding.
Ideal Candidate Profile
You are:
Technically rigorous
Comfortable operating with ambiguity
Capable of balancing research innovation with engineering discipline
Excited by real-world constraints (compute, power, lighting, weather)
Able to move from model training to deployed system integration
Pay: $60,000.00 - $150,000.00 per year
Benefits:
Dental insurance
Health insurance
Paid time off
Stock options
Vision insurance
Work Location: Remote





