

Compunnel Inc.
Machine Learning Engineer with GEN AI Experience
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Machine Learning Engineer with GEN AI experience, offering a long-term contract in Durham, NC, or Boston, MA. Requires 10+ years in software engineering, 3-5 years in ML, and expertise in RAG, vector databases, and AWS.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
November 14, 2025
π - Duration
Unknown
-
ποΈ - Location
Hybrid
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Durham, NC
-
π§ - Skills detailed
#S3 (Amazon Simple Storage Service) #Microservices #Monitoring #BI (Business Intelligence) #Oracle #ML (Machine Learning) #AWS Lambda #Knowledge Graph #Snowflake #Cloud #Deployment #Kubernetes #Docker #SageMaker #Prometheus #Observability #Neo4J #Grafana #RDF (Resource Description Framework) #AI (Artificial Intelligence) #Lambda (AWS Lambda) #DevOps #AWS (Amazon Web Services) #Databases #GitHub #Scala #Langchain
Role description
Please find the position details below:
Job Title: Machine Learning Engineer with GEN AI experience
Location: Durham, NC or Boston MA, Merrimack NH & Smithfield RI ( 2 weeks remote , 2 weeks onsite)
Duration: Long term Contract with possibility of Conversion
Interview: 2 rounds, 1st β 60 mins Technical Panel, 2ND - 30 mins Manager call
What the Client is Looking For:
Core Technical Fit
β’ Software Engineering Strength (10+ yrs) β APIs, microservices, cloud deployments.
β’ Machine Learning Engineering (3β5 yrs) β experience building, deploying, and maintaining ML or GenAI solutions.
β’ RAG Expertise β must have built and deployed Retrieval-Augmented Generation pipelines.
β’ Vector Database Experience β FAISS, Pinecone, Weaviate, Milvus.
β’ Agent Frameworks β LangChain, CrewAI, LangGraph, AutoGen.
β’ Cloud Native Skills (AWS) β S3, Lambda, ECS, SageMaker.
β’ DevOps β Docker, Kubernetes, GitHub Actions, CI/CD.
β’ Observability β Prometheus, Grafana, OpenTelemetry (nice-to-have).
β’ Data & Knowledge Graphs β Snowflake, Oracle, Neo4j, RDF/SPARQL.
β’ AI Ethics & Governance β understanding of Responsible AI.
What the Project Is About:
This is a Machine Learning Engineering project within AI/ML division, focusing on integrating and productionizing AI/GenAI solutions.
Team Context:
β’ 15-member team: 6 are AI/ML specialists; 9 are on the BI (Business Intelligence) side.
β’ This is the first AI delivery team in the organization β meaning they are defining AI/ML standards, pipelines, and best practices for future teams.
Project Focus:
The goal is to:
β’ Deploy and scale AI/ML models (especially Generative AI and RAG-based solutions) into production.
β’ Integrate agentic or multi-agent systems using frameworks like LangChain, CrewAI, LangGraph, or AutoGen.
β’ Build cloud-native ML/GenAI pipelines on AWS (Lambda, ECS, S3, SageMaker).
β’ Establish data retrieval and augmentation systems (RAG) using vector databases (FAISS, Pinecone, Weaviate, Milvus).
β’ Develop monitoring, observability, and CI/CD practices for deployed ML models.
β’ Promote Responsible AI practices across AI ecosystem.
Essentially, this is a hands-on engineering role (not research-oriented) focused on:
βTurning research and experimental models into scalable, production-grade AI systems.β
Please find the position details below:
Job Title: Machine Learning Engineer with GEN AI experience
Location: Durham, NC or Boston MA, Merrimack NH & Smithfield RI ( 2 weeks remote , 2 weeks onsite)
Duration: Long term Contract with possibility of Conversion
Interview: 2 rounds, 1st β 60 mins Technical Panel, 2ND - 30 mins Manager call
What the Client is Looking For:
Core Technical Fit
β’ Software Engineering Strength (10+ yrs) β APIs, microservices, cloud deployments.
β’ Machine Learning Engineering (3β5 yrs) β experience building, deploying, and maintaining ML or GenAI solutions.
β’ RAG Expertise β must have built and deployed Retrieval-Augmented Generation pipelines.
β’ Vector Database Experience β FAISS, Pinecone, Weaviate, Milvus.
β’ Agent Frameworks β LangChain, CrewAI, LangGraph, AutoGen.
β’ Cloud Native Skills (AWS) β S3, Lambda, ECS, SageMaker.
β’ DevOps β Docker, Kubernetes, GitHub Actions, CI/CD.
β’ Observability β Prometheus, Grafana, OpenTelemetry (nice-to-have).
β’ Data & Knowledge Graphs β Snowflake, Oracle, Neo4j, RDF/SPARQL.
β’ AI Ethics & Governance β understanding of Responsible AI.
What the Project Is About:
This is a Machine Learning Engineering project within AI/ML division, focusing on integrating and productionizing AI/GenAI solutions.
Team Context:
β’ 15-member team: 6 are AI/ML specialists; 9 are on the BI (Business Intelligence) side.
β’ This is the first AI delivery team in the organization β meaning they are defining AI/ML standards, pipelines, and best practices for future teams.
Project Focus:
The goal is to:
β’ Deploy and scale AI/ML models (especially Generative AI and RAG-based solutions) into production.
β’ Integrate agentic or multi-agent systems using frameworks like LangChain, CrewAI, LangGraph, or AutoGen.
β’ Build cloud-native ML/GenAI pipelines on AWS (Lambda, ECS, S3, SageMaker).
β’ Establish data retrieval and augmentation systems (RAG) using vector databases (FAISS, Pinecone, Weaviate, Milvus).
β’ Develop monitoring, observability, and CI/CD practices for deployed ML models.
β’ Promote Responsible AI practices across AI ecosystem.
Essentially, this is a hands-on engineering role (not research-oriented) focused on:
βTurning research and experimental models into scalable, production-grade AI systems.β






