

Quantum World Technologies Inc.
Generative AI Lead
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Generative AI Lead, offering a remote contract position. Key requirements include expertise in LangChain, Claude Code, and Python/TypeScript, with experience deploying agent-based systems in production. Contract length and pay rate are unspecified.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
April 15, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Remote
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
United States
-
🧠 - Skills detailed
#GIT #Kafka (Apache Kafka) #CLI (Command-Line Interface) #Observability #Azure #GCP (Google Cloud Platform) #TypeScript #AWS (Amazon Web Services) #Debugging #Cloud #DevOps #REST (Representational State Transfer) #Strategy #AI (Artificial Intelligence) #Deployment #Consulting #Python #Containers #Langchain #Jupyter #Datasets #API (Application Programming Interface)
Role description
Job Title: Gen AI Leads
location : Remote
Key Responsibilities
Delivery & Architecture
• Own end-to-end delivery of AI-native programs - from architecture through production deployment
• Design and build multi-agent orchestration systems using LangChain, LangGraph, CrewAI, or equivalent
• Integrate agent systems with enterprise surfaces: APIs, ERPs, CRMs, data platforms - not toy datasets
• Define agent topology: tool routing, memory strategy, state machines, fallback handling
Agentic Coding & Development
• Run agentic coding workflows using Claude Code, Cursor, OpenAI Codex, or equivalent CLI tools
• Lead projects where AI writes significant portions of the codebase - and you guide, review, and ship it
• Work with CLAUDE.md, shared context frameworks, and multi-session agent setups for team use
• Debug non-deterministic agent outputs systematically - not by gut feel
Client & Stakeholder Engagement
• Translate business problems into agent architectures for global CXO-level stakeholders
• Run discovery workshops, solution reviews, and delivery cadences with client teams
• Prepare and present technical proposals, POC plans, and roadmaps - own the story end-to-end
Team & Practice
• Mentor junior AI engineers; raise AI engineering quality across the delivery team
• Stay current: evaluate new models, frameworks, and tooling before the hype catches up
• Contribute to internal knowledge bases, reusable frameworks, and accelerators
Skills
Agent Orchestration
LangChain, LangGraph, CrewAI - not just conceptual
Agentic Coding Tools
Claude Code CLI, Cursor, OpenAI Codex, Copilot
RAG & Vector Stores
Chroma, Weaviate, Pinecone - knows where RAG breaks
LLM APIs & SDKs
Anthropic, OpenAI, Gemini - prompt design, tool use
Python / TypeScript
Primary languages for agent + backend development
LangSmith / Observability
Tracing, evaluation, debugging agent runs
Cloud Platforms
Azure, AWS, GCP (at least one) - deployment, infra, managed services
API & System Integration
REST, gRPC, Kafka - enterprise integration patterns
MCP / Shared Context
Model Context Protocol, CLAUDE.md, Beads
Agent Evaluation
Testing non-deterministic outputs, guardrails, evals
CI/CD & DevOps
Git, containers, pipelines - agents need to ship
Client Communication
Can present architecture to a CXO without jargon
What You Must Have Actually Done
Not just what you know. What you have shipped.
• Deployed 2–3 agent-based systems in production - stateful, multi-step, real users
• Used LangGraph for multi-agent orchestration with memory, tool routing, and state management
• Built projects where AI (Claude Code, Codex, Cursor) wrote significant portions of the code
• Implemented RAG pipelines end-to-end - chunking, embedding, retrieval, re-ranking, evaluation
• Integrated agents with real enterprise APIs - not just OpenAI playground or sample data
• Debugged a production agent failure - and fixed it without blaming the model
• Can articulate when NOT to use agents - that is how we know you have built things
Bonus - Real Differentiators
• Experience with Claude Code CLI in team environments (CLAUDE.md, shared context, multi-session flows)
• Familiarity with LangSmith for agent tracing, evaluation pipelines, and debugging at scale
• Has shipped something using MCP (Model Context Protocol) or similar shared-context tooling
• QA/testing mindset for agents - systematic evaluation of non-deterministic outputs
• Background in IT services or consulting - managing client expectations while building
• Experience with SLMs, fine-tuning, or on-device/edge agent deployment
What We Are Not Looking For
• Someone who lists LLMs on a resume but has only called the API in a Jupyter notebook
• AI enthusiasts whose hands-on experience is less than a year old
• People who explain everything in terms of frameworks they have never deployed
• Consultants who can only narrate what others have built
Job Title: Gen AI Leads
location : Remote
Key Responsibilities
Delivery & Architecture
• Own end-to-end delivery of AI-native programs - from architecture through production deployment
• Design and build multi-agent orchestration systems using LangChain, LangGraph, CrewAI, or equivalent
• Integrate agent systems with enterprise surfaces: APIs, ERPs, CRMs, data platforms - not toy datasets
• Define agent topology: tool routing, memory strategy, state machines, fallback handling
Agentic Coding & Development
• Run agentic coding workflows using Claude Code, Cursor, OpenAI Codex, or equivalent CLI tools
• Lead projects where AI writes significant portions of the codebase - and you guide, review, and ship it
• Work with CLAUDE.md, shared context frameworks, and multi-session agent setups for team use
• Debug non-deterministic agent outputs systematically - not by gut feel
Client & Stakeholder Engagement
• Translate business problems into agent architectures for global CXO-level stakeholders
• Run discovery workshops, solution reviews, and delivery cadences with client teams
• Prepare and present technical proposals, POC plans, and roadmaps - own the story end-to-end
Team & Practice
• Mentor junior AI engineers; raise AI engineering quality across the delivery team
• Stay current: evaluate new models, frameworks, and tooling before the hype catches up
• Contribute to internal knowledge bases, reusable frameworks, and accelerators
Skills
Agent Orchestration
LangChain, LangGraph, CrewAI - not just conceptual
Agentic Coding Tools
Claude Code CLI, Cursor, OpenAI Codex, Copilot
RAG & Vector Stores
Chroma, Weaviate, Pinecone - knows where RAG breaks
LLM APIs & SDKs
Anthropic, OpenAI, Gemini - prompt design, tool use
Python / TypeScript
Primary languages for agent + backend development
LangSmith / Observability
Tracing, evaluation, debugging agent runs
Cloud Platforms
Azure, AWS, GCP (at least one) - deployment, infra, managed services
API & System Integration
REST, gRPC, Kafka - enterprise integration patterns
MCP / Shared Context
Model Context Protocol, CLAUDE.md, Beads
Agent Evaluation
Testing non-deterministic outputs, guardrails, evals
CI/CD & DevOps
Git, containers, pipelines - agents need to ship
Client Communication
Can present architecture to a CXO without jargon
What You Must Have Actually Done
Not just what you know. What you have shipped.
• Deployed 2–3 agent-based systems in production - stateful, multi-step, real users
• Used LangGraph for multi-agent orchestration with memory, tool routing, and state management
• Built projects where AI (Claude Code, Codex, Cursor) wrote significant portions of the code
• Implemented RAG pipelines end-to-end - chunking, embedding, retrieval, re-ranking, evaluation
• Integrated agents with real enterprise APIs - not just OpenAI playground or sample data
• Debugged a production agent failure - and fixed it without blaming the model
• Can articulate when NOT to use agents - that is how we know you have built things
Bonus - Real Differentiators
• Experience with Claude Code CLI in team environments (CLAUDE.md, shared context, multi-session flows)
• Familiarity with LangSmith for agent tracing, evaluation pipelines, and debugging at scale
• Has shipped something using MCP (Model Context Protocol) or similar shared-context tooling
• QA/testing mindset for agents - systematic evaluation of non-deterministic outputs
• Background in IT services or consulting - managing client expectations while building
• Experience with SLMs, fine-tuning, or on-device/edge agent deployment
What We Are Not Looking For
• Someone who lists LLMs on a resume but has only called the API in a Jupyter notebook
• AI enthusiasts whose hands-on experience is less than a year old
• People who explain everything in terms of frameworks they have never deployed
• Consultants who can only narrate what others have built





