

Generative AI Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Generative AI Engineer based in Mountain View, CA, for 12 months, offering a competitive pay rate. Key skills include Python, LangChain, cloud LLM providers, and workflow orchestration. Experience in RAG pipelines and prompt engineering is essential.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
560
-
ποΈ - Date discovered
September 24, 2025
π - Project duration
More than 6 months
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Mountain View, CA
-
π§ - Skills detailed
#Security #Observability #Airflow #Azure #AI (Artificial Intelligence) #Langchain #Monitoring #Python #AWS (Amazon Web Services) #Cloud
Role description
GenAI Engineer
Location: Mountain View , CA
Duration: 12 months
The project is focused on building production-grade GenAI solutions with emphasis on:
β’ RAG pipelines leveraging LLMs
β’ Prompt engineering (system/tool prompts, function calling, versioning with evals)
β’ Evaluation & observability (ground truth setup, confusion metrics, LLM-as-judge with human review, cost & latency monitoring)
β’ Retrieval strategies & prompt patterns (context management, hallucination mitigation)
β’ LangChain/LlamaIndex (or equivalent) proficiency
β’ Cloud LLM providers (Azure OpenAI, AWS Bedrock, Vertex AI)
β’ Workflow orchestration (Airflow, Dagster)
β’ Security & privacy (PII handling, RBAC)
β’ Python expertise with strong practical experience
Success criteria:
β’ A production-ready prompt + RAG pipeline system with measurable quality lift (accuracy/confusion metrics) and reduced hallucinations, backed by a repeatable evaluation harness.
GenAI Engineer
Location: Mountain View , CA
Duration: 12 months
The project is focused on building production-grade GenAI solutions with emphasis on:
β’ RAG pipelines leveraging LLMs
β’ Prompt engineering (system/tool prompts, function calling, versioning with evals)
β’ Evaluation & observability (ground truth setup, confusion metrics, LLM-as-judge with human review, cost & latency monitoring)
β’ Retrieval strategies & prompt patterns (context management, hallucination mitigation)
β’ LangChain/LlamaIndex (or equivalent) proficiency
β’ Cloud LLM providers (Azure OpenAI, AWS Bedrock, Vertex AI)
β’ Workflow orchestration (Airflow, Dagster)
β’ Security & privacy (PII handling, RBAC)
β’ Python expertise with strong practical experience
Success criteria:
β’ A production-ready prompt + RAG pipeline system with measurable quality lift (accuracy/confusion metrics) and reduced hallucinations, backed by a repeatable evaluation harness.