

Wall Street Consulting Services LLC
GEN AI Lead
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a GEN AI Lead with a contract length of "unknown" and a pay rate of "unknown." Located remotely, it requires 14+ years in software/ML engineering, Python proficiency, and expertise in LLM/SLM/GenAI solutions.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
April 7, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Unknown
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Warren, NJ
-
🧠 - Skills detailed
#Compliance #Model Deployment #DevOps #Data Quality #Databases #Docker #Regression #Kubernetes #Security #ML (Machine Learning) #TypeScript #GitHub #Batch #Datasets #Containers #Scala #Azure #Infrastructure as Code (IaC) #AI (Artificial Intelligence) #Azure DevOps #Data Pipeline #Observability #Deployment #AWS (Amazon Web Services) #Terraform #Langchain #Cloud #MLflow #Prometheus #Python #GCP (Google Cloud Platform) #Quality Assurance
Role description
Job Title
GenAI / SLM Lead Engineer
About the Role
We are looking for a GenAI / Small Language Model (SLM) Lead Engineer to design, deploy, and maintain agentic AI solutions that are safe, scalable, and business‑ready. You will own end‑to‑end delivery—from prompt and agent design to data pipelines, model deployment, observability, and rigorous validation—partnering with product, architecture, security, and QA to ship AI features that perform reliably in production.
• Update - Python Coding Experience is required. Someone who is well versed with understanding and coding Python, make changes add/change rules specially insurance related and redeploy not just create code and deploy - deep expertise in writing, understand, changing, inventing, debug and deploy AI agent using python.
What You’ll Do
SLM Design and Fine Tunning
Collect, clean, and preprocess domain-specific datasets for SLM training and fine-tuning.
Ensure data quality, diversity, and compliance with privacy and security standards.
Fine-tune small language models on curated datasets using techniques like LoRA, adapters, or
parameter-efficient tuning.
Optimize hyperparameters for performance, latency, and resource efficiency.
Agent & Prompt Implementation
Help design and implement agent orchestration (single and multi‑agent) and function/tool use
strategies.
Craft, version, and optimize prompts and system instructions for accuracy, coherence, and
domain alignment.
Integrate external tools/APIs and establish content‑safety guardrails (e.g., policy enforcement, PII
redaction, jailbreak prevention).
Implementation, Testing & Maintenance
Build resilient agent workflows and services; harden reliability with retries, fallbacks, circuit
breakers.
Develop automated tests for prompts, tools, and agent behaviors; maintain regression suites and
golden datasets.
Operate AI services in production: performance tuning, cost optimization, incident response, and
iterative improvement.
Data & MLOps
Design and manage data pipelines for fine‑tuning and retrieval (RAG), including cleansing,
labeling, and governance.
Monitor drift, quality, latency, and safety signals; implement model/agent observability and
alerting.
Quality Assurance & Risk
Run structured evaluations of agent outputs (functional, coherence, safety, bias); track
precision/recall and hallucination rates.
Perform risk assessments for agent behaviors and tool actions; document mitigations and
approval workflows.
Collaborate with security/compliance to meet regulatory, privacy, and usage‑policy requirements.
Minimum Qualifications
14 + years in software/ML engineering, with 2+ years building LLM/SLM/GenAI solutions in
production.
Proficiency in Python (and/or TypeScript) and modern AI orchestration frameworks (e.g., Microsoft
Agent Framework, Google Agent Development Kit, LangChain, Semantic Kernel).
Hands‑on with retrieval‑augmented generation (RAG), function calling, prompt optimization, and
agent design patterns.
Experience building data pipelines (batch/stream), and managing datasets for training/fine‑tuning and
evaluation.
Practical understanding of AI guardrails: content filtering, safety policies, redaction, rate limiting, and
misuse prevention.
Strong willingness to learn advanced agent orchestration and MLOps practices.
Preferred Qualifications
MLOps fluency: model packaging, CI/CD, experiment tracking (e.g., MLflow), deployment on
cloud/container platforms.
IaC (e.g., Terraform/Bicep) and DevOps tooling (e.g., GitHub Actions/Azure DevOps); strong grasp of
observability.
Experience with multi‑agent systems, toolformer patterns, and complex orchestration graphs.
Knowledge of vector databases and retrieval systems; evaluation frameworks (e.g., Ragas, DeepEval)
and custom metrics.
Familiarity with privacy, compliance, and model risk management practices for AI.
Background in tuning open‑source and hosted models; comfort with hybrid cloud environments.
Tools & Technologies
Python; TypeScript; MAF/Google ADK/LangChain/Semantic Kernel; Vector DBs and frameworks (e.g.,
Qdrant/FAISS/Pinecone); CI/CD (GitHub Actions/Azure DevOps); IaC (Terraform/Bicep); Observability
(OpenTelemetry/Prometheus); Experiment tracking (MLflow); Cloud AI services (e.g., Azure AI FGoundry, Azure OpenAI, GCP Vertex AI, AWS Bedrock); Containers (Docker/Kubernetes).
Working Model
Partner with Product, Architecture, Security, and QA to plan, design, and ship safe AI features.
Contribute to internal prompt standards, evaluation datasets, and reuseable components.
Document designs, decisions, and risks; mentor peers and champion responsible AI practices.
Job Title
GenAI / SLM Lead Engineer
About the Role
We are looking for a GenAI / Small Language Model (SLM) Lead Engineer to design, deploy, and maintain agentic AI solutions that are safe, scalable, and business‑ready. You will own end‑to‑end delivery—from prompt and agent design to data pipelines, model deployment, observability, and rigorous validation—partnering with product, architecture, security, and QA to ship AI features that perform reliably in production.
• Update - Python Coding Experience is required. Someone who is well versed with understanding and coding Python, make changes add/change rules specially insurance related and redeploy not just create code and deploy - deep expertise in writing, understand, changing, inventing, debug and deploy AI agent using python.
What You’ll Do
SLM Design and Fine Tunning
Collect, clean, and preprocess domain-specific datasets for SLM training and fine-tuning.
Ensure data quality, diversity, and compliance with privacy and security standards.
Fine-tune small language models on curated datasets using techniques like LoRA, adapters, or
parameter-efficient tuning.
Optimize hyperparameters for performance, latency, and resource efficiency.
Agent & Prompt Implementation
Help design and implement agent orchestration (single and multi‑agent) and function/tool use
strategies.
Craft, version, and optimize prompts and system instructions for accuracy, coherence, and
domain alignment.
Integrate external tools/APIs and establish content‑safety guardrails (e.g., policy enforcement, PII
redaction, jailbreak prevention).
Implementation, Testing & Maintenance
Build resilient agent workflows and services; harden reliability with retries, fallbacks, circuit
breakers.
Develop automated tests for prompts, tools, and agent behaviors; maintain regression suites and
golden datasets.
Operate AI services in production: performance tuning, cost optimization, incident response, and
iterative improvement.
Data & MLOps
Design and manage data pipelines for fine‑tuning and retrieval (RAG), including cleansing,
labeling, and governance.
Monitor drift, quality, latency, and safety signals; implement model/agent observability and
alerting.
Quality Assurance & Risk
Run structured evaluations of agent outputs (functional, coherence, safety, bias); track
precision/recall and hallucination rates.
Perform risk assessments for agent behaviors and tool actions; document mitigations and
approval workflows.
Collaborate with security/compliance to meet regulatory, privacy, and usage‑policy requirements.
Minimum Qualifications
14 + years in software/ML engineering, with 2+ years building LLM/SLM/GenAI solutions in
production.
Proficiency in Python (and/or TypeScript) and modern AI orchestration frameworks (e.g., Microsoft
Agent Framework, Google Agent Development Kit, LangChain, Semantic Kernel).
Hands‑on with retrieval‑augmented generation (RAG), function calling, prompt optimization, and
agent design patterns.
Experience building data pipelines (batch/stream), and managing datasets for training/fine‑tuning and
evaluation.
Practical understanding of AI guardrails: content filtering, safety policies, redaction, rate limiting, and
misuse prevention.
Strong willingness to learn advanced agent orchestration and MLOps practices.
Preferred Qualifications
MLOps fluency: model packaging, CI/CD, experiment tracking (e.g., MLflow), deployment on
cloud/container platforms.
IaC (e.g., Terraform/Bicep) and DevOps tooling (e.g., GitHub Actions/Azure DevOps); strong grasp of
observability.
Experience with multi‑agent systems, toolformer patterns, and complex orchestration graphs.
Knowledge of vector databases and retrieval systems; evaluation frameworks (e.g., Ragas, DeepEval)
and custom metrics.
Familiarity with privacy, compliance, and model risk management practices for AI.
Background in tuning open‑source and hosted models; comfort with hybrid cloud environments.
Tools & Technologies
Python; TypeScript; MAF/Google ADK/LangChain/Semantic Kernel; Vector DBs and frameworks (e.g.,
Qdrant/FAISS/Pinecone); CI/CD (GitHub Actions/Azure DevOps); IaC (Terraform/Bicep); Observability
(OpenTelemetry/Prometheus); Experiment tracking (MLflow); Cloud AI services (e.g., Azure AI FGoundry, Azure OpenAI, GCP Vertex AI, AWS Bedrock); Containers (Docker/Kubernetes).
Working Model
Partner with Product, Architecture, Security, and QA to plan, design, and ship safe AI features.
Contribute to internal prompt standards, evaluation datasets, and reuseable components.
Document designs, decisions, and risks; mentor peers and champion responsible AI practices.






