

The AES Group
AI Engineer Level II
β - Featured Role | Apply direct with Data Freelance Hub
This role is for an AI Engineer Level II in Washington, DC, offering a contract length of "unknown" at a pay rate of "unknown." Key skills include GenAI architecture, Azure and AWS services, and strong proficiency in C# and Python. Required certifications include Microsoft Azure AI Fundamentals and AWS Machine Learning Specialty, along with 5+ years of software engineering experience and 2+ years in GenAI/LLM applications.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
February 5, 2026
π - Duration
Unknown
-
ποΈ - Location
On-site
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Washington, DC
-
π§ - Skills detailed
#S3 (Amazon Simple Storage Service) #MLflow #AWS Machine Learning #Data Pipeline #Metadata #Scala #Azure DevOps #Data Enrichment #TensorFlow #Databricks #Langchain #"ETL (Extract #Transform #Load)" #API (Application Programming Interface) #AWS (Amazon Web Services) #Agile #AWS SageMaker #ADF (Azure Data Factory) #ML (Machine Learning) #Normalization #AWS EMR (Amazon Elastic MapReduce) #Monitoring #AI (Artificial Intelligence) #Hugging Face #Python #C# #Data Science #A/B Testing #Security #Docker #Azure #Cloud #Hadoop #Logging #Deployment #Kubernetes #.Net #Vault #Batch #Redis #Azure Data Factory #Lambda (AWS Lambda) #Spark (Apache Spark) #DevOps #SageMaker #TypeScript
Role description
Role: AI Engineer β Level II
Location: Washington, DC - Onsite
Position Summary
As an AI Engineer (Level II), youβll design and implement enterprise-grade AI systems with a focus on Retrieval-Augmented Generation (RAG), agentic AI, and cloud-native ML pipelines. You'll work cross-functionally to operationalize secure, scalable solutions across Azure and AWS platforms, contributing to production-ready, multi-modal GenAI applications.
Key Responsibilities
AI Architecture & Delivery
β’ Design and deploy RAG pipelines using Azure AI/Search and vector DBs (Redis, FAISS, HNSW).
β’ Develop conversational AI systems with prompt lifecycle management, telemetry, and guardrails.
β’ Integrate LLMs like Azure OpenAI, Llama, Claude, and OSS models across vision and speech domains.
Infrastructure & Orchestration
β’ Implement Model Context Protocol (MCP) servers with RBAC, schema versioning, validation, and audit trails.
β’ Deploy Azure AI Agent Service patterns: agent registry, policy enforcement, and telemetry logging.
β’ Use Azure Batch and AWS EMR for parallel inferencing and distributed feature processing.
Data Pipeline Engineering
β’ Build and manage ingestion pipelines: document normalization, metadata enrichment, PII redaction, SLA monitoring.
β’ Operate scalable vectorization pipelines with drift detection and quality gates.
β’ Use Azure Data Factory and Databricks; AWS EMR for large-scale Hadoop/Spark workloads.
Agentic AI Development
β’ Implement secure tool-calling and multi-agent orchestration using Semantic Kernel, AutoGen, Agent Framework, CrewAI, Agno, and LangChain.
β’ Apply governance, telemetry, and lifecycle management across agent runtimes with MCP controls.
Model Ops & Evaluation
β’ Fine-tune and evaluate OSS and proprietary models; conduct A/B tests and latency/cost analysis.
β’ Implement CI/CD pipelines with security scans and validation for AI/LLM workloads.
Software Engineering Core
β’ Proficiency in CS fundamentals: algorithms, distributed systems, concurrency, networking.
β’ Experience with SDLC excellence: clean architecture, SOLID, testing pyramids (unit, integration, E2E).
β’ Secure AI app development: input validation, secret hygiene, RBAC, sandboxed functions.
β’ Performance engineering: latency tuning, token optimization, vector index profiling.
Cloud AI Tech Stack
Azure: Azure OpenAI, AI/Search, AML, AKS, Azure Functions, Key Vault, ADF, Databricks, Azure Batch
AWS: SageMaker, Bedrock, Lambda, EMR, Comprehend, API Gateway, S3, EKS
Vector DBs: Azure AI Search, Redis, FAISS/HNSW
Frameworks: Semantic Kernel, AutoGen, Microsoft Agent Framework, CrewAI, Agno, LangChain
Inference: Docker/Ollama, vLLM, GPU provisioning, quantization (GGUF)
Qualifications
Education: Bachelorβs in CS, Engineering, or equivalent hands-on expertise
Experience: 5+ years in software engineering; 2+ years in GenAI/LLM applications (RAG, agents, safety, eval)
Certifications (Required)
β’ Microsoft Certified: Azure AI Fundamentals (AI-900)
β’ Microsoft Certified: Azure Data Fundamentals (DP-900)
β’ Responsible AI certifications
β’ AWS Machine Learning Specialty
β’ TensorFlow Developer
β’ Kubernetes CKA or CKAD
β’ SAFe Agile Software Engineering
Preferred
β’ Azure AI Engineer Associate (AI-102)
β’ Azure Data Scientist Associate (DP-100)
β’ Azure Solutions Architect (AZ-305)
β’ Azure Developer Associate (AZ-204)
Required Skills/Abilities
β’ GenAI architecture mastery: RAG, vector DBs, embeddings, transformer internals, multi-modal pipelines.
β’ Agentic systems: Azure AI Agent Service patterns, MCP servers, registry/broker/governance, secure tool-calling.
β’ Languages: C# and Python (production-grade), .Net, plus TypeScript for service/UI when needed.
β’ Azure & AWS services (see Knowledge Requirements) with hands-on implementation and operations.
β’ Model ops: eval suites, safety tooling, fine-tuning, guardrails, traceability.
β’ Business & delivery: solution architecture, stakeholder alignment, roadmap planning, measurable impact.
Desired Skills/Abilities (not Required But a Plus)
β’ LangChain, Hugging Face, MLflow; Kubernetes + GPU scheduling; vector search tuning (HNSW/IVF).
β’ Responsible AI: policy mapping, red-team playbooks, incident response for AI.
β’ Hybrid/multi-cloud deployments using Azure Arc and AWS Outposts; CI/CD for AI workloads across Azure DevOps and AWS CodePipeline.
Step into a high-impact role where AI meets cloud scalability. Apply now and bring cutting-edge solutions to life.
Role: AI Engineer β Level II
Location: Washington, DC - Onsite
Position Summary
As an AI Engineer (Level II), youβll design and implement enterprise-grade AI systems with a focus on Retrieval-Augmented Generation (RAG), agentic AI, and cloud-native ML pipelines. You'll work cross-functionally to operationalize secure, scalable solutions across Azure and AWS platforms, contributing to production-ready, multi-modal GenAI applications.
Key Responsibilities
AI Architecture & Delivery
β’ Design and deploy RAG pipelines using Azure AI/Search and vector DBs (Redis, FAISS, HNSW).
β’ Develop conversational AI systems with prompt lifecycle management, telemetry, and guardrails.
β’ Integrate LLMs like Azure OpenAI, Llama, Claude, and OSS models across vision and speech domains.
Infrastructure & Orchestration
β’ Implement Model Context Protocol (MCP) servers with RBAC, schema versioning, validation, and audit trails.
β’ Deploy Azure AI Agent Service patterns: agent registry, policy enforcement, and telemetry logging.
β’ Use Azure Batch and AWS EMR for parallel inferencing and distributed feature processing.
Data Pipeline Engineering
β’ Build and manage ingestion pipelines: document normalization, metadata enrichment, PII redaction, SLA monitoring.
β’ Operate scalable vectorization pipelines with drift detection and quality gates.
β’ Use Azure Data Factory and Databricks; AWS EMR for large-scale Hadoop/Spark workloads.
Agentic AI Development
β’ Implement secure tool-calling and multi-agent orchestration using Semantic Kernel, AutoGen, Agent Framework, CrewAI, Agno, and LangChain.
β’ Apply governance, telemetry, and lifecycle management across agent runtimes with MCP controls.
Model Ops & Evaluation
β’ Fine-tune and evaluate OSS and proprietary models; conduct A/B tests and latency/cost analysis.
β’ Implement CI/CD pipelines with security scans and validation for AI/LLM workloads.
Software Engineering Core
β’ Proficiency in CS fundamentals: algorithms, distributed systems, concurrency, networking.
β’ Experience with SDLC excellence: clean architecture, SOLID, testing pyramids (unit, integration, E2E).
β’ Secure AI app development: input validation, secret hygiene, RBAC, sandboxed functions.
β’ Performance engineering: latency tuning, token optimization, vector index profiling.
Cloud AI Tech Stack
Azure: Azure OpenAI, AI/Search, AML, AKS, Azure Functions, Key Vault, ADF, Databricks, Azure Batch
AWS: SageMaker, Bedrock, Lambda, EMR, Comprehend, API Gateway, S3, EKS
Vector DBs: Azure AI Search, Redis, FAISS/HNSW
Frameworks: Semantic Kernel, AutoGen, Microsoft Agent Framework, CrewAI, Agno, LangChain
Inference: Docker/Ollama, vLLM, GPU provisioning, quantization (GGUF)
Qualifications
Education: Bachelorβs in CS, Engineering, or equivalent hands-on expertise
Experience: 5+ years in software engineering; 2+ years in GenAI/LLM applications (RAG, agents, safety, eval)
Certifications (Required)
β’ Microsoft Certified: Azure AI Fundamentals (AI-900)
β’ Microsoft Certified: Azure Data Fundamentals (DP-900)
β’ Responsible AI certifications
β’ AWS Machine Learning Specialty
β’ TensorFlow Developer
β’ Kubernetes CKA or CKAD
β’ SAFe Agile Software Engineering
Preferred
β’ Azure AI Engineer Associate (AI-102)
β’ Azure Data Scientist Associate (DP-100)
β’ Azure Solutions Architect (AZ-305)
β’ Azure Developer Associate (AZ-204)
Required Skills/Abilities
β’ GenAI architecture mastery: RAG, vector DBs, embeddings, transformer internals, multi-modal pipelines.
β’ Agentic systems: Azure AI Agent Service patterns, MCP servers, registry/broker/governance, secure tool-calling.
β’ Languages: C# and Python (production-grade), .Net, plus TypeScript for service/UI when needed.
β’ Azure & AWS services (see Knowledge Requirements) with hands-on implementation and operations.
β’ Model ops: eval suites, safety tooling, fine-tuning, guardrails, traceability.
β’ Business & delivery: solution architecture, stakeholder alignment, roadmap planning, measurable impact.
Desired Skills/Abilities (not Required But a Plus)
β’ LangChain, Hugging Face, MLflow; Kubernetes + GPU scheduling; vector search tuning (HNSW/IVF).
β’ Responsible AI: policy mapping, red-team playbooks, incident response for AI.
β’ Hybrid/multi-cloud deployments using Azure Arc and AWS Outposts; CI/CD for AI workloads across Azure DevOps and AWS CodePipeline.
Step into a high-impact role where AI meets cloud scalability. Apply now and bring cutting-edge solutions to life.






