Data Scientist

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Scientist specializing as a Principal GenAI Scientist in McLean, VA, with a 6+ month W2 contract. Requires a PhD, 10+ years in AI/ML, and expertise in Generative AI, LLMs, and AWS tools.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
-
πŸ—“οΈ - Date discovered
September 25, 2025
πŸ•’ - Project duration
More than 6 months
-
🏝️ - Location type
On-site
-
πŸ“„ - Contract type
W2 Contractor
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
McLean, VA
-
🧠 - Skills detailed
#MLflow #Data Science #Base #Deployment #Langchain #SageMaker #Kubernetes #AWS (Amazon Web Services) #"ETL (Extract #Transform #Load)" #Transformers #GitHub #Azure #Agile #Leadership #Scala #ML Ops (Machine Learning Operations) #API (Application Programming Interface) #AI (Artificial Intelligence) #Libraries #ML (Machine Learning) #Jupyter #Spark (Apache Spark) #Databases #PySpark #Python
Role description
πŸ’Ό Data Scientist Specialist – Principal GenAI Scientist (Consultant) πŸ“ McLean, VA | Full-Time Onsite (M–F) | W2 Contract Only ⏳ Contract | 2 Open Positions | Potential Long-Term Position Summary: Our financial client is seeking a highly experienced Principal GenAI Scientist with deep expertise in Generative AI, AI Agents, and Agentic Workflows to design and implement cutting-edge enterprise-grade GenAI applications. This role requires strong technical leadership, hands-on development, and collaboration across engineering, product, and data teams. You will work on LLMs, RAG, Graph RAG, MCP, A2A, multimodal AI pipelines, and AWS-based deployments, helping the enterprise shape the next generation of AI-driven business solutions. Key Responsibilities: β€’ Architect and implement scalable AI Agents, Agentic Workflows, and GenAI applications for complex business use cases. β€’ Fine-tune and optimize lightweight LLMs; evaluate/adapt models such as Claude (Anthropic), Azure OpenAI, and open-source LLMs. β€’ Design/deploy RAG and Graph RAG systems using vector databases (AWS Knowledge Base, Elastic, Mongo Atlas). β€’ Curate enterprise data using AWS Bedrock connectors and integrate with knowledge bases. β€’ Implement MCP (Model Context Protocol) and A2A (Agent-to-Agent) communication workflows. β€’ Develop Jupyter-based notebooks using SageMaker, MLFlow/Kubeflow on Kubernetes (EKS). β€’ Collaborate with cross-functional teams to build full-stack GenAI solutions with API integrations. β€’ Establish guardrails, validation procedures, and safety protocols for production-ready GenAI deployments. β€’ Design ingestion pipelines for structured/unstructured data (PDF, video, audio) with semantic chunking and privacy controls. β€’ Build multimodal pipelines using Spark/PySpark for automated ETL/ELT workflows. β€’ Implement embeddings pipelines for RAG architectures with vector stores. Required Qualifications: β€’ PhD in AI/Data Science (or equivalent experience). β€’ 10+ years in AI/ML, including 3+ years applied GenAI/LLM solutions. β€’ Hands-on expertise in prompt engineering, fine-tuning, RAG, Graph RAG, vector databases, and multimodal AI models. β€’ Strong Python development with ML libraries (Transformers, LangChain, etc.). β€’ Experience with AWS AI/ML tools (SageMaker, Bedrock, MLFlow on EKS). β€’ Proven track record in GenAI architectural patterns, evaluation frameworks, and bias mitigation. β€’ Ability to collaborate in cross-functional agile teams. β€’ GitHub repository link required for each candidate submission. Preferred Qualifications: β€’ Published contributions or patents in AI/ML/LLM. β€’ Experience with AI governance and ethical deployment frameworks. β€’ Familiarity with CI/CD for ML Ops and scalable inference APIs. Additional Details: β€’ πŸ“Œ Location: McLean, VA (Onsite M–F) β€’ πŸ“Œ Type: W2 Contract Only