

Senior Data Scientist
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Scientist with a contract length of "unknown", offering a pay rate of "$X/hour". Key skills include machine learning, GenAI, Python, Jupyter, and cloud-native development. Requires 10+ years in AI/ML, with 3+ years in GenAI.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
-
🗓️ - Date discovered
August 29, 2025
🕒 - Project duration
Unknown
-
🏝️ - Location type
Unknown
-
📄 - Contract type
Unknown
-
🔒 - Security clearance
Unknown
-
📍 - Location detailed
McLean, VA
-
🧠 - Skills detailed
#MLflow #"ETL (Extract #Transform #Load)" #Langchain #AWS SageMaker #Data Science #Programming #Libraries #Agile #AI (Artificial Intelligence) #Databases #Kubernetes #AWS (Amazon Web Services) #Jupyter #Transformers #Spark (Apache Spark) #Deployment #Python #Apache Spark #ML (Machine Learning) #Scala #PySpark #Cloud #Azure #Data Engineering #SageMaker #API (Application Programming Interface) #Base
Role description
Must Have Qualifications: Must have hands on experience with machine learning transitioned into GenAI. Rag, Python- Jupyter, other Software knowledge, using agents in workflows, strong understanding of data. Preferred: Built AI agent, MCP, A2A, Graph Rag, deployed Gen AI applications to production.
Responsibilities
• Architect and implement scalable AI Agents, Agentic Workflows and GenAI applications to address diverse and complex business use cases.
• Develop, fine-tune, and optimize lightweight LLMs; lead the evaluation and adaptation of models such as Claude (Anthropic), Azure OpenAI, and open-source alternatives.
• Design and deploy Retrieval-Augmented Generation (RAG) and Graph RAG systems using vector databases and knowledge bases.
• Curate enterprise data using connectors integrated with AWS Bedrock's Knowledge Base/Elastic
• Implement solutions leveraging MCP (Model Context Protocol) and A2A (Agent-to-Agent) communication.
• Build and maintain Jupyter-based notebooks using platforms like SageMaker and MLFlow/Kubeflow on Kubernetes (EKS).
• Collaborate with cross-functional teams of UI and microservice engineers, designers, and data engineers to build full-stack Gen AI experiences.
• Integrate GenAI solutions with enterprise platforms via API-based methods and GenAI standardized patterns.
• Establish and enforce validation procedures with Evaluation Frameworks, bias mitigation, safety protocols, and guardrails for production-ready deployment.
• Design & build robust ingestion pipelines that extract, chunk, enrich, and anonymize data from PDFs, video, and audio sources for use in LLM-powered workflows—leveraging best practices like semantic chunking and privacy controls
• Orchestrate multimodal pipelines
•
• using scalable frameworks (e.g., Apache Spark, PySpark) for automated ETL/ELT workflows appropriate for unstructured media
• Implement embeddings drives—map media content to vector representations using embedding models, and integrate with vector stores (AWS KnowledgeBase/Elastic/Mongo Atlas) to support RAG architectures
•
• Required Qualifications:
•
•
• 10+ years of experience in AI/ML, with 3+ years in applied GenAI or LLM-based solutions.
• Deep expertise in prompt engineering, fine-tuning, RAG, GraphRAG, vector databases (e.g., AWS KnowledgeBase / Elastic), and multi-modal models.
• Proven experience with cloud-native AI development (AWS SageMaker, Bedrock, MLFlow on EKS).
• Strong programming skills in Python and ML libraries (Transformers, LangChain, etc.).
• Deep understanding of Gen AI system patterns and architectural best practices, Evaluation Frameworks
• Demonstrated ability to work in cross-functional agile teams.
Must Have Qualifications: Must have hands on experience with machine learning transitioned into GenAI. Rag, Python- Jupyter, other Software knowledge, using agents in workflows, strong understanding of data. Preferred: Built AI agent, MCP, A2A, Graph Rag, deployed Gen AI applications to production.
Responsibilities
• Architect and implement scalable AI Agents, Agentic Workflows and GenAI applications to address diverse and complex business use cases.
• Develop, fine-tune, and optimize lightweight LLMs; lead the evaluation and adaptation of models such as Claude (Anthropic), Azure OpenAI, and open-source alternatives.
• Design and deploy Retrieval-Augmented Generation (RAG) and Graph RAG systems using vector databases and knowledge bases.
• Curate enterprise data using connectors integrated with AWS Bedrock's Knowledge Base/Elastic
• Implement solutions leveraging MCP (Model Context Protocol) and A2A (Agent-to-Agent) communication.
• Build and maintain Jupyter-based notebooks using platforms like SageMaker and MLFlow/Kubeflow on Kubernetes (EKS).
• Collaborate with cross-functional teams of UI and microservice engineers, designers, and data engineers to build full-stack Gen AI experiences.
• Integrate GenAI solutions with enterprise platforms via API-based methods and GenAI standardized patterns.
• Establish and enforce validation procedures with Evaluation Frameworks, bias mitigation, safety protocols, and guardrails for production-ready deployment.
• Design & build robust ingestion pipelines that extract, chunk, enrich, and anonymize data from PDFs, video, and audio sources for use in LLM-powered workflows—leveraging best practices like semantic chunking and privacy controls
• Orchestrate multimodal pipelines
•
• using scalable frameworks (e.g., Apache Spark, PySpark) for automated ETL/ELT workflows appropriate for unstructured media
• Implement embeddings drives—map media content to vector representations using embedding models, and integrate with vector stores (AWS KnowledgeBase/Elastic/Mongo Atlas) to support RAG architectures
•
• Required Qualifications:
•
•
• 10+ years of experience in AI/ML, with 3+ years in applied GenAI or LLM-based solutions.
• Deep expertise in prompt engineering, fine-tuning, RAG, GraphRAG, vector databases (e.g., AWS KnowledgeBase / Elastic), and multi-modal models.
• Proven experience with cloud-native AI development (AWS SageMaker, Bedrock, MLFlow on EKS).
• Strong programming skills in Python and ML libraries (Transformers, LangChain, etc.).
• Deep understanding of Gen AI system patterns and architectural best practices, Evaluation Frameworks
• Demonstrated ability to work in cross-functional agile teams.