Data Scientist- Gen AI

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Scientist-Gen AI, offering a contract of "length" at a pay rate of "$/hour". The position is remote, requiring 7+ years of Data Science experience, strong Python and SQL skills, and proven GenAI product delivery.
🌎 - Country
United Kingdom
💱 - Currency
£ GBP
-
💰 - Day rate
-
🗓️ - Date discovered
September 13, 2025
🕒 - Project duration
Unknown
-
🏝️ - Location type
Unknown
-
📄 - Contract type
Unknown
-
🔒 - Security clearance
Unknown
-
📍 - Location detailed
London Area, United Kingdom
-
🧠 - Skills detailed
#Cloud #Azure #Data Governance #SQL (Structured Query Language) #React #Strategy #Streamlit #GitHub #A/B Testing #Snowflake #Pandas #Data Ethics #REST (Representational State Transfer) #PyTorch #BigQuery #Classification #dbt (data build tool) #Deployment #REST API #Observability #Python #MLflow #Automation #GCP (Google Cloud Platform) #AWS (Amazon Web Services) #Langchain #Airflow #Security #Docker #Statistics #Transformers #"ETL (Extract #Transform #Load)" #ML (Machine Learning) #AI (Artificial Intelligence) #Data Pipeline #Data Science
Role description
The role We’re hiring a Data Scientist with strong Generative-AI experience to design, build, and ship AI-powered tools end-to-end. You’ll work in a small, multi-disciplinary team and take ownership from discovery to deployment: scoping use-cases, building prototypes, hardening them for production, and putting the right evaluation and governance around them. What you’ll do • Build GenAI tools end-to-end (independently): chat/assistants, document Q&A (RAG), summarisation, classification, extraction, and workflow/agent automations. • Own evaluation & safety: create offline/online eval sets, measure faithfulness/hallucination, bias, safety, latency and cost; add guardrails and red-teaming. • Productionise: package as services/APIs or lightweight apps (e.g., Streamlit/Gradio/React), containerise, and integrate via CI/CD. • Data pipelines: design chunking/embedding strategies, pick vector stores, manage prompt/versioning, and monitor drift & quality. • Model strategy: select and mix providers (hosted and open-source), fine-tune where it’s sensible, and optimise for cost/perf/privacy. • Stakeholder enablement: translate problems into measurable KPIs, run discovery, document clearly, and hand over maintainable solutions. • Good practice: apply data ethics, security and privacy by design; align to service standards and accessibility where relevant. Tech you’ll likely use • Python (pandas, PyTorch/Transformers), SQL • LLM frameworks: LangChain, LlamaIndex (or similar) • Vector DBs: FAISS / pgvector / Pinecone (or similar) • Cloud & Dev: Azure/AWS/GCP, Docker, REST APIs, GitHub Actions/CI • Data & MLOps: BigQuery/Snowflake, MLflow/DVC, dbt/Airflow (nice to have) • Front ends (for internal tools): Streamlit / Gradio / basic React Must-have experience • 7+ years in Data Science/ML, including hands-on delivery of GenAI products (not just PoCs). • Proven ability to ship independently: from idea → prototype → secure, supportable production tool. • Strong Python & SQL; solid software engineering habits (testing, versioning, CI/CD). • Practical LLM skills: prompt design, RAG, tool/function calling, evaluation & guardrails, and prompt/model observability. • Sound grasp of statistics/experimentation (A/B tests, hypothesis testing) and communicating impact to non-technical audiences. • Data governance, privacy and secure handling of sensitive data. Nice to have • Experience in regulated or public-sector-like environments. • Azure OpenAI / Vertex AI / Bedrock; lightweight fine-tuning/LoRA. • Front-end skills to craft usable internal UIs. How to apply Send your CV (referencing DS-GENAI) to the Recruitment Team. Shortlisted candidates will complete a brief technical exercise or portfolio walk-through focusing on a GenAI tool you built and shipped.