Alignerr

Data Science Expert - AI Content Specialist

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is a Data Science Expert - AI Content Specialist, offering a flexible remote contract for 10–40 hours/week at an hourly rate. Key skills required include machine learning, data engineering, and proficiency in Python/R and SQL; a Master's or PhD is preferred.
🌎 - Country
United Kingdom
💱 - Currency
£ GBP
-
💰 - Day rate
640
-
🗓️ - Date
April 28, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Remote
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
United Kingdom
-
🧠 - Skills detailed
#ML (Machine Learning) #Deployment #Data Engineering #Unsupervised Learning #R #Data Quality #Datasets #Statistics #PyTorch #Deep Learning #Python #Data Science #Computer Science #SQL (Structured Query Language) #Supervised Learning #Hadoop #AI (Artificial Intelligence) #Model Deployment #NLP (Natural Language Processing) #Spark (Apache Spark) #TensorFlow #Big Data #SQL Queries #Libraries
Role description
Data Science Expert – AI Content Specialist About The Role What if your expertise in machine learning, statistical inference, and data engineering could directly shape how the world's most advanced AI systems think and reason? We're looking for Data Science Experts to stress-test cutting-edge language models, author gold-standard solutions, and help push AI reasoning to its limits. This is a fully remote, flexible contract role built for serious data scientists who want meaningful, intellectually stimulating work — on their own schedule. • Organization: Alignerr • Type: Hourly Contract • Location: Remote • Commitment: 10–40 hours/week What You'll Do • Design Advanced Challenges — Create complex, domain-specific data science problems spanning hyperparameter optimization, Bayesian inference, cross-validation strategies, dimensionality reduction, and more • Author Ground-Truth Solutions — Develop rigorous, step-by-step technical responses including Python/R scripts, SQL queries, and mathematical derivations that serve as the definitive "golden" benchmark • Audit AI-Generated Code — Evaluate outputs from models leveraging Scikit-Learn, PyTorch, TensorFlow, and other leading libraries for correctness, efficiency, and best practices • Refine AI Reasoning — Identify and document failure modes such as data leakage, overfitting, and improper handling of imbalanced datasets, then provide structured feedback to improve model logic • Document Model Weaknesses — Systematically probe AI reasoning across machine learning theory, statistical inference, neural network architectures, and data engineering pipelines Who You Are • Pursuing or holding a Master's or PhD in Data Science, Statistics, Computer Science, or a related quantitative field • Strong foundational knowledge across supervised/unsupervised learning, deep learning, big data technologies (Spark, Hadoop), or NLP • Able to communicate complex algorithmic concepts and statistical results clearly and precisely in writing • Highly detail-oriented when reviewing code syntax, mathematical notation, and statistical conclusions • Self-motivated and comfortable working independently on task-based assignments • No prior AI or annotation experience required Nice to Have • Experience with data annotation, data quality workflows, or AI evaluation systems • Familiarity with production-level data science practices such as MLOps or CI/CD for model deployment • Prior work auditing or benchmarking machine learning pipelines Why Join Us • Work directly with industry-leading large language models at the frontier of AI research • Fully remote and asynchronous — work when and where it suits you • Freelance autonomy with the structure of meaningful, intellectually challenging work • Contribute to AI development that has a real and lasting impact on how technology reasons through complex problems • Potential for ongoing work and contract extension as new projects launch