AI Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for an AI Engineer on a 1-year contract in Seattle, WA (50% onsite), offering competitive pay. Required skills include 5+ years in data engineering, strong Python and SQL proficiency, and experience in regulated industries like pharma or biotech.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
840
-
πŸ—“οΈ - Date discovered
August 9, 2025
πŸ•’ - Project duration
More than 6 months
-
🏝️ - Location type
Hybrid
-
πŸ“„ - Contract type
Unknown
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
Seattle, WA
-
🧠 - Skills detailed
#"ETL (Extract #Transform #Load)" #Transformers #SQL (Structured Query Language) #dbt (data build tool) #Compliance #Security #Python #Documentation #Classification #Microsoft Power BI #Cloud #Time Series #Visualization #Forecasting #Data Pipeline #Version Control #Snowflake #AI (Artificial Intelligence) #BigQuery #Data Warehouse #Data Science #Data Quality #Airflow #Tableau #Data Engineering #Data Integrity #Datasets #AWS (Amazon Web Services) #DataOps #Regression #Scala #Statistics #BI (Business Intelligence) #NLP (Natural Language Processing) #Langchain #ML (Machine Learning) #Azure #Hugging Face #Deep Learning
Role description
Job Title: AI Analytics Engineer Location: Seattle, WA - 50% onsite Schedule: Standard Working Hours Type: 1 year contract - possible extension Responsibilities β€’ Design and build scalable data pipelines using python, SQL and cloud-based tools (Azure, AWS) β€’ Develop and maintain analytics models and data transformation using dbt, airflow and data warehouses (ex: Snowflake, BigQuery) β€’ Build and manage dashboards and data visualizations using web-based tools, Tableau, Power BI β€’ Ensure data quality, governance and security compliance across engineering workflows β€’ Responsible for ingestion, integration and delivery of data products and insights across multiple platforms applying and maintaining data integrity and governance rules β€’ Knowledge of deep learning methods for NLP (quantitative area of study, Comp Science preferred) β€’ Support adhoc analytics initiatives for clinical trials, commercial ops and digital health systems β€’ Utilizes supervised or unsupervised methods, learning from vast amounts of unlabeled data to drive insight β€’ Experience working with unstructured text β€’ Develop high quality analytical and statistical models, insights, patterns, visualizations, that can be used to improve decision making in manufacturing operations β€’ Responsible for documentation of all technical work both within and outside of formal document management systems Required Qualifications β€’ 5+ Yrs of experience in data engineering, analytics or applied machine learning β€’ Strong python and SQL Skills, experience in ETL Orchestration tools (Airflow, Prefect. Domino) ability to manipulate and analyze complex datasets. β€’ Proficiency in modern data stack (dbt, snowflake/Big Query, version control, CI/CD pipelines) β€’ Experience and familiarity in DataOps/MLOps frameworks, AI/LLM tooling such as OpenAI ,Langchain, Hugging face Transformers β€’ Strong communication and stakeholder engagement skills β€’ Experience in regulated industry (Pharma, biotech or life sciences) is highly preferred β€’ Solid grasp of data science fundamentals: causal inference, regression, classification, experimental design Nice to Have β€’ Knowledge of GxP, HIPAA or compliance standards β€’ experience with data modelling in clinical trials, real world evidence (RWE) or commercial analytics β€’ Familiarity with NLP, time series forecasting, or image analytics in healthcare. β€’ Master’s degree or PhD in Public Health, Economics, Statistics, Computer/Data Science, or related field Per manager, the Top 3 – 5 Must Have’s a candidate should have are: β€’ Experience in Data engineering, analytics or applied machine learning β€’ Strong python and SQL Skills, experience in ETL Orchestration tools (Airflow, Prefect. Domino) ability to manipulate and analyze complex datasets.