ProSearch

AI Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for an AI Data Engineer, a fully remote position with a contract length of unspecified duration. The pay rate is also unspecified. Key skills include advanced SQL, Python, and R, with a requirement of at least four years of Data Engineering experience.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
November 18, 2025
πŸ•’ - Duration
Unknown
-
🏝️ - Location
Remote
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
United States
-
🧠 - Skills detailed
#PySpark #Programming #Azure #Hadoop #Terraform #Data Science #SQL (Structured Query Language) #Kubernetes #GIT #AWS (Amazon Web Services) #Data Lake #Data Modeling #AI (Artificial Intelligence) #ML Ops (Machine Learning Operations) #Python #"ETL (Extract #Transform #Load)" #Cloud #Data Lakehouse #NoSQL #Databricks #Delta Lake #Big Data #dbt (data build tool) #Storage #Scala #Monitoring #Docker #Disaster Recovery #Data Architecture #Metadata #Data Governance #PyTorch #Spark (Apache Spark) #Data Quality #Data Engineering #Data Warehouse #Documentation #R #Model Validation #Infrastructure as Code (IaC) #S3 (Amazon Simple Storage Service) #YAML (YAML Ain't Markup Language) #ML (Machine Learning) #Data Pipeline #GitHub #Data Analysis #Databases #Datasets #Snowflake #TensorFlow
Role description
We have partnered with a leading technology research organization to hire an AI Data Engineer. In this role, you will build scalable data pipelines, partner closely with Data Scientists and ML Engineers, and ensure the organization’s AI/ML models are fueled by high-quality, well-structured data. This is a fully remote opportunity to contribute to impactful AI initiatives that support scientific innovation, clinical solutions, and operational excellence. About the Role As an AI Data Engineer, you will support data science model validation, analytics workloads, and machine learning operations by building high-quality feature tables, analytical datasets, and automated workflows. You’ll collaborate with senior data staff, product owners, and AI/ML scientists to deliver reliable data assets that enhance model performance and accelerate R&D innovation. You will work across core data streams, including discovery, imaging, clinical, and operational, and contribute to the pipelines that power next-generation AI products in veterinary and animal health. Top Required Skills β€’ SQL (advanced) β€’ Python β€’ R Nice-to-Have Skills β€’ dbt Core β€’ Databricks β€’ Data analysis experience Technology Stack Python β€’ Databricks β€’ dbt Core β€’ Hadoop β€’ TensorFlow β€’ PyTorch β€’ PySpark β€’ Snowflake β€’ AWS What You’ll Do β€’ Build scalable, reliable, distributed data pipelines to support machine learning operations and analytics workloads. β€’ Partner with data scientists, ML engineers, analysts, and data product owners to understand requirements and deliver high-quality solutions. β€’ Work with modern cloud and ML stacks, including Databricks, Snowflake, AWS, and Azure. β€’ Use Databricks (pipelines, workflows, asset bundles) to streamline engineering processes. β€’ Apply dbt Core for transformations, documentation, testing, and semantic consistency. β€’ Maintain code quality using SQL/YAML linters (SQLFluff) and enforce standards through GitHub Actions CI/CD. β€’ Develop solutions for data quality issues such as missing, duplicate, and inconsistent data. β€’ Contribute to data warehouse, data lake, data lakehouse, and data mesh architectural patterns. β€’ Build pipelines in Python to integrate diverse data types: structured tables, text documents, images, and more. β€’ Implement CI/CD systems and IaC tools like Terraform or AWS CloudFormation. β€’ Support data systems across the full lifecycle: exploration, production, monitoring, disaster recovery, and optimization. β€’ Stay current on advanced data engineering practices, including emerging technologies like Generative AI. What You Bring You have a relevant technical degree and at least four (4) years of Data Engineering experience. You are experienced with: β€’ Cloud platforms (preferably AWS) β€’ Big data technologies: Spark, Databricks, Delta Lake β€’ Git and Git-based workflows β€’ dbt Core and modern data modeling β€’ SQL and NoSQL databases β€’ Cloud object storage (e.g., S3) β€’ Containerization (Docker, Kubernetes, AWS ECS) β€’ Building, testing, and maintaining fault-tolerant data pipelines β€’ Understanding data architecture concepts: warehouse, lake, lakehouse, mesh You’re also eager to deepen your knowledge of AI/ML techniques, and it’s a plus if you have: β€’ Experience developing APIs or web applications β€’ Certifications in data engineering or AI/ML Leveling Guide (Intermediate) β€’ Build metadata and schemas based on logical models β€’ Write scripts for physical data layout and load test data β€’ Design and validate schemas β€’ Use ER modeling tools for intermediate tasks β€’ Adhere to data governance, naming conventions, testing principles β€’ Resolve moderately complex data problems β€’ Provide SQL and Python scripts for tuning and validation β€’ Write intermediate-level database programming scripts β€’ Contribute independently to team projects and semantic layer enhancements β€’ Suggest improvements to standards and processes β€’ Take new perspectives on solving moderately complex problems Why This Role Matters Your work will directly impact: β€’ The performance of AI/ML models β€’ The accuracy, reliability, and timeliness of analytics β€’ The innovation of new data streams from R&D pipelines β€’ The quality and discoverability of curated datasets β€’ How the organization advances clinical AI technologies Join Us If you are an analytical, collaborative, and forward-thinking AI Data Engineer looking for a remote opportunity that combines modern data engineering with applied machine learning, we encourage you to apply. Your expertise will help shape the next generation of AI-driven products and scientific innovation.