

The Zig
Senior Data Engineer (AI Systems & Data Platforms)
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer (AI Systems & Data Platforms) with a contract length of "unknown", offering a pay rate of "unknown". Key skills include Python, SQL, and experience with data pipelines and modern data platforms like Azure and Snowflake.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
April 10, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Unknown
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
United States
-
🧠 - Skills detailed
#Databricks #AI (Artificial Intelligence) #Data Pipeline #Databases #"ETL (Extract #Transform #Load)" #Datasets #Snowflake #Data Engineering #Scala #Data Modeling #Data Quality #Azure #SQL (Structured Query Language) #Python
Role description
Why this role exists
We build AI systems that drive real business outcomes. This role exists to design and build the data systems that power those outcomes — not dashboards, but operational intelligence and AI.
What you will own
• Design and implement scalable data pipelines
• Write production-grade code for ingestion and transformation
• Build data models for analytics and AI use cases
• Work within Microsoft Fabric / Azure ecosystem
• Ensure data quality, performance, and reliability
• Prepare data for AI use (RAG, embeddings, semantic layers)
What you will build (real outcomes)
• Reliable pipelines powering business workflows
• Data models enabling AI agents and decision systems
• Systems that reduce manual effort and improve efficiency
• AI-ready datasets deployable in weeks
What we’re looking for
• Strong coding ability (Python and SQL)
• Experience building data pipelines end-to-end
• Deep understanding of data modeling
• Experience with modern data platforms (Fabric, Databricks, Snowflake)
• Ability to work with messy enterprise data
What will make you stand out
• Experience supporting AI/LLM systems
• Familiarity with vector databases and embeddings
• Bias toward shipping quickly and iterating in production
Why this role exists
We build AI systems that drive real business outcomes. This role exists to design and build the data systems that power those outcomes — not dashboards, but operational intelligence and AI.
What you will own
• Design and implement scalable data pipelines
• Write production-grade code for ingestion and transformation
• Build data models for analytics and AI use cases
• Work within Microsoft Fabric / Azure ecosystem
• Ensure data quality, performance, and reliability
• Prepare data for AI use (RAG, embeddings, semantic layers)
What you will build (real outcomes)
• Reliable pipelines powering business workflows
• Data models enabling AI agents and decision systems
• Systems that reduce manual effort and improve efficiency
• AI-ready datasets deployable in weeks
What we’re looking for
• Strong coding ability (Python and SQL)
• Experience building data pipelines end-to-end
• Deep understanding of data modeling
• Experience with modern data platforms (Fabric, Databricks, Snowflake)
• Ability to work with messy enterprise data
What will make you stand out
• Experience supporting AI/LLM systems
• Familiarity with vector databases and embeddings
• Bias toward shipping quickly and iterating in production






