

Algoworks
ELT/EDI Lead Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for an "ELT/EDI Lead Engineer" with a contract length of "unknown" and a pay rate of "unknown." It requires 8+ years of data engineering experience, expertise in EDI/X12 transactions, and proficiency in Snowflake and AI tools.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
March 24, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Unknown
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
United States
-
🧠 - Skills detailed
#AI (Artificial Intelligence) #JSON (JavaScript Object Notation) #Data Warehouse #Leadership #XML (eXtensible Markup Language) #Classification #Azure #Data Modeling #DevOps #Azure SQL #Scala #Data Pipeline #GitHub #"ETL (Extract #Transform #Load)" #Documentation #Normalization #BI (Business Intelligence) #Data Quality #Data Engineering #SQL (Structured Query Language) #Snowflake #Data Layers
Role description
ELT / EDI Lead Engineer (Supply Chain Data Platform – AI-Assisted)
Role Overview
We are seeking an experienced ELT / EDI Lead Engineer to lead the design and implementation of data pipelines for a supply chain transaction analytics platform. This role will focus on ingesting, normalizing, and modeling EDI-driven transaction data (purchase orders, invoices, shipment notices, acknowledgements) to enable reliable reporting and operational insights.
You will play a critical role in defining how data from multiple trading partners and systems is standardized and interpreted, ensuring consistency in transaction lifecycle tracking, error handling, and KPI reporting.
This is a hands-on leadership role where you will guide data engineers while actively contributing to pipeline development and data modeling.
Key Responsibilities
EDI & Supply Chain Data Ownership
• Lead ingestion and normalization of EDI/X12 transaction data (850, 855, 856, 810, 997) across multiple source systems
• Define consistent interpretation of transaction lifecycle states (received, processed, failed, delayed, acknowledged, etc.)
• Standardize data across different trading partners with varying schemas and formats
• Work closely with business stakeholders to define supply chain KPIs (transaction success rates, processing delays, error patterns)
ELT Pipeline & Data Modeling
• Design and implement ELT pipelines using Snowflake, Azure data services, or similar platforms
• Define and enforce Bronze (raw), Silver (cleaned), and Gold (analytics-ready) data layers
• Develop transformation logic for structured and semi-structured data (XML, JSON, EDI payloads)
• Ensure pipelines are scalable, reliable, and optimized for performance
• Guide data engineers on best practices for pipeline development and data modeling
AI-Assisted Development & Optimization
• Use AI tools such as Cursor and GitHub Copilot to accelerate SQL development, transformation logic, and pipeline design
• Leverage LLM-based tools to analyze EDI schemas, summarize structure differences, and assist in normalization design
• Use Snowflake Cortex (where applicable) for query optimization, classification (e.g., error grouping), and performance improvements
• Apply AI tools to rapidly prototype transformation logic and refine through manual validation
• Ensure all AI-generated outputs are thoroughly reviewed for correctness, especially in business-critical transaction logic
Data Quality, Semantics & Governance
• Define data validation rules for transaction completeness, accuracy, and consistency
• Ensure alignment between source data and reporting outputs through reconciliation logic
• Establish semantic consistency across KPIs and reporting layers
• Identify and resolve issues related to schema inconsistencies, missing data, and transformation errors
Leadership & Collaboration
• Provide hands-on technical leadership to a team of data engineers
• Review code, transformation logic, and pipeline implementations
• Collaborate with QA, BI, and DevOps teams to ensure end-to-end data quality and delivery
• Act as a key technical point of contact for stakeholders and delivery leadership
Required Skills & Experience
Core Technical Skills
• 8+ years of experience in data engineering, with strong focus on ELT/ETL pipelines
• Deep expertise in EDI/X12 transactions (850, 855, 856, 810, 997) and transaction lifecycles
• Strong hands-on experience with Snowflake, Azure SQL, or similar data warehouse platforms
• Advanced SQL skills for complex transformations and performance optimization
• Experience working with semi-structured data (XML, JSON, EDI formats)
• Strong understanding of data modeling and medallion architecture (Bronze/Silver/Gold)
AI-Assisted Engineering (Mandatory)
• Hands-on experience using AI tools such as:
• Cursor (SQL and transformation development)
• GitHub Copilot (code generation and optimization)
• LLM-based tools (schema analysis, documentation, and design support)
• Ability to use AI tools for:
• Accelerating SQL development and transformation logic
• Analyzing complex EDI schemas and identifying patterns
• Generating and refining pipeline logic
• Strong ability to validate and refine AI-generated outputs, especially for business-critical transaction data
• Experience working in environments where AI is used to improve productivity while maintaining strict quality standards
ELT / EDI Lead Engineer (Supply Chain Data Platform – AI-Assisted)
Role Overview
We are seeking an experienced ELT / EDI Lead Engineer to lead the design and implementation of data pipelines for a supply chain transaction analytics platform. This role will focus on ingesting, normalizing, and modeling EDI-driven transaction data (purchase orders, invoices, shipment notices, acknowledgements) to enable reliable reporting and operational insights.
You will play a critical role in defining how data from multiple trading partners and systems is standardized and interpreted, ensuring consistency in transaction lifecycle tracking, error handling, and KPI reporting.
This is a hands-on leadership role where you will guide data engineers while actively contributing to pipeline development and data modeling.
Key Responsibilities
EDI & Supply Chain Data Ownership
• Lead ingestion and normalization of EDI/X12 transaction data (850, 855, 856, 810, 997) across multiple source systems
• Define consistent interpretation of transaction lifecycle states (received, processed, failed, delayed, acknowledged, etc.)
• Standardize data across different trading partners with varying schemas and formats
• Work closely with business stakeholders to define supply chain KPIs (transaction success rates, processing delays, error patterns)
ELT Pipeline & Data Modeling
• Design and implement ELT pipelines using Snowflake, Azure data services, or similar platforms
• Define and enforce Bronze (raw), Silver (cleaned), and Gold (analytics-ready) data layers
• Develop transformation logic for structured and semi-structured data (XML, JSON, EDI payloads)
• Ensure pipelines are scalable, reliable, and optimized for performance
• Guide data engineers on best practices for pipeline development and data modeling
AI-Assisted Development & Optimization
• Use AI tools such as Cursor and GitHub Copilot to accelerate SQL development, transformation logic, and pipeline design
• Leverage LLM-based tools to analyze EDI schemas, summarize structure differences, and assist in normalization design
• Use Snowflake Cortex (where applicable) for query optimization, classification (e.g., error grouping), and performance improvements
• Apply AI tools to rapidly prototype transformation logic and refine through manual validation
• Ensure all AI-generated outputs are thoroughly reviewed for correctness, especially in business-critical transaction logic
Data Quality, Semantics & Governance
• Define data validation rules for transaction completeness, accuracy, and consistency
• Ensure alignment between source data and reporting outputs through reconciliation logic
• Establish semantic consistency across KPIs and reporting layers
• Identify and resolve issues related to schema inconsistencies, missing data, and transformation errors
Leadership & Collaboration
• Provide hands-on technical leadership to a team of data engineers
• Review code, transformation logic, and pipeline implementations
• Collaborate with QA, BI, and DevOps teams to ensure end-to-end data quality and delivery
• Act as a key technical point of contact for stakeholders and delivery leadership
Required Skills & Experience
Core Technical Skills
• 8+ years of experience in data engineering, with strong focus on ELT/ETL pipelines
• Deep expertise in EDI/X12 transactions (850, 855, 856, 810, 997) and transaction lifecycles
• Strong hands-on experience with Snowflake, Azure SQL, or similar data warehouse platforms
• Advanced SQL skills for complex transformations and performance optimization
• Experience working with semi-structured data (XML, JSON, EDI formats)
• Strong understanding of data modeling and medallion architecture (Bronze/Silver/Gold)
AI-Assisted Engineering (Mandatory)
• Hands-on experience using AI tools such as:
• Cursor (SQL and transformation development)
• GitHub Copilot (code generation and optimization)
• LLM-based tools (schema analysis, documentation, and design support)
• Ability to use AI tools for:
• Accelerating SQL development and transformation logic
• Analyzing complex EDI schemas and identifying patterns
• Generating and refining pipeline logic
• Strong ability to validate and refine AI-generated outputs, especially for business-critical transaction data
• Experience working in environments where AI is used to improve productivity while maintaining strict quality standards






