

Gentis Solutions
Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with strong Azure and Databricks experience, focusing on supply chain data. It is a remote contract position with a duration of over 6 months, requiring expertise in SQL, Python, and data pipeline development.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
560
-
ποΈ - Date
April 24, 2026
π - Duration
More than 6 months
-
ποΈ - Location
Remote
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Cincinnati Metropolitan Area
-
π§ - Skills detailed
#Data Pipeline #Data Catalog #Data Quality #Data Engineering #Data Modeling #Data Migration #Data Lake #Data Governance #PySpark #"ETL (Extract #Transform #Load)" #Azure cloud #GitHub #Agile #Terraform #Data Strategy #Scala #SQL (Structured Query Language) #Azure #BI (Business Intelligence) #Scrum #Complex Queries #Python #Databricks #Strategy #Cloud #Migration #Microsoft Power BI #Azure Databricks #Spark (Apache Spark) #Alation
Role description
Data Engineer β Azure / Databricks (Supply Chain Data)
Location: Remote (or Cincinnati, OH preferred)
Duration: Contract (with potential extension)
Start Date: ASAP
Overview
We are seeking a Data Engineer with strong Azure and Databricks experience to support Krogerβs Supply Chain Data Strategy. This role focuses on building cloud-first data solutions within a large-scale warehouse data environment.
You will play a key role in a multi-year (2β5 year) transformation initiative, including the implementation of Manhattan (WMS), helping design and optimize data pipelines that power enterprise-wide supply chain insights.
Key Responsibilities
β’ Design, build, and optimize data pipelines using Azure Databricks
β’ Develop data solutions leveraging Azure Data Lake (Gen2)
β’ Write and optimize complex queries using SQL and PySpark/Python
β’ Collaborate on data modeling, mapping, and transformation
β’ Support data migration and cloud transformation initiatives
β’ Build and maintain data catalogs and governance frameworks
β’ Partner with cross-functional teams (engineering, PMs, analytics)
β’ Ensure adherence to data quality, standards, and governance
β’ Contribute to long-term initiatives including Manhattan WMS implementation
Required Skills
β’ Strong hands-on experience with:
β’ Azure Data Lake (Gen2)
β’ Azure Databricks
β’ Advanced proficiency in:
β’ SQL
β’ Python / PySpark
β’ Experience building and optimizing scalable data pipelines
β’ Experience with data modeling, mapping, and transformation
β’ Familiarity with data cataloging/governance tools (Unity Catalog, Alation)
β’ Experience working in Agile/Scrum environments
β’ Strong collaboration and communication skills
Preferred Qualifications
β’ Experience in supply chain, logistics, or warehouse data domains
β’ Experience with Manhattan WMS or similar systems
β’ Experience in large enterprise data environments
β’ Experience with:
β’ Power BI
β’ Terraform
β’ GitHub Actions
β’ Experience working with analytics partners
Ideal Candidate
β’ Data Engineer with strong Azure cloud and Databricks expertise
β’ Proven track record working with enterprise-scale data platforms
β’ Experience supporting supply chain or warehouse data systems
β’ Strong background in data pipeline development and optimization
β’ Comfortable working in fast-paced Agile environments
β’ Experience with data governance and cataloging tools
Disqualifiers
β’ No experience with Azure Databricks
β’ No SQL or data engineering background
β’ Candidates focused only on BI/reporting (no pipeline development)
Team & Environment
β’ Agile/Scrum-based warehouse data engineering team
β’ High-impact role supporting enterprise supply chain operations
β’ Collaborative environment with engineers, PMs, and analytics teams
β’ Fast-paced, large-scale enterprise transformation initiative
Interview Process
β’ Round 1: Hiring Manager (30 minutes)
β’ Round 2: Panel Interview (Tech Lead, PM, Engineer)
β’ Hiring Manager reviewing candidates this week
Data Engineer β Azure / Databricks (Supply Chain Data)
Location: Remote (or Cincinnati, OH preferred)
Duration: Contract (with potential extension)
Start Date: ASAP
Overview
We are seeking a Data Engineer with strong Azure and Databricks experience to support Krogerβs Supply Chain Data Strategy. This role focuses on building cloud-first data solutions within a large-scale warehouse data environment.
You will play a key role in a multi-year (2β5 year) transformation initiative, including the implementation of Manhattan (WMS), helping design and optimize data pipelines that power enterprise-wide supply chain insights.
Key Responsibilities
β’ Design, build, and optimize data pipelines using Azure Databricks
β’ Develop data solutions leveraging Azure Data Lake (Gen2)
β’ Write and optimize complex queries using SQL and PySpark/Python
β’ Collaborate on data modeling, mapping, and transformation
β’ Support data migration and cloud transformation initiatives
β’ Build and maintain data catalogs and governance frameworks
β’ Partner with cross-functional teams (engineering, PMs, analytics)
β’ Ensure adherence to data quality, standards, and governance
β’ Contribute to long-term initiatives including Manhattan WMS implementation
Required Skills
β’ Strong hands-on experience with:
β’ Azure Data Lake (Gen2)
β’ Azure Databricks
β’ Advanced proficiency in:
β’ SQL
β’ Python / PySpark
β’ Experience building and optimizing scalable data pipelines
β’ Experience with data modeling, mapping, and transformation
β’ Familiarity with data cataloging/governance tools (Unity Catalog, Alation)
β’ Experience working in Agile/Scrum environments
β’ Strong collaboration and communication skills
Preferred Qualifications
β’ Experience in supply chain, logistics, or warehouse data domains
β’ Experience with Manhattan WMS or similar systems
β’ Experience in large enterprise data environments
β’ Experience with:
β’ Power BI
β’ Terraform
β’ GitHub Actions
β’ Experience working with analytics partners
Ideal Candidate
β’ Data Engineer with strong Azure cloud and Databricks expertise
β’ Proven track record working with enterprise-scale data platforms
β’ Experience supporting supply chain or warehouse data systems
β’ Strong background in data pipeline development and optimization
β’ Comfortable working in fast-paced Agile environments
β’ Experience with data governance and cataloging tools
Disqualifiers
β’ No experience with Azure Databricks
β’ No SQL or data engineering background
β’ Candidates focused only on BI/reporting (no pipeline development)
Team & Environment
β’ Agile/Scrum-based warehouse data engineering team
β’ High-impact role supporting enterprise supply chain operations
β’ Collaborative environment with engineers, PMs, and analytics teams
β’ Fast-paced, large-scale enterprise transformation initiative
Interview Process
β’ Round 1: Hiring Manager (30 minutes)
β’ Round 2: Panel Interview (Tech Lead, PM, Engineer)
β’ Hiring Manager reviewing candidates this week





