

Jobs via Dice
Senior Data Engineer Blue Ash, OH
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer in Blue Ash, OH, on a fully onsite C2C contract. Requires strong expertise in Azure Databricks, Spark, Python, SQL, and distributed data pipeline optimization. Senior data engineering experience is essential.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
March 24, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Corp-to-Corp (C2C)
-
🔒 - Security
Unknown
-
📍 - Location detailed
Cincinnati, OH
-
🧠 - Skills detailed
#AI (Artificial Intelligence) #GIT #Delta Lake #Azure #Python #Programming #Cloud #Agile #Data Pipeline #GitHub #Data Security #Azure Databricks #Documentation #Strategy #Data Catalog #Security #Data Processing #Data Governance #Version Control #Data Strategy #Data Engineering #SQL (Structured Query Language) #Infrastructure as Code (IaC) #Monitoring #Terraform #PySpark #Databricks #Spark (Apache Spark)
Role description
Dice is the leading career destination for tech experts at every stage of their careers. Our client, V-CENTRIX-US LLC, is seeking the following. Apply via Dice today!
We are seeking a Senior Data Engineer with deep expertise in Azure Databricks, Spark, Python, SQL, and distributed data pipeline optimization. This is a fully onsite C2C contract role based in Blue Ash, OH.
Role: Senior Data Engineer
Type: Contract (C2C)
Location & Onsite: Blue Ash, OH (5 days onsite)
Visa: Any visa acceptable
Interview Process: In-person onsite
Team Details: 10 team members; work independently with peer programming sessions throughout the day
Top Skills: Azure Databricks, Python, Spark
Soft Skills: Problem-solving, attention to detail, ability to work independently and collaboratively in an agile team
VERY IMPORTANT DETAILS:
• Work location must be local
• Interviews will be in person, onsite
• Candidates must be willing to come onsite for their interview and work fully onsite with the team
• Prescreening includes 3 video questions; candidates must answer using their own knowledge and experience, no AI-generated responses
• Include a link to the candidate''s LinkedIn profile with the submittal
Requirements:
• Senior experience as a Data Engineer
• Strong experience with Azure Databricks, Spark, Python
• Strong SQL skills and database experience
• Experience monitoring and optimizing Databricks clusters or workflows
• Experience working with Azure data services and integrating them with Databricks and enterprise data platforms
• Experience building and optimizing distributed data processing systems (partitions, joins, shuffles, cluster performance)
• Experience with data pipeline development using tools such as Delta Live Tables (DLT) or Databricks SQL
• Experience with orchestration, messaging services, or serverless components (e.g., Azure Functions)
• Experience with version control and CI/CD tools such as GitHub and GitHub Actions
• Experience using Terraform for cloud infrastructure provisioning
• Familiarity with SDLC and modern data engineering best practices
• Strong organizational skills with the ability to manage multiple priorities and work independently
Nice to Have:
• Experience with data governance, lineage, or cataloging tools (Purview, Unity Catalog)
Responsibilities:
• Analyze, design, and develop enterprise data solutions using Azure, Databricks, Spark, Python, SQL
• Develop, optimize, and maintain Spark/PySpark data pipelines, addressing performance issues such as data skew, partitioning, caching, and shuffle optimization
• Build and support Delta Lake tables and data models for analytical and operational use cases
• Apply reusable design patterns, data standards, and architectural guidelines, including collaboration with 84.51° when needed
• Use Terraform to provision and manage cloud and Databricks resources (Infrastructure as Code)
• Implement and maintain CI/CD workflows using GitHub and GitHub Actions
• Manage Git-based workflows for Databricks notebooks, jobs, and data engineering artifacts
• Troubleshoot failures and improve reliability across Databricks jobs, clusters, and data pipelines
• Apply cloud computing skills to deploy fixes, upgrades, and enhancements in Azure environments
• Collaborate with engineering teams to enhance tools, systems, development processes, and data security
• Participate in the development and communication of data strategy, standards, and roadmaps
• Create architectural diagrams, interface specifications, and design documentation
• Promote reuse of data assets and contribute to enterprise data catalog practices
• Provide timely support and communication to stakeholders and end users
• Mentor team members on data engineering best practices and emerging technologies
This is an excellent opportunity for a highly technical, senior candidate to join data engineering team and work on cutting-edge Azure Databricks and Spark solutions.
Dice is the leading career destination for tech experts at every stage of their careers. Our client, V-CENTRIX-US LLC, is seeking the following. Apply via Dice today!
We are seeking a Senior Data Engineer with deep expertise in Azure Databricks, Spark, Python, SQL, and distributed data pipeline optimization. This is a fully onsite C2C contract role based in Blue Ash, OH.
Role: Senior Data Engineer
Type: Contract (C2C)
Location & Onsite: Blue Ash, OH (5 days onsite)
Visa: Any visa acceptable
Interview Process: In-person onsite
Team Details: 10 team members; work independently with peer programming sessions throughout the day
Top Skills: Azure Databricks, Python, Spark
Soft Skills: Problem-solving, attention to detail, ability to work independently and collaboratively in an agile team
VERY IMPORTANT DETAILS:
• Work location must be local
• Interviews will be in person, onsite
• Candidates must be willing to come onsite for their interview and work fully onsite with the team
• Prescreening includes 3 video questions; candidates must answer using their own knowledge and experience, no AI-generated responses
• Include a link to the candidate''s LinkedIn profile with the submittal
Requirements:
• Senior experience as a Data Engineer
• Strong experience with Azure Databricks, Spark, Python
• Strong SQL skills and database experience
• Experience monitoring and optimizing Databricks clusters or workflows
• Experience working with Azure data services and integrating them with Databricks and enterprise data platforms
• Experience building and optimizing distributed data processing systems (partitions, joins, shuffles, cluster performance)
• Experience with data pipeline development using tools such as Delta Live Tables (DLT) or Databricks SQL
• Experience with orchestration, messaging services, or serverless components (e.g., Azure Functions)
• Experience with version control and CI/CD tools such as GitHub and GitHub Actions
• Experience using Terraform for cloud infrastructure provisioning
• Familiarity with SDLC and modern data engineering best practices
• Strong organizational skills with the ability to manage multiple priorities and work independently
Nice to Have:
• Experience with data governance, lineage, or cataloging tools (Purview, Unity Catalog)
Responsibilities:
• Analyze, design, and develop enterprise data solutions using Azure, Databricks, Spark, Python, SQL
• Develop, optimize, and maintain Spark/PySpark data pipelines, addressing performance issues such as data skew, partitioning, caching, and shuffle optimization
• Build and support Delta Lake tables and data models for analytical and operational use cases
• Apply reusable design patterns, data standards, and architectural guidelines, including collaboration with 84.51° when needed
• Use Terraform to provision and manage cloud and Databricks resources (Infrastructure as Code)
• Implement and maintain CI/CD workflows using GitHub and GitHub Actions
• Manage Git-based workflows for Databricks notebooks, jobs, and data engineering artifacts
• Troubleshoot failures and improve reliability across Databricks jobs, clusters, and data pipelines
• Apply cloud computing skills to deploy fixes, upgrades, and enhancements in Azure environments
• Collaborate with engineering teams to enhance tools, systems, development processes, and data security
• Participate in the development and communication of data strategy, standards, and roadmaps
• Create architectural diagrams, interface specifications, and design documentation
• Promote reuse of data assets and contribute to enterprise data catalog practices
• Provide timely support and communication to stakeholders and end users
• Mentor team members on data engineering best practices and emerging technologies
This is an excellent opportunity for a highly technical, senior candidate to join data engineering team and work on cutting-edge Azure Databricks and Spark solutions.






