

Senior Python Engineer - Data Ingestion & Databricks - Remote/Hybrid
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Python Engineer specializing in Data Ingestion & Databricks, offering a contract-to-hire position. Key skills include Python, Databricks, Delta Lake, CI/CD practices, and metadata-driven architecture. Location is remote/hybrid, primarily for candidates in MN, SD, ND, IA, and WI.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
June 12, 2025
π - Project duration
Unknown
-
ποΈ - Location type
Remote
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
United States
-
π§ - Skills detailed
#Scala #Data Ingestion #GitHub #Airflow #Code Reviews #Docker #Data Lake #Automated Testing #Containers #Argo #Metadata #Azure #Databricks #Scrum #Libraries #Documentation #Pytest #Terraform #Delta Lake #Agile #Python #Kubernetes #Data Pipeline
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Dice is the leading career destination for tech experts at every stage of their careers. Our client, Smart Caliber Technology, is seeking the following. Apply via Dice today!
Job Title: Senior Python Engineer - Data Ingestion & Databricks
Location: Remote / Hybrid
Duration: Contract to Hire
Looking for a candidate currently in MN Hybrid
Would consider someone from South Dakota, North Dakota, Iowa and WI for remote work.
About The Role
β’ We are seeking a strong Python engineer with a passion for clean, scalable code and a mindset rooted in metadata-driven architecture. This role is ideal for someone who thrives in a collaborative environment, enjoys setting engineering standards, and has hands-on experience with Databricks and Delta Lake, preferably in an Azure ecosystem.
β’ You'll work closely with technical leads to shape the development culture, contribute to architectural decisions, and help build robust, reusable Python libraries that power our data platform.
Key Responsibilities
β’ Design, develop, and maintain Python libraries with a focus on packaging, distribution, and reusability.
β’ Champion metadata-driven development practices to build flexible and scalable systems.
β’ Collaborate with team leads to define and enforce coding standards, including code reviews and documentation.
β’ Implement and maintain CI/CD pipelines with tools like linters, type checkers (e.g., mypy), and automated testing frameworks (e.g., pytest).
β’ Develop and optimize data workflows on Databricks, leveraging Data Lake and best practices for performance and scalability.
β’ Communicate technical decisions clearly and confidently to both technical and non-technical stakeholders.
β’ Mentor junior engineers and contribute to a culture of continuous improvement.
Required Qualifications
β’ Proven experience in packaging and distributing Python libraries (e.g., setuptools, poetry, pip, uv).
β’ Strong understanding of metadata-driven architecture and its application in software or data systems.
β’ Familiarity with CI/CD practices in Python projects, including automated testing, linting, and type checking.
β’ Experience with Github and Github Actions for CI/CD.
β’ Hands-on experience with Databricks and Delta Lake, ideally within an Azure environment.
β’ Excellent communication skills with the ability to explain and justify technical decisions.
β’ A collaborative, opinionated mindset with a drive to lead by example.
β’ Understanding of Scrum and Agile methodologies.
Nice to Have
β’ Experience with data ingestion frameworks or data pipeline orchestration tools (e.g. Airflow).
β’ Familiarity with containers (Docker, Kubernetes, Helm)
β’ CI/CD build and deploy architecture (Tekton/Argo)
β’ Familiarity with infrastructure-as-code (e.g. Terraform).
β’ Experience with managing Databricks infrastructure-as-code (clusters, policies, etc.)
β’ Contributions to open-source Python projects or internal tooling libraries.
Best Regards,
Chetna
β’ D
β’ Fax
Truth Lies in Heart