

Arkhya Tech. Inc.
Databricks Data Engineer with DevOps Skills
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a "Databricks Data Engineer with DevOps Skills" in Los Angeles, CA (Hybrid), offering a contract length of FTE/CTH. Key skills include PySpark, SQL, Azure cloud services, and strong DevOps expertise. Certifications in Databricks or Azure Data Engineering are optional.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
February 5, 2026
π - Duration
Unknown
-
ποΈ - Location
Hybrid
-
π - Contract
W2 Contractor
-
π - Security
Unknown
-
π - Location detailed
Los Angeles, CA
-
π§ - Skills detailed
#Version Control #Storage #Spark SQL #Delta Lake #Data Governance #ADLS (Azure Data Lake Storage) #Databases #Libraries #Data Pipeline #PySpark #Scala #Data Processing #Azure DevOps #Databricks #Data Engineering #"ETL (Extract #Transform #Load)" #GitLab #ADF (Azure Data Factory) #Data Quality #Triggers #Azure cloud #Monitoring #BI (Business Intelligence) #Azure Blob Storage #GIT #Data Security #Observability #Security #Data Analysis #Azure #Cloud #Logging #Deployment #SQL (Structured Query Language) #Compliance #Datasets #Vault #Spark (Apache Spark) #DevOps #"ACID (Atomicity #Consistency #Isolation #Durability)" #Terraform
Role description
Job Title: Databricks Data Engineer with DevOps Skills
Location : Los Angeles CA (Hybrid)
Hire type : FTE / CTH
Job Summary
We are looking for an experienced Databricks Data Engineer with strong DevOps expertise to join our data engineering team. The ideal candidate will design, build, and optimize large-scale data pipelines on the Databricks Lakehouse platform while implementing robust CI/CD and deployment practices. This role requires strong skills in PySpark, SQL, Azure cloud services, and modern DevOps tooling. You will collaborate with cross-functional teams to deliver scalable, secure, and highβperformance data solutions.
Technical Skills
β’ Strong hands-on experience with Databricks, including:
β’ Delta Lake
β’ Unity Catalog
β’ Lakehouse Architecture
β’ Delta Live Pipelines
β’ Databricks Runtime
β’ Table Triggers
β’ Proficiency in PySpark, Spark, and advanced SQL.
β’ Expertise with Azure cloud services (ADLS, ADF, Key Vault, Functions, etc.).
β’ Experience with relational databases and data warehousing concepts.
β’ Strong understanding of DevOps tools:
β’ Git/GitLab
β’ CI/CD pipelines
β’ Databricks Asset Bundles
β’ Familiarity with infrastructure-as-code (Terraform is a plus).
Key Responsibilities
1. Data Pipeline Development
β’ Design, build, and maintain scalable ETL/ELT pipelines using Databricks.
β’ Develop data processing workflows using PySpark/Spark and SQL for largeβvolume datasets.
β’ Integrate data from ADLS, Azure Blob Storage, and relational/non-relational data sources.
β’ Implement Delta Lake best practices including schema evolution, ACID transactions, OPTIMIZE, ZORDER, and performance tuning.
1. DevOps & CI/CD
β’ Implement CI/CD pipelines for Databricks using Git, GitLab, Azure DevOps, or similar tools.
β’ Build and manage automated deployments using Databricks Asset Bundles.
β’ Manage version control for notebooks, workflows, libraries, and configuration artifacts.
β’ Automate cluster configuration, job creation, and environment provisioning.
1. Collaboration & Business Support
β’ Work with data analysts and BI teams to prepare datasets for reporting and dashboarding.
β’ Collaborate with product owners, business partners, and engineering teams to translate requirements into scalable data solutions.
β’ Document data flows, architecture, and deployment processes.
1. Performance & Optimization
β’ Tune Databricks clusters, jobs, and pipelines for cost efficiency and high performance.
β’ Monitor workflows, debug failures, and ensure pipeline stability and reliability.
β’ Implement job instrumentation and observability using logging/monitoring tools.
1. Governance & Security
β’ Implement and manage data governance using Unity Catalog.
β’ Enforce access controls, data security, and compliance with enterprise policies.
β’ Ensure best practices around data quality, lineage, and auditability.
Preferred Experience
β’ Knowledge of streaming technologies like Structured Streaming or Spark Streaming.
β’ Experience building real-time or near real-time pipelines.
β’ Exposure to advanced Databricks runtime configurations and tuning.
Certifications (Optional)
β’ Databricks Certified Data Engineer Associate / Professional
β’ Azure Data Engineer Associate
Job Title: Databricks Data Engineer with DevOps Skills
Location : Los Angeles CA (Hybrid)
Hire type : FTE / CTH
Job Summary
We are looking for an experienced Databricks Data Engineer with strong DevOps expertise to join our data engineering team. The ideal candidate will design, build, and optimize large-scale data pipelines on the Databricks Lakehouse platform while implementing robust CI/CD and deployment practices. This role requires strong skills in PySpark, SQL, Azure cloud services, and modern DevOps tooling. You will collaborate with cross-functional teams to deliver scalable, secure, and highβperformance data solutions.
Technical Skills
β’ Strong hands-on experience with Databricks, including:
β’ Delta Lake
β’ Unity Catalog
β’ Lakehouse Architecture
β’ Delta Live Pipelines
β’ Databricks Runtime
β’ Table Triggers
β’ Proficiency in PySpark, Spark, and advanced SQL.
β’ Expertise with Azure cloud services (ADLS, ADF, Key Vault, Functions, etc.).
β’ Experience with relational databases and data warehousing concepts.
β’ Strong understanding of DevOps tools:
β’ Git/GitLab
β’ CI/CD pipelines
β’ Databricks Asset Bundles
β’ Familiarity with infrastructure-as-code (Terraform is a plus).
Key Responsibilities
1. Data Pipeline Development
β’ Design, build, and maintain scalable ETL/ELT pipelines using Databricks.
β’ Develop data processing workflows using PySpark/Spark and SQL for largeβvolume datasets.
β’ Integrate data from ADLS, Azure Blob Storage, and relational/non-relational data sources.
β’ Implement Delta Lake best practices including schema evolution, ACID transactions, OPTIMIZE, ZORDER, and performance tuning.
1. DevOps & CI/CD
β’ Implement CI/CD pipelines for Databricks using Git, GitLab, Azure DevOps, or similar tools.
β’ Build and manage automated deployments using Databricks Asset Bundles.
β’ Manage version control for notebooks, workflows, libraries, and configuration artifacts.
β’ Automate cluster configuration, job creation, and environment provisioning.
1. Collaboration & Business Support
β’ Work with data analysts and BI teams to prepare datasets for reporting and dashboarding.
β’ Collaborate with product owners, business partners, and engineering teams to translate requirements into scalable data solutions.
β’ Document data flows, architecture, and deployment processes.
1. Performance & Optimization
β’ Tune Databricks clusters, jobs, and pipelines for cost efficiency and high performance.
β’ Monitor workflows, debug failures, and ensure pipeline stability and reliability.
β’ Implement job instrumentation and observability using logging/monitoring tools.
1. Governance & Security
β’ Implement and manage data governance using Unity Catalog.
β’ Enforce access controls, data security, and compliance with enterprise policies.
β’ Ensure best practices around data quality, lineage, and auditability.
Preferred Experience
β’ Knowledge of streaming technologies like Structured Streaming or Spark Streaming.
β’ Experience building real-time or near real-time pipelines.
β’ Exposure to advanced Databricks runtime configurations and tuning.
Certifications (Optional)
β’ Databricks Certified Data Engineer Associate / Professional
β’ Azure Data Engineer Associate





