

Optomi
Databricks Architect
β - Featured Role | Apply direct with Data Freelance Hub
This role is a Databricks Architect contract position lasting 6 months, 100% remote. Candidates must have strong skills in Databricks SQL, Delta Lake, Unity Catalog, PySpark, and Python, with experience in data pipeline optimization and governance. W2 only.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
640
-
ποΈ - Date
October 1, 2025
π - Duration
More than 6 months
-
ποΈ - Location
Remote
-
π - Contract
W2 Contractor
-
π - Security
Unknown
-
π - Location detailed
New York, United States
-
π§ - Skills detailed
#Spark (Apache Spark) #Delta Lake #Migration #SQL (Structured Query Language) #Data Quality #Databricks #Data Modeling #Python #Scala #Security #"ACID (Atomicity #Consistency #Isolation #Durability)" #Compliance #Data Security #Data Engineering #Data Migration #ML (Machine Learning) #Programming #"ETL (Extract #Transform #Load)" #Data Pipeline #PySpark #Data Architecture #Data Management #Data Governance #Metadata #Batch
Role description
Databricks Architect & Data Engineer (2 openings)
β’ 100% Remote - client based out of NY
β’ W2 Only - must be able to work W2, no C2C or sponsorship
β’ Type: Contract - 6 months with high likelihood to extend.
Optomi, in partnership with a big 4 leader, is seeking both a Databricks Architect and a Databricks Engineer. They are looking for experience in building, optimizing, and managing modern data pipelines and analytics platforms. The ideal candidate will be hands-on with Databricks SQL (DBSQL), Delta Lake, Unity Catalog, PySpark, and Python, and will have strong expertise in performance tuning and large-scale data engineering best practices. This role requires close collaboration with client stakeholders, architects, and business teams.
Key ResponsibilitiesDesign, build, and optimize ETL/ELT pipelines on Databricks Lakehouse platform using PySpark, Delta Lake, and Databricks SQL.
β’ Implement data modeling and manage large-scale data sets for analytics, reporting, and machine learning workloads.
β’ Configure and manage Unity Catalog for centralized data governance, security, and lineage.
β’ Write high-performance PySpark and SQL code ensuring scalability and cost efficiency.
β’ Apply performance tuning techniques for queries, jobs, and pipelines to optimize compute utilization.
β’ Collaborate with data architects, analysts, and business teams to understand requirements and deliver reliable data solutions.
β’ Establish best practices for data quality, lineage, and metadata management.
β’ Work on data migration, ingestion frameworks, and streaming/batch data pipelines.
β’ Ensure compliance with data security, governance, and privacy standards.
Core Databricks Expertise:
β’ Databricks SQL (DBSQL) β advanced query development and performance optimization.
β’ Delta Lake β ACID transactions, schema evolution, time travel, and optimization.
β’ Unity Catalog β access controls, lineage, and catalog/schema/table management.
β’ Programming & Data Engineering:
β’ Strong coding skills in PySpark and Python.
β’ Experience with data pipelines, transformations, and orchestration.
Databricks Architect & Data Engineer (2 openings)
β’ 100% Remote - client based out of NY
β’ W2 Only - must be able to work W2, no C2C or sponsorship
β’ Type: Contract - 6 months with high likelihood to extend.
Optomi, in partnership with a big 4 leader, is seeking both a Databricks Architect and a Databricks Engineer. They are looking for experience in building, optimizing, and managing modern data pipelines and analytics platforms. The ideal candidate will be hands-on with Databricks SQL (DBSQL), Delta Lake, Unity Catalog, PySpark, and Python, and will have strong expertise in performance tuning and large-scale data engineering best practices. This role requires close collaboration with client stakeholders, architects, and business teams.
Key ResponsibilitiesDesign, build, and optimize ETL/ELT pipelines on Databricks Lakehouse platform using PySpark, Delta Lake, and Databricks SQL.
β’ Implement data modeling and manage large-scale data sets for analytics, reporting, and machine learning workloads.
β’ Configure and manage Unity Catalog for centralized data governance, security, and lineage.
β’ Write high-performance PySpark and SQL code ensuring scalability and cost efficiency.
β’ Apply performance tuning techniques for queries, jobs, and pipelines to optimize compute utilization.
β’ Collaborate with data architects, analysts, and business teams to understand requirements and deliver reliable data solutions.
β’ Establish best practices for data quality, lineage, and metadata management.
β’ Work on data migration, ingestion frameworks, and streaming/batch data pipelines.
β’ Ensure compliance with data security, governance, and privacy standards.
Core Databricks Expertise:
β’ Databricks SQL (DBSQL) β advanced query development and performance optimization.
β’ Delta Lake β ACID transactions, schema evolution, time travel, and optimization.
β’ Unity Catalog β access controls, lineage, and catalog/schema/table management.
β’ Programming & Data Engineering:
β’ Strong coding skills in PySpark and Python.
β’ Experience with data pipelines, transformations, and orchestration.