Optomi

Databricks Architect

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is a Databricks Architect contract position lasting 6 months, 100% remote. Candidates must have strong skills in Databricks SQL, Delta Lake, Unity Catalog, PySpark, and Python, with experience in data pipeline optimization and governance. W2 only.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
640
-
πŸ—“οΈ - Date
October 1, 2025
πŸ•’ - Duration
More than 6 months
-
🏝️ - Location
Remote
-
πŸ“„ - Contract
W2 Contractor
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
New York, United States
-
🧠 - Skills detailed
#Spark (Apache Spark) #Delta Lake #Migration #SQL (Structured Query Language) #Data Quality #Databricks #Data Modeling #Python #Scala #Security #"ACID (Atomicity #Consistency #Isolation #Durability)" #Compliance #Data Security #Data Engineering #Data Migration #ML (Machine Learning) #Programming #"ETL (Extract #Transform #Load)" #Data Pipeline #PySpark #Data Architecture #Data Management #Data Governance #Metadata #Batch
Role description
Databricks Architect & Data Engineer (2 openings) β€’ 100% Remote - client based out of NY β€’ W2 Only - must be able to work W2, no C2C or sponsorship β€’ Type: Contract - 6 months with high likelihood to extend. Optomi, in partnership with a big 4 leader, is seeking both a Databricks Architect and a Databricks Engineer. They are looking for experience in building, optimizing, and managing modern data pipelines and analytics platforms. The ideal candidate will be hands-on with Databricks SQL (DBSQL), Delta Lake, Unity Catalog, PySpark, and Python, and will have strong expertise in performance tuning and large-scale data engineering best practices. This role requires close collaboration with client stakeholders, architects, and business teams. Key ResponsibilitiesDesign, build, and optimize ETL/ELT pipelines on Databricks Lakehouse platform using PySpark, Delta Lake, and Databricks SQL. β€’ Implement data modeling and manage large-scale data sets for analytics, reporting, and machine learning workloads. β€’ Configure and manage Unity Catalog for centralized data governance, security, and lineage. β€’ Write high-performance PySpark and SQL code ensuring scalability and cost efficiency. β€’ Apply performance tuning techniques for queries, jobs, and pipelines to optimize compute utilization. β€’ Collaborate with data architects, analysts, and business teams to understand requirements and deliver reliable data solutions. β€’ Establish best practices for data quality, lineage, and metadata management. β€’ Work on data migration, ingestion frameworks, and streaming/batch data pipelines. β€’ Ensure compliance with data security, governance, and privacy standards. Core Databricks Expertise: β€’ Databricks SQL (DBSQL) – advanced query development and performance optimization. β€’ Delta Lake – ACID transactions, schema evolution, time travel, and optimization. β€’ Unity Catalog – access controls, lineage, and catalog/schema/table management. β€’ Programming & Data Engineering: β€’ Strong coding skills in PySpark and Python. β€’ Experience with data pipelines, transformations, and orchestration.