

Dataflix
Data Architect - Locals to Orange County, CA
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is a Data Architect contract position based in Irvine, CA, requiring 6+ years of data engineering experience, strong skills in Informatica and Databricks, and expertise in ETL migration. Databricks certification preferred; salary is negotiable.
🌎 - Country
United States
💱 - Currency
Unknown
-
💰 - Day rate
Unknown
-
🗓️ - Date
November 27, 2025
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Irvine, CA
-
🧠 - Skills detailed
#Azure DevOps #Delta Lake #EDW (Enterprise Data Warehouse) #Metadata #Airflow #Scala #Data Management #Leadership #"ETL (Extract #Transform #Load)" #Version Control #Logging #Data Lineage #Project Management #SQL Server #Scripting #Azure #Migration #Databricks #Data Engineering #Data Pipeline #Spark (Apache Spark) #Shell Scripting #Unix #BI (Business Intelligence) #Datasets #PySpark #Data Migration #"ACID (Atomicity #Consistency #Isolation #Durability)" #Data Transformations #DevOps #Data Architecture #Informatica #SQL (Structured Query Language) #Jenkins #GitHub #BitBucket #Informatica PowerCenter #Data Security #Security
Role description
Overview
Dataflix is seeking a Lead Data Engineer / Data Architect to lead the hands-on migration of existing ETL workflows from Informatica PowerCenter to the Databricks Lakehouse Platform. This role combines technical leadership, architecture design, and hands-on development, ensuring successful delivery of a large-scale enterprise data migration. The ideal candidate has deep experience with Informatica and Databricks, a strong grasp of enterprise data warehousing, and the ability to guide and mentor a technical team while contributing directly to design and development.
Key Responsibilities
Analyze legacy ETL pipelines, identify dependencies, and plan phased migration sequencing.
Redesign and optimize existing SQL logic, stored procedures, and transformations for Databricks.
Architect and implement end-to-end data pipelines across Source Staging EDW BI layers.
Apply Databricks Delta Lake best practices for partitioning, schema evolution, and performance optimization.
Implement data validation, error handling, and logging frameworks for migration integrity.
Oversee orchestration and scheduling using Databricks Workflows, Airflow, or equivalent tools.
Ensure performance tuning, scalability, and security across all migrated components.
Collaborate with project management, architecture, and QA teams to ensure successful delivery.
Mentor junior engineers, perform code/design reviews, and enforce development standards.
Required Qualifications
6+ years of experience in data engineering, ETL development, or data architecture.
3+ years of hands-on experience with Databricks, Spark, or PySpark.
Strong background in Informatica PowerCenter and SQL Server.
Experience designing and implementing end-to-end data pipelines across Source, Staging, EDW, and BI layers.
Solid understanding of data warehousing and dimensional modeling (facts, dimensions, SCD1/SCD2).
Ability to analyze legacy Informatica ETL workflows, identify dependencies, and plan migration sequencing.
Strong experience optimizing Spark/Databricks performance through partitioning, caching, and query tuning.
Understanding of data validation, reconciliation, and QA processes for data migration and cutover.
Hands-on expertise in Unix shell scripting, stored procedures, and large-scale data transformations.
Skilled in Databricks Delta Lake operations (ACID transactions, merge, optimize, vacuum) and job performance tuning.
Knowledge of data security, RBAC, and access control within Databricks or similar enterprise data environments.
Experience with CI/CD pipelines and version control tools (GitHub, Bitbucket, Jenkins, or Azure DevOps).
Proven ability to lead and mentor technical teams while contributing as a hands-on developer.
Preferred Qualifications
Experience modernizing on-prem ETL to Databricks or other modern data platforms.
Exposure to metadata management, data lineage, and governance frameworks.
Familiarity with orchestration tools (Airflow, Control-M, Databricks Workflows).
Databricks certification (Associate or Professional Data Engineer) preferred.
Experience with hospitality or travel industry datasets is a plus.
Soft Skills
Strong problem-solving, accountability and analytical mindset.
Excellent communication and cross-functional collaboration skills.
Self-driven and proactive, with the ability to balance leadership and hands-on execution.
Location
Irvine, United States
Salary
negotiable
Job Type
Contract
Date Posted
November 19th, 2025
Overview
Dataflix is seeking a Lead Data Engineer / Data Architect to lead the hands-on migration of existing ETL workflows from Informatica PowerCenter to the Databricks Lakehouse Platform. This role combines technical leadership, architecture design, and hands-on development, ensuring successful delivery of a large-scale enterprise data migration. The ideal candidate has deep experience with Informatica and Databricks, a strong grasp of enterprise data warehousing, and the ability to guide and mentor a technical team while contributing directly to design and development.
Key Responsibilities
Analyze legacy ETL pipelines, identify dependencies, and plan phased migration sequencing.
Redesign and optimize existing SQL logic, stored procedures, and transformations for Databricks.
Architect and implement end-to-end data pipelines across Source Staging EDW BI layers.
Apply Databricks Delta Lake best practices for partitioning, schema evolution, and performance optimization.
Implement data validation, error handling, and logging frameworks for migration integrity.
Oversee orchestration and scheduling using Databricks Workflows, Airflow, or equivalent tools.
Ensure performance tuning, scalability, and security across all migrated components.
Collaborate with project management, architecture, and QA teams to ensure successful delivery.
Mentor junior engineers, perform code/design reviews, and enforce development standards.
Required Qualifications
6+ years of experience in data engineering, ETL development, or data architecture.
3+ years of hands-on experience with Databricks, Spark, or PySpark.
Strong background in Informatica PowerCenter and SQL Server.
Experience designing and implementing end-to-end data pipelines across Source, Staging, EDW, and BI layers.
Solid understanding of data warehousing and dimensional modeling (facts, dimensions, SCD1/SCD2).
Ability to analyze legacy Informatica ETL workflows, identify dependencies, and plan migration sequencing.
Strong experience optimizing Spark/Databricks performance through partitioning, caching, and query tuning.
Understanding of data validation, reconciliation, and QA processes for data migration and cutover.
Hands-on expertise in Unix shell scripting, stored procedures, and large-scale data transformations.
Skilled in Databricks Delta Lake operations (ACID transactions, merge, optimize, vacuum) and job performance tuning.
Knowledge of data security, RBAC, and access control within Databricks or similar enterprise data environments.
Experience with CI/CD pipelines and version control tools (GitHub, Bitbucket, Jenkins, or Azure DevOps).
Proven ability to lead and mentor technical teams while contributing as a hands-on developer.
Preferred Qualifications
Experience modernizing on-prem ETL to Databricks or other modern data platforms.
Exposure to metadata management, data lineage, and governance frameworks.
Familiarity with orchestration tools (Airflow, Control-M, Databricks Workflows).
Databricks certification (Associate or Professional Data Engineer) preferred.
Experience with hospitality or travel industry datasets is a plus.
Soft Skills
Strong problem-solving, accountability and analytical mindset.
Excellent communication and cross-functional collaboration skills.
Self-driven and proactive, with the ability to balance leadership and hands-on execution.
Location
Irvine, United States
Salary
negotiable
Job Type
Contract
Date Posted
November 19th, 2025






