

New York Technology Partners
Sr. Databricks Solution Architect with Healthcare Domain (Only W2 or Selfcorp)
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Sr. Databricks Solution Architect with Healthcare Domain, offering a contract-to-hire position. Pay rate is competitive. Key skills include Azure Databricks, ADF, CI/CD, and Terraform. Requires 5+ years in cloud data engineering and healthcare experience.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
December 27, 2025
🕒 - Duration
Unknown
-
🏝️ - Location
Remote
-
📄 - Contract
W2 Contractor
-
🔒 - Security
Unknown
-
📍 - Location detailed
United States
-
🧠 - Skills detailed
#GitHub #Infrastructure as Code (IaC) #Code Reviews #Alation #Automation #Logging #Python #SQL (Structured Query Language) #Databricks #Data Quality #Presto #Security #AI (Artificial Intelligence) #dbt (data build tool) #Deployment #Compliance #Azure DevOps #Azure Data Factory #Documentation #Terraform #Quality Assurance #Version Control #Scala #Monitoring #Automated Testing #Observability #SQL Server #Azure #Metadata #Data Engineering #Cloud #Anomaly Detection #ADF (Azure Data Factory) #Azure Databricks #Kubernetes #"ETL (Extract #Transform #Load)" #Strategy #Storage #DevOps
Role description
Job Title: Sr. Databricks Solution Architect with Healthcare Domain (Only W2 or Selfcorp)
Location: Remote (Must be able to come in the office when needed - Washington, DC)
Position Type: Contract to Hire
Job Description:
Top Skills Needed:
• Deep hands-on expertise with Databricks platform architecture and governance
• Unity Catalog, workspaces, external locations, compute, access controls, cluster governance.
• Reliability engineering, monitoring, and operational hardening of the Lakehouse
• Observability, alerting, DR readiness, backup/restore, performance tuning, incident response.
• Strong experience with ADF, CI/CD, and Terraform for orchestrating and managing the Lakehouse
• Pipeline orchestration, IaC, DevOps, environment promotion, compute policies.
Typical Day-to-Day:
• Design how the Databricks Lakehouse should work including the structure, tools, standards, and best practices
• Guide engineering teams on how to build pipelines and use Databricks correctly
• Solve technical issues when data jobs fail or performance slows
• Work with stakeholders to understand data needs and deliver solutions
• Set standards for security, governance, naming conventions, and architecture
• Ensure the Databricks platform is stable, reliable, and always available
• Build and maintain monitoring, alerting, logging, and health dashboards
• Strengthen and fix ingestion pipelines (ADF → landing → raw → curated)
• Improve data quality checks, anomaly detection, and pipeline reliability
• Manage CI/CD pipelines and deployment processes using Azure DevOps or GitHub
• Use Terraform (IaC) to deploy and manage Databricks and Azure infrastructure
• Partner with Security and FinOps on access controls, compliance, and cost governance
• Mentor the Data Engineer and support distributed data engineering teams across the organization
Key Responsibilities
1. Lakehouse Architecture & Platform Administration
(Approximately 60% of role when combined with mentoring & code review)
• Serve as the primary architect and administrator for the Azure Databricks Lakehouse (Unity Catalog, workspaces, external locations, compute, access controls).
• Lead execution of a Minimal Viable Hardening Roadmap for the platform, prioritizing:
• High availability and DR readiness
• Backup/restore patterns for data and metadata
• Platform observability and operational metrics
• Secure and maintainable catalog/namespace structure
• Robust and proactive data quality assurance
• Implement and evolve naming conventions, environment strategies, and platform standards that enable long-term maintainability and safe scaling.
• Act as the Lakehouse-facing counterpart to Enterprise Architecture and Security, collaborating on network architecture, identity & access, compliance controls, and platform guardrails.
1. Reliability, Monitoring, and Incident Management
• Design, implement, and maintain comprehensive monitoring and alerting for Lakehouse platform components, ingestion jobs, key data assets, and system health indicators.
• Oversee end-to-end reliability engineering, including capacity planning, throughput tuning, job performance optimization, and preventative maintenance (e.g., IR updates, compute policy reviews).
• Participate in — and help shape — the on-call rotation for high-priority incidents affecting production workloads, including rapid diagnosis and mitigation during off-hours as needed.
• Develop and maintain incident response runbooks, escalation pathways, stakeholder communication protocols, and operational readiness checklists.
• Lead or participate in post-incident Root Cause Analyses, ensuring durable remediation and preventing recurrence.
• Conduct periodic DR and failover simulations, validating RPO/RTO and documenting improvements.
• This role is foundational to ensuring 24/7/365 availability and timely delivery of mission-critical data for clinical, financial, operational, and analytical needs.
1. Pipeline Reliability, Ingestion Patterns & Data Quality
• Strengthen and standardize ingestion pipelines (ADF → landing → raw → curated), including watermarking, incremental logic, backfills, and retry/cancel/resume patterns.
• Collaborate with the Data Engineer to modernize logging, automated anomaly detection, pipeline health dashboards, and DQ validation automation.
• Provide architectural guidance, code reviews, mentoring, and best-practice patterns to distributed engineering teams across MedStar.
• Support stabilization of existing ingestion and transformation pipelines across clinical (notes, OHDSI), financial, operational, and quality use cases.
1. DevOps, CI/CD, and Infrastructure as Code
• Administer and improve CI/CD pipelines using Azure DevOps or GitHub Enterprise.
• Support automated testing, environment promotion, and rollback patterns for Databricks and dbt assets.
• Maintain and extend Terraform (or adopt Terraform from another IaC background) for Databricks, storage, networking, compute policies, and related infrastructure.
• Promote version control standards, branching strategies, and deployment governance across data engineering teams.
1. Security, FinOps, and Guardrails Partnership
• Partner with Enterprise Architecture and Security on platform access controls, identity strategy, encryption, networking, and compliance.
• Implement and enforce cost tagging, compute policies, and alerts supporting FinOps transparency and cost governance.
• Collaborate with the team defining agentic coding guardrails, ensuring the Lakehouse platform supports safe & compliant use of AI-assisted code generation and execution.
• Help assess and optimize serverless SQL, serverless Python, and job compute patterns for cost-efficiency and reliability.
1. Mentorship, Collaboration, & Distributed Enablement
• Mentor the mid-level Data Engineer on Databricks, ADF, dbt, observability, DevOps, Terraform, and operational engineering patterns.
• Provide guidance, design patterns, and code review support to multiple distributed data engineering teams (Finance, MCPI, Safety/Risk, Quality, Digital Transformation, etc.).
• Lead platform knowledge-sharing efforts through documentation, workshops, and best-practice guidance.
• Demonstrate strong collaboration skills, balancing independence with alignment across teams.
1. Optional / Nice-to-Have: OHDSI Platform Support
(Not required for hiring; can be learned on the job.)
• Assist with or support operational administration of the OHDSI/OMOP stack (Atlas, WebAPI, vocabularies, Kubernetes deployments).
• Collaborate with partners to ensure the OHDSI platform is secure, maintainable, and well-integrated with the Lakehouse.
Required Qualifications
• 5+ years in cloud data engineering, platform engineering, or solution architecture.
• Strong hands-on expertise in Azure Databricks:
• Unity Catalog
• Workspaces & external locations
• SQL/Python notebooks & Jobs
• Cluster/warehouse governance
• Solid working experience with Azure Data Factory (pipelines, IRs, linked services).
• Strong SQL and Python engineering skills.
• Experience with CI/CD in Azure DevOps or GitHub Enterprise.
• Experience with Terraform or another IaC framework, and willingness to adopt Terraform.
• Demonstrated ability to design or support monitoring, alerting, logging, or reliability systems.
• Strong communication, collaboration, and problem-solving skills.
Preferred Qualifications (Optional)
• Advanced Terraform experience.
• Familiarity with healthcare, HIPAA, PHI, or regulated environments.
• Experience with Purview or enterprise cataloging.
• Exposure to OHDSI/OMOP.
• Experience optimizing or refactoring legacy ingestion pipelines.
• Experience supporting secure, controlled AI/agentic execution environments.
• Experience with EPIC EHR data exchange and/or EPIC Caboodle or Cogito analytics suite.
Personal Attributes
• Hands-on, pragmatic, and operationally minded.
• Comfortable leading both architecture and implementation.
• Collaborative and mentorship-oriented; thrives in small core teams with broad influence.
• Values platform stability, observability, and hardening over shiny features.
• Curious and adaptable, especially with emerging AI-assisted engineering patterns.
• Ability to remain calm and effective during incidents and high-pressure situations.
If you believe you are qualified for this position and are currently in the job market or interested in making a change, please email me the resume along with contact details at roshni@nytpcorp.com
Job Title: Sr. Databricks Solution Architect with Healthcare Domain (Only W2 or Selfcorp)
Location: Remote (Must be able to come in the office when needed - Washington, DC)
Position Type: Contract to Hire
Job Description:
Top Skills Needed:
• Deep hands-on expertise with Databricks platform architecture and governance
• Unity Catalog, workspaces, external locations, compute, access controls, cluster governance.
• Reliability engineering, monitoring, and operational hardening of the Lakehouse
• Observability, alerting, DR readiness, backup/restore, performance tuning, incident response.
• Strong experience with ADF, CI/CD, and Terraform for orchestrating and managing the Lakehouse
• Pipeline orchestration, IaC, DevOps, environment promotion, compute policies.
Typical Day-to-Day:
• Design how the Databricks Lakehouse should work including the structure, tools, standards, and best practices
• Guide engineering teams on how to build pipelines and use Databricks correctly
• Solve technical issues when data jobs fail or performance slows
• Work with stakeholders to understand data needs and deliver solutions
• Set standards for security, governance, naming conventions, and architecture
• Ensure the Databricks platform is stable, reliable, and always available
• Build and maintain monitoring, alerting, logging, and health dashboards
• Strengthen and fix ingestion pipelines (ADF → landing → raw → curated)
• Improve data quality checks, anomaly detection, and pipeline reliability
• Manage CI/CD pipelines and deployment processes using Azure DevOps or GitHub
• Use Terraform (IaC) to deploy and manage Databricks and Azure infrastructure
• Partner with Security and FinOps on access controls, compliance, and cost governance
• Mentor the Data Engineer and support distributed data engineering teams across the organization
Key Responsibilities
1. Lakehouse Architecture & Platform Administration
(Approximately 60% of role when combined with mentoring & code review)
• Serve as the primary architect and administrator for the Azure Databricks Lakehouse (Unity Catalog, workspaces, external locations, compute, access controls).
• Lead execution of a Minimal Viable Hardening Roadmap for the platform, prioritizing:
• High availability and DR readiness
• Backup/restore patterns for data and metadata
• Platform observability and operational metrics
• Secure and maintainable catalog/namespace structure
• Robust and proactive data quality assurance
• Implement and evolve naming conventions, environment strategies, and platform standards that enable long-term maintainability and safe scaling.
• Act as the Lakehouse-facing counterpart to Enterprise Architecture and Security, collaborating on network architecture, identity & access, compliance controls, and platform guardrails.
1. Reliability, Monitoring, and Incident Management
• Design, implement, and maintain comprehensive monitoring and alerting for Lakehouse platform components, ingestion jobs, key data assets, and system health indicators.
• Oversee end-to-end reliability engineering, including capacity planning, throughput tuning, job performance optimization, and preventative maintenance (e.g., IR updates, compute policy reviews).
• Participate in — and help shape — the on-call rotation for high-priority incidents affecting production workloads, including rapid diagnosis and mitigation during off-hours as needed.
• Develop and maintain incident response runbooks, escalation pathways, stakeholder communication protocols, and operational readiness checklists.
• Lead or participate in post-incident Root Cause Analyses, ensuring durable remediation and preventing recurrence.
• Conduct periodic DR and failover simulations, validating RPO/RTO and documenting improvements.
• This role is foundational to ensuring 24/7/365 availability and timely delivery of mission-critical data for clinical, financial, operational, and analytical needs.
1. Pipeline Reliability, Ingestion Patterns & Data Quality
• Strengthen and standardize ingestion pipelines (ADF → landing → raw → curated), including watermarking, incremental logic, backfills, and retry/cancel/resume patterns.
• Collaborate with the Data Engineer to modernize logging, automated anomaly detection, pipeline health dashboards, and DQ validation automation.
• Provide architectural guidance, code reviews, mentoring, and best-practice patterns to distributed engineering teams across MedStar.
• Support stabilization of existing ingestion and transformation pipelines across clinical (notes, OHDSI), financial, operational, and quality use cases.
1. DevOps, CI/CD, and Infrastructure as Code
• Administer and improve CI/CD pipelines using Azure DevOps or GitHub Enterprise.
• Support automated testing, environment promotion, and rollback patterns for Databricks and dbt assets.
• Maintain and extend Terraform (or adopt Terraform from another IaC background) for Databricks, storage, networking, compute policies, and related infrastructure.
• Promote version control standards, branching strategies, and deployment governance across data engineering teams.
1. Security, FinOps, and Guardrails Partnership
• Partner with Enterprise Architecture and Security on platform access controls, identity strategy, encryption, networking, and compliance.
• Implement and enforce cost tagging, compute policies, and alerts supporting FinOps transparency and cost governance.
• Collaborate with the team defining agentic coding guardrails, ensuring the Lakehouse platform supports safe & compliant use of AI-assisted code generation and execution.
• Help assess and optimize serverless SQL, serverless Python, and job compute patterns for cost-efficiency and reliability.
1. Mentorship, Collaboration, & Distributed Enablement
• Mentor the mid-level Data Engineer on Databricks, ADF, dbt, observability, DevOps, Terraform, and operational engineering patterns.
• Provide guidance, design patterns, and code review support to multiple distributed data engineering teams (Finance, MCPI, Safety/Risk, Quality, Digital Transformation, etc.).
• Lead platform knowledge-sharing efforts through documentation, workshops, and best-practice guidance.
• Demonstrate strong collaboration skills, balancing independence with alignment across teams.
1. Optional / Nice-to-Have: OHDSI Platform Support
(Not required for hiring; can be learned on the job.)
• Assist with or support operational administration of the OHDSI/OMOP stack (Atlas, WebAPI, vocabularies, Kubernetes deployments).
• Collaborate with partners to ensure the OHDSI platform is secure, maintainable, and well-integrated with the Lakehouse.
Required Qualifications
• 5+ years in cloud data engineering, platform engineering, or solution architecture.
• Strong hands-on expertise in Azure Databricks:
• Unity Catalog
• Workspaces & external locations
• SQL/Python notebooks & Jobs
• Cluster/warehouse governance
• Solid working experience with Azure Data Factory (pipelines, IRs, linked services).
• Strong SQL and Python engineering skills.
• Experience with CI/CD in Azure DevOps or GitHub Enterprise.
• Experience with Terraform or another IaC framework, and willingness to adopt Terraform.
• Demonstrated ability to design or support monitoring, alerting, logging, or reliability systems.
• Strong communication, collaboration, and problem-solving skills.
Preferred Qualifications (Optional)
• Advanced Terraform experience.
• Familiarity with healthcare, HIPAA, PHI, or regulated environments.
• Experience with Purview or enterprise cataloging.
• Exposure to OHDSI/OMOP.
• Experience optimizing or refactoring legacy ingestion pipelines.
• Experience supporting secure, controlled AI/agentic execution environments.
• Experience with EPIC EHR data exchange and/or EPIC Caboodle or Cogito analytics suite.
Personal Attributes
• Hands-on, pragmatic, and operationally minded.
• Comfortable leading both architecture and implementation.
• Collaborative and mentorship-oriented; thrives in small core teams with broad influence.
• Values platform stability, observability, and hardening over shiny features.
• Curious and adaptable, especially with emerging AI-assisted engineering patterns.
• Ability to remain calm and effective during incidents and high-pressure situations.
If you believe you are qualified for this position and are currently in the job market or interested in making a change, please email me the resume along with contact details at roshni@nytpcorp.com






