

METRIX IT SOLUTIONS INC
Data Architect (Databricks & Snowflake)
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Architect (Databricks & Snowflake) in Boston, MA, on a contract basis. Requires 15+ years of experience, expertise in Databricks, Snowflake, and Azure, and strong skills in Spark, ETL/ELT, and data governance.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
May 14, 2026
π - Duration
Unknown
-
ποΈ - Location
Hybrid
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Boston, MA
-
π§ - Skills detailed
#Snowflake #Shell Scripting #Data Warehouse #Airflow #Spark SQL #Cloud #Azure Databricks #Data Pipeline #Azure #Data Governance #GCP (Google Cloud Platform) #Leadership #Scripting #Spark (Apache Spark) #Linux #GIT #PySpark #Compliance #Deployment #Python #"ETL (Extract #Transform #Load)" #SQL (Structured Query Language) #Talend #Computer Science #Data Analysis #Apache Spark #Data Modeling #AWS (Amazon Web Services) #Scala #Security #ADF (Azure Data Factory) #Data Architecture #ADLS (Azure Data Lake Storage) #Azure DevOps #Databricks #Azure cloud #Informatica #Kafka (Apache Kafka) #Unix #Data Ingestion #Data Processing #Delta Lake #DevOps #Jenkins #Migration #Data Engineering #Batch
Role description
Role: Data Architect (Databricks & Snowflake)
Location: Boston, MA - Hybrid (as needed) - The customer is seeking local candidates only (in and around Boston) to ensure the consultant can visit the office when required.
Contract
LOCAL Profile Only
Ideally 15+ years experience required
Digital : Snowflake
Job Summary
We are seeking a highly experienced Senior Data Engineer / Data Architect with deep expertise in Databricks, Snowflake, and Azure cloud data platforms. The ideal candidate will have extensive experience designing and implementing scalable data pipelines, Lakehouse architectures, and real-time data processing solutions, particularly in regulated domains such as Life Sciences or Healthcare.
This role requires strong proficiency in Spark (PySpark), Delta Lake, Medallion architecture, and cloud-native data engineering practices, along with a solid background in data warehouse modernization and performance optimization.
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Key Responsibilities
β’ Design and implement end-to-end data engineering pipelines using Azure Databricks, ADLS Gen2, and Snowflake.
β’ Develop scalable ETL/ELT pipelines using PySpark, Spark SQL, Python, and Talend.
β’ Build and maintain Lakehouse architecture using Delta Lake and Medallion (Bronze, Silver, Gold) layers.
β’ Implement real-time and batch data ingestion pipelines, including streaming using Spark Structured Streaming.
β’ Design and enforce data governance, access control, and lineage using Unity Catalog.
β’ Optimize Spark workloads through partitioning, caching, broadcast joins, and cluster tuning to improve performance and reduce cloud costs.
β’ Architect and manage CI/CD pipelines using Azure DevOps, Jenkins, and Git for automated deployments.
β’ Integrate multiple data sources and systems, ensuring high-quality, reliable, and scalable data delivery.
β’ Collaborate with cross-functional teams including data analysts, scientists, and business stakeholders to support analytics and reporting needs.
β’ Support data warehouse modernization initiatives, including migration from legacy systems to cloud platforms.
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Required Qualifications
β’ 10+ years (ideally 15β20+) of experience in data engineering or data architecture.
β’ Strong expertise in:
o Databricks & Delta Lake
o Snowflake Data Warehouse
o Apache Spark (PySpark, Spark SQL)
β’ Hands-on experience with Azure Cloud (ADLS Gen2, Azure Databricks, ADF).
β’ Proficiency in Python and SQL for data engineering.
β’ Experience with ETL/ELT tools such as Talend or Informatica.
β’ Strong knowledge of data modeling, CDC (Change Data Capture), and incremental loading techniques.
β’ Experience working in Linux/Unix environments with shell scripting.
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Preferred Qualifications
β’ Knowledge of data governance, compliance, and regulatory standards (e.g., IDMP).
β’ Exposure to real-time data streaming technologies (Kafka, Kinesis).
β’ Experience with multi-cloud environments (AWS, GCP).
β’ Familiarity with workflow orchestration tools such as Airflow or Databricks Workflows.
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Key Skills
β’ Data Engineering & Architecture
β’ Lakehouse & Data Warehousing
β’ Spark Performance Optimization
β’ Cloud Data Platforms (Azure)
β’ ETL/ELT Pipeline Development
β’ Data Governance & Security
β’ CI/CD & DevOps Practices
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Education
β’ Bachelorβs degree in Information Technology, Computer Science, or a related field.
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Nice-to-Have Traits
β’ Strong analytical and problem-solving skills
β’ Ability to work in enterprise-scale, complex environments
β’ Experience working with global stakeholders and cross-functional teams
β’ Leadership capability with mentoring experience
Role: Data Architect (Databricks & Snowflake)
Location: Boston, MA - Hybrid (as needed) - The customer is seeking local candidates only (in and around Boston) to ensure the consultant can visit the office when required.
Contract
LOCAL Profile Only
Ideally 15+ years experience required
Digital : Snowflake
Job Summary
We are seeking a highly experienced Senior Data Engineer / Data Architect with deep expertise in Databricks, Snowflake, and Azure cloud data platforms. The ideal candidate will have extensive experience designing and implementing scalable data pipelines, Lakehouse architectures, and real-time data processing solutions, particularly in regulated domains such as Life Sciences or Healthcare.
This role requires strong proficiency in Spark (PySpark), Delta Lake, Medallion architecture, and cloud-native data engineering practices, along with a solid background in data warehouse modernization and performance optimization.
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Key Responsibilities
β’ Design and implement end-to-end data engineering pipelines using Azure Databricks, ADLS Gen2, and Snowflake.
β’ Develop scalable ETL/ELT pipelines using PySpark, Spark SQL, Python, and Talend.
β’ Build and maintain Lakehouse architecture using Delta Lake and Medallion (Bronze, Silver, Gold) layers.
β’ Implement real-time and batch data ingestion pipelines, including streaming using Spark Structured Streaming.
β’ Design and enforce data governance, access control, and lineage using Unity Catalog.
β’ Optimize Spark workloads through partitioning, caching, broadcast joins, and cluster tuning to improve performance and reduce cloud costs.
β’ Architect and manage CI/CD pipelines using Azure DevOps, Jenkins, and Git for automated deployments.
β’ Integrate multiple data sources and systems, ensuring high-quality, reliable, and scalable data delivery.
β’ Collaborate with cross-functional teams including data analysts, scientists, and business stakeholders to support analytics and reporting needs.
β’ Support data warehouse modernization initiatives, including migration from legacy systems to cloud platforms.
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Required Qualifications
β’ 10+ years (ideally 15β20+) of experience in data engineering or data architecture.
β’ Strong expertise in:
o Databricks & Delta Lake
o Snowflake Data Warehouse
o Apache Spark (PySpark, Spark SQL)
β’ Hands-on experience with Azure Cloud (ADLS Gen2, Azure Databricks, ADF).
β’ Proficiency in Python and SQL for data engineering.
β’ Experience with ETL/ELT tools such as Talend or Informatica.
β’ Strong knowledge of data modeling, CDC (Change Data Capture), and incremental loading techniques.
β’ Experience working in Linux/Unix environments with shell scripting.
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Preferred Qualifications
β’ Knowledge of data governance, compliance, and regulatory standards (e.g., IDMP).
β’ Exposure to real-time data streaming technologies (Kafka, Kinesis).
β’ Experience with multi-cloud environments (AWS, GCP).
β’ Familiarity with workflow orchestration tools such as Airflow or Databricks Workflows.
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Key Skills
β’ Data Engineering & Architecture
β’ Lakehouse & Data Warehousing
β’ Spark Performance Optimization
β’ Cloud Data Platforms (Azure)
β’ ETL/ELT Pipeline Development
β’ Data Governance & Security
β’ CI/CD & DevOps Practices
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Education
β’ Bachelorβs degree in Information Technology, Computer Science, or a related field.
\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_
Nice-to-Have Traits
β’ Strong analytical and problem-solving skills
β’ Ability to work in enterprise-scale, complex environments
β’ Experience working with global stakeholders and cross-functional teams
β’ Leadership capability with mentoring experience






