METRIX IT SOLUTIONS INC

Data Architect (Databricks & Snowflake)

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Architect (Databricks & Snowflake) in Boston, MA, on a contract basis. Requires 15+ years of experience, expertise in Databricks, Snowflake, and Azure, and strong skills in Spark, ETL/ELT, and data governance.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
May 14, 2026
πŸ•’ - Duration
Unknown
-
🏝️ - Location
Hybrid
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
Boston, MA
-
🧠 - Skills detailed
#Snowflake #Shell Scripting #Data Warehouse #Airflow #Spark SQL #Cloud #Azure Databricks #Data Pipeline #Azure #Data Governance #GCP (Google Cloud Platform) #Leadership #Scripting #Spark (Apache Spark) #Linux #GIT #PySpark #Compliance #Deployment #Python #"ETL (Extract #Transform #Load)" #SQL (Structured Query Language) #Talend #Computer Science #Data Analysis #Apache Spark #Data Modeling #AWS (Amazon Web Services) #Scala #Security #ADF (Azure Data Factory) #Data Architecture #ADLS (Azure Data Lake Storage) #Azure DevOps #Databricks #Azure cloud #Informatica #Kafka (Apache Kafka) #Unix #Data Ingestion #Data Processing #Delta Lake #DevOps #Jenkins #Migration #Data Engineering #Batch
Role description
Role: Data Architect (Databricks & Snowflake) Location: Boston, MA - Hybrid (as needed) - The customer is seeking local candidates only (in and around Boston) to ensure the consultant can visit the office when required. Contract LOCAL Profile Only Ideally 15+ years experience required Digital : Snowflake Job Summary We are seeking a highly experienced Senior Data Engineer / Data Architect with deep expertise in Databricks, Snowflake, and Azure cloud data platforms. The ideal candidate will have extensive experience designing and implementing scalable data pipelines, Lakehouse architectures, and real-time data processing solutions, particularly in regulated domains such as Life Sciences or Healthcare. This role requires strong proficiency in Spark (PySpark), Delta Lake, Medallion architecture, and cloud-native data engineering practices, along with a solid background in data warehouse modernization and performance optimization. \_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_ Key Responsibilities β€’ Design and implement end-to-end data engineering pipelines using Azure Databricks, ADLS Gen2, and Snowflake. β€’ Develop scalable ETL/ELT pipelines using PySpark, Spark SQL, Python, and Talend. β€’ Build and maintain Lakehouse architecture using Delta Lake and Medallion (Bronze, Silver, Gold) layers. β€’ Implement real-time and batch data ingestion pipelines, including streaming using Spark Structured Streaming. β€’ Design and enforce data governance, access control, and lineage using Unity Catalog. β€’ Optimize Spark workloads through partitioning, caching, broadcast joins, and cluster tuning to improve performance and reduce cloud costs. β€’ Architect and manage CI/CD pipelines using Azure DevOps, Jenkins, and Git for automated deployments. β€’ Integrate multiple data sources and systems, ensuring high-quality, reliable, and scalable data delivery. β€’ Collaborate with cross-functional teams including data analysts, scientists, and business stakeholders to support analytics and reporting needs. β€’ Support data warehouse modernization initiatives, including migration from legacy systems to cloud platforms. \_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_ Required Qualifications β€’ 10+ years (ideally 15–20+) of experience in data engineering or data architecture. β€’ Strong expertise in: o Databricks & Delta Lake o Snowflake Data Warehouse o Apache Spark (PySpark, Spark SQL) β€’ Hands-on experience with Azure Cloud (ADLS Gen2, Azure Databricks, ADF). β€’ Proficiency in Python and SQL for data engineering. β€’ Experience with ETL/ELT tools such as Talend or Informatica. β€’ Strong knowledge of data modeling, CDC (Change Data Capture), and incremental loading techniques. β€’ Experience working in Linux/Unix environments with shell scripting. \_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_ Preferred Qualifications β€’ Knowledge of data governance, compliance, and regulatory standards (e.g., IDMP). β€’ Exposure to real-time data streaming technologies (Kafka, Kinesis). β€’ Experience with multi-cloud environments (AWS, GCP). β€’ Familiarity with workflow orchestration tools such as Airflow or Databricks Workflows. \_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_ Key Skills β€’ Data Engineering & Architecture β€’ Lakehouse & Data Warehousing β€’ Spark Performance Optimization β€’ Cloud Data Platforms (Azure) β€’ ETL/ELT Pipeline Development β€’ Data Governance & Security β€’ CI/CD & DevOps Practices \_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_ Education β€’ Bachelor’s degree in Information Technology, Computer Science, or a related field. \_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_ Nice-to-Have Traits β€’ Strong analytical and problem-solving skills β€’ Ability to work in enterprise-scale, complex environments β€’ Experience working with global stakeholders and cross-functional teams β€’ Leadership capability with mentoring experience