

Zillion Technologies, Inc.
Snowflake Data Engineer/architect(Only W2/ Hybrid - NJ)
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Snowflake Data Engineer/Architect in Berkeley Heights, NJ (hybrid) with a contract length of "unknown" and pay rate of "unknown." Requires 7–10 years of experience, expert-level SQL and Snowflake skills, and cloud platform experience.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
March 3, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Hybrid
-
📄 - Contract
W2 Contractor
-
🔒 - Security
Unknown
-
📍 - Location detailed
Berkeley Heights, NJ
-
🧠 - Skills detailed
#Microsoft Azure #"ETL (Extract #Transform #Load)" #Azure #Data Science #AWS (Amazon Web Services) #Data Warehouse #Snowflake #Python #Leadership #EDW (Enterprise Data Warehouse) #Migration #Data Engineering #Scala #Data Architecture #Data Migration #Security #Cloud #GCP (Google Cloud Platform) #Computer Science #dbt (data build tool) #Strategy #SQL (Structured Query Language)
Role description
Position Title: Snowflake Data Engineer/ Architect
Location: Berkeley Heights, NJ(hybrid)
Experience Level: Senior / Principal
Industry Focus: Financial Services, Healthcare, Manufacturing, Fortune 500 Enterprise
Only w2 candidates can apply for this.
Position Overview
We are seeking a highly experienced Snowflake Data Engineer /Architect to design, build, and optimize enterprise-grade data solutions in complex, large-scale environments. This role requires deep expertise in Snowflake, cloud data platforms, and modern ELT architectures, combined with strong business acumen and leadership capabilities.
You will play a key role in architecting scalable data warehouse solutions, leading data migration initiatives, and driving measurable business outcomes for enterprise clients.
This is not a hands-on developer-only role — we are looking for a strategic technical leader who can translate complex business problems into high-impact, scalable data solutions.
Key Responsibilities
Architecture & Engineering
• Design and implement enterprise data warehouse solutions using Snowflake.
• Architect and optimize large-scale ETL/ELT pipelines.
• Lead data migration initiatives from legacy systems to modern cloud platforms.
• Develop and maintain scalable data models using SQL, Python, and dbt.
• Implement performance optimization strategies achieving measurable improvements.
• Troubleshoot and resolve production data issues in mission-critical environments.
Cloud & Platform Strategy
• Design and implement data solutions across Azure, AWS, or GCP.
• Define best practices for cloud data architecture and governance.
• Ensure scalability, reliability, and security of enterprise data platforms.
Leadership & Stakeholder Engagement
• Translate complex business requirements into technical data solutions.
• Partner with executive stakeholders and cross-functional teams.
• Lead and mentor junior data engineers.
• Drive delivery across multi-stakeholder enterprise environments.
• Demonstrate measurable ROI through cost savings, performance gains, or SLA improvements.
Required Qualifications
• 7–10 years of relevant data engineering experience (minimum 5 years for exceptional candidates).
• Expert-level SQL proficiency.
• Expert-level Snowflake experience.
• Strong Python skills in a data engineering context.
• Experience with Azure, AWS, or GCP cloud platforms.
• Hands-on experience with dbt.
• Enterprise data warehouse design experience.
• Large-scale ETL/ELT pipeline implementation.
• Data migration project leadership experience.
• Experience mentoring or leading engineering teams.
• Bachelor’s degree in Computer Science, Engineering, or related technical field.
Preferred Qualifications
• Snowflake SnowPro Core certification (strongly preferred).
• Snowflake SnowPro Advanced Architect certification.
• Microsoft Azure Data or Solutions Architect certification.
• dbt certification.
• Master’s degree (MBA, M.S. in Data Science, Information Science, or related field).
• Experience in Financial Services, Healthcare, Manufacturing, or Fortune 500 enterprises.
• Demonstrated 30–90% performance optimization improvements.
• Proven record of measurable business impact (cost savings, system performance, SLA improvements).
Position Title: Snowflake Data Engineer/ Architect
Location: Berkeley Heights, NJ(hybrid)
Experience Level: Senior / Principal
Industry Focus: Financial Services, Healthcare, Manufacturing, Fortune 500 Enterprise
Only w2 candidates can apply for this.
Position Overview
We are seeking a highly experienced Snowflake Data Engineer /Architect to design, build, and optimize enterprise-grade data solutions in complex, large-scale environments. This role requires deep expertise in Snowflake, cloud data platforms, and modern ELT architectures, combined with strong business acumen and leadership capabilities.
You will play a key role in architecting scalable data warehouse solutions, leading data migration initiatives, and driving measurable business outcomes for enterprise clients.
This is not a hands-on developer-only role — we are looking for a strategic technical leader who can translate complex business problems into high-impact, scalable data solutions.
Key Responsibilities
Architecture & Engineering
• Design and implement enterprise data warehouse solutions using Snowflake.
• Architect and optimize large-scale ETL/ELT pipelines.
• Lead data migration initiatives from legacy systems to modern cloud platforms.
• Develop and maintain scalable data models using SQL, Python, and dbt.
• Implement performance optimization strategies achieving measurable improvements.
• Troubleshoot and resolve production data issues in mission-critical environments.
Cloud & Platform Strategy
• Design and implement data solutions across Azure, AWS, or GCP.
• Define best practices for cloud data architecture and governance.
• Ensure scalability, reliability, and security of enterprise data platforms.
Leadership & Stakeholder Engagement
• Translate complex business requirements into technical data solutions.
• Partner with executive stakeholders and cross-functional teams.
• Lead and mentor junior data engineers.
• Drive delivery across multi-stakeholder enterprise environments.
• Demonstrate measurable ROI through cost savings, performance gains, or SLA improvements.
Required Qualifications
• 7–10 years of relevant data engineering experience (minimum 5 years for exceptional candidates).
• Expert-level SQL proficiency.
• Expert-level Snowflake experience.
• Strong Python skills in a data engineering context.
• Experience with Azure, AWS, or GCP cloud platforms.
• Hands-on experience with dbt.
• Enterprise data warehouse design experience.
• Large-scale ETL/ELT pipeline implementation.
• Data migration project leadership experience.
• Experience mentoring or leading engineering teams.
• Bachelor’s degree in Computer Science, Engineering, or related technical field.
Preferred Qualifications
• Snowflake SnowPro Core certification (strongly preferred).
• Snowflake SnowPro Advanced Architect certification.
• Microsoft Azure Data or Solutions Architect certification.
• dbt certification.
• Master’s degree (MBA, M.S. in Data Science, Information Science, or related field).
• Experience in Financial Services, Healthcare, Manufacturing, or Fortune 500 enterprises.
• Demonstrated 30–90% performance optimization improvements.
• Proven record of measurable business impact (cost savings, system performance, SLA improvements).





