Yochana

Data Lead Engineer-Fremont, CA-onsite

โญ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Lead Engineer in Fremont, CA, with a contract length of "unknown" and a pay rate of "unknown." Candidates should have 12+ years of experience and expertise in Azure Data Factory, PySpark, and cloud-native solutions in banking, healthcare, and retail.
๐ŸŒŽ - Country
United States
๐Ÿ’ฑ - Currency
$ USD
-
๐Ÿ’ฐ - Day rate
Unknown
-
๐Ÿ—“๏ธ - Date
April 24, 2026
๐Ÿ•’ - Duration
Unknown
-
๐Ÿ๏ธ - Location
On-site
-
๐Ÿ“„ - Contract
Unknown
-
๐Ÿ”’ - Security
Unknown
-
๐Ÿ“ - Location detailed
Fremont, CA
-
๐Ÿง  - Skills detailed
#Synapse #Triggers #AWS (Amazon Web Services) #AWS Glue #HDFS (Hadoop Distributed File System) #Data Engineering #Data Modeling #Azure SQL #Delta Lake #ADLS (Azure Data Lake Storage) #Redshift #PySpark #"ETL (Extract #Transform #Load)" #Snowflake #Scala #SQL (Structured Query Language) #Azure #Spark SQL #Apache Spark #Apache Airflow #Airflow #AWS Kinesis #Python #Indexing #Databricks #Cloud #PostgreSQL #Oracle #Azure Data Factory #Spark (Apache Spark) #Azure Databricks #Data Processing #Kafka (Apache Kafka) #ADF (Azure Data Factory) #Data Architecture
Role description
Position: Data Lead Engineer Location: Fremont, CA-onsite Experience: 12+ Years Primary Skills: Azure Data Factory (ADF), Azure Databricks, PySpark, Azure Synapse, Azure SQL, Python, Spark SQL Candidate Summary: Accomplished Data Architect / Senior Data Engineer with 12+ years of experience designing and modernizing enterprise data platforms across banking, healthcare, and retail domains. Strong expertise in building scalable, cloud-native data solutions across Azure and AWS ecosystems. Key Highlights: โ€ข Architected scalable ETL/ELT pipelines using Python, PySpark, Databricks, AWS Glue, and Azure Data Factory for high-volume data processing โ€ข Led enterprise-level data architecture initiatives including data modeling, governance, and cloud platform design โ€ข Designed Medallion (Bronze/Silver/Gold) Lakehouse architectures using Delta Lake, ADLS Gen2, Snowflake, Redshift, and Synapse โ€ข Built large-scale distributed data processing solutions using Apache Spark, Databricks, EMR, Hive, and HDFS handling billions of records โ€ข Orchestrated workflows using Apache Airflow, Databricks Workflows, AWS Step Functions, and ADF triggers โ€ข Strong expertise in Advanced SQL (CTEs, window functions, performance tuning, indexing, partitioning) across Snowflake, PostgreSQL, and Oracle โ€ข Developed real-time streaming solutions using Kafka, AWS Kinesis, Azure Event Hub, and Service Bus