Senior Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer with a contract length of "unknown" and a pay rate of "unknown." Required skills include proficiency in Python or Scala, strong SQL experience, and hands-on expertise with AWS, GCP, or Azure.
🌎 - Country
United States
💱 - Currency
$ USD
💰 - Day rate
Unknown
Unknown
🗓️ - Date discovered
April 25, 2025
🕒 - Project duration
Unknown
🏝️ - Location type
Unknown
📄 - Contract type
Unknown
🔒 - Security clearance
Unknown
📍 - Location detailed
San Jose, CA
🧠 - Skills detailed
#"ETL (Extract #Transform #Load)" #Infrastructure as Code (IaC) #Data Modeling #Code Reviews #BigQuery #Kafka (Apache Kafka) #Storage #Compliance #Leadership #S3 (Amazon Simple Storage Service) #SQL (Structured Query Language) #Data Architecture #Azure #Data Quality #Spark (Apache Spark) #Synapse #GCP (Google Cloud Platform) #Python #AWS (Amazon Web Services) #Cloud #Dataflow #Data Governance #Security #Data Warehouse #DevOps #Data Lake #Data Pipeline #Data Science #Scala #GDPR (General Data Protection Regulation) #Airflow #Data Engineering #Redshift
Role description

Key Responsibilities:

   • Design, develop, and maintain scalable and reliable data pipelines using modern data engineering tools and frameworks.

   • Build ETL/ELT pipelines for structured and unstructured data from multiple sources.

   • Implement and manage data lake, data warehouse, and real-time streaming solutions.

   • Work closely with Data Scientists, Analysts, and Product teams to understand data requirements and deliver solutions.

   • Ensure data quality, integrity, and governance across all pipelines.

   • Optimize performance and cost-effectiveness of data processes in the cloud.

   • Lead architecture and design discussions for new data initiatives.

   • Mentor junior data engineers and contribute to best practices and code reviews.

Required Skills:

   • Proficiency in Python or Scala for data engineering.

   • Strong experience with SQL, Spark, Kafka, and Airflow.

   • Deep understanding of ETL/ELT processes, data modeling, and data architecture.

   • Hands-on experience with cloud data platforms such as AWS (Redshift, Glue, EMR, S3), GCP (BigQuery, Dataflow), or Azure (Synapse, Data Factory).

   • Experience with data lakes, data warehouses, and columnar storage formats (Parquet, ORC).

   • Knowledge of DevOps practices, CI/CD for data pipelines, and infrastructure as code.

   • Experience with data governance, security, and compliance (GDPR, HIPAA, etc.).

   • Excellent problem-solving, communication, and leadership skills.