Lead GCP Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead GCP Data Engineer in Boston, MA (Hybrid) with a contract length of "unknown" and a pay rate of "unknown." Candidates must have 6+ years in Python, PySpark, GCP, SQL, and relevant GCP certification.
🌎 - Country
United States
💱 - Currency
$ USD
💰 - Day rate
Unknown
Unknown
🗓️ - Date discovered
May 15, 2025
🕒 - Project duration
Unknown
🏝️ - Location type
Hybrid
📄 - Contract type
Unknown
🔒 - Security clearance
Unknown
📍 - Location detailed
Boston, MA
🧠 - Skills detailed
#Data Quality #Data Engineering #Deployment #Business Analysis #DevOps #Scala #Spark (Apache Spark) #Data Pipeline #EDW (Enterprise Data Warehouse) #Pandas #Cloud #GIT #SQL (Structured Query Language) #BigQuery #Documentation #PySpark #Data Mart #Data Accuracy #Python #"ETL (Extract #Transform #Load)" #Data Warehouse #Automation #GCP (Google Cloud Platform) #Version Control #Code Reviews #Snowflake #Storage
Role description
Lead GCP Data Engineer Boston MA (Hybrid ) We are looking for a skilled and motivated Lead Data Engineer with strong expertise in Python, PySpark, and Pandas, along with hands-on experience in Google Cloud Platform (GCP) services. The ideal candidate should have a deep understanding of data warehousing concepts, dimension modeling, and SQL, with proven experience building and supporting enterprise-level data solutions in a cloud environment.. Key Responsibilities: • Design, develop, and maintain data pipelines and ETL workflows using Python and PySpark. • Build and manage Enterprise Data Warehouses and Data Marts, ensuring high performance and scalability on GCP. • Work on descriptive analytics and reporting by transforming and querying data using BigQuery SQL. • Conduct peer code reviews, contribute to technical design documents, and help write unit tests and integration test cases. • Collaborate with business analysts and stakeholders to understand data requirements and ensure data accuracy. • Support, monitor, and troubleshoot data pipelines and systems in a production environment. • Participate in deployment and CI/CD processes using GIT and GCP DevOps tools. • Maintain detailed documentation and contribute to a knowledge repository for team enablement and capability building. • Demonstrate strong problem-solving and communication skills in a fast-paced, collaborative setting. • Stay updated with evolving cloud technologies and suggest improvements to enhance performance and efficiency. Required Skills & Qualifications: • 6+ years of experience working with Python, PySpark, and GCP. • Hands-on experience with data transformation, data quality checks, and data validation in large-scale systems. • Strong expertise in SQL (including advanced SQL concepts) and data warehousing methodologies. • Hands-on experience with BigQuery, DataProc, and Cloud Storage in GCP. • Understanding of dimensional modeling, star/snowflake schemas, and data marts. • Experience with data validation, data quality, and ETL automation in large-scale environments. • Relevant GCP certification (e.g., Google Professional Data Engineer) is preferred. • Familiarity with version control systems like GIT and CI/CD processes