

GCP Senior Data Lead
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a GCP Senior Data Lead in San Jose, CA, for 12+ months at a contract rate. Requires 12+ years of experience, expertise in Google BigQuery, SQL, Python, and AI/ML data preparation, with a strong focus on GCP services.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
June 14, 2025
π - Project duration
More than 6 months
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
San Jose, CA
-
π§ - Skills detailed
#SQL (Structured Query Language) #Deployment #Spark (Apache Spark) #Data Pipeline #Version Control #GIT #PyTorch #Data Governance #Compliance #Python #TensorFlow #Programming #Monitoring #ML (Machine Learning) #NoSQL #Data Engineering #Airflow #Security #Cloud #Data Lifecycle #GCP (Google Cloud Platform) #Data Accuracy #Data Warehouse #Data Quality #Computer Science #Storage #"ETL (Extract #Transform #Load)" #Datasets #Scala #Automation #AI (Artificial Intelligence) #BigQuery #Data Processing #Dataflow #Keras #Batch #Hadoop #ERWin #Java #Apache Airflow #Data Science #Big Data
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Job Title: GCP Senior Data Lead with Google BigQuery, SQL, Python, Google dataflow, CA Erwin Data Modeler
Job Location: San Jose, CA (100% Onsite)
Job Duration: 12+ months
Minimum years of experience required: 12+ years
Type of Hire: Contract
JOB DETAILS:
We are seeking a highly skilled and experienced Data Engineer with a strong background in AI/ML to design, build, and optimize robust data pipelines and infrastructure on the Google Cloud Platform (GCP). The ideal candidate will be passionate about leveraging data to power machine learning initiatives, ensuring data quality, accessibility, and scalability for advanced analytics and AI applications.
Responsibilities:
Data Pipeline Development: Design, build, and maintain scalable, efficient, and reliable ETL/ELT data pipelines for batch and real-time processing using GCP services (e.g., Dataflow, Dataproc, Cloud Composer, Pub/Sub).
AI/ML Data Preparation: Collaborate closely with Data Scientists and Machine Learning Engineers to understand data requirements for model training, evaluation, and serving. Prepare, transform, and curate large, diverse datasets (structured, unstructured, streaming) to optimize them for AI/ML workloads.
GCP Ecosystem Expertise: Leverage a wide range of GCP data and AI/ML services, including:
Data Warehousing & Storage: BigQuery (for analytics and BigQuery ML), Cloud Storage, Cloud SQL, Cloud Bigtable.
Data Processing: Dataflow, Dataproc (Spark, Hadoop), Cloud Composer (Apache Airflow), Data Fusion.
AI/ML Services: Vertex AI (for model training, deployment, MLOps, Pipelines, Workbench, AutoML), AI Platform, TensorFlow Enterprise, Keras, PyTorch.
Data Governance & Quality: Implement and enforce data quality, security, and governance standards throughout the data lifecycle, ensuring data accuracy, consistency, and compliance with regulations.
Performance Optimization: Monitor, troubleshoot, and optimize the performance and cost-effectiveness of data pipelines and AI/ML infrastructure.
Automation & MLOps: Automate data processes, develop CI/CD pipelines for data and ML models, and contribute to MLOps best practices for seamless deployment and monitoring of AI/ML solutions.
Collaboration & Communication: Work effectively with cross-functional teams, including Data Scientists, Analysts, Software Engineers, and Product Managers, to understand data needs and deliver impactful solutions.
Innovation & Research: Stay up-to-date with the latest advancements in data engineering, AI/ML, and GCP technologies, continuously exploring and recommending new tools and approaches.
Qualifications:
Bachelor's or Master's degree in Computer Science, Data Engineering, or a related quantitative field.
Proven experience as a Data Engineer, with a strong focus on building data solutions for AI/ML applications.
In-depth knowledge and hands-on experience with Google Cloud Platform (GCP) data services (BigQuery, Dataflow, Dataproc, Cloud Storage, Cloud Composer, etc.).
Strong proficiency in programming languages such as Python (essential), and experience with Scala or Java is a plus.
Expertise in SQL and experience with various database technologies (relational, NoSQL, data warehouses).
Familiarity with machine learning concepts, algorithms, and workflows (e.g., feature engineering, model training, evaluation, deployment).
Experience with machine learning frameworks like TensorFlow or PyTorch.
Understanding of distributed systems, big data technologies, and real-time data processing.
Experience with version control systems (e.g., Git) and CI/CD practices.
Excellent problem-solving, analytical, and communication skills.
Google Cloud Professional Data Engineer certification is a strong plus.