

Principal GCP Data Engineer_Remote_USC/GC or H4EAD_ONLY ON W2
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Principal GCP Data Engineer, remote, with a contract length of "unknown". Pay rate is "unknown". Requires 5-6 years of experience, expertise in Airflow, SnapLogic, Python, SQL, and Google BigQuery, preferably with healthcare experience.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
June 27, 2025
π - Project duration
Unknown
-
ποΈ - Location type
Remote
-
π - Contract type
W2 Contractor
-
π - Security clearance
Unknown
-
π - Location detailed
Eagan, MN
-
π§ - Skills detailed
#Data Modeling #Alteryx #SQL (Structured Query Language) #Apache Beam #Visualization #Migration #Data Engineering #Kafka (Apache Kafka) #Dataflow #Java #Data Pipeline #Redshift #DataStage #"ETL (Extract #Transform #Load)" #BI (Business Intelligence) #GCP (Google Cloud Platform) #Airflow #Data Ingestion #Spark (Apache Spark) #Cloud #Python #BigQuery
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Principal GCP Data Engineer
Remote
USC/GC or H4EAD
ONLY ON W2
Required Skills : Needs experience in Airflow or Cloud Composer orchestration, development of new DAGs from scratch. Development of data ingestion and ETL pipelines from scratch. Using SnapLogic primarily for data pipelines and integrations, but also Python, SQL, Dataflow, Spark. Needs experience in data warehousing, Google BigQuery. Understanding of analytics as a whole, how data moves from source, warehouse, semantic or reporting layer, models, and reporting/BI, but their hands-on focus will be around building data pipelines and orchestration.
Years of experience: Principal 5-6 years roughly - More about what they've done i.e. healthcare experience, exposure to more than one project, communication.
Remote: yes, 9-3 CST are core hours
Pain Points: candidates who do not have the correct focus area (building the data pipelines), candidates who have not built something from scratch; either they are maintaining existing data workflows or processes in production, but not building.
Overview: Have multiple needs for Associate Data Engineer, Senior Data Engineer, and Principal Data Engineer to support multiple ongoing initiatives including migration of data to GCP, building out the GCP BigQuery warehouse (and moving data from Redshift into it), sunsetting legacy ETL (Datastage) and replacing with various new data ingestion solutions.
β’ Experience in SnapLogic is strongly preferred; they have found this to be more amenable to upskilling.
β’ Needs experience in Airflow or Cloud Composer orchestration, development of new DAGs from scratch.
β’ Development of data ingestion and ETL pipelines from scratch. Using SnapLogic primarily for data pipelines and integrations, but also Python, SQL, Dataflow, Spark.
β’ Needs experience in data warehousing, Google BigQuery.
β’ Not responsible for building out visualizations; another team handles
β’ Will be supporting data modeling, but not owning. Should have some experience in data modeling, data warehousing fundamentals.
β’ Understanding of analytics as a whole, how data moves from source, warehouse, semantic or reporting layer, models, and reporting/BI, but their hands-on focus will be around building data pipelines and orchestration.
β’ Proactive communicators, inquisitive people, problem-solvers, unafraid to make suggestions, ask questions. "Order taker" and "heads down" types of Engineers will not be a culture fit for the team.
β’ They have a list of other technologies in their environment in smaller amounts/more dispersed--any would be a "nice to have": i.e. Kafka, Java, Apache Beam, Alteryx, etc.