

Tek Dallas Inc.
Python with GCP | Only W2
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Python Data Engineer with GCP experience, focusing on data marts and API development. Contract length is unspecified, with a pay rate of "unknown." Key skills include Python, BigQuery, PySpark, and data warehousing.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
March 25, 2026
π - Duration
Unknown
-
ποΈ - Location
Unknown
-
π - Contract
W2 Contractor
-
π - Security
Unknown
-
π - Location detailed
United States
-
π§ - Skills detailed
#Data Aggregation #PySpark #AWS Databases #Cybersecurity #Data Mart #ML (Machine Learning) #GCP (Google Cloud Platform) #REST API #Scala #AI (Artificial Intelligence) #Oracle #Big Data #BigQuery #Python #REST (Representational State Transfer) #Databases #Data Processing #AWS (Amazon Web Services) #Kubernetes #Data Pipeline #Cloud #Data Migration #Migration #Hadoop #Lean #Data Engineering #Spark (Apache Spark) #SQL (Structured Query Language) #Data Warehouse #API (Application Programming Interface) #Automation #Data Ingestion #Security #Programming
Role description
Company Description
TekDallas specializes in delivering innovative technology solutions with a strong emphasis on cybersecurity, including Zero Trust Architectures and Risk-Based Authentication. The company is a leader in AI, machine learning, data analytics, and cloud migration, offering tailored solutions that empower businesses to securely innovate and remain resilient. Serving industries such as healthcare, finance, retail, and manufacturing, TekDallas integrates advanced analytics and automation to optimize operations. Committed to a security-first approach, TekDallas helps organizations enhance productivity, protect data, and achieve long-term success.
Role Description
The still need strong Data Engineerβs who are hands on in Python but put an emphasis on their GCP, BQ and API skills.Β Awesome work here team and a couple more wins for us to track down.
β’ New Use Cases & Skill Requirements
β’ Data Marts & Mini Data Warehouses:
β’ Building data marts using the same GCP stack.
β’ These are like mini data warehouses.
β’ Involve dealing with numbers and large volumes of claims data.
β’ Require aggregations and stuff.
β’ Specific Skill Needs:
β’ Someone with good BigQuery PySpark skills.
β’ Experience in data warehousing.
β’ Experience building data warehouses.
β’ Background in working with transaction data (not just master data).
β’ Experience with facts and aggregations.
β’ Provider Portal & API Development:
β’ Building a portal that shows holistic data for a provider (UI is a separate team using Appian).
β’ Involves API work.
β’ Building Python APIs to read analytics databases like BigQuery (and potentially AWS databases).
β’ Specific Skill Needs (for APIs):
β’ Someone with API building with Python.
β’ Strong Python REST API developers.
β’ Resource Allocation: Don't necessarily need one person with all skills; looking for resources for each area.
β’ Data Aggregation Focus: Candidates will heavily lean on the data side for building data marts where data is already aggregated.
β’ Challenges & Collaboration
β’ Data Requirements:
β’ Concern about the network and provider team giving the right requirements for consolidating data.
β’ Need to reach out to multiple teams as the data is very different.
β’ Difficulty in defining metrics like "claims paid data" (e.g., bill charges vs. actual check amount).
β’ May need to supplement expertise from another domain (e.g., someone with knowledge of provider data, claim data).
Overview: At a high level, they have migrated from Hadoop to GCP for data processing. Have a GCP data environment, predominantly for big data applications on the cloud. Seeking 3-5 Senior Level Data Engineers with strong Python skills to support ongoing data migration and ingestion efforts.
β’ Source systems: get data from multiple external channels - provider data, healthcare groups, hospitals, etc. send data and their platform processes it and provides to operation systems.
β’ Their end state is not a data warehouse for analytics - but the data directly feeds applications.
β’ Currently, a lot of the data ingestion is being done manually and they are looking to automate.
β’ Their data pipelines are PySpark, Scala/Spark run on Dataproc for larger volumes.
β’ Python, Google Cloud Functions to execute the scripts with Google Kubernetes Engine (GKE)
β’ Should have experience in working with denormalized data types, both structured and unstructured data.
β’ Using Cloud SQL as relational cloud database, but okay with others i.e. Oracle, Postgres.
β’ Building AI use cases as well - 5-6 use cases on their plate right now, including AI for data pipeline builds. Folks who can have some background in developing AI applications would be the ideal profile. If we found a strong ML or AI candidate with Python programming skills, they could potentially find a space for them.
Must Have:
β’ Strong hands-on Python programming
β’ Spark/PySpark
β’ GCP (BigQuery, Dataproc, Google Cloud Functions, GKE, Cloud SQL - would not consider all must haves, just general awareness of the GCP ecosystem and data services)
β’ Experience working with various data types and structures
Nice to Have:
β’ AI experience - building AI systems, models, or building inference pipelines and processing data "for AI"
Email id : mark@tekdallas.com
Company Description
TekDallas specializes in delivering innovative technology solutions with a strong emphasis on cybersecurity, including Zero Trust Architectures and Risk-Based Authentication. The company is a leader in AI, machine learning, data analytics, and cloud migration, offering tailored solutions that empower businesses to securely innovate and remain resilient. Serving industries such as healthcare, finance, retail, and manufacturing, TekDallas integrates advanced analytics and automation to optimize operations. Committed to a security-first approach, TekDallas helps organizations enhance productivity, protect data, and achieve long-term success.
Role Description
The still need strong Data Engineerβs who are hands on in Python but put an emphasis on their GCP, BQ and API skills.Β Awesome work here team and a couple more wins for us to track down.
β’ New Use Cases & Skill Requirements
β’ Data Marts & Mini Data Warehouses:
β’ Building data marts using the same GCP stack.
β’ These are like mini data warehouses.
β’ Involve dealing with numbers and large volumes of claims data.
β’ Require aggregations and stuff.
β’ Specific Skill Needs:
β’ Someone with good BigQuery PySpark skills.
β’ Experience in data warehousing.
β’ Experience building data warehouses.
β’ Background in working with transaction data (not just master data).
β’ Experience with facts and aggregations.
β’ Provider Portal & API Development:
β’ Building a portal that shows holistic data for a provider (UI is a separate team using Appian).
β’ Involves API work.
β’ Building Python APIs to read analytics databases like BigQuery (and potentially AWS databases).
β’ Specific Skill Needs (for APIs):
β’ Someone with API building with Python.
β’ Strong Python REST API developers.
β’ Resource Allocation: Don't necessarily need one person with all skills; looking for resources for each area.
β’ Data Aggregation Focus: Candidates will heavily lean on the data side for building data marts where data is already aggregated.
β’ Challenges & Collaboration
β’ Data Requirements:
β’ Concern about the network and provider team giving the right requirements for consolidating data.
β’ Need to reach out to multiple teams as the data is very different.
β’ Difficulty in defining metrics like "claims paid data" (e.g., bill charges vs. actual check amount).
β’ May need to supplement expertise from another domain (e.g., someone with knowledge of provider data, claim data).
Overview: At a high level, they have migrated from Hadoop to GCP for data processing. Have a GCP data environment, predominantly for big data applications on the cloud. Seeking 3-5 Senior Level Data Engineers with strong Python skills to support ongoing data migration and ingestion efforts.
β’ Source systems: get data from multiple external channels - provider data, healthcare groups, hospitals, etc. send data and their platform processes it and provides to operation systems.
β’ Their end state is not a data warehouse for analytics - but the data directly feeds applications.
β’ Currently, a lot of the data ingestion is being done manually and they are looking to automate.
β’ Their data pipelines are PySpark, Scala/Spark run on Dataproc for larger volumes.
β’ Python, Google Cloud Functions to execute the scripts with Google Kubernetes Engine (GKE)
β’ Should have experience in working with denormalized data types, both structured and unstructured data.
β’ Using Cloud SQL as relational cloud database, but okay with others i.e. Oracle, Postgres.
β’ Building AI use cases as well - 5-6 use cases on their plate right now, including AI for data pipeline builds. Folks who can have some background in developing AI applications would be the ideal profile. If we found a strong ML or AI candidate with Python programming skills, they could potentially find a space for them.
Must Have:
β’ Strong hands-on Python programming
β’ Spark/PySpark
β’ GCP (BigQuery, Dataproc, Google Cloud Functions, GKE, Cloud SQL - would not consider all must haves, just general awareness of the GCP ecosystem and data services)
β’ Experience working with various data types and structures
Nice to Have:
β’ AI experience - building AI systems, models, or building inference pipelines and processing data "for AI"
Email id : mark@tekdallas.com





