

Katalyst CRO
Data Engineer (Data Pipelines & Modeling)
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer (Data Pipelines & Modeling) with a contract length of "unknown," offering a pay rate of "unknown." Key skills include Python, SQL, and Apache Spark, with a focus on data engineering and pipeline development, preferably in healthcare.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
April 19, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Unknown
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Warrendale, PA
-
🧠 - Skills detailed
#Python #GitHub #Cloud #"ETL (Extract #Transform #Load)" #Spark (Apache Spark) #GCP (Google Cloud Platform) #Data Pipeline #Forecasting #Data Quality #Big Data #Jenkins #Computer Science #Azure #Apache Spark #Database Migration #SQL (Structured Query Language) #Data Accuracy #AWS (Amazon Web Services) #Data Engineering #Data Ingestion #Migration #Databases
Role description
Responsibilities
• Design and implement robust data ingestion pipelines from multiple sources (APIs, databases, files, streaming systems).
• Support C4C offline database migration, ensuring data accuracy and consistency.
• Integrate data from enterprise systems into centralized data platforms.
• Design and implement data models for Workforce planning.
• Service operations forecasting.
• Develop optimized schemas for reporting and analytics.
• Ensure data quality, integrity, and consistency across models.
Requirements:
• Strong experience in data engineering and pipeline development.
• Proficiency in Python / SQL.
• Hands-on experience with Apache Spark or similar big data tools.
• Strong understanding of ETL/ELT concepts and data warehousing.
• Ability to work independently and in cross-functional teams.
• Bachelor's / Master's in Computer Science, IT, or related field.
Good to Have:
• Exposure to CI/CD tools like Jenkins or GitHub Actions.
• Knowledge of cloud platforms (AWS / Azure / GCP).
• Experience in healthcare or regulated environments.
Responsibilities
• Design and implement robust data ingestion pipelines from multiple sources (APIs, databases, files, streaming systems).
• Support C4C offline database migration, ensuring data accuracy and consistency.
• Integrate data from enterprise systems into centralized data platforms.
• Design and implement data models for Workforce planning.
• Service operations forecasting.
• Develop optimized schemas for reporting and analytics.
• Ensure data quality, integrity, and consistency across models.
Requirements:
• Strong experience in data engineering and pipeline development.
• Proficiency in Python / SQL.
• Hands-on experience with Apache Spark or similar big data tools.
• Strong understanding of ETL/ELT concepts and data warehousing.
• Ability to work independently and in cross-functional teams.
• Bachelor's / Master's in Computer Science, IT, or related field.
Good to Have:
• Exposure to CI/CD tools like Jenkins or GitHub Actions.
• Knowledge of cloud platforms (AWS / Azure / GCP).
• Experience in healthcare or regulated environments.






