CloudIngest

Data Engineer – GMP- (Only W2)

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer – GMP in Indianapolis, IN, with a contract length of "unknown" and a pay rate of "unknown". Requires strong experience in GMP, data pipelines, AWS Glue, and Power BI. Work is onsite 3 days a week.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
February 6, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
W2 Contractor
-
🔒 - Security
Unknown
-
📍 - Location detailed
Indianapolis, IN
-
🧠 - Skills detailed
#Data Lineage #API (Application Programming Interface) #Microsoft Power BI #BI (Business Intelligence) #Monitoring #Lambda (AWS Lambda) #Data Lake #AWS S3 (Amazon Simple Storage Service) #SaaS (Software as a Service) #S3 (Amazon Simple Storage Service) #Data Warehouse #DMS (Data Migration Service) #"ETL (Extract #Transform #Load)" #Batch #Data Governance #Metadata #REST API #Data Catalog #Azure #Data Pipeline #Oracle #Data Quality #REST (Representational State Transfer) #Azure Data Factory #AWS Glue #Compliance #AWS (Amazon Web Services) #Security #Classification #AWS DMS (AWS Database Migration Service) #Data Engineering #ADF (Azure Data Factory) #Data Ingestion
Role description
Data Engineer -Location: Indianapolis, IN (3 days onsite/week)- Only W2 Only W2 Role Overview: The Data Engineer will be responsible for designing, building, and maintaining data pipelines that extract data from source systems, transform it through the medallion architecture layers, and prepare it for consumption by the analytics layer. This role will work closely with the Power BI Developer to ensure data is properly structured, documented, and accessible for reporting and analytics. GMP experience is a must. Resource will need to work from Indianapolis client office at least 3 days a week. Key Responsibilities Phase 1 Deliverables • Following EDB standard, design and establish AWS S3 bucket structure for data lake (Bronze/Silver/Gold zones) with Red CCI security controls • Build new data pipelines to extract data from the selected SaaS-based HSE systems • Enhance/extend data pipelines using the dataset from AWS EDB Asset related data products and digital solutions • MES selection to be finalized by the end of Jan 2027 - Build data ingestion pipeline for selected MES via REST API (Lambda-based extraction). If a different MES is chosen, change course accordingly • Data pipeline to SaaS based HSE tools to integrate data into the data warehouse • Implement CDC pipeline for LabVantage LIMS using AWS DMS from Oracle database (or Azure Data Factory) • Develop Bronze-to-Silver transformations using AWS Glue or Azure Data Factory depending on data domains • Configure AWS Glue Data Catalog with appropriate metadata and Red CCI classification tags Phase 2 Deliverables • Connect to EDB marketplace for enterprise reference data • Build Silver-to-Gold transformations creating batch-centric data products • Implement data quality checks and monitoring dashboards • Develop orchestration workflows using AWS Step Functions • Support Power BI Developer in validating OneLake shortcut connectivity to Gold zone • Document data lineage, schema definitions, and pipeline architecture Ongoing Responsibilities • Monitor pipeline health, troubleshoot failures, and optimize performance • Collaborate with Power BI Developer on data model requirements and data quality issues • Support data governance and security compliance reviews • Respond to ad-hoc data requests from engineers • Coordinate with enterprise EDB team on data sharing agreements and standards