Insight Global

Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is a Senior Data Engineer for a 6-month contract in Charlotte (hybrid), paying $40-65/hr. Requires 6+ years of AWS data engineering, Python, PySpark, and experience with AWS Glue, Amazon EMR, and Apache Airflow.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
520
-
πŸ—“οΈ - Date
December 11, 2025
πŸ•’ - Duration
More than 6 months
-
🏝️ - Location
Hybrid
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
Charlotte, NC
-
🧠 - Skills detailed
#AWS Glue #Amazon EMR (Amazon Elastic MapReduce) #SQL (Structured Query Language) #PySpark #Python #Datasets #Data Management #MDM (Master Data Management) #Apache Airflow #ML (Machine Learning) #Data Processing #Data Engineering #Scala #AWS (Amazon Web Services) #Spark (Apache Spark) #Data Pipeline #"ETL (Extract #Transform #Load)" #Airflow #Data Quality
Role description
Senior Data Engineer Charlotte- Hybrid (4 days onsite 1 day in person) 6 months extending 2 Openings - 40-65/hr based on experience level This position is for a Senior AWS Data Engineer working on a Master Data Management (MDM) project. The goal of the project is to create a single, trusted view of business data by cleaning up duplicate and inconsistent information from multiple sources. You’ll be building scalable data pipelines on AWS, improving data quality, and working on advanced features like entity resolution and machine learning-assisted matching. It’s a hands-on role where you’ll own production-grade pipelines and work with large datasets. If you enjoy solving complex data challenges and making systems more efficient, this is a great fit. Main Responsibilities: β€’ Build and maintain data pipelines on AWS β€’ Develop ETL jobs using AWS Glue (PySpark) and Amazon EMR β€’ Orchestrate workflows using Apache Airflow β€’ Support full and incremental data processing β€’ Implement data matching, deduplication, and entity resolution β€’ Monitor, troubleshoot, and support production pipelines β€’ Partner with analytics and business teams Must Haves β€’ 6+ years in AWS data engineering experience β€’ Python and PySpark development β€’ Hands-on with AWS Glue and Amazon EMR β€’ Experience using Apache Airflow β€’ Strong SQL skills β€’ Experience working with large datasets β€’ Familiarity with ML concepts for data quality or matching Nice to have β€’ Experience with entity resolution, fuzzy matching, or deduplication β€’ Experience with AWS Entity Resolution β€’ Experience in Business MDM programs