

Brooksource
Data Engineer
โญ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Quality Engineer on a 12-month contract, paying "rate" and requiring on-site work in Charlotte, NC. Key skills include AWS Glue, PySpark, Python, Kafka, and advanced SQL with experience in data validation and reconciliation.
๐ - Country
United States
๐ฑ - Currency
$ USD
-
๐ฐ - Day rate
520
-
๐๏ธ - Date
April 24, 2026
๐ - Duration
More than 6 months
-
๐๏ธ - Location
On-site
-
๐ - Contract
Unknown
-
๐ - Security
Unknown
-
๐ - Location detailed
Charlotte Metro
-
๐ง - Skills detailed
#Data Pipeline #Data Quality #AWS (Amazon Web Services) #AWS Glue #Data Engineering #Data Modeling #PySpark #"ETL (Extract #Transform #Load)" #DMS (Data Migration Service) #Data Reconciliation #Lambda (AWS Lambda) #Deployment #SQL (Structured Query Language) #Monitoring #AWS Lambda #AWS DMS (AWS Database Migration Service) #BI (Business Intelligence) #Documentation #Logging #Python #Data Layers #Cloud #PostgreSQL #Batch #Observability #Qlik #Aurora PostgreSQL #Spark (Apache Spark) #Kafka (Apache Kafka) #Data Architecture #Aurora #Datasets
Role description
Senior Data Quality Engineer
12 โ month contract (high likelihood of extension or full-time conversion)
On โ site 3 days a week in Charlotte, NC
Role Overview
We are seeking a Senior Data Quality Engineer to design, implement, and maintain automated data quality validations across our enterprise data engineering ecosystem. This role focuses on ensuring the accuracy, completeness, consistency, and timeliness of data flowing through both batch and streaming pipelines built on AWS.
You will work closely with data engineers, analytics teams, and business stakeholders to embed data quality controls into pipelines built with AWS Glue, PySpark, Kafka, AWS DMS, Lambda, and Aurora PostgreSQL, supporting trusted analytics and reporting in Qlik.
Key Responsibilities
Data Quality Engineering
โข Design and implement automated data quality checks across ingestion, transformation, and consumption layers.
โข Define and enforce data quality rules for key dimensions such as completeness, validity, uniqueness, consistency, and timeliness.
โข Build reusable Python and PySpark frameworks for validating large-scale datasets.
Batch & Streaming Validation
โข Embed data quality validations into AWS Glue (PySpark) batch pipelines.
โข Implement real-time or near-real-time validations for Kafka-based streaming pipelines, including schema validation, duplicate detection, and latency checks.
โข Monitor and validate event-time vs. processing-time behavior for streaming data.
CDC & Ingestion Quality
โข Validate AWS DMS change data capture pipelines, ensuring accuracy between source systems and downstream targets.
โข Perform reconciliation checks (row counts, aggregates, checksums) between source and target systems.
โข Detect and alert on data gaps, duplication, or schema drift in CDC pipelines.
Data Stores & Analytics Readiness
โข Write advanced SQL-based data quality checks against Amazon Aurora PostgreSQL and curated data layers.
โข Ensure data delivered to Qlik meets defined quality thresholds and freshness SLAs.
โข Validate semantic consistency and completeness of datasets used for reporting and dashboards.
Monitoring, Alerting & Incident Management
โข Implement data quality monitoring, logging, and alerting using AWS Lambda, CloudWatch, and pipeline metrics.
โข Create dashboards and alerts for data quality failures and SLA breaches.
โข Perform root-cause analysis of data quality incidents and drive long-term remediation.
Standards, Governance & Collaboration
โข Partner with data engineers to embed quality gates into CI/CD and deployment workflows.
โข Contribute to data quality standards, documentation, and operational runbooks.
โข Act as a subject-matter expert for data quality best practices across batch and streaming architectures.
Required Qualifications
โข 6+ years of experience in data engineering, analytics engineering, or data quality engineering.
โข Strong hands-on experience with AWS Glue, PySpark, and Python.
โข Experience validating batch and streaming data pipelines.
โข Practical knowledge of Kafka for streaming ingestion and validation use cases.
โข Experience working with AWS DMS for CDC pipelines and data reconciliation.
โข Advanced SQL skills and experience with Amazon Aurora PostgreSQL.
โข Experience implementing serverless workflows using AWS Lambda.
โข Understanding of data modeling concepts and multi-layer data architectures.
โข Strong analytical and problem-solving skills with attention to detail.
โข Ability to communicate data quality issues clearly to technical and non-technical stakeholders.
Preferred Qualifications
โข Experience supporting BI tools such as Qlik or similar analytics platforms.
โข Familiarity with data observability concepts and quality metrics.
โข Knowledge of schema management and schema evolution in streaming systems.
โข Experience in regulated or highly governed data environments.
โข Exposure to CI/CD pipelines and Infrastructure-as-Code practices.
What Success Looks Like
โข Critical datasets have automated, repeatable data quality validations.
โข Data quality issues are detected early and resolved before impacting analytics.
โข Streaming and batch pipelines meet defined quality and freshness SLAs.
โข Business users trust analytics and reporting outputs with minimal manual intervention.
Senior Data Quality Engineer
12 โ month contract (high likelihood of extension or full-time conversion)
On โ site 3 days a week in Charlotte, NC
Role Overview
We are seeking a Senior Data Quality Engineer to design, implement, and maintain automated data quality validations across our enterprise data engineering ecosystem. This role focuses on ensuring the accuracy, completeness, consistency, and timeliness of data flowing through both batch and streaming pipelines built on AWS.
You will work closely with data engineers, analytics teams, and business stakeholders to embed data quality controls into pipelines built with AWS Glue, PySpark, Kafka, AWS DMS, Lambda, and Aurora PostgreSQL, supporting trusted analytics and reporting in Qlik.
Key Responsibilities
Data Quality Engineering
โข Design and implement automated data quality checks across ingestion, transformation, and consumption layers.
โข Define and enforce data quality rules for key dimensions such as completeness, validity, uniqueness, consistency, and timeliness.
โข Build reusable Python and PySpark frameworks for validating large-scale datasets.
Batch & Streaming Validation
โข Embed data quality validations into AWS Glue (PySpark) batch pipelines.
โข Implement real-time or near-real-time validations for Kafka-based streaming pipelines, including schema validation, duplicate detection, and latency checks.
โข Monitor and validate event-time vs. processing-time behavior for streaming data.
CDC & Ingestion Quality
โข Validate AWS DMS change data capture pipelines, ensuring accuracy between source systems and downstream targets.
โข Perform reconciliation checks (row counts, aggregates, checksums) between source and target systems.
โข Detect and alert on data gaps, duplication, or schema drift in CDC pipelines.
Data Stores & Analytics Readiness
โข Write advanced SQL-based data quality checks against Amazon Aurora PostgreSQL and curated data layers.
โข Ensure data delivered to Qlik meets defined quality thresholds and freshness SLAs.
โข Validate semantic consistency and completeness of datasets used for reporting and dashboards.
Monitoring, Alerting & Incident Management
โข Implement data quality monitoring, logging, and alerting using AWS Lambda, CloudWatch, and pipeline metrics.
โข Create dashboards and alerts for data quality failures and SLA breaches.
โข Perform root-cause analysis of data quality incidents and drive long-term remediation.
Standards, Governance & Collaboration
โข Partner with data engineers to embed quality gates into CI/CD and deployment workflows.
โข Contribute to data quality standards, documentation, and operational runbooks.
โข Act as a subject-matter expert for data quality best practices across batch and streaming architectures.
Required Qualifications
โข 6+ years of experience in data engineering, analytics engineering, or data quality engineering.
โข Strong hands-on experience with AWS Glue, PySpark, and Python.
โข Experience validating batch and streaming data pipelines.
โข Practical knowledge of Kafka for streaming ingestion and validation use cases.
โข Experience working with AWS DMS for CDC pipelines and data reconciliation.
โข Advanced SQL skills and experience with Amazon Aurora PostgreSQL.
โข Experience implementing serverless workflows using AWS Lambda.
โข Understanding of data modeling concepts and multi-layer data architectures.
โข Strong analytical and problem-solving skills with attention to detail.
โข Ability to communicate data quality issues clearly to technical and non-technical stakeholders.
Preferred Qualifications
โข Experience supporting BI tools such as Qlik or similar analytics platforms.
โข Familiarity with data observability concepts and quality metrics.
โข Knowledge of schema management and schema evolution in streaming systems.
โข Experience in regulated or highly governed data environments.
โข Exposure to CI/CD pipelines and Infrastructure-as-Code practices.
What Success Looks Like
โข Critical datasets have automated, repeatable data quality validations.
โข Data quality issues are detected early and resolved before impacting analytics.
โข Streaming and batch pipelines meet defined quality and freshness SLAs.
โข Business users trust analytics and reporting outputs with minimal manual intervention.






