Senior Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer in Dallas, TX, offering a long-term contract at an undisclosed pay rate. Candidates must have 10+ years of IT experience, extensive data migration, testing experience, and proficiency in SQL, Python, AWS, and ETL tools.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
-
πŸ—“οΈ - Date discovered
September 11, 2025
πŸ•’ - Project duration
Unknown
-
🏝️ - Location type
On-site
-
πŸ“„ - Contract type
Unknown
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
Dallas, TX
-
🧠 - Skills detailed
#Data Modeling #Kafka (Apache Kafka) #Data Migration #Python #JSON (JavaScript Object Notation) #HiveQL #S3 (Amazon Simple Storage Service) #Cloud #Data Warehouse #DataStage #Storage #Teradata #Data Lake #Pig #Sybase #HDFS (Hadoop Distributed File System) #MongoDB #REST API #Migration #Unix #Deployment #API (Application Programming Interface) #Data Engineering #Hadoop #Data Integration #Jenkins #SQL Queries #ML (Machine Learning) #Redis #Spark (Apache Spark) #System Testing #Agile #"ETL (Extract #Transform #Load)" #Data Science #GitLab #SageMaker #Docker #Oracle #Zookeeper #Strategy #SSIS (SQL Server Integration Services) #Athena #Data Quality #SQL Server #R #Integration Testing #Data Architecture #REST (Representational State Transfer) #Linux #Scala #SQL (Structured Query Language) #Teradata SQL #Java #Automation #Informatica #Programming #Regression #Scripting #AWS (Amazon Web Services) #EC2 #Redshift #Alteryx #DevOps
Role description
JOB DESCRIPTION: TITLE: DATA QUALITY ENGINEER Location: Dallas, TX (NEED ONLY LOCALS – ONE ROUND OF FACE TO FACE AVAILABILITY IS MUST) Duration: Long term Interview: Mode One Teams call, One Face to Face interview MUST HAVE 10+ Years experience over all IT experience…! Data Engineer Responsibilities: (SDET) β€’ Work with business stakeholders, Business Systems Analysts and Developers to ensure quality delivery of software. β€’ Interact with key business functions to confirm data quality policies and governed attributes. β€’ Follow quality management best practices and processes to bring consistency and completeness to integration service testing β€’ Designing and managing the testing AWS environments of data workflows during development and deployment of data products β€’ Provide assistance to the team in Test Estimation & Test Planning β€’ Design, development of Reports and dashboards. β€’ Analyzing and evaluating data sources, data volume, and business rules. β€’ Proficiency with SQL, familiarity with Python, Scala, Athena, EMR, Redshift and AWS. β€’ No SQL data and unstructured data experience. β€’ Extensive experience in programming tools like Map Reduce to HIVEQL β€’ Experience in data science platforms like SageMaker/Machine Learning Studio/ H2O. β€’ Should be well versed with the Data flow and Test Strategy for Cloud/ On Prem ETL Testing. β€’ Interpret and analyses data from various source systems to support data integration and data reporting needs. β€’ Experience in testing Database Application to validate source to destination data movement and transformation. β€’ Work with team leads to prioritize business and information needs. β€’ Develop complex SQL scripts (Primarily Advanced SQL) for Cloud ETL and On prem. β€’ Develop and summarize Data Quality analysis and dashboards. β€’ Knowledge of Data modeling and Data warehousing concepts with emphasis on Cloud/ On Prem ETL. β€’ Execute testing of data analytic and data integration on time and within budget. β€’ Work with team leads to prioritize business and information needs β€’ Troubleshoot & determine best resolution for data issues and anomalies β€’ Experience in Functional Testing, Regression Testing, System Testing, Integration Testing & End to End testing. β€’ Has deep understanding of data architecture & data modeling best practices and guidelines for different data and analytic platforms Requirements: β€’ Extensive Experience in Data migration is a must ( Teradata to Redshift preferred) β€’ Extensive testing Experience with SQL/Unix/Linux scripting is a must β€’ Extensive experience testing Cloud/On Prem ETL (e.g. Abinitio, Informatica, SSIS, Datastage, Alteryx, Glu) β€’ Extensive experience DBMS like Oracle, Teradata, SQL Server, DB2, Redshift, Postgres and Sybase. β€’ Extensive experience using Python scripting and AWS and Cloud Technologies. β€’ Extensive experience using Athena, EMR, Redshift, AWS, and Cloud Technologies β€’ Experienced in large-scale application development testing – Cloud/ On Prem Data warehouse, Data Lake, Data science β€’ Experience with multi-year, large-scale projects β€’ Expert technical skills with hands-on testing experience using SQL queries. β€’ Extensive experience with both data migration and data transformation testing β€’ Extensive experience DBMS like Oracle, Teradata, SQL Server, DB2, Redshift, Postgres and Sybase. β€’ Extensive testing Experience with SQL/Unix/Linux. β€’ Extensive experience testing Cloud/On Prem ETL (e.g. Abinitio, Informatica, SSIS, Datastage, Alteryx, Glu) β€’ Extensive experience using Python scripting and AWS and Cloud Technologies. β€’ Extensive experience using Athena, EMR , Redshift and AWS and Cloud Technologies β€’ API/Rest Assured automation, building reusable frameworks, and good technical expertise/acumen β€’ Java/Java Script - Implement core Java, Integration, Core Java and API. β€’ Functional/UI/ Selenium - BDD/Cucumber, Specflow, Data Validation/Kafka, BigData, also automation experience using Cypress. β€’ AWS/Cloud - Jenkins/ Gitlab/ EC2 machine, S3 and building Jenkins and CI/CD pipelines, SouceLabs. β€’ API/Rest API - Rest API and Micro Services using JSON, SoapUI β€’ Extensive experience in DevOps/Data Ops space. β€’ Strong experience in working with DevOps and build pipelines. β€’ Strong experience of AWS data services including Redshift, Glue, Kinesis, Kafka (MSK) and EMR/ Spark, Sage Maker etc… β€’ Experience with technologies like Kubeflow, EKS, Docker β€’ Extensive experience using No SQL data and unstructured data experience like MongoDB, Cassandra, Redis, ZooKeeper. β€’ Extensive experience in Map reduce using tools like Hadoop, Hive, Pig, Kafka, S4, Map R. β€’ Experience using Jenkins and Gitlab β€’ Experience using both Waterfall and Agile methodologies. β€’ Experience in testing storage tools like S3, HDFS β€’ Experience with one or more industry-standard defect or Test Case management Tools β€’ Great communication skills (regularly interacts with cross functional team members)