

Big Data Developer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Big Data Developer in Irving, TX, with a contract length of unspecified duration and a pay rate of "unknown." Requires 7-10 years of experience, expertise in Spark, Scala, ETL processes, and cloud environments. Local candidates only.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
July 17, 2025
π - Project duration
Unknown
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Irving, TX
-
π§ - Skills detailed
#Data Quality #Monitoring #AI (Artificial Intelligence) #Python #Data Pipeline #Data Engineering #Data Ingestion #Data Profiling #Java #Spark SQL #Quality Assurance #Teradata #Shell Scripting #Computer Science #Scripting #Spark (Apache Spark) #Big Data #GCP (Google Cloud Platform) #"ETL (Extract #Transform #Load)" #Kafka (Apache Kafka) #Impala #SQL (Structured Query Language) #Data Migration #Project Management #GIT #Jira #Leadership #Cloud #AWS (Amazon Web Services) #Scala #Web Services #HBase #Jenkins #Sqoop (Apache Sqoop) #API (Application Programming Interface) #Libraries #Datasets #Programming #Migration #Strategy #Hadoop #Automation #BI (Business Intelligence) #Data Modeling #UAT (User Acceptance Testing) #Data Integrity
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Big Data Engineer
Visa : H4 EAD , TN , GC , GC EAD and USC
Targeted Years of Experience: 7-10 years
Location β Irving TX 75038 United States
Need only local to Texas
Job Responsibilities:
As a part of our AI & D team, the Data Quality Assurance Engineer will be responsible for leading development and validation activities of Big Data products and applications which run on the large Hadoop and Teradata clusters. The qualified engineer will bring technical leadership for developing and testing ETL process, migrating different applications to cloud, developing data validation tools used for performing quality assessments and measurements on different data sets that feed Big Data products.
The candidate will be involved in,
Lead in design, development and testing of data ingestion pipelines, perform end to end validation of ETL process for various datasets that are being ingested into the big data platform.
Perform data migration and conversion validation activities on different applications and platforms.
Provide the technical leadership on data profiling/analysis, discovery, analysis, suitability and coverage of data, and identify the various data types, formats, and data quality issues which exist within a given data source.
Contribute to development of transformation logic, interfaces and reports as needed to meet project requirements.
Participate in discussion for technical architecture, data modeling and ETL standards, collaborate with Product Managers, Architects and Senior Developers to establish the physical application framework (e.g. libraries, modules, execution environments)
Lead in design and develop validation framework and integrated automated test suites to validate end to end data pipeline flow, data transformation rules, and data integrity.
Develop tools to measure the data quality and visualize the anomaly pattern in source and processed data.
Assist Manager in project planning, validation strategy development
Provide support in User acceptance testing and production validation activities.
Provide technical recommendations for identifying data validation tools, recommend new technologies to improve the validation process.
Evaluate existing methodologies and processes and recommend improvements.
Work with the stakeholders, Product Management, Data and Design, Architecture teams and executives to call out issues, guide and contribute to the resolutions discussions.
Required Skills:
8+ years of Software development and testing experience.
4+ years of Working experience on tools like Spark, HBase, Hive, Sqoop, Impala, Kafka, Flume, Oozie, MapReduce, etc.
4+ years of programming experience in Scala, Java or Python
Experience in technical leading and mentoring the teams
Experience with developing and testing ETL, real-time data-processing and Analytics Application Systems.
Strong knowledge in Spark SQL, Scala code development in big data Hadoop environment and/or BI/DW development experiences.
Strong knowledge in shell scripting Experience in Web Services - API development and testing.
Experience with development and automated framework in a CI/CD environment.
Experience with cloud environments - AWS or GCP is a plus.
Knowledge of GIT/Jenkins and pipeline automation is a must.
A solid understanding of common software development practices and tools.
Strong analytical skills with a methodical approach to problem solving applied to the Big Data domain
Good organizational skills and strong written and verbal communication skills.
Desired Skills:
Working experience on large migration Projects is a big plus.
Working experience on Google Cloud platform is a big plus
Development experience for tools and utilities for monitoring and alert set etc.
Familiarity with project Management and bug tracking tools, i.e., JIRA or a similar tool.
EDUCATION/CERTIFICATIONS:
Please indicate whether education and/or certifications are required or desired.
Bachelor's Degree in computer science, or engineering