

PSI (Proteam Solutions)
Big Data Developer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Big Data Developer on a long-term project (12+ months) requiring U.S. citizenship or permanent residency. Key skills include 4+ years in Hadoop, 2+ years with Apache Spark, proficiency in Java, Scala, or Python, and strong data architecture knowledge.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
680
-
ποΈ - Date
December 2, 2025
π - Duration
More than 6 months
-
ποΈ - Location
Unknown
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Columbus, Ohio Metropolitan Area
-
π§ - Skills detailed
#Agile #DevOps #Scala #S3 (Amazon Simple Storage Service) #Leadership #Lambda (AWS Lambda) #Data Engineering #Datasets #Big Data #Spark (Apache Spark) #Data Quality #Python #Kubernetes #Documentation #AWS (Amazon Web Services) #Batch #Monitoring #Hadoop #Data Architecture #Data Modeling #"ETL (Extract #Transform #Load)" #Airflow #Java #Computer Science #Security #Compliance #Apache Spark #Data Pipeline #Data Science #Automation #Data Processing #Cloud
Role description
The Big Data Developer will play a critical role on the Big Data engineering team, designing and implementing large-scale data processing systems that power scientific research and innovation. The ideal candidate has hands-on experience building enterprise-grade data pipelines, working with distributed systems, and optimizing data processing workflows.
This is a long-term project (12+ months) with high visibility, cutting-edge tools, and opportunities to influence technical direction.
β’
β’ U.S. CITIZENSHIP OR PERMANENT RESIDENCY REQUIRED. VISA OR EAD STATUS NOT ACCEPTABLE FOR THIS POSITION. NO EXCEPTIONS
β’
β’ What Youβll Do
β Data Pipeline Design & Development
β’ Design, build, and deploy scalable data pipelines for ingesting, processing, transforming, and storing high-volume datasets.
β’ Implement streaming and batch-processing solutions using Hadoop, Spark, and cloud-based tools.
β Data Architecture & Engineering
β’ Develop and maintain data architecture and data flow models.
β’ Ensure data reliability, accuracy, and integrity across all environments.
β’ Support data warehousing strategies and best practices.
β Data Quality, Security & Compliance
β’ Implement automated data validation, error handling, and monitoring.
β’ Ensure compliance with internal security controls and regulatory standards.
β’ Partner with governance teams to enforce data quality and security guidelines.
β Cross-Functional Collaboration
β’ Work closely with data scientists, analysts, product teams, and application developers.
β’ Translate business requirements into robust technical solutions.
β’ Participate in Agile ceremonies and contribute to technical design discussions.
β Performance Optimization
β’ Tune Spark applications, Hadoop jobs, and distributed data systems for performance and cost efficiency.
β’ Troubleshoot bottlenecks and implement improvements to system performance.
β Technical Leadership
β’ Provide mentorship to junior developers and contribute to coding standards, best practices, and technical documentation.
Required Skills & Qualifications
β’ 4+ years of Big Data Development experience in Hadoop ecosystems
β’ 2+ years of hands-on development with Apache Spark
β’ Proficiency in Java, Scala, or Python
β’ Strong understanding of distributed systems, ETL, data warehousing, and data modeling concepts
β’ Experience with large-scale datasets, performance tuning, and troubleshooting
β’ Strong problem-solving, communication, and collaboration skills
β’ Bachelorβs degree in Computer Science, Engineering, or related discipline
Preferred Skills
β’ Experience working with AWS cloud services (EMR, S3, Lambda, Glue, etc.)
β’ Experience with Spark 3.x or 4.x
β’ Exposure to Kubernetes, Airflow, or similar orchestration tools
β’ Familiarity with CI/CD and DevOps automation for data engineering
Why This Opportunity Stands Out
β’ Long-term project stability (12+ months, likely extension)
β’ Ability to work on high-impact scientific and research-driven datasets
β’ Hands-on cloud modernization (AWS) and next-generation big data tooling
β’ Collaborative and innovative engineering culture
The Big Data Developer will play a critical role on the Big Data engineering team, designing and implementing large-scale data processing systems that power scientific research and innovation. The ideal candidate has hands-on experience building enterprise-grade data pipelines, working with distributed systems, and optimizing data processing workflows.
This is a long-term project (12+ months) with high visibility, cutting-edge tools, and opportunities to influence technical direction.
β’
β’ U.S. CITIZENSHIP OR PERMANENT RESIDENCY REQUIRED. VISA OR EAD STATUS NOT ACCEPTABLE FOR THIS POSITION. NO EXCEPTIONS
β’
β’ What Youβll Do
β Data Pipeline Design & Development
β’ Design, build, and deploy scalable data pipelines for ingesting, processing, transforming, and storing high-volume datasets.
β’ Implement streaming and batch-processing solutions using Hadoop, Spark, and cloud-based tools.
β Data Architecture & Engineering
β’ Develop and maintain data architecture and data flow models.
β’ Ensure data reliability, accuracy, and integrity across all environments.
β’ Support data warehousing strategies and best practices.
β Data Quality, Security & Compliance
β’ Implement automated data validation, error handling, and monitoring.
β’ Ensure compliance with internal security controls and regulatory standards.
β’ Partner with governance teams to enforce data quality and security guidelines.
β Cross-Functional Collaboration
β’ Work closely with data scientists, analysts, product teams, and application developers.
β’ Translate business requirements into robust technical solutions.
β’ Participate in Agile ceremonies and contribute to technical design discussions.
β Performance Optimization
β’ Tune Spark applications, Hadoop jobs, and distributed data systems for performance and cost efficiency.
β’ Troubleshoot bottlenecks and implement improvements to system performance.
β Technical Leadership
β’ Provide mentorship to junior developers and contribute to coding standards, best practices, and technical documentation.
Required Skills & Qualifications
β’ 4+ years of Big Data Development experience in Hadoop ecosystems
β’ 2+ years of hands-on development with Apache Spark
β’ Proficiency in Java, Scala, or Python
β’ Strong understanding of distributed systems, ETL, data warehousing, and data modeling concepts
β’ Experience with large-scale datasets, performance tuning, and troubleshooting
β’ Strong problem-solving, communication, and collaboration skills
β’ Bachelorβs degree in Computer Science, Engineering, or related discipline
Preferred Skills
β’ Experience working with AWS cloud services (EMR, S3, Lambda, Glue, etc.)
β’ Experience with Spark 3.x or 4.x
β’ Exposure to Kubernetes, Airflow, or similar orchestration tools
β’ Familiarity with CI/CD and DevOps automation for data engineering
Why This Opportunity Stands Out
β’ Long-term project stability (12+ months, likely extension)
β’ Ability to work on high-impact scientific and research-driven datasets
β’ Hands-on cloud modernization (AWS) and next-generation big data tooling
β’ Collaborative and innovative engineering culture






