PSI (Proteam Solutions)

Big Data Developer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Big Data Developer on a long-term project (12+ months) requiring U.S. citizenship or permanent residency. Key skills include 4+ years in Hadoop, 2+ years with Apache Spark, proficiency in Java, Scala, or Python, and strong data architecture knowledge.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
680
-
πŸ—“οΈ - Date
December 2, 2025
πŸ•’ - Duration
More than 6 months
-
🏝️ - Location
Unknown
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
Columbus, Ohio Metropolitan Area
-
🧠 - Skills detailed
#Agile #DevOps #Scala #S3 (Amazon Simple Storage Service) #Leadership #Lambda (AWS Lambda) #Data Engineering #Datasets #Big Data #Spark (Apache Spark) #Data Quality #Python #Kubernetes #Documentation #AWS (Amazon Web Services) #Batch #Monitoring #Hadoop #Data Architecture #Data Modeling #"ETL (Extract #Transform #Load)" #Airflow #Java #Computer Science #Security #Compliance #Apache Spark #Data Pipeline #Data Science #Automation #Data Processing #Cloud
Role description
The Big Data Developer will play a critical role on the Big Data engineering team, designing and implementing large-scale data processing systems that power scientific research and innovation. The ideal candidate has hands-on experience building enterprise-grade data pipelines, working with distributed systems, and optimizing data processing workflows. This is a long-term project (12+ months) with high visibility, cutting-edge tools, and opportunities to influence technical direction. β€’ β€’ U.S. CITIZENSHIP OR PERMANENT RESIDENCY REQUIRED. VISA OR EAD STATUS NOT ACCEPTABLE FOR THIS POSITION. NO EXCEPTIONS β€’ β€’ What You’ll Do βœ” Data Pipeline Design & Development β€’ Design, build, and deploy scalable data pipelines for ingesting, processing, transforming, and storing high-volume datasets. β€’ Implement streaming and batch-processing solutions using Hadoop, Spark, and cloud-based tools. βœ” Data Architecture & Engineering β€’ Develop and maintain data architecture and data flow models. β€’ Ensure data reliability, accuracy, and integrity across all environments. β€’ Support data warehousing strategies and best practices. βœ” Data Quality, Security & Compliance β€’ Implement automated data validation, error handling, and monitoring. β€’ Ensure compliance with internal security controls and regulatory standards. β€’ Partner with governance teams to enforce data quality and security guidelines. βœ” Cross-Functional Collaboration β€’ Work closely with data scientists, analysts, product teams, and application developers. β€’ Translate business requirements into robust technical solutions. β€’ Participate in Agile ceremonies and contribute to technical design discussions. βœ” Performance Optimization β€’ Tune Spark applications, Hadoop jobs, and distributed data systems for performance and cost efficiency. β€’ Troubleshoot bottlenecks and implement improvements to system performance. βœ” Technical Leadership β€’ Provide mentorship to junior developers and contribute to coding standards, best practices, and technical documentation. Required Skills & Qualifications β€’ 4+ years of Big Data Development experience in Hadoop ecosystems β€’ 2+ years of hands-on development with Apache Spark β€’ Proficiency in Java, Scala, or Python β€’ Strong understanding of distributed systems, ETL, data warehousing, and data modeling concepts β€’ Experience with large-scale datasets, performance tuning, and troubleshooting β€’ Strong problem-solving, communication, and collaboration skills β€’ Bachelor’s degree in Computer Science, Engineering, or related discipline Preferred Skills β€’ Experience working with AWS cloud services (EMR, S3, Lambda, Glue, etc.) β€’ Experience with Spark 3.x or 4.x β€’ Exposure to Kubernetes, Airflow, or similar orchestration tools β€’ Familiarity with CI/CD and DevOps automation for data engineering Why This Opportunity Stands Out β€’ Long-term project stability (12+ months, likely extension) β€’ Ability to work on high-impact scientific and research-driven datasets β€’ Hands-on cloud modernization (AWS) and next-generation big data tooling β€’ Collaborative and innovative engineering culture