Prophecy Technologies

SRE / Hadoop Admin

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior SRE/Hadoop Admin with 10+ years of experience managing large-scale Hadoop infrastructures. Contract length is unspecified, pay rate is "unknown," and it requires on-site work. Key skills include Hadoop ecosystem expertise, Linux administration, automation, and leadership.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
October 14, 2025
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Fountain Valley, CA
-
🧠 - Skills detailed
#Monitoring #Infrastructure as Code (IaC) #Data Governance #HDFS (Hadoop Distributed File System) #Hadoop #Security #Prometheus #Deployment #Kubernetes #Compliance #YARN (Yet Another Resource Negotiator) #Scripting #Ambari #Docker #Python #Disaster Recovery #Automation #OpenSearch #Grafana #Spark (Apache Spark) #Leadership #Metadata #GitLab #Linux #Data Management #Bash #Observability #NiFi (Apache NiFi)
Role description
Purpose: Seeking a highly experienced Senior or Lead Platform Engineer/Site Reliability Engineer (SRE)/Hadoop Admin to manage and enhance our petabyte-scale, on-premises data platform. This platform is built using the open-source Hadoop ecosystem. The ideal candidate possesses in-depth technical expertise, a solid understanding of distributed systems, and extensive experience in operating and optimizing large-scale data infrastructures. This role requires a hands-on technical leader who can drive platform innovation, ensure high availability and reliability, and mentor team members in best practices for performance, automation, and resiliency. Essential Functions: • Own and operate the end-to-end infrastructure of a large-scale, on-prem Hadoop-based data platform, ensuring high availability and reliability. • Design, implement, and maintain core platform components, including Hadoop, Hive, Spark, NiFi, Iceberg, ELK, OpenSearch and Ambari. • Automate infrastructure management, monitoring, and deployments using CI/CD pipelines (GitLab) and scripting. • Implement and enforce security controls, access management, and compliance standards. • Perform system upgrades, patching, performance tuning, and troubleshooting across platform components Basic Requirements: • 10+ years of experience in Platform Engineering, Site Reliability Engineering, or similar roles, with proven success managing large-scale, distributed Hadoop infrastructure. • Deep expertise in the Hadoop ecosystem, including HDFS, YARN, Hive, Spark, NiFi, Ambari, and Iceberg. • Strong Linux system administration skills (CentOS/Rocky preferred), including system tuning, performance optimization, and troubleshooting. • Proficiency in containerization and orchestration using Docker and Kubernetes. • Solid experience with automation and Infrastructure as Code, leveraging tools like GitLab CI/CD and scripting in Python and bash. • Practical knowledge of monitoring and observability tools (e.g., Prometheus, Grafana, OpenTelemetry) and understanding of system health, alerting, and telemetry. • Familiarity with networking concepts, security protocols, and data compliance requirements. • Experience managing petabyte-scale data platforms and implementing disaster recovery strategies. • Understanding of data governance, metadata management, and operational best practices. • Demonstrated ability to lead technical projects, mentor engineers, and collaborate effectively with cross-functional teams. • Excellent problem-solving, communication, and leadership skills. Regards,