Senior Data Architect

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is a Senior Data Architect/Data Engineer for a 6-month remote contract, requiring PST hours. Key skills include expertise in MapR, Kubernetes, Spark, Airflow, and Kafka. Strong programming skills in Python and Bash are essential.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
-
πŸ—“οΈ - Date discovered
June 4, 2025
πŸ•’ - Project duration
More than 6 months
-
🏝️ - Location type
Remote
-
πŸ“„ - Contract type
Unknown
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
United States
-
🧠 - Skills detailed
#Presto #Automated Testing #Compliance #Spark (Apache Spark) #Batch #Data Quality #Data Engineering #GitHub #Programming #Data Integration #Jenkins #API (Application Programming Interface) #Data Pipeline #Deployment #Scala #Data Processing #Storage #Python #Data Architecture #Visualization #Kafka (Apache Kafka) #Bash #Airflow #Scripting #Kubernetes #Data Lake #Data Storage #Apache Iceberg
Role description
World Wide Technology is looking for a Data Architect / Data Engineer. This job is part of WWT’s Strategic Resourcing services. The candidate will be supporting a WWT customer and will be employed by one of WWT’s preferred partners. Role: Data Architect / Data Engineer This REMOTE role is 6 month contract with possibility of extension, candidate needs to sit in PST hours. This person needs to be a hands on Engineer with an Architecture background as well. Overview: We are seeking a skilled and versatile Data Architect / Data Engineer to design, build, and optimize data platforms and pipelines within a distributed environment. The ideal candidate will possess deep expertise in managing large-scale data systems, data integration, modern data engineering practices, and pipeline orchestration. You will play a key role in architecting and engineering scalable, high-performance data solutions that drive business insights and innovation. Key Responsibilities: β€’ Design, implement, and manage scalable data architectures on distributed platforms (e.g., MapR, HPE Unified Analytics and Data Fabric). β€’ Develop, optimize, and maintain robust data pipelines using tools such as Spark, Airflow, and EzPresto. β€’ Configure and maintain Kafka architecture, MapR Streams, and related technologies to support real-time and batch data processing. β€’ Implement Change Data Capture (CDC) mechanisms and integrate data using APIs and streaming techniques. β€’ Monitor, tune, and troubleshoot distributed data clusters including MapR and Kubernetes environments. β€’ Develop and maintain CI/CD pipelines using Jenkins and integrate with GitHub for automated testing and deployment. β€’ Collaborate with cross-functional teams to ensure data quality, governance, and compliance standards are met. β€’ Leverage tools such as Iceberg and Superset for data storage optimization and visualization. Required Skills & Experience: β€’ Strong experience with distributed data platforms, including MapR and Kubernetes. β€’ Proficient in data pipeline tools and frameworks: Spark, Airflow, EzPresto. β€’ Solid programming and scripting skills: Python, Bash. β€’ Expertise in Kafka architecture and operations. β€’ Experience with CI/CD development workflows using Jenkins and GitHub. β€’ Knowledge and use of Apache Iceberg for data lake management. β€’ Familiarity with data architecture best practices, including CDC and API-based integrations. Preferred / Nice to Have: β€’ Experience with HPE Unified Analytics and Data Fabric. β€’ Familiarity with MapR Streams and Superset for real-time analytics and dashboarding.