

Senior Data Architect
β - Featured Role | Apply direct with Data Freelance Hub
This role is a Senior Data Architect/Data Engineer for a 6-month remote contract, requiring PST hours. Key skills include expertise in MapR, Kubernetes, Spark, Airflow, and Kafka. Strong programming skills in Python and Bash are essential.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
June 4, 2025
π - Project duration
More than 6 months
-
ποΈ - Location type
Remote
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
United States
-
π§ - Skills detailed
#Presto #Automated Testing #Compliance #Spark (Apache Spark) #Batch #Data Quality #Data Engineering #GitHub #Programming #Data Integration #Jenkins #API (Application Programming Interface) #Data Pipeline #Deployment #Scala #Data Processing #Storage #Python #Data Architecture #Visualization #Kafka (Apache Kafka) #Bash #Airflow #Scripting #Kubernetes #Data Lake #Data Storage #Apache Iceberg
Role description
World Wide Technology is looking for a Data Architect / Data Engineer. This job is part of WWTβs Strategic Resourcing services. The candidate will be supporting a WWT customer and will be employed by one of WWTβs preferred partners.
Role: Data Architect / Data Engineer
This REMOTE role is 6 month contract with possibility of extension, candidate needs to sit in PST hours. This person needs to be a hands on Engineer with an Architecture background as well.
Overview:
We are seeking a skilled and versatile Data Architect / Data Engineer to design, build, and optimize data platforms and pipelines within a distributed environment. The ideal candidate will possess deep expertise in managing large-scale data systems, data integration, modern data engineering practices, and pipeline orchestration. You will play a key role in architecting and engineering scalable, high-performance data solutions that drive business insights and innovation.
Key Responsibilities:
β’ Design, implement, and manage scalable data architectures on distributed platforms (e.g., MapR, HPE Unified Analytics and Data Fabric).
β’ Develop, optimize, and maintain robust data pipelines using tools such as Spark, Airflow, and EzPresto.
β’ Configure and maintain Kafka architecture, MapR Streams, and related technologies to support real-time and batch data processing.
β’ Implement Change Data Capture (CDC) mechanisms and integrate data using APIs and streaming techniques.
β’ Monitor, tune, and troubleshoot distributed data clusters including MapR and Kubernetes environments.
β’ Develop and maintain CI/CD pipelines using Jenkins and integrate with GitHub for automated testing and deployment.
β’ Collaborate with cross-functional teams to ensure data quality, governance, and compliance standards are met.
β’ Leverage tools such as Iceberg and Superset for data storage optimization and visualization.
Required Skills & Experience:
β’ Strong experience with distributed data platforms, including MapR and Kubernetes.
β’ Proficient in data pipeline tools and frameworks: Spark, Airflow, EzPresto.
β’ Solid programming and scripting skills: Python, Bash.
β’ Expertise in Kafka architecture and operations.
β’ Experience with CI/CD development workflows using Jenkins and GitHub.
β’ Knowledge and use of Apache Iceberg for data lake management.
β’ Familiarity with data architecture best practices, including CDC and API-based integrations.
Preferred / Nice to Have:
β’ Experience with HPE Unified Analytics and Data Fabric.
β’ Familiarity with MapR Streams and Superset for real-time analytics and dashboarding.
World Wide Technology is looking for a Data Architect / Data Engineer. This job is part of WWTβs Strategic Resourcing services. The candidate will be supporting a WWT customer and will be employed by one of WWTβs preferred partners.
Role: Data Architect / Data Engineer
This REMOTE role is 6 month contract with possibility of extension, candidate needs to sit in PST hours. This person needs to be a hands on Engineer with an Architecture background as well.
Overview:
We are seeking a skilled and versatile Data Architect / Data Engineer to design, build, and optimize data platforms and pipelines within a distributed environment. The ideal candidate will possess deep expertise in managing large-scale data systems, data integration, modern data engineering practices, and pipeline orchestration. You will play a key role in architecting and engineering scalable, high-performance data solutions that drive business insights and innovation.
Key Responsibilities:
β’ Design, implement, and manage scalable data architectures on distributed platforms (e.g., MapR, HPE Unified Analytics and Data Fabric).
β’ Develop, optimize, and maintain robust data pipelines using tools such as Spark, Airflow, and EzPresto.
β’ Configure and maintain Kafka architecture, MapR Streams, and related technologies to support real-time and batch data processing.
β’ Implement Change Data Capture (CDC) mechanisms and integrate data using APIs and streaming techniques.
β’ Monitor, tune, and troubleshoot distributed data clusters including MapR and Kubernetes environments.
β’ Develop and maintain CI/CD pipelines using Jenkins and integrate with GitHub for automated testing and deployment.
β’ Collaborate with cross-functional teams to ensure data quality, governance, and compliance standards are met.
β’ Leverage tools such as Iceberg and Superset for data storage optimization and visualization.
Required Skills & Experience:
β’ Strong experience with distributed data platforms, including MapR and Kubernetes.
β’ Proficient in data pipeline tools and frameworks: Spark, Airflow, EzPresto.
β’ Solid programming and scripting skills: Python, Bash.
β’ Expertise in Kafka architecture and operations.
β’ Experience with CI/CD development workflows using Jenkins and GitHub.
β’ Knowledge and use of Apache Iceberg for data lake management.
β’ Familiarity with data architecture best practices, including CDC and API-based integrations.
Preferred / Nice to Have:
β’ Experience with HPE Unified Analytics and Data Fabric.
β’ Familiarity with MapR Streams and Superset for real-time analytics and dashboarding.