

Red Oak Technologies
Data Engineer [Python, Spark, API, Data Pipelines] Contract On-Site at Cupertino, CA
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with expertise in Python, Spark, SQL, and API development. The contract is on-site in Cupertino, CA, with a focus on building data pipelines and ensuring data quality. Bachelor's degree required; cloud experience preferred.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
560
-
ποΈ - Date
April 28, 2026
π - Duration
Unknown
-
ποΈ - Location
On-site
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
Cupertino, CA
-
π§ - Skills detailed
#Data Processing #Databases #Azure #Apache Spark #AWS (Amazon Web Services) #Data Engineering #Data Pipeline #Automation #Cloud #Data Manipulation #Data Accuracy #API (Application Programming Interface) #Data Quality #Data Governance #Data Extraction #GCP (Google Cloud Platform) #Data Ingestion #Kafka (Apache Kafka) #Python #Data Science #Computer Science #PySpark #SQL (Structured Query Language) #"ETL (Extract #Transform #Load)" #Programming #Security #Spark (Apache Spark) #Data Analysis #SQL Queries #Scala
Role description
Data Engineer [Python, Spark, API, Data Pipelines] Contract On-Site at Cupertino, CA
Position Overview: We are looking for a passionate and skilled Data Engineer with expertise in Spark, Python, SQL, and API development to design, develop, and maintain end-to-end data solutions. The ideal candidate will work closely with cross-functional teams to build scalable data pipelines, ensure data quality, and enable analytics and reporting.
Key Responsibilities:
Design, develop, and maintain scalable data pipelines using Apache Spark and Python.
Develop and optimize SQL queries for data extraction, transformation, and loading (ETL).
Build and implement end-to-end API integrations for data ingestion and dissemination.
Collaborate with data analysts, data scientists, and business stakeholders to understand data requirements.
Ensure data accuracy, integrity, and security across all platforms.
Monitor and troubleshoot data pipeline issues, ensuring high availability and performance.
Document data workflows, architecture, and best practices.
Qualifications:
Bachelorβs or Masterβs degree in Computer Science, Data Engineering, or related field.
Proven experience with Spark (PySpark) for large-scale data processing.
Strong programming skills in Python for data manipulation and automation.
Extensive experience with SQL and relational databases.
Hands-on experience developing and consuming APIs (RESTful/SOAP).
Knowledge of data warehousing concepts and tools.
Familiarity with cloud platforms (AWS, Azure, GCP) is a plus.
Excellent problem-solving skills and attention to detail.
Strong communication and teamwork skills.
Preferred Skills:
Experience with streaming data technologies (Kafka, Kinesis).
Knowledge of data governance and security best practices.
Experience with CI/CD pipelines for data workflows.
Data Engineer [Python, Spark, API, Data Pipelines] Contract On-Site at Cupertino, CA
Position Overview: We are looking for a passionate and skilled Data Engineer with expertise in Spark, Python, SQL, and API development to design, develop, and maintain end-to-end data solutions. The ideal candidate will work closely with cross-functional teams to build scalable data pipelines, ensure data quality, and enable analytics and reporting.
Key Responsibilities:
Design, develop, and maintain scalable data pipelines using Apache Spark and Python.
Develop and optimize SQL queries for data extraction, transformation, and loading (ETL).
Build and implement end-to-end API integrations for data ingestion and dissemination.
Collaborate with data analysts, data scientists, and business stakeholders to understand data requirements.
Ensure data accuracy, integrity, and security across all platforms.
Monitor and troubleshoot data pipeline issues, ensuring high availability and performance.
Document data workflows, architecture, and best practices.
Qualifications:
Bachelorβs or Masterβs degree in Computer Science, Data Engineering, or related field.
Proven experience with Spark (PySpark) for large-scale data processing.
Strong programming skills in Python for data manipulation and automation.
Extensive experience with SQL and relational databases.
Hands-on experience developing and consuming APIs (RESTful/SOAP).
Knowledge of data warehousing concepts and tools.
Familiarity with cloud platforms (AWS, Azure, GCP) is a plus.
Excellent problem-solving skills and attention to detail.
Strong communication and teamwork skills.
Preferred Skills:
Experience with streaming data technologies (Kafka, Kinesis).
Knowledge of data governance and security best practices.
Experience with CI/CD pipelines for data workflows.






