

Optomi
Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with 6+ years of experience, offering a hybrid contract in Austin, Texas or Cupertino, California. Pay rate is competitive. Key skills include Python, SQL, Spark, Kafka, Docker, Kubernetes, and cloud platforms.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
424
-
🗓️ - Date
April 18, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Hybrid
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Austin, Texas Metropolitan Area
-
🧠 - Skills detailed
#Airflow #Data Pipeline #Snowflake #Data Modeling #Azure #AWS (Amazon Web Services) #PySpark #Databricks #Observability #Batch #Scala #Data Quality #"ETL (Extract #Transform #Load)" #Docker #Python #Cloud #Data Engineering #ML (Machine Learning) #GCP (Google Cloud Platform) #Programming #SQL (Structured Query Language) #Kubernetes #Kafka (Apache Kafka) #Spark (Apache Spark)
Role description
Open to both Austin, Texas and Cupertino, California locations!
Description:
• We are seeking experienced Data Engineers to join a high-impact team focused on building and optimizing modern data platforms. This role is ideal for hands-on engineers who thrive in fast-paced environments and can independently design, develop, and maintain scalable data solutions.
• You will play a critical role in enabling data-driven decision-making by developing robust pipelines, ensuring data quality, and supporting advanced analytics and machine learning initiatives. This is a highly technical position requiring strong coding ability, problem-solving skills, and real-world experience with large-scale data systems.
Qualifications:
• 6+ years of hands-on data engineering experience
• Strong programming skills in Python and SQL (Scala is highly preferred)
• Spark / PySpark
• Kafka
• Airflow or similar orchestration tools
• Docker & Kubernetes (required)
• Spark architecture and performance tuning
• Data warehousing concepts and best practices
• Pipeline design and optimization
• Experience with cloud platforms (AWS, Azure, or GCP)
• Ability to troubleshoot complex data and system issues
• Strong communication skills and ability to work independently
Preferred Qualifications:
• Experience with data modeling and lakehouse architectures
• Familiarity with CI/CD, data observability, and infrastructure-as-code
• Exposure to ML pipelines or advanced analytics workflows
• Experience with tools such as Snowflake, Databricks, or similar platforms
Responsibilities:
• Design and build scalable batch and near real-time data pipelines
• Develop and optimize ETL/ELT workflows for performance and cost efficiency
• Work with Spark-based systems, including understanding job execution and optimization
• Design and implement data models (e.g., star schema, medallion architecture)
• Support machine learning and advanced data use cases (feature engineering, retraining workflows, etc.)
• Ensure data quality, governance, privacy, and system reliability
• Troubleshoot production issues and handle real-world incident management scenarios
• Collaborate on data pipeline orchestration and job scheduling
• Continuously improve system performance and scalability
Open to both Austin, Texas and Cupertino, California locations!
Description:
• We are seeking experienced Data Engineers to join a high-impact team focused on building and optimizing modern data platforms. This role is ideal for hands-on engineers who thrive in fast-paced environments and can independently design, develop, and maintain scalable data solutions.
• You will play a critical role in enabling data-driven decision-making by developing robust pipelines, ensuring data quality, and supporting advanced analytics and machine learning initiatives. This is a highly technical position requiring strong coding ability, problem-solving skills, and real-world experience with large-scale data systems.
Qualifications:
• 6+ years of hands-on data engineering experience
• Strong programming skills in Python and SQL (Scala is highly preferred)
• Spark / PySpark
• Kafka
• Airflow or similar orchestration tools
• Docker & Kubernetes (required)
• Spark architecture and performance tuning
• Data warehousing concepts and best practices
• Pipeline design and optimization
• Experience with cloud platforms (AWS, Azure, or GCP)
• Ability to troubleshoot complex data and system issues
• Strong communication skills and ability to work independently
Preferred Qualifications:
• Experience with data modeling and lakehouse architectures
• Familiarity with CI/CD, data observability, and infrastructure-as-code
• Exposure to ML pipelines or advanced analytics workflows
• Experience with tools such as Snowflake, Databricks, or similar platforms
Responsibilities:
• Design and build scalable batch and near real-time data pipelines
• Develop and optimize ETL/ELT workflows for performance and cost efficiency
• Work with Spark-based systems, including understanding job execution and optimization
• Design and implement data models (e.g., star schema, medallion architecture)
• Support machine learning and advanced data use cases (feature engineering, retraining workflows, etc.)
• Ensure data quality, governance, privacy, and system reliability
• Troubleshoot production issues and handle real-world incident management scenarios
• Collaborate on data pipeline orchestration and job scheduling
• Continuously improve system performance and scalability






