

Creospan Inc.
Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer on a long-term contract in the San Francisco Bay Area or Chicago, IL. Required skills include Python, SQL, R, and experience with large-scale data processing and ETL frameworks. W2 candidates only.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
Unknown
-
ποΈ - Date
November 1, 2025
π - Duration
Unknown
-
ποΈ - Location
On-site
-
π - Contract
W2 Contractor
-
π - Security
Unknown
-
π - Location detailed
San Francisco Bay Area
-
π§ - Skills detailed
#Data Processing #Datasets #Scala #"ETL (Extract #Transform #Load)" #Scripting #Airflow #DevOps #Automation #Spark (Apache Spark) #Luigi #Data Engineering #Data Pipeline #Data Quality #Data Ingestion #Python #R #Monitoring #SQL (Structured Query Language) #Process Automation #Presto
Role description
Creospan is a growing tech collective of makers, shakers, and problem solvers, offering solutions today that will propel businesses into a better tomorrow. βTomorrowβs ideas, built today!β In addition to being able to work alongside equally brilliant and motivated developers, our consultants appreciate the opportunity to learn and apply new skills and methodologies to different clients and industries.
β’
β’
β’
β’
β’
β’ NO C2C/3RD PARTY, LOOKING FOR W2 CANDIDATES ONLY, must be able to work in the US without sponsorship now or in the future
β’
β’
β’ Data Engineer
San Francisco Bay Area and Chicago, IL
Long term Contract
Responsibilities
β’ Design and implement scalable data ingestion frameworks capable of processing petabyte-scale datasets.
β’ Develop automation scripts and orchestration workflows that transform manual data operations into self-service processes.
β’ Operationalize ingestion processes by creating clear, repeatable runbooks and automation tools.
β’ Collaborate with cross-functional teams to understand data sources, integration requirements, and performance bottlenecks.
β’ Build monitoring, alerting, and reporting tools to ensure reliability and data quality across ingestion pipelines.
β’ Continuously optimize system performance and resource utilization for large-scale data workflows.
Required Qualifications
β’ Proficiency in Python, SQL, and R, with a strong scripting and automation background.
β’ Hands-on experience working with large-scale data processing (terabytes to petabytes).
β’ Strong understanding of data pipeline orchestration, ETL frameworks, and workflow automation.
β’ Experience with distributed data systems (e.g., Spark, Presto, Hive, or similar).
β’ Ability to translate manual ingestion processes into automated, self-service tools.
β’ Excellent problem-solving skills and attention to detail.
Preferred Qualifications
β’ Experience working in large enterprise data environments.
β’ Familiarity with Airflow, Luigi, or other orchestration frameworks.
β’ Background in data engineering, DevOps, or process automation.
β’ Strong communication skills and ability to document and operationalize complex systems.
Creospan is a growing tech collective of makers, shakers, and problem solvers, offering solutions today that will propel businesses into a better tomorrow. βTomorrowβs ideas, built today!β In addition to being able to work alongside equally brilliant and motivated developers, our consultants appreciate the opportunity to learn and apply new skills and methodologies to different clients and industries.
β’
β’
β’
β’
β’
β’ NO C2C/3RD PARTY, LOOKING FOR W2 CANDIDATES ONLY, must be able to work in the US without sponsorship now or in the future
β’
β’
β’ Data Engineer
San Francisco Bay Area and Chicago, IL
Long term Contract
Responsibilities
β’ Design and implement scalable data ingestion frameworks capable of processing petabyte-scale datasets.
β’ Develop automation scripts and orchestration workflows that transform manual data operations into self-service processes.
β’ Operationalize ingestion processes by creating clear, repeatable runbooks and automation tools.
β’ Collaborate with cross-functional teams to understand data sources, integration requirements, and performance bottlenecks.
β’ Build monitoring, alerting, and reporting tools to ensure reliability and data quality across ingestion pipelines.
β’ Continuously optimize system performance and resource utilization for large-scale data workflows.
Required Qualifications
β’ Proficiency in Python, SQL, and R, with a strong scripting and automation background.
β’ Hands-on experience working with large-scale data processing (terabytes to petabytes).
β’ Strong understanding of data pipeline orchestration, ETL frameworks, and workflow automation.
β’ Experience with distributed data systems (e.g., Spark, Presto, Hive, or similar).
β’ Ability to translate manual ingestion processes into automated, self-service tools.
β’ Excellent problem-solving skills and attention to detail.
Preferred Qualifications
β’ Experience working in large enterprise data environments.
β’ Familiarity with Airflow, Luigi, or other orchestration frameworks.
β’ Background in data engineering, DevOps, or process automation.
β’ Strong communication skills and ability to document and operationalize complex systems.






