

Data Engineer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer in Greenwood Village, CO, with a W2 contract for 8+ years of experience. Key skills include AWS, Spark, PySpark, and SQL. Familiarity with Kafka, GitLab, and Terraform is required.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
August 9, 2025
π - Project duration
Unknown
-
ποΈ - Location type
On-site
-
π - Contract type
W2 Contractor
-
π - Security clearance
Unknown
-
π - Location detailed
Greenwood Village, CO
-
π§ - Skills detailed
#"ETL (Extract #Transform #Load)" #Monitoring #SQL (Structured Query Language) #Kafka (Apache Kafka) #Data Transformations #Python #Terraform #AWS EMR (Amazon Elastic MapReduce) #PySpark #JSON (JavaScript Object Notation) #Cloud #DevOps #Data Pipeline #AI (Artificial Intelligence) #EC2 #S3 (Amazon Simple Storage Service) #Airflow #Programming #GitLab #Data Engineering #AWS (Amazon Web Services) #Scala #Spark (Apache Spark) #ML (Machine Learning) #Deployment #Data Processing
Role description
Job Role: Data Engineer
Location: Greenwood Village, CO (Onsite)
Experience: 8 + Years
Contract: W2 Only/ No C2C
Visa: All Visa except OPTs
Client: Kforce / Confidential
Kforce has a client in Greenwood Village, CO that is seeking a skilled Data Engineer to support a high-impact data engineering initiative focused on AWS infrastructure, Spark-based transformations, and orchestration tools. This individual contributor role requires hands-on experience building data pipelines, processing large-scale JSON messages, and deploying solutions in cloud environments. The ideal candidate has a solid foundation in data engineering within enterprise environments, strong Spark/PySpark expertise, and experience with AWS-native services and modern orchestration tools.
Responsibilities:
Β· Build and optimize scalable data pipelines to process and transform structured and semi-structured data using PySpark and SparkSQL
Β· Work with JSON objects to parse messages and send MDK messages via Kafka to S3
Β· Create and manage data endpoints with defined schemas, orchestrated through RabbitMQ and Kafka
Β· Execute data transformations from JSON to RDDs using Spark on AWS EMR/EC2
Β· Support orchestration through AWS Step Functions with a future transition to Airflow
Β· Use SQL to extract, transform, and load data into reporting and dashboarding tools
Β· Collaborate with DevOps and CI/CD teams to maintain GitLab pipelines and automate infrastructure deployment using Terraform
Β· Maintain version-controlled ETL code in GitLab and participate in testing, deployment, and monitoring workflows
Β· Participate in cross-functional engineering discussions, aligning with best practices and timelines
Primary Skills:
Β· 5 years of experience in data engineering with strong hands-on development skills
Β· Proficiency in Spark and PySpark, including SparkSQL for distributed data processing
Β· Strong programming experience in Python and solid SQL query skills
Β· Experience working with AWS services such as EMR, EC2, and S3
Β· Familiarity with JSON-based messaging systems, message parsing, and stream processing
Β· Exposure to Kafka, MSK, or RabbitMQ for message queueing and data delivery
Β· Experience working with GitLab and Terraform in a CI/CD environment
Β· Ability to work independently and contribute effectively in a fast-paced team setting
Preferred Skills:
Β· Background in enterprise-level data engineering environments
Β· Experience with Airflow for workflow orchestration
Β· Familiarity with AI/ML pipelines or advanced analytics systems
Β· Understanding of ETL lifecycle management, including staging, deployment, and teardown processes
Job Role: Data Engineer
Location: Greenwood Village, CO (Onsite)
Experience: 8 + Years
Contract: W2 Only/ No C2C
Visa: All Visa except OPTs
Client: Kforce / Confidential
Kforce has a client in Greenwood Village, CO that is seeking a skilled Data Engineer to support a high-impact data engineering initiative focused on AWS infrastructure, Spark-based transformations, and orchestration tools. This individual contributor role requires hands-on experience building data pipelines, processing large-scale JSON messages, and deploying solutions in cloud environments. The ideal candidate has a solid foundation in data engineering within enterprise environments, strong Spark/PySpark expertise, and experience with AWS-native services and modern orchestration tools.
Responsibilities:
Β· Build and optimize scalable data pipelines to process and transform structured and semi-structured data using PySpark and SparkSQL
Β· Work with JSON objects to parse messages and send MDK messages via Kafka to S3
Β· Create and manage data endpoints with defined schemas, orchestrated through RabbitMQ and Kafka
Β· Execute data transformations from JSON to RDDs using Spark on AWS EMR/EC2
Β· Support orchestration through AWS Step Functions with a future transition to Airflow
Β· Use SQL to extract, transform, and load data into reporting and dashboarding tools
Β· Collaborate with DevOps and CI/CD teams to maintain GitLab pipelines and automate infrastructure deployment using Terraform
Β· Maintain version-controlled ETL code in GitLab and participate in testing, deployment, and monitoring workflows
Β· Participate in cross-functional engineering discussions, aligning with best practices and timelines
Primary Skills:
Β· 5 years of experience in data engineering with strong hands-on development skills
Β· Proficiency in Spark and PySpark, including SparkSQL for distributed data processing
Β· Strong programming experience in Python and solid SQL query skills
Β· Experience working with AWS services such as EMR, EC2, and S3
Β· Familiarity with JSON-based messaging systems, message parsing, and stream processing
Β· Exposure to Kafka, MSK, or RabbitMQ for message queueing and data delivery
Β· Experience working with GitLab and Terraform in a CI/CD environment
Β· Ability to work independently and contribute effectively in a fast-paced team setting
Preferred Skills:
Β· Background in enterprise-level data engineering environments
Β· Experience with Airflow for workflow orchestration
Β· Familiarity with AI/ML pipelines or advanced analytics systems
Β· Understanding of ETL lifecycle management, including staging, deployment, and teardown processes