

Cliff Services Inc
Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with 7+ years of experience in Scala, AWS, and Apache Spark. Contract length is unspecified, with a pay rate of "unknown." Locations include McLean, VA; Richmond, VA; and Chicago, IL.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
480
-
🗓️ - Date
December 11, 2025
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Chicago, IL
-
🧠 - Skills detailed
#Kubernetes #Cloud #Docker #Programming #SQL (Structured Query Language) #EC2 #Data Ingestion #IAM (Identity and Access Management) #NoSQL #Data Architecture #Version Control #Lambda (AWS Lambda) #Apache Spark #Data Processing #Data Engineering #Scala #AWS (Amazon Web Services) #Kafka (Apache Kafka) #Spark (Apache Spark) #S3 (Amazon Simple Storage Service) #Data Modeling #Data Pipeline #"ETL (Extract #Transform #Load)" #Airflow #Data Lake #Hadoop
Role description
Job Title: Data Engineer (Scala + AWS + Spark)
Locations: McLean, VA / Richmond, VA / Chicago, IL
Experience: 7+ Years
Visa: H1B
Interview Process: In-Person / Face-to-Face (F2F)
Job Description
We are seeking a highly skilled Data Engineer with strong expertise in Scala, AWS, and Apache Spark. The ideal candidate will have 7+ years of hands-on experience building scalable data pipelines, distributed processing systems, and cloud-native data solutions.
Key Responsibilities
• Design, build, and optimize large-scale data pipelines using Scala and Spark.
• Develop and maintain ETL/ELT workflows across AWS services.
• Work on distributed data processing using Spark, Hadoop, or similar.
• Build data ingestion, transformation, cleansing, and validation routines.
• Optimize pipeline performance and ensure reliability in production environments.
• Collaborate with cross-functional teams to understand requirements and deliver robust solutions.
• Implement CI/CD best practices, testing, and version control.
• Troubleshoot and resolve issues in complex data flow systems.
Required Skills & Experience
• 7+ years of Data Engineering experience.
• Strong programming experience with Scala (must-have).
• Hands-on experience with Apache Spark (core, SQL, streaming).
• Solid experience with AWS cloud services (Glue, EMR, Lambda, S3, EC2, IAM, etc.).
• High proficiency in SQL and relational/noSQL data stores.
• Strong understanding of data modeling, data architecture, and distributed systems.
• Experience with workflow orchestration tools (Airflow, Step Functions, etc.).
• Strong communication and problem-solving skills.
Preferred Skills
• Experience with Kafka, Kinesis, or other streaming platforms.
• Knowledge of containerization tools like Docker or Kubernetes.
• Background in data warehousing or modern data lake architectures.
Job Title: Data Engineer (Scala + AWS + Spark)
Locations: McLean, VA / Richmond, VA / Chicago, IL
Experience: 7+ Years
Visa: H1B
Interview Process: In-Person / Face-to-Face (F2F)
Job Description
We are seeking a highly skilled Data Engineer with strong expertise in Scala, AWS, and Apache Spark. The ideal candidate will have 7+ years of hands-on experience building scalable data pipelines, distributed processing systems, and cloud-native data solutions.
Key Responsibilities
• Design, build, and optimize large-scale data pipelines using Scala and Spark.
• Develop and maintain ETL/ELT workflows across AWS services.
• Work on distributed data processing using Spark, Hadoop, or similar.
• Build data ingestion, transformation, cleansing, and validation routines.
• Optimize pipeline performance and ensure reliability in production environments.
• Collaborate with cross-functional teams to understand requirements and deliver robust solutions.
• Implement CI/CD best practices, testing, and version control.
• Troubleshoot and resolve issues in complex data flow systems.
Required Skills & Experience
• 7+ years of Data Engineering experience.
• Strong programming experience with Scala (must-have).
• Hands-on experience with Apache Spark (core, SQL, streaming).
• Solid experience with AWS cloud services (Glue, EMR, Lambda, S3, EC2, IAM, etc.).
• High proficiency in SQL and relational/noSQL data stores.
• Strong understanding of data modeling, data architecture, and distributed systems.
• Experience with workflow orchestration tools (Airflow, Step Functions, etc.).
• Strong communication and problem-solving skills.
Preferred Skills
• Experience with Kafka, Kinesis, or other streaming platforms.
• Knowledge of containerization tools like Docker or Kubernetes.
• Background in data warehousing or modern data lake architectures.






