Cliff Services Inc

Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with 7+ years of experience in Scala, AWS, and Apache Spark. Contract length is unspecified, with a pay rate of "unknown." Locations include McLean, VA; Richmond, VA; and Chicago, IL.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
480
-
🗓️ - Date
December 11, 2025
🕒 - Duration
Unknown
-
🏝️ - Location
On-site
-
📄 - Contract
Unknown
-
🔒 - Security
Unknown
-
📍 - Location detailed
Chicago, IL
-
🧠 - Skills detailed
#Kubernetes #Cloud #Docker #Programming #SQL (Structured Query Language) #EC2 #Data Ingestion #IAM (Identity and Access Management) #NoSQL #Data Architecture #Version Control #Lambda (AWS Lambda) #Apache Spark #Data Processing #Data Engineering #Scala #AWS (Amazon Web Services) #Kafka (Apache Kafka) #Spark (Apache Spark) #S3 (Amazon Simple Storage Service) #Data Modeling #Data Pipeline #"ETL (Extract #Transform #Load)" #Airflow #Data Lake #Hadoop
Role description
Job Title: Data Engineer (Scala + AWS + Spark) Locations: McLean, VA / Richmond, VA / Chicago, IL Experience: 7+ Years Visa: H1B Interview Process: In-Person / Face-to-Face (F2F) Job Description We are seeking a highly skilled Data Engineer with strong expertise in Scala, AWS, and Apache Spark. The ideal candidate will have 7+ years of hands-on experience building scalable data pipelines, distributed processing systems, and cloud-native data solutions. Key Responsibilities • Design, build, and optimize large-scale data pipelines using Scala and Spark. • Develop and maintain ETL/ELT workflows across AWS services. • Work on distributed data processing using Spark, Hadoop, or similar. • Build data ingestion, transformation, cleansing, and validation routines. • Optimize pipeline performance and ensure reliability in production environments. • Collaborate with cross-functional teams to understand requirements and deliver robust solutions. • Implement CI/CD best practices, testing, and version control. • Troubleshoot and resolve issues in complex data flow systems. Required Skills & Experience • 7+ years of Data Engineering experience. • Strong programming experience with Scala (must-have). • Hands-on experience with Apache Spark (core, SQL, streaming). • Solid experience with AWS cloud services (Glue, EMR, Lambda, S3, EC2, IAM, etc.). • High proficiency in SQL and relational/noSQL data stores. • Strong understanding of data modeling, data architecture, and distributed systems. • Experience with workflow orchestration tools (Airflow, Step Functions, etc.). • Strong communication and problem-solving skills. Preferred Skills • Experience with Kafka, Kinesis, or other streaming platforms. • Knowledge of containerization tools like Docker or Kubernetes. • Background in data warehousing or modern data lake architectures.