

Mastech Digital
Big Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Big Data Engineer with a long-term contract, remote work in Atlanta, GA. Requires 6–10 years of experience, expertise in Apache Spark, Scala, Google BigQuery, PostgreSQL, and SQL optimization. Bachelor’s or Master’s degree in a related field is needed.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
Unknown
-
🗓️ - Date
October 25, 2025
🕒 - Duration
Unknown
-
🏝️ - Location
Remote
-
📄 - Contract
W2 Contractor
-
🔒 - Security
Unknown
-
📍 - Location detailed
Atlanta, GA
-
🧠 - Skills detailed
#Data Storage #Debugging #Monitoring #Scala #Security #Apache Spark #Storage #Schema Design #Big Data #Kafka (Apache Kafka) #Data Science #Cloud #Data Warehouse #Distributed Computing #Datasets #Data Pipeline #SQL Queries #BigQuery #PostgreSQL #Data Quality #Spark (Apache Spark) #Data Engineering #Data Processing #Computer Science #DevOps #SQL (Structured Query Language) #Airflow #"ETL (Extract #Transform #Load)" #GCP (Google Cloud Platform) #Data Modeling #Databases #Data Governance #Database Performance
Role description
Title: : Big Data Engineer
Duration: Long term
Location: Atlanta, GA (Remote)
(ONLY W2)
Job Description:
We are seeking a highly skilled Big Data Engineer at the senior or staff level to design, develop, and optimize large-scale data platforms and processing systems. The ideal candidate will have a strong background in Apache Spark, Scala, Google BigQuery, PostgreSQL, and Apache Parquet, with hands-on experience building performant, scalable, and reliable data pipelines.This role requires an individual who can write efficient SQL queries, optimize performance across distributed architectures, and collaborate with cross-functional teams to deliver end-to-end data solutions. Experience in CPU localization or performance tuning at the hardware level is a plus.
Key Responsibilities
Design, implement, and maintain highly scalable data processing pipelines using Apache Spark and Scala.
Develop optimized data warehouse and analytics solutions using Google BigQuery.
Architect and manage PostgreSQL databases, ensuring efficient schema design and data partitioning.
Work extensively with Apache Parquet for data serialization and efficient big data storage solutions.
Write high-performance SQL queries and perform tuning on complex datasets.
Collaborate with data scientists, application developers, and product teams to deliver data-driven capabilities.
Optimize data processing and CPU utilization through strong understanding of distributed systems.
Implement best practices for ETL/ELT processes, data quality, monitoring, and security.
Benchmark system performance and recommend improvements for scalability and efficiency.
Required Skills and Qualifications
Bachelor’s or Master’s degree in Computer Science, Engineering, or a related technical field.
6–10 years of experience in Data Engineering or Big Data roles.
Expert-level proficiency in Apache Spark and Scala.
Strong working experience with Google BigQuery and PostgreSQL.
Hands-on expertise with Apache Parquet, ETL frameworks, and distributed computing.
Strong knowledge of SQL optimization and database performance tuning.
Familiarity with CPU localization, cluster performance tuning, and hardware-level optimization techniques.
Experience working in cloud-based environments (preferably Google Cloud Platform).
Excellent communication skills and ability to work in distributed teams.
Nice to Have
Experience with Airflow, Kafka, or other orchestration and streaming tools.
Knowledge of data modeling, data governance, and DevOps CI/CD principles for data systems.
Strong analytical, problem-solving, and debugging abilities.
Title: : Big Data Engineer
Duration: Long term
Location: Atlanta, GA (Remote)
(ONLY W2)
Job Description:
We are seeking a highly skilled Big Data Engineer at the senior or staff level to design, develop, and optimize large-scale data platforms and processing systems. The ideal candidate will have a strong background in Apache Spark, Scala, Google BigQuery, PostgreSQL, and Apache Parquet, with hands-on experience building performant, scalable, and reliable data pipelines.This role requires an individual who can write efficient SQL queries, optimize performance across distributed architectures, and collaborate with cross-functional teams to deliver end-to-end data solutions. Experience in CPU localization or performance tuning at the hardware level is a plus.
Key Responsibilities
Design, implement, and maintain highly scalable data processing pipelines using Apache Spark and Scala.
Develop optimized data warehouse and analytics solutions using Google BigQuery.
Architect and manage PostgreSQL databases, ensuring efficient schema design and data partitioning.
Work extensively with Apache Parquet for data serialization and efficient big data storage solutions.
Write high-performance SQL queries and perform tuning on complex datasets.
Collaborate with data scientists, application developers, and product teams to deliver data-driven capabilities.
Optimize data processing and CPU utilization through strong understanding of distributed systems.
Implement best practices for ETL/ELT processes, data quality, monitoring, and security.
Benchmark system performance and recommend improvements for scalability and efficiency.
Required Skills and Qualifications
Bachelor’s or Master’s degree in Computer Science, Engineering, or a related technical field.
6–10 years of experience in Data Engineering or Big Data roles.
Expert-level proficiency in Apache Spark and Scala.
Strong working experience with Google BigQuery and PostgreSQL.
Hands-on expertise with Apache Parquet, ETL frameworks, and distributed computing.
Strong knowledge of SQL optimization and database performance tuning.
Familiarity with CPU localization, cluster performance tuning, and hardware-level optimization techniques.
Experience working in cloud-based environments (preferably Google Cloud Platform).
Excellent communication skills and ability to work in distributed teams.
Nice to Have
Experience with Airflow, Kafka, or other orchestration and streaming tools.
Knowledge of data modeling, data governance, and DevOps CI/CD principles for data systems.
Strong analytical, problem-solving, and debugging abilities.






