TPA technologies

Senior Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer with 6–10 years of experience in data engineering, focusing on Apache Airflow and Apache Spark. Contract length is unspecified, with a W2 pay rate. Location is remote. A Bachelor's degree and AWS experience are required.
🌎 - Country
United States
💱 - Currency
$ USD
-
💰 - Day rate
592
-
🗓️ - Date
May 12, 2026
🕒 - Duration
Unknown
-
🏝️ - Location
Unknown
-
📄 - Contract
W2 Contractor
-
🔒 - Security
Unknown
-
📍 - Location detailed
United States
-
🧠 - Skills detailed
#Athena #Data Lake #Hadoop #Datasets #Batch #Data Processing #Data Pipeline #Apache Spark #Redshift #Security #Version Control #Airflow #Agile #Cloud #Big Data #Spark (Apache Spark) #Data Engineering #Data Orchestration #Scrum #DevOps #AWS (Amazon Web Services) #Data Architecture #Kafka (Apache Kafka) #Data Quality #Apache Airflow #Scala #Computer Science #PySpark #Data Catalog #Data Warehouse #SQL (Structured Query Language)
Role description
No third party recruiters please Must be able to work W2 Job Summary: We are seeking an experienced Data Engineer with 6–10 years of expertise in designing, building, and maintaining scalable, high-performance data pipelines and processing frameworks. The ideal candidate will have strong hands-on experience with orchestration using Apache Airflow and distributed data processing with Apache Spark (EMR). This role requires a solid understanding of big data architecture, data engineering best practices, and a commitment to delivering efficient, reliable, and maintainable data solutions that align with both business and technical requirements. \_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_ Roles and Responsibilities: • Design, build, and manage scalable and reliable data pipelines using Apache Airflow. • Develop and optimize large-scale data processing workflows using Apache Spark, including both batch and Structured Streaming. • Collaborate with data architects, analysts, and business stakeholders to translate data requirements into efficient engineering solutions. • Ensure the quality, performance, and security of data processes and systems. • Monitor, troubleshoot, and optimize data workflows and job executions. • Document solutions, workflows, and technical designs. \_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_ Qualifications Required: • Bachelor’s degree in Computer Science, Information Systems, Engineering, or a related field. • 6–10 years of experience in data engineering or related roles. • Strong experience with Apache Airflow for data orchestration and workflow management. • Proven expertise in building and tuning distributed data processing applications using Apache Spark (PySpark), including both Structured Streaming and Batch. • Experience with AWS cloud-based data ecosystems, particularly Athena and Redshift. • Proficient in SQL (Athena version) and experienced in working with large datasets from various sources (structured and unstructured). • Experience with data lakes, data warehouses, and batch/streaming data architectures. • Familiarity with CI/CD pipelines, version control, and DevOps practices in a data engineering context. • Strong problem-solving and communication skills, with the ability to work both independently and collaboratively. \_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_\_ Preferred Qualifications: • Familiarity with additional big data technologies (e.g., Hadoop, Kafka). • Experience working in Agile/Scrum environments. • Knowledge of data quality frameworks and validation engines. • Experience with data catalog tools.