Senior Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer in Orlando, FL, on a contract basis. Key skills required include Python, PySpark, AWS Glue, and Apache Spark. Experience in healthcare data and ETL/ELT pipeline design is essential.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
-
πŸ—“οΈ - Date discovered
August 9, 2025
πŸ•’ - Project duration
Unknown
-
🏝️ - Location type
On-site
-
πŸ“„ - Contract type
Unknown
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
Orlando, FL
-
🧠 - Skills detailed
#"ETL (Extract #Transform #Load)" #Data Modeling #SQL (Structured Query Language) #AWS Glue #Storage #Normalization #Linux #Kafka (Apache Kafka) #Delta Lake #Python #Terraform #PySpark #Automation #Cloud #DevOps #Spark SQL #Data Pipeline #GitHub #Version Control #GIT #S3 (Amazon Simple Storage Service) #Data Quality #Data Lake #Airflow #Batch #Data Engineering #Datasets #IAM (Identity and Access Management) #AWS (Amazon Web Services) #DataOps #Apache Airflow #Lambda (AWS Lambda) #Scala #Spark (Apache Spark) #Infrastructure as Code (IaC) #Unix #Apache Spark #Data Lakehouse #Data Processing
Role description
Job Title: Senior Data Engineer Location: Orlando, FL (Complete Onsite Role) Duration: Contract Job Description: We are looking for a Senior Data Engineer to design, build, and optimize large-scale data processing systems supporting healthcare analytics and operational reporting. This role will involve working closely with DataOps, DevOps, and QA teams to enable scalable and reliable data pipelines. Key Responsibilities: Design and implement ETL/ELT pipelines using Python and PySpark Develop scalable data workflows using Apache Spark and AWS Glue Collaborate with QA and DevOps to integrate CI/CD and testing automation Manage data lake structures and ensure data quality, lineage, and auditability Optimize and monitor performance of batch and streaming pipelines Build infrastructure as code (IaC) using tools like Terraform, GitHub Actions Work across structured, semi-structured, and unstructured healthcare datasets Required Technical Skills: Core & Deep Knowledge Assessment: Python PySpark SQL (including Window functions and CASE) AWS Glue, S3, Lambda Apache Spark Apache Airflow Delta Lake/Data Lakehouse Architecture CI/CD (Terraform, GitHub Actions) ETL/ELT pipeline design and optimization Basic Overall Knowledge Assessment: Kafka Data modeling and normalization Unix/Linux Infrastructure as Code (IaC) Cloud storage, IAM, and networking fundamentals (AWS) Git version control Healthcare data domain knowledge Please Note: Centraprise is an equal opportunity employer. Applicants must be authorized to work in the U.S. U.S. citizens and Green Card holders are strongly encouraged to apply.