Lead PySpark Developer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead PySpark Developer with a contract length of "unknown" and a pay rate of "unknown." Key skills required include 7+ years in AWS, 10+ years in big data, and strong experience with PySpark, SQL, and ETL workflows.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
520
-
πŸ—“οΈ - Date discovered
June 1, 2025
πŸ•’ - Project duration
Unknown
-
🏝️ - Location type
Unknown
-
πŸ“„ - Contract type
W2 Contractor
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
Owings Mills, MD
-
🧠 - Skills detailed
#Data Modeling #Snowflake #Cloud #Data Security #SQL (Structured Query Language) #NoSQL #Compliance #Apache Spark #PySpark #Big Data #dbt (data build tool) #"ETL (Extract #Transform #Load)" #Data Science #Spark (Apache Spark) #Distributed Computing #DevOps #PostgreSQL #Docker #Databases #Deployment #AWS (Amazon Web Services) #Airflow #Scala #Security #Version Control #Kubernetes #Python #Data Engineering #Unit Testing
Role description
Job Description β€’ 7+ years of experience in Amazon Web Service(AWS) Cloud Computing. β€’ 10+ years of experience in big data and distributed computing. β€’ Very Strong hands-on experience with PySpark, Apache Spark, and Python. β€’ Strong Hands on experience with SQL and NoSQL databases (DB2, PostgreSQL, Snowflake, etc.). β€’ Proficiency in data modeling and ETL workflows. β€’ Proficiency with workflow schedulers like Airflow. β€’ Hands on experience with AWS cloud-based data platforms. β€’ Experience in DevOps, CI/CD pipelines, and containerization (Docker, Kubernetes) is a plus. β€’ Strong problem-solving skills and ability to lead a team β€’ DBT, AWS Astronomer β€’ Lead the design, development, and deployment of PySpark-based big data solutions. β€’ Architect and optimize ETL pipelines for structured and unstructured data. β€’ Collaborate with Client, data engineers, data scientists, and business teams to understand requirements and provide scalable solutions. β€’ Optimize Spark performance through partitioning, caching, and tuning. β€’ Implement best practices in data engineering (CI/CD, version control, unit testing). β€’ Work with cloud platforms like AWS. β€’ Ensure data security, governance, and compliance. β€’ Mentor junior developers and review code for best practices and efficiency. Additional Notes β€’ Please submit the candidate's resume in PDF format. β€’ Please Note: This position is not available under a Corp-to-Corp (C2C) employment structure. Only candidates authorized to work on a W2 basis will be considered at this time. Skills: pyspark,docker,containerization,sql,cloud computing,nosql,apache spark,big data,airflow,data modeling,dbt,devops,distributed computing,etl workflows,amazon web service (aws),python,ci/cd,kubernetes,aws astronomer,aws cloud computing