

Lead PySpark Developer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead PySpark Developer with a contract length of "unknown" and a pay rate of "unknown." Key skills required include 7+ years in AWS, 10+ years in big data, and strong experience with PySpark, SQL, and ETL workflows.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
520
-
ποΈ - Date discovered
June 1, 2025
π - Project duration
Unknown
-
ποΈ - Location type
Unknown
-
π - Contract type
W2 Contractor
-
π - Security clearance
Unknown
-
π - Location detailed
Owings Mills, MD
-
π§ - Skills detailed
#Data Modeling #Snowflake #Cloud #Data Security #SQL (Structured Query Language) #NoSQL #Compliance #Apache Spark #PySpark #Big Data #dbt (data build tool) #"ETL (Extract #Transform #Load)" #Data Science #Spark (Apache Spark) #Distributed Computing #DevOps #PostgreSQL #Docker #Databases #Deployment #AWS (Amazon Web Services) #Airflow #Scala #Security #Version Control #Kubernetes #Python #Data Engineering #Unit Testing
Role description
Job Description
β’ 7+ years of experience in Amazon Web Service(AWS) Cloud Computing.
β’ 10+ years of experience in big data and distributed computing.
β’ Very Strong hands-on experience with PySpark, Apache Spark, and Python.
β’ Strong Hands on experience with SQL and NoSQL databases (DB2, PostgreSQL, Snowflake, etc.).
β’ Proficiency in data modeling and ETL workflows.
β’ Proficiency with workflow schedulers like Airflow.
β’ Hands on experience with AWS cloud-based data platforms.
β’ Experience in DevOps, CI/CD pipelines, and containerization (Docker, Kubernetes) is a plus.
β’ Strong problem-solving skills and ability to lead a team
β’ DBT, AWS Astronomer
β’ Lead the design, development, and deployment of PySpark-based big data solutions.
β’ Architect and optimize ETL pipelines for structured and unstructured data.
β’ Collaborate with Client, data engineers, data scientists, and business teams to understand requirements and provide scalable solutions.
β’ Optimize Spark performance through partitioning, caching, and tuning.
β’ Implement best practices in data engineering (CI/CD, version control, unit testing).
β’ Work with cloud platforms like AWS.
β’ Ensure data security, governance, and compliance.
β’ Mentor junior developers and review code for best practices and efficiency.
Additional Notes
β’ Please submit the candidate's resume in PDF format.
β’ Please Note: This position is not available under a Corp-to-Corp (C2C) employment structure. Only candidates authorized to work on a W2 basis will be considered at this time.
Skills: pyspark,docker,containerization,sql,cloud computing,nosql,apache spark,big data,airflow,data modeling,dbt,devops,distributed computing,etl workflows,amazon web service (aws),python,ci/cd,kubernetes,aws astronomer,aws cloud computing
Job Description
β’ 7+ years of experience in Amazon Web Service(AWS) Cloud Computing.
β’ 10+ years of experience in big data and distributed computing.
β’ Very Strong hands-on experience with PySpark, Apache Spark, and Python.
β’ Strong Hands on experience with SQL and NoSQL databases (DB2, PostgreSQL, Snowflake, etc.).
β’ Proficiency in data modeling and ETL workflows.
β’ Proficiency with workflow schedulers like Airflow.
β’ Hands on experience with AWS cloud-based data platforms.
β’ Experience in DevOps, CI/CD pipelines, and containerization (Docker, Kubernetes) is a plus.
β’ Strong problem-solving skills and ability to lead a team
β’ DBT, AWS Astronomer
β’ Lead the design, development, and deployment of PySpark-based big data solutions.
β’ Architect and optimize ETL pipelines for structured and unstructured data.
β’ Collaborate with Client, data engineers, data scientists, and business teams to understand requirements and provide scalable solutions.
β’ Optimize Spark performance through partitioning, caching, and tuning.
β’ Implement best practices in data engineering (CI/CD, version control, unit testing).
β’ Work with cloud platforms like AWS.
β’ Ensure data security, governance, and compliance.
β’ Mentor junior developers and review code for best practices and efficiency.
Additional Notes
β’ Please submit the candidate's resume in PDF format.
β’ Please Note: This position is not available under a Corp-to-Corp (C2C) employment structure. Only candidates authorized to work on a W2 basis will be considered at this time.
Skills: pyspark,docker,containerization,sql,cloud computing,nosql,apache spark,big data,airflow,data modeling,dbt,devops,distributed computing,etl workflows,amazon web service (aws),python,ci/cd,kubernetes,aws astronomer,aws cloud computing