

Lead PySpark Developer (7234-1)
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead PySpark Developer with a 12-month contract, offering a competitive pay rate. Key requirements include 7+ years in AWS, 10+ years in big data, strong PySpark and SQL/NoSQL experience, and proficiency in ETL workflows.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
568
-
ποΈ - Date discovered
July 3, 2025
π - Project duration
Unknown
-
ποΈ - Location type
Unknown
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Owings Mills, MD
-
π§ - Skills detailed
#SQL (Structured Query Language) #Security #Data Science #Cloud #DevOps #Version Control #Deployment #AWS (Amazon Web Services) #Docker #Python #Data Security #Data Engineering #Kubernetes #Data Modeling #Big Data #Spark (Apache Spark) #"ETL (Extract #Transform #Load)" #dbt (data build tool) #Unit Testing #Snowflake #Compliance #Airflow #Apache Spark #Databases #NoSQL #PySpark #Scala #PostgreSQL #Distributed Computing
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Job Description
β’ 7+ years of experience in Amazon Web Service(AWS) Cloud Computing.
β’ 10+ years of experience in big data and distributed computing.
β’ Very Strong hands-on experience with PySpark, Apache Spark, and Python.
β’ Strong Hands on experience with SQL and NoSQL databases (DB2, PostgreSQL, Snowflake, etc.).
β’ Proficiency in data modeling and ETL workflows.
β’ Proficiency with workflow schedulers like Airflow.
β’ Hands on experience with AWS cloud-based data platforms.
β’ Experience in DevOps, CI/CD pipelines, and containerization (Docker, Kubernetes) is a plus.
β’ Strong problem-solving skills and ability to lead a team
β’ DBT, AWS Astronomer
β’ Lead the design, development, and deployment of PySpark-based big data solutions.
β’ Architect and optimize ETL pipelines for structured and unstructured data.
β’ Collaborate with Client, data engineers, data scientists, and business teams to understand requirements and provide scalable solutions.
β’ Optimize Spark performance through partitioning, caching, and tuning.
β’ Implement best practices in data engineering (CI/CD, version control, unit testing).
β’ Work with cloud platforms like AWS.
β’ Ensure data security, governance, and compliance.
β’ Mentor junior developers and review code for best practices and efficiency.
Must Have
β’ 7+ years of experience in Amazon Web Service(AWS) Cloud Computing.
β’ 10+ years of experience in big data and distributed computing.
β’ Experience with PySpark, Apache Spark, and Python.
β’ Experience with SQL and NoSQL databases (DB2, PostgreSQL, Snowflake, etc.).
β’ Hands on experience with AWS cloud-based data platforms.
β’ Experience in DevOps, CI/CD pipelines, and containerization (Docker, Kubernetes) is a plus.
Skills: cloud computing,nosql,cd,pyspark developer,python,pyspark,ci/cd,apache spark,big data,airflow,etl workflows,sql,snowflake,ci,kubernetes,aws,postgresql,aws astronomer,distributed computing,docker,db2,devops,data modeling,dbt,aws cloud computing