

Lead PySpark Developer
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Lead PySpark Developer in Great River, offering a contract length of "unknown" and a pay rate of "unknown." Requires 10 years of big data experience, proficiency in PySpark, SQL, AWS, and strong leadership skills.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
June 3, 2025
π - Project duration
Unknown
-
ποΈ - Location type
On-site
-
π - Contract type
Unknown
-
π - Security clearance
Unknown
-
π - Location detailed
Portland, OR
-
π§ - Skills detailed
#Spark (Apache Spark) #Airflow #Big Data #AWS (Amazon Web Services) #Cloud #Data Security #DevOps #Docker #Data Modeling #PostgreSQL #"ETL (Extract #Transform #Load)" #Snowflake #Version Control #Databases #SQL (Structured Query Language) #Deployment #Python #Compliance #Scala #Data Engineering #Distributed Computing #Kubernetes #Data Science #dbt (data build tool) #NoSQL #Security #Unit Testing #Apache Spark #PySpark
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
Join our dynamic team in Great River, where innovation meets opportunity. We are seeking an experienced professional with a passion for big data and distributed computing. If you are a proactive individual with a strong background in PySpark and Apache Spark, this role is for you.
Key Responsibilities
β’ Lead the design, development, and deployment of PySpark-based big data solutions.
β’ Architect and optimize ETL pipelines for both structured and unstructured data.
β’ Collaborate with data engineers, data scientists, and business teams to understand requirements and deliver scalable solutions.
β’ Optimize Spark performance through partitioning, caching, and tuning.
β’ Implement best practices in data engineering, including CI/CD, version control, and unit testing.
β’ Work with cloud platforms like AWS to enhance data solutions.
β’ Ensure data security, governance, and compliance.
β’ Mentor junior developers and review code for best practices and efficiency.
Qualifications
β’ 10 years of experience in big data and distributed computing.
β’ Strong hands-on experience with PySpark, Apache Spark, and Python.
β’ Proficient in SQL and NoSQL databases such as DB2, PostgreSQL, and Snowflake.
β’ Experience in data modeling and ETL workflows.
β’ Proficiency with workflow schedulers like Airflow.
β’ Hands-on experience with AWS cloud-based data platforms.
β’ Experience in DevOps, CI/CD pipelines, and containerization (Docker, Kubernetes) is a plus.
β’ Strong problem-solving skills and ability to lead a team.
β’ Experience with DBT and AWS Astronomer is advantageous.
We are committed to fostering a diverse and inclusive workplace. We welcome applications from all qualified individuals, regardless of gender, race, or background. For immediate consideration please click APPLY or contact me directly at priyanka@thegreatriverfinancial.com