

Python and PySpark
β - Featured Role | Apply direct with Data Freelance Hub
This role is a W2 contract position for a Python and PySpark developer in McLean, focusing on ETL pipeline optimization, large-scale data processing, and AWS EMR. Key skills include Python, PySpark, IICS, and distributed computing.
π - Country
United States
π± - Currency
$ USD
-
π° - Day rate
-
ποΈ - Date discovered
July 11, 2025
π - Project duration
Unknown
-
ποΈ - Location type
On-site
-
π - Contract type
W2 Contractor
-
π - Security clearance
Unknown
-
π - Location detailed
McLean, VA
-
π§ - Skills detailed
#Scala #Cloud #PySpark #Big Data #AWS EMR (Amazon Elastic MapReduce) #Data Science #Datasets #IICS (Informatica Intelligent Cloud Services) #AWS (Amazon Web Services) #Spark (Apache Spark) #Automation #Distributed Computing #"ETL (Extract #Transform #Load)" #Python #Data Processing #Data Ingestion #Data Engineering
Role description
Heading 1
Heading 2
Heading 3
Heading 4
Heading 5
Heading 6
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.
Block quote
Ordered list
- Item 1
- Item 2
- Item 3
Unordered list
- Item A
- Item B
- Item C
Bold text
Emphasis
Superscript
Subscript
HI Everyone ,
Title : Python and PySpark
Location: McLean
Contract : w2 Contract
Job Description:
Develop and optimize ETL pipelines using Python, PySpark, PySpark Notebooks with AWS EMR.
Should have good understanding of Spark Resilient Distributed Datasets (RDD), Data Frame, and Datasets.
Work with large-scale datasets and build distributed computing solutions.
Design and implement data ingestion, transformation, and processing workflows using IICS jobs.
Write efficient and scalable Python code for data processing.
Collaborate with data engineers, data scientists, and business teams to deliver insights.
Optimize performance and cost efficiency for big data solutions.
Implement best practices for CI/CD, testing, and automation in a cloud environment.
Monitor job performance, troubleshooting failures, and tuning queries.
Thanks,
Mahesh VG
vmahesh@galaxyitech.com
NOTE: Please share your updated resume to vmahesh@galaxyitech.com or can reach out me at 480-407-6915.