Kelly Science, Engineering, Technology & Telecom

Senior Data Engineer (Databricks / PySpark)

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer (Databricks / PySpark) on a W2 contract until Oct 2026, paying $64 - $71 per hour. Candidates must have 5+ years in Python and 3+ years in Databricks/PySpark, with strong communication skills.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
568
-
πŸ—“οΈ - Date
March 3, 2026
πŸ•’ - Duration
More than 6 months
-
🏝️ - Location
On-site
-
πŸ“„ - Contract
W2 Contractor
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
Urbandale, IA
-
🧠 - Skills detailed
#"ETL (Extract #Transform #Load)" #Lambda (AWS Lambda) #Data Science #Delta Lake #Data Processing #AWS (Amazon Web Services) #PySpark #Libraries #AI (Artificial Intelligence) #Python #Databricks #Datasets #Code Reviews #Data Engineering #Scala #Spatial Data #GitHub #Data Pipeline #Batch #S3 (Amazon Simple Storage Service) #Cloud #Spark (Apache Spark) #IAM (Identity and Access Management) #Kafka (Apache Kafka) #Terraform #Debugging
Role description
Important information: To be immediately considered, please send an updated version of your resume to somp767@kellyservices.com. Title: Sr. Software Engineer Pay Rate: $64 - $71 per hour Location: San Francisco Bay Area or the Des Moines Metro Area, Duration: Until Oct/2026 Type – W2 contract (No C2C) We will use noted Glider test. Location: Onsite β€” Candidates may be based in either the San Francisco Bay Area or the Des Moines Metro Area, depending on proximity. The engineering team is distributed across both locations. (Please designate on resume which location candidate would prefer to work). Interview process: 2 rounds. 1st round: talent assessment and technical questions. 2nd round: review assessment, additional technical round Description: We are seeking a highly technical and self-directed Senior Software Engineer to contribute to the development of data processing pipelines for a new AI-enabled data analytics product targeted at Large Ag customers. Ideal candidates will have: β€’ 5+ years of professional software development experience using Python β€’ 3+ years of hands-on experience with Databricks and PySpark in production environments We are looking for mid-career professionals with a proven track record of deploying cloud-native solutions in fast-paced software delivery environments. In addition to technical expertise, successful candidates will demonstrate: β€’ Strong communication skills, with the ability to clearly articulate technical concepts to both technical and non-technical stakeholders (this is extremely important - please vet out accordingly) β€’ The ability to work effectively with limited supervision in a distributed team environment β€’ A disciplined engineering approach: breaking down work into small, reviewable increments, authoring focused pull requests, and iterating toward solutions progressively rather than in large, delayed batches Key Responsibilities: β€’ Author and optimize PySpark Databricks ETL and streaming jobs to ensure efficient, scalable, and reliable data processing workflows β€’ Design and implement Databricks-native solutions β€” including Delta Live Tables, Structured Streaming, and Vector Search β€” to process large-scale datasets for analytical and operational use cases β€’ Build and maintain CI/CD pipelines using GitHub Actions, with a strong emphasis on code quality, test coverage, and incremental delivery β€’ Contribute infrastructure-as-code using Terraform β€’ Support field testing and customer operations by debugging and resolving data issues β€’ Work closely with data scientists to productionize prototypes and proof-of-concept models Required Skills & Experience: β€’ Excellent coding skills in Python with experience deploying production-grade software β€’ Deep professional experience building Databricks workflows, optimizing PySpark queries, and working with Delta Lake β€’ Hands-on experience with modern Databricks capabilities, particularly Structured Streaming, Delta Live Tables, and Vector Search β€’ Demonstrated proficiency with GitHub: authoring well-scoped pull requests, conducting code reviews, and managing collaborative branching workflows β€’ Solid understanding of cloud computing fundamentals, with working knowledge of AWS services such as S3, Lambda, and IAM Preferred Experience: β€’ Experience with event-driven architectures and streaming data pipelines (e.g., Kafka, Kinesis) β€’ Prior experience in cross-functional teams involving product, data science, and backend engineering β€’ Experience working with geospatial data and related libraries (beneficial but not required)