iVoyant

Senior Python Data Engineer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Python Data Engineer on a contract for "length" with a pay rate of "rate" located "location." Key skills include Python, Pandas, FastAPI, data modeling, and infrastructure-as-code. Experience with Snowflake and ETL frameworks is preferred.
🌎 - Country
United States
πŸ’± - Currency
$ USD
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
November 19, 2025
πŸ•’ - Duration
Unknown
-
🏝️ - Location
Unknown
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
United States
-
🧠 - Skills detailed
#FastAPI #Scala #Angular #dbt (data build tool) #Data Processing #Data Lake #Libraries #Automation #Terraform #AI (Artificial Intelligence) #Airflow #Data Engineering #Snowflake #Apache Airflow #"ETL (Extract #Transform #Load)" #Storage #API (Application Programming Interface) #Data Modeling #Migration #Pandas #Data Pipeline #AWS (Amazon Web Services) #Data Integration #Datasets #Python #GitHub
Role description
About the Project: Join a dynamic engineering team working on a high-impact tax reporting platform for the 2025 fiscal season. The core goal is to modernize and significantly accelerate the generation of Excel-based reports, reducing report generation time from minutes to seconds. This involves ingesting data from multiple upstream systems, transforming it using efficient data processing libraries, and serving it through APIs. The team emphasizes scalability, maintainability, and developer productivity, incorporating spec-driven development powered by AI tools. You’ll be contributing to the backend architecture and data pipeline powering this transformation, helping to evolve a high-performance system that's central to the client’s reporting domain. Responsibilities: β€’ Design, build, and maintain high-performance data processing pipelines using Python libraries (Pandas required; Polars nice-to-have). β€’ Develop and expose RESTful APIs using FastAPI or similar frameworks. β€’ Consume and process normalized Parquet files from multiple upstream sources to generate dynamic Excel reports. β€’ Contribute to a spec-driven development workflow (using GitHub Copilot, Claude, etc.) to scaffold and generate API/data pipeline code. β€’ Optimize report generation logic for speed and scalability, currently targeting sub-20 second response times. β€’ Integrate with messaging and storage mechanisms (e.g., Service Bus, Storage Accounts). β€’ Collaborate on infrastructure-as-code automation using Bicep similar tools (Terraform, CDK). β€’ Participate in design discussions for future migration to Snowflake and/or a data lake architecture. β€’ Contribute to CI/CD pipelines using GitHub Actions. Required Skills and Experience: β€’ Strong proficiency in Python for data processing with hands-on expertise in Pandas. β€’ Ability to quickly learn new frameworks such as Polars if needed. β€’ Experience building backend services or APIs using frameworks like FastAPI. β€’ Solid understanding of data modeling principles (Star Schema) and handling normalized datasets. β€’ Familiarity with enterprise messaging patterns and data integration from various sources (API-based and file-based). β€’ Experience working with GitHub and CI/CD pipelines (GitHub Actions or similar). β€’ Infrastructure-as-Code experience with Bicep or comparable tools (Terraform, AWS CDK). β€’ Comfort with spec-driven development and leveraging AI tools like GitHub Copilot for scaffolding. Nice-to-Have / Preferred Qualifications: β€’ Experience with Polars (not required). β€’ Experience with orchestration tools like Apache Airflow or ETL frameworks like dbt. β€’ Exposure to Snowflake (streams, tasks, stored procedures). β€’ Experience working with DuckDB and/or DocTV/OpenTI XLarge for report generation. β€’ Knowledge of Angular or frontend plugins for Excel is a plus but not required. β€’ Familiarity with async workflows and distributed processing concepts.