Python Quantitative Developer

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Python Quantitative Developer with a hybrid working arrangement in London. The contract length is unspecified, and pay is DOE. Key skills include Python programming, AWS/Azure experience, and financial data knowledge.
🌎 - Country
United Kingdom
πŸ’± - Currency
Β£ GBP
-
πŸ’° - Day rate
896
-
πŸ—“οΈ - Date discovered
August 29, 2025
πŸ•’ - Project duration
Unknown
-
🏝️ - Location type
Hybrid
-
πŸ“„ - Contract type
Unknown
-
πŸ”’ - Security clearance
Unknown
-
πŸ“ - Location detailed
London Area, United Kingdom
-
🧠 - Skills detailed
#"ETL (Extract #Transform #Load)" #Snowflake #Programming #Documentation #Version Control #Libraries #Automation #Airflow #AWS (Amazon Web Services) #Data Transformations #Deployment #NumPy #Python #Scala #Apache Airflow #Data Pipeline #Cloud #Data Architecture #Pandas #Azure #Data Engineering #Data Analysis
Role description
Python Quantitative Developer Rate - DOE Hybrid working - London Role Overview: The Quantitative Developer will design, implement, and maintain analytical models and processes, with a strong focus on Python development, leveraging statistical libraries, and utilising cloud platforms such as AWS, Azure and Snowflake for scalable deployment. Key Responsibilities: β€’ Develop and optimise quantitative models and analytical tools in Python β€’ Leverage statistical g packages (e.g., pandas, NumPy) for data analysis and modelling β€’ Work closely with portfolio managers, researchers, and technology teams to deploy robust, scalable solutions β€’ Design and implement data pipelines and workflows utilising AWS or Azure infrastructures β€’ Ensure best practices in code quality, version control, and deployment automation β€’ Provide technical support and documentation for developed solutions Required Skills and Experience: β€’ Proven experience in Python programming for quantitative or analytical applications β€’ Strong proficiency with relevant statistical libraries Pyth β€’ Practical experience utilising AWS and/or Azure for deployment and workflow orchestration β€’ Solid understanding of financial data and modelling techniques (preferred) β€’ Excellent analytical, communication, and problem-solving skills β€’ Experience with data engineering & ETL tools such as Apache Airflow or custom ETL scripts. β€’ Strong problem-solving skills with a keen analytical mindset especially in handling large data sets and complex data transformations. β€’ Strong experience in setting up an underlying test framework for complex data calculations and processing. β€’ Strong familiarity with engineering and data architecture principles. β€’ Prior experience in a front office or financial services environment (desirable)