Gazelle Global

Data Engineer (Pyspark)

⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer (Pyspark) with a contract length of "unknown," offering a pay rate of "unknown." Required skills include Azure, Spark, Python, and data pipeline development, with a focus on financial market datasets and compliance protocols.
🌎 - Country
United Kingdom
πŸ’± - Currency
Β£ GBP
-
πŸ’° - Day rate
Unknown
-
πŸ—“οΈ - Date
November 20, 2025
πŸ•’ - Duration
Unknown
-
🏝️ - Location
Unknown
-
πŸ“„ - Contract
Unknown
-
πŸ”’ - Security
Unknown
-
πŸ“ - Location detailed
London Area, United Kingdom
-
🧠 - Skills detailed
#Programming #Documentation #Microsoft Azure #Scala #"ETL (Extract #Transform #Load)" #GitLab #PySpark #Python #Dataflow #Azure #Data Processing #Unit Testing #Compliance #Security #Data Engineering #SQL (Structured Query Language) #Datasets #Spark (Apache Spark) #Databases #Spark SQL #Batch #Semantic Models #NoSQL #Agile #Data Pipeline #Java
Role description
Senior Data Engineer (Developer) – Pyspark We are supporting a leading global financial markets infrastructure and data provider as they modernise and scale their core data engineering capabilities. This role sits at the centre of their transformation programme, delivering high-quality data pipelines, models, and platforms that underpin critical services across the business. Key Responsibilities Design, build, and optimise scalable data pipelines for both batch and streaming workloads: β€’ Develop dataflows and semantic models aligned to analytics and reporting needs β€’ Implement complex transformations and performance-focused data processing logic β€’ Apply data validation, cleansing, and profiling techniques to ensure accuracy and consistency β€’ Implement access controls, data masking, and compliance-aligned security protocols β€’ Tune workloads and optimise performance across Spark, Fabric, and Azure components β€’ Translate business requirements into technical solutions through close collaboration with analysts and stakeholders β€’ Maintain clear documentation and contribute to internal knowledge repositories Essential Skills Strong experience developing within Microsoft Azure and Microsoft Fabric: β€’ Proficiency in Spark programming including DataFrames, RDDs, and Spark SQL β€’ Python and PySpark development experience, including notebook-based workflows β€’ Hands-on experience with Spark streaming and batch processing β€’ Delta table optimisation and Fabric Spark job development β€’ Solid Java programming and OOP understanding β€’ Experience working with relational and NoSQL databases β€’ Familiarity with GitLab, unit testing, and CI/CD pipelines β€’ Strong troubleshooting ability and experience working in Agile environments β€’ Excellent communication skills with stakeholder-facing experience β€’ Practical experience building ETL workflows, lakehouse architectures, dataflows, and semantic models β€’ Exposure to time-series data, financial market feeds, transactional records, and risk-related datasets