

Gazelle Global
Data Engineer (Pyspark)
β - Featured Role | Apply direct with Data Freelance Hub
This role is for a Senior Data Engineer (Pyspark) with a contract length of "unknown," offering a pay rate of "unknown." Required skills include Azure, Spark, Python, and data pipeline development, with a focus on financial market datasets and compliance protocols.
π - Country
United Kingdom
π± - Currency
Β£ GBP
-
π° - Day rate
Unknown
-
ποΈ - Date
November 20, 2025
π - Duration
Unknown
-
ποΈ - Location
Unknown
-
π - Contract
Unknown
-
π - Security
Unknown
-
π - Location detailed
London Area, United Kingdom
-
π§ - Skills detailed
#Programming #Documentation #Microsoft Azure #Scala #"ETL (Extract #Transform #Load)" #GitLab #PySpark #Python #Dataflow #Azure #Data Processing #Unit Testing #Compliance #Security #Data Engineering #SQL (Structured Query Language) #Datasets #Spark (Apache Spark) #Databases #Spark SQL #Batch #Semantic Models #NoSQL #Agile #Data Pipeline #Java
Role description
Senior Data Engineer (Developer) β Pyspark
We are supporting a leading global financial markets infrastructure and data provider as they modernise and scale their core data engineering capabilities. This role sits at the centre of their transformation programme, delivering high-quality data pipelines, models, and platforms that underpin critical services across the business.
Key Responsibilities
Design, build, and optimise scalable data pipelines for both batch and streaming workloads:
β’ Develop dataflows and semantic models aligned to analytics and reporting needs
β’ Implement complex transformations and performance-focused data processing logic
β’ Apply data validation, cleansing, and profiling techniques to ensure accuracy and consistency
β’ Implement access controls, data masking, and compliance-aligned security protocols
β’ Tune workloads and optimise performance across Spark, Fabric, and Azure components
β’ Translate business requirements into technical solutions through close collaboration with analysts and stakeholders
β’ Maintain clear documentation and contribute to internal knowledge repositories
Essential Skills
Strong experience developing within Microsoft Azure and Microsoft Fabric:
β’ Proficiency in Spark programming including DataFrames, RDDs, and Spark SQL
β’ Python and PySpark development experience, including notebook-based workflows
β’ Hands-on experience with Spark streaming and batch processing
β’ Delta table optimisation and Fabric Spark job development
β’ Solid Java programming and OOP understanding
β’ Experience working with relational and NoSQL databases
β’ Familiarity with GitLab, unit testing, and CI/CD pipelines
β’ Strong troubleshooting ability and experience working in Agile environments
β’ Excellent communication skills with stakeholder-facing experience
β’ Practical experience building ETL workflows, lakehouse architectures, dataflows, and semantic models
β’ Exposure to time-series data, financial market feeds, transactional records, and risk-related datasets
Senior Data Engineer (Developer) β Pyspark
We are supporting a leading global financial markets infrastructure and data provider as they modernise and scale their core data engineering capabilities. This role sits at the centre of their transformation programme, delivering high-quality data pipelines, models, and platforms that underpin critical services across the business.
Key Responsibilities
Design, build, and optimise scalable data pipelines for both batch and streaming workloads:
β’ Develop dataflows and semantic models aligned to analytics and reporting needs
β’ Implement complex transformations and performance-focused data processing logic
β’ Apply data validation, cleansing, and profiling techniques to ensure accuracy and consistency
β’ Implement access controls, data masking, and compliance-aligned security protocols
β’ Tune workloads and optimise performance across Spark, Fabric, and Azure components
β’ Translate business requirements into technical solutions through close collaboration with analysts and stakeholders
β’ Maintain clear documentation and contribute to internal knowledge repositories
Essential Skills
Strong experience developing within Microsoft Azure and Microsoft Fabric:
β’ Proficiency in Spark programming including DataFrames, RDDs, and Spark SQL
β’ Python and PySpark development experience, including notebook-based workflows
β’ Hands-on experience with Spark streaming and batch processing
β’ Delta table optimisation and Fabric Spark job development
β’ Solid Java programming and OOP understanding
β’ Experience working with relational and NoSQL databases
β’ Familiarity with GitLab, unit testing, and CI/CD pipelines
β’ Strong troubleshooting ability and experience working in Agile environments
β’ Excellent communication skills with stakeholder-facing experience
β’ Practical experience building ETL workflows, lakehouse architectures, dataflows, and semantic models
β’ Exposure to time-series data, financial market feeds, transactional records, and risk-related datasets





