

Data Engineer
⭐ - Featured Role | Apply direct with Data Freelance Hub
This role is for a Data Engineer with a contract length of "unknown" and a pay rate of "unknown," located in "unknown." Key skills include GCP, SQL, data modeling, and experience with data visualization tools. A GCP data engineering certification is a merit.
🌎 - Country
United Kingdom
💱 - Currency
£ GBP
💰 - Day rate
Unknown
Unknown
🗓️ - Date discovered
April 29, 2025
🕒 - Project duration
Unknown
🏝️ - Location type
Unknown
📄 - Contract type
Unknown
🔒 - Security clearance
Unknown
📍 - Location detailed
London Area, United Kingdom
🧠 - Skills detailed
#Data Processing #Programming #Cloud #DAX #dbt (data build tool) #GCP (Google Cloud Platform) #ML (Machine Learning) #Data Engineering #Databases #Java #Deployment #DevOps #Dataflow #Microsoft Excel #Scala #"ETL (Extract #Transform #Load)" #BigQuery #Power Pivot #Terraform #Data Pipeline #API (Application Programming Interface) #NoSQL #Python #Spark (Apache Spark) #SQL (Structured Query Language) #RDBMS (Relational Database Management System) #Visualization #Automation
Role description
Requirement
• Several years of relevant work experience
• Take end-to-end responsibility to build, optimize and support of existing and new data products towards the defined target vision
• Be a champion of DevOps mindset and principles and able to manage CI/CD pipelines and terraform as well as Cloud infrastructure, in our context, it is GCP (Google Cloud Platform).
• Evaluate and drive continuous improvement and reducing technical debt in the teams
• Design and implement efficient data models, data pipelines that support analytical requirements. Good understanding of different data modelling techniques and trade-off
• Should have experience with Data Modelling
• Experience in data query languages (SQL or similar). Knowledge of ETL processes and tool
• Experience in data centric and API programming (for automation) using one of more programming languages Python, Java /or Scala.
• Knowledge of NoSQL and RDBMS databases
• Experience in different data formats (Avro, Parquet)
• Have a collaborative and co-creative mindset with excellent communication skills
• Motivated to work in an environment that allows you to work and take decisions independently
• Experience in working with data visualization tools
• Experience in GCP tools – Cloud Function, Dataflow, Dataproc and Bigquery
• Experience in data processing framework – Beam, Spark, Hive, Flink
• GCP data engineering certification is a merit
• Have hands on experience in Analytical tools such as powerBI or similar visualization tools
• Exhibit understanding in creating intermediate-level DAX measures to enhance data models and visualizations
• Have understanding of Microsoft excel functions such as: power pivot, power query. Tabular Editor, DAX etc.
• Fluent in English both written and verbal
The candidate:
• Partner with retail business units (e.g., merchandising, supply chain, stores, digital) to design and deliver domain-aligned data products that power analytics and machine learning initiatives.
• Translate complex retail business needs into technical requirements and proactively identify opportunities for data-driven innovation.
• Lead the development and deployment of data mesh architecture, ensuring federated governance, discoverability, and self-serve capabilities.
• Design and build scalable data pipelines using GCP (BigQuery, Dataflow, Cloud Composer, Cloud Functions) and orchestrate transformations using DBT.
• Develop modular, reusable DBT models for core retail metrics such as inventory accuracy, sales trends, promotions performance, and customer loyalty.
Requirement
• Several years of relevant work experience
• Take end-to-end responsibility to build, optimize and support of existing and new data products towards the defined target vision
• Be a champion of DevOps mindset and principles and able to manage CI/CD pipelines and terraform as well as Cloud infrastructure, in our context, it is GCP (Google Cloud Platform).
• Evaluate and drive continuous improvement and reducing technical debt in the teams
• Design and implement efficient data models, data pipelines that support analytical requirements. Good understanding of different data modelling techniques and trade-off
• Should have experience with Data Modelling
• Experience in data query languages (SQL or similar). Knowledge of ETL processes and tool
• Experience in data centric and API programming (for automation) using one of more programming languages Python, Java /or Scala.
• Knowledge of NoSQL and RDBMS databases
• Experience in different data formats (Avro, Parquet)
• Have a collaborative and co-creative mindset with excellent communication skills
• Motivated to work in an environment that allows you to work and take decisions independently
• Experience in working with data visualization tools
• Experience in GCP tools – Cloud Function, Dataflow, Dataproc and Bigquery
• Experience in data processing framework – Beam, Spark, Hive, Flink
• GCP data engineering certification is a merit
• Have hands on experience in Analytical tools such as powerBI or similar visualization tools
• Exhibit understanding in creating intermediate-level DAX measures to enhance data models and visualizations
• Have understanding of Microsoft excel functions such as: power pivot, power query. Tabular Editor, DAX etc.
• Fluent in English both written and verbal
The candidate:
• Partner with retail business units (e.g., merchandising, supply chain, stores, digital) to design and deliver domain-aligned data products that power analytics and machine learning initiatives.
• Translate complex retail business needs into technical requirements and proactively identify opportunities for data-driven innovation.
• Lead the development and deployment of data mesh architecture, ensuring federated governance, discoverability, and self-serve capabilities.
• Design and build scalable data pipelines using GCP (BigQuery, Dataflow, Cloud Composer, Cloud Functions) and orchestrate transformations using DBT.
• Develop modular, reusable DBT models for core retail metrics such as inventory accuracy, sales trends, promotions performance, and customer loyalty.