← All Jobs
Posted Mar 31, 2026

Senior Data Engineer

Apply Now

This is a remote position.

Job Title: Senior Data Engineer
Experience: 7–9 Years
Location: Remote
Notice Period: Immediate Joiners Only

About the Engagement

We are looking for experienced Senior Data Engineers to join a large-scale, multi-year Data Mapping and DataOps Platform modernisation programme for a major government ministry.

This strategic initiative focuses on transforming how data is managed, governed, and leveraged as a core enterprise asset.

Key Focus Areas

  • Data Engineering: Designing, building, and optimising scalable data pipelines to enhance stability, traceability, and refresh frequency
  • Data Mapping & Metadata Capture: Onboarding critical financial systems into a centralised enterprise data catalogue and supporting governance initiatives

Work will be delivered in an Agile/Scrum model, collaborating closely with data product managers and cross-functional teams.

Platform Overview

The platform is built on Microsoft Azure and Microsoft Fabric, with an existing in-house Finance Data Catalogue (FDC).

This role will contribute to scaling the platform into a production-grade DataOps ecosystem.

Key Responsibilities

  • Design, build, optimise, and maintain scalable data pipelines
  • Develop and manage ELT pipelines, orchestration, and automation using Python
  • Capture and onboard metadata into enterprise data catalogues
  • Perform data analysis to identify patterns, anomalies, and quality improvements
  • Contribute to data product design, including data models and structures
  • Design and implement data models for analytics and reporting
  • Build and maintain CI/CD pipelines for automated deployment
  • Develop automation scripts using Python, Bash, and PowerShell
  • Develop and maintain RESTful APIs for integration and interoperability
  • Implement data quality checks and validation frameworks
  • Set up monitoring, alerting, and performance tracking
  • Develop dashboards and reporting solutions using Microsoft Fabric & Power BI
  • Collaborate with stakeholders to translate business needs into technical solutions
  • Ensure compliance with data governance, privacy, and security standards
  • Maintain technical documentation and communicate with both technical and business teams

Mandatory Requirements

Candidates must have recent (last 48 months) hands-on production experience in:

  • Python (data pipelines, scripting, automation)
  • SQL (data transformation, modelling)
  • Microsoft Azure & Microsoft Fabric
  • ELT Pipeline Development
  • Git (version control & collaboration)
  • Metadata Management & Data Catalogue onboarding
  • Data Product Design
  • Agile/Scrum methodologies
  • Data Analysis & Data Quality Assessment

Technical Skills Required

Cloud & Platform

  • Microsoft Azure
  • Microsoft Fabric

Data Engineering

  • Python, SQL
  • Bash, PowerShell

Pipelines & DevOps

  • ELT pipelines
  • CI/CD pipelines

Data Modelling

  • Schema design
  • Metadata management
  • Data product design

API Development

  • RESTful APIs

Analytics & Reporting

  • Power BI
  • Microsoft Fabric Reporting

Version Control

  • Git, GitHub

Methodology

  • Agile / Scrum

 Nice to Have

  • Advanced Power BI (DAX, dashboarding, modelling)
  • Data visualization beyond standard reporting
  • Experience with REST API integrations
  • Exposure to:
    • Azure Synapse Analytics
    • Azure Data Lake
    • Azure DevOps
  • Experience in government/public sector or regulated environments
  • Familiarity with data governance frameworks
  • Experience in financial systems modernisation programs