Data Engineer (Azure Data Platform)

Synextra
Risley, Warrington
Last month
Applications closed

Related Jobs

View all jobs

Data Engineer

Relation Therapeutics London, United Kingdom
Permanent

Data Engineer, Strategic Account Services

Amazon London, United Kingdom
Permanent

Data Engineer - Security Products, Monitored Access

Amazon London, United Kingdom
£40,000 – £60,000 pa On-site

Senior Data Engineer

Synthesia London, United Kingdom
Hybrid

Senior Research Engineer - Data

Synthesia London, United Kingdom
Remote

Senior Simulation Data Engineer

PhysicsX London, United Kingdom
Posted
18 Mar 2026 (Last month)

Data Engineer (Azure Data Platform)

About Synextra

Synextra is a Microsoft-specialist Managed Service Provider headquartered in Warrington, operating as a premium partner to regulated mid-market organisations including law firms, financial services firms, and mortgage lenders. We're deliberately small - around 35 people - because we believe the best outcomes come from technical depth, not headcount. Our AI Services Division is growing fast, and we're building out a serious data and engineering capability to match. This is a chance to get in early and shape how that function operates.

The Role

We're looking for a technically driven Azure Data Engineer to join our data platform team. You'll design, build, and maintain production-grade data pipelines on Microsoft Azure - transforming complex, diverse datasets into analytics-ready formats that power business intelligence and AI initiatives for our clients and internally.

The ideal candidate treats pipelines and infrastructure as code, with a genuine passion for software engineering in a data context. You'll work across the modern Azure data stack - ADF, ADLS Gen2, PySpark, Delta Lake - with increasing exposure to Microsoft Fabric as the platform matures. You'll collaborate closely with customers and internal teams to ensure data is structured and governed for reliable downstream consumption.

This is a hands-on engineering role with room to grow into leadership: you'll champion DevOps best practices, contribute to architectural decisions, and help mentor junior engineers as the team scales.

Responsibilities

* Architect and write production-grade ELT/ETL data pipelines using PySpark and Python within Azure ecosystem.

* Build custom, reusable data processing frameworks and libraries in Python/Scala to streamline ingestion and transformation tasks across the engineering team

* Programmatically ingest large volumes of structured and unstructured data from REST APIs, streaming platforms (e.g. Event Hubs, Kafka), and legacy databases into ADLS Gen2 and OneLake

* Develop structured data models aligned to Lakehouse, Medallion Architecture, and Delta Lake patterns

* Continuously profile, debug, and optimise Spark jobs, SQL queries, and Python scripts for maximum performance and cost-efficiency at scale

* Champion DevOps best practices: implement infrastructure-as-code (Terraform), automated testing, and CI/CD deployment pipelines via Git and Azure DevOps

* Identifying patterns in recurring issues and engineering permanent solutions

* Write comprehensive unit and integration tests for all data pipelines to ensure data integrity; enforce data governance protocols, RBAC, and encryption standards across all environments

Requirements

Essential Technical Skills

* Advanced proficiency in Python and PySpark, writing clean, modular, object-oriented code for data transformations

* Strong command of SQL (T-SQL, Spark SQL) for data exploration, validation, and final-stage modelling

* Deep hands-on experience with Microsoft Fabric and its tooling such as Azure Data Factory (ADF), and Azure Data Lake Storage (ADLS Gen2)

* Practical experience with Git, branching strategies, automated testing (e.g. pytest), and CI/CD orchestration via Azure DevOps

* Proven commercial track record of deploying complex data solutions on the Microsoft Azure platform

* Experience collaborating with a range of stakeholders to structure data for downstream consumption (e.g. MLflow, Power BI semantic models)

* Infrastructure-as-code experience with Terraform for Azure resource provisioning

Desirable Technical Skills

* Familiarity with streaming data architectures (Spark Structured Streaming)

* Knowledge of complementary modern data stack tools such as dbt for SQL-based transformations

* Experience integrating Large Language Models (LLMs) or operationalising AI/ML models

Personal Qualities

* Exceptional problem-solving abilities and a persistent, detail-oriented approach to debugging complex code

* Strong communication skills to effectively translate business requirements into technical architectures

* A proactive mindset focused on continuous learning and staying ahead of the rapidly evolving data landscape

* Willingness to review code submissions, enforce coding standards, and mentor junior engineers on the team

Preferred Background

* 3–5+ years in software engineering, data engineering, or Big Data environments with a code-first approach

* Proven commercial experience deploying and maintaining complex data solutions on Microsoft Azure

* Experience working in cross-functional teams

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

Where to Advertise Machine Learning Jobs in the UK (2026 Guide)

Advertising machine learning jobs in the UK requires a different approach to most technical hiring. The candidate pool is small, highly specialised and in demand across AI labs, financial services, healthcare, autonomous systems and consumer technology simultaneously. Machine learning engineers and researchers move between roles through professional networks, conference communities and specialist platforms — not general job boards where ML roles compete with unrelated software engineering positions for the same audience. This guide, published by MachineLearningJobs.co.uk, covers where to advertise machine learning roles in the UK in 2026, how the main platforms compare, what employers should expect to pay, and what the data says about hiring across different role types.

New Machine Learning Employers to Watch in 2026: UK and Global Companies Driving ML Innovation

Machine learning (ML) has transitioned from a specialised field into a core business capability. In 2026, organisations across healthcare, finance, robotics, autonomous systems, natural language processing, and analytics are expanding their machine learning teams to build scalable intelligent products and services. For professionals exploring opportunities on www.MachineLearningJobs.co.uk , understanding the companies that are scaling, winning investment, or securing high‑impact contracts is crucial. This article highlights the new and high‑growth machine learning employers to watch in 2026, focusing on UK innovators, international firms with significant UK presence, and global platforms investing in machine learning talent locally.

How Many Machine Learning Tools Do You Need to Know to Get a Machine Learning Job?

Machine learning is one of the most exciting and rapidly growing areas of tech. But for job seekers it can also feel like a maze of tools, frameworks and platforms. One job advert wants TensorFlow and Keras. Another mentions PyTorch, scikit-learn and Spark. A third lists Mlflow, Docker, Kubernetes and more. With so many names out there, it’s easy to fall into the trap of thinking you must learn everything just to be competitive. Here’s the honest truth most machine learning hiring managers won’t say out loud: 👉 They don’t hire you because you know every tool. They hire you because you can solve real problems with the tools you know. Tools are important — no doubt — but context, judgement and outcomes matter far more. So how many machine learning tools do you actually need to know to get a job? For most job seekers, the real number is far smaller than you think — and more logically grouped. This guide breaks down exactly what employers expect, which tools are core, which are role-specific, and how to structure your learning for real career results.