Senior Staff Data Engineer

Warner Bros. Discovery
London
3 days ago
Create job alert

This job is with Warner Bros. Discovery, an inclusive employer and a member of myGwork – the largest global platform for the LGBTQ+ business community. Please do not contact the recruiter directly.

Welcome to Warner Bros. Discovery… the stuff dreams are made of.
Who We Are…
When we say, “the stuff dreams are made of,” we’re not just referring to the world of wizards, dragons and superheroes, or even to the wonders of Planet Earth. Behind WBD’s vast portfolio of iconic content and beloved brands, are the storytellers bringing our characters to life, the creators bringing them to your living rooms and the dreamers creating what’s next…

From brilliant creatives, to technology trailblazers, across the globe, WBD offers career defining opportunities, thoughtfully curated benefits, and the tools to explore and grow into your best selves. Here you are supported, here you are celebrated, here you can thrive.

Warner Bros. has been entertaining audiences for more than 90 years through the world’s most-loved characters and franchises. Warner Bros. employs people all over the world in a wide variety of disciplines. We're always on the lookout for energetic, creative people to join our team.

Your New Role...

We are seeking an exceptional Senior Staff Data Engineer to lead the design, development, and scaling of the data and platform systems that power our experimentation, adaptive optimisation, and automated decisioning ecosystem. This is a high-impact, hands-on technical leadership role that will shape how experimentation data is collected, processed, served, and operationalised across millions of users worldwide.

This role will act as a force multiplier for all Labs initiatives and, in particular, will support the development and productionization of the new Canvas Optimisation system.

As a senior technical leader, you will define the architecture and long-term strategy for experimentation data infrastructure, ensure reliable and cost-efficient data pipelines, and partner closely with Data Science, Engineering, Product, and Analytics teams to scale our platform from hundreds to thousands of concurrent experiments and bandits.

This role would be tasked with reducing Labs data processing costs by ~25% over the year through architectural optimisation, storage strategy improvements, compute efficiency, and intelligent data lifecycle management.

Your Role Accountabilities...

Scale Experimentation Data Platforms
Architect and lead the design of scalable, reliable data pipelines supporting large-scale A/B testing, multivariate testing, and adaptive experimentation.

Build and maintain systems that support real-time and batch experiment telemetry ingestion, feature logging, exposure tracking, and outcome measurement.

Design data models and storage strategies optimised for:
Experiment analysis latency

Cost efficiency

Long-term reproducibility

Governance and auditability

Enable production-grade pipelines for statistical methods such as CUPED, regression adjustment, and other variance-reduction workflows (in partnership with Data Science).

Enable Adaptive & Bandit Systems at Scale
Build data infrastructure that supports multi-armed bandit decisioning systems, including:
Low-latency reward signal pipelines

Feature and context streaming

Policy logging and replay data stores

Partner with scientists to productionize bandit frameworks (e.g., Thompson Sampling, epsilon-greedy, UCB) via reliable data services and APIs.

Design systems enabling off-policy evaluation (OPE), replay simulation datasets, and long-term policy evaluation storage.

Canvas Optimization & Personalization Infrastructure
Lead data system architecture supporting the Canvas Optimisation platform, including:

Artwork / creative performance telemetry

Impression → engagement attribution pipelines

Near real-time reward computation

Global rollout observability and monitoring

Ensure high availability, correctness, and explainability of decisioning-support data feeds.

Cost Optimisation & Efficiency Leadership
Drive initiatives to reduce overall Labs data processing costs by ~25%, including:
Query and job optimisation across compute platforms (Databricks, Spark, etc.)

Storage tiering and retention policy optimisation

Data compaction, partitioning, and indexing strategies

Eliminating redundant or low-value pipeline stages

Establish cost observability dashboards and cost-to-value monitoring frameworks.

Platform, Reliability & Developer Experience
Build reusable data services, libraries, and APIs that:
Simplify experiment onboarding

Standardize telemetry schemas

Enable self-service data access

Define and enforce SLAs for critical experimentation and decisioning datasets.

Lead data quality frameworks including anomaly detection, reconciliation, and automated validation.

Technical Leadership & Mentorship
Serve as a technical thought leader across data platform architecture for experimentation and optimization.

Mentor data engineers and platform engineers on distributed systems design, pipeline reliability, and performance optimization.

Influence cross-org roadmaps spanning experimentation science, platform engineering, and product personalization systems.

Qualifications and Experience... BS/MS in Computer Science, Engineering, or related field (or equivalent industry experience).

10+ years building and operating large-scale distributed data systems.

Deep experience with:
Streaming and batch data architectures

Experiment telemetry systems

Data modelling for analytics and decisioning

Hands-on experience with modern data stack technologies (e.g., Spark, Airflow, Redshift, Databricks, Snowflake, Delta Lake, etc.).

Strong programming mindset with a focus on building and enforcing standards that ensure code is maintainable, readable, and extensible over time.

Strong communication skills and ability to partner across DS, Engineering, Product, and Analytics.

Preferred Qualifications
Experience supporting experimentation or personalisation platforms at scale.

Familiarity with adaptive experimentation and bandit system data requirements.

Experience with cost optimisation of large-scale cloud data platforms.

Experience operating global, multi-region data systems.

Hybrid Working - This role is advertised as a Hybrid work model, that combines remote and in-office work, following our current company policy and to be agreed with your Line Manager. Subject to any applicable laws, WBD / your Line Manager reserves the right to change this working agreement where this is essential to business needs and upon reasonable notice to you.

How We Get Things Done…

This last bit is probably the most important! Here at WBD, our guiding principles are the core values by which we operate and are central to how we get things done. You can find them at www.wbd.com/guiding-principles/  along with some insights from the team on what they mean and how they show up in their day to day. We hope they resonate with you and look forward to discussing them during your interview.

Championing Inclusion at WBD
Warner Bros. Discovery embraces the opportunity to build a workforce that reflects a wide array of perspectives, backgrounds and experiences. Being an equal opportunity employer means that we take seriously our responsibility to consider qualified candidates on the basis of merit, regardless of sex, gender identity, ethnicity, age, sexual orientation, religion or belief, marital status, pregnancy, parenthood, disability or any other category protected by law.

If you’re a qualified candidate with a disability and you require adjustments or accommodations during the job application and/or recruitment process, please visit our accessibility page for instructions to submit your request.

Related Jobs

View all jobs

Senior Staff Data Engineer

Staff Data Engineer

Staff Data Engineer

Senior Data Engineer (AWS, Airflow, Python)

Senior Data Engineer (AWS, Airflow, Python)

Senior Data Engineer (AWS, Airflow, Python)

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

How Many Machine Learning Tools Do You Need to Know to Get a Machine Learning Job?

Machine learning is one of the most exciting and rapidly growing areas of tech. But for job seekers it can also feel like a maze of tools, frameworks and platforms. One job advert wants TensorFlow and Keras. Another mentions PyTorch, scikit-learn and Spark. A third lists Mlflow, Docker, Kubernetes and more. With so many names out there, it’s easy to fall into the trap of thinking you must learn everything just to be competitive. Here’s the honest truth most machine learning hiring managers won’t say out loud: 👉 They don’t hire you because you know every tool. They hire you because you can solve real problems with the tools you know. Tools are important — no doubt — but context, judgement and outcomes matter far more. So how many machine learning tools do you actually need to know to get a job? For most job seekers, the real number is far smaller than you think — and more logically grouped. This guide breaks down exactly what employers expect, which tools are core, which are role-specific, and how to structure your learning for real career results.

What Hiring Managers Look for First in Machine Learning Job Applications (UK Guide)

Whether you’re applying for machine learning engineer, applied scientist, research scientist, ML Ops or data scientist roles, hiring managers scan applications quickly — often making decisions before they’ve read beyond the top third of your CV. In the competitive UK market, it’s not enough to list skills. You must send clear signals of relevance, delivery, impact, reasoning and readiness for production — and do it within the first few lines of your CV or portfolio. This guide walks you through exactly what hiring managers look for first in machine learning applications, how they evaluate CVs and portfolios, and what you can do to improve your chances of getting shortlisted at every stage — from your CV and LinkedIn profile to your cover letter and project portfolio.

MLOps Jobs in the UK: The Complete Career Guide for Machine Learning Professionals

Machine learning has moved from experimentation to production at scale. As a result, MLOps jobs have become some of the most in-demand and best-paid roles in the UK tech market. For job seekers with experience in machine learning, data science, software engineering or cloud infrastructure, MLOps represents a powerful career pivot or progression. This guide is designed to help you understand what MLOps roles involve, which skills employers are hiring for, how to transition into MLOps, salary expectations in the UK, and how to land your next role using specialist platforms like MachineLearningJobs.co.uk.