Lead Data Engineer

Core Consultants
London
4 months ago
Applications closed

Related Jobs

View all jobs

Lead Data Engineer

Lead Data Engineer

Lead Data Engineer

Lead Data Engineer - Snowflake, DBT, Airflow - London - £100k

Lead Data Engineer - Manchester - Hybrid - £75k - £80k

Lead data Engineer - Financial Markets - Day rate

Role Profile: Senior Data Engineering Lead – Internal Data Ingestion Platform



Role Overview



The Senior Data Engineering Lead will spearhead the design, implementation, and technical oversight of an advanced internal data ingestion platform, enabling scalable and efficient ingestion of diverse datasets. The role involves leading a team of data engineers, driving the development of configurable Spark ETL pipelines, and leveraging modern data architecture principles to support downstream analytics and visualisation requirements.



Core Responsibilities



  • Team Leadership and Development:
  • Recruit, train, and manage a high-performing data engineering team.
  • Foster a collaborative environment that promotes innovation and technical excellence.
  • Establish best practices for data engineering processes and mentor team members to develop their skills.
  • Platform Ownership and Technical Oversight:
  • Act as the product owner for the internal data ingestion platform, ensuring alignment with business needs and long-term scalability.
  • Provide technical oversight for the design and development of configurable Spark ETL pipelines that consume data from Kafka streams.
  • Leverage Airflow for pipeline orchestration, ensuring efficient scheduling and execution of workflows.
  • Manage the hosting and deployment of the platform on Azure Kubernetes Service (AKS).
  • Data Processing and Storage:
  • Oversee the integration of preprocessing tasks in Databricks, ensuring seamless transformation of raw data.
  • Implement a hybrid storage architecture, utilising Snowflake, on-premise SQL warehouses, and Delta Lake to meet varied storage and access needs.
  • Maintain the Medallion architecture (bronze/silver/gold layers) to organise and optimise data for downstream analytics.
  • Metadata and Discoverability:
  • Design and implement metadata capture processes to improve data discovery for internal consultants.
  • Ensure that metadata aids in understanding data lineage, quality, and context for effective utilisation.
  • Enablement for Data Visualisation and Analytics:
  • Ensure the data platform provides well-organised datasets to downstream consultants for use with tools like PowerBI.
  • Optimise pipelines to deliver high-quality, query-ready datasets that align with analytical and visualisation needs.
  • Technical Innovation and Optimisation:
  • Continuously evaluate and incorporate emerging technologies and practices to enhance the platform’s performance and scalability.
  • Optimise the Spark-based ETL processes for handling large-scale data efficiently and reliably.
  • Ensure pipelines are built with fault tolerance, monitoring, and logging to minimise downtime.



Key Skills and Qualifications



  • Technical Expertise:
  • Extensive experience with Apache Spark, Kafka, and Airflow for data ingestion and orchestration.
  • Proficiency in designing and deploying solutions on Azure Kubernetes Service (AKS).
  • Deep understanding of Databricks for data preprocessing and transformation.
  • Hands-on experience with Snowflake, SQL warehouses, and Delta Lake for data storage.
  • Knowledge of the Medallion architecture (bronze/silver/gold layers) and its practical applications.
  • Team Leadership and Collaboration:
  • Proven ability to lead and develop data engineering teams.
  • Strong collaboration skills, working with data analysts, consultants, and cross-functional teams.
  • Excellent communication skills to bridge technical and business requirements.
  • Problem Solving and Scalability:
  • Expertise in designing scalable systems to handle diverse datasets efficiently.
  • Strong analytical and troubleshooting skills to optimise pipelines and resolve integration challenges.
  • Metadata and Discoverability:
  • Familiarity with metadata management tools and techniques to enhance data discoverability and usability.
  • Understanding of data lineage and quality management.



Preferred Qualifications:



  • Experience in hybrid cloud and on-premise data environments.
  • Hands-on experience with PowerBI and similar data visualisation tools.
  • Familiarity with advanced orchestration and automation frameworks beyond Airflow.



Success Metrics:



  • Fully operational data ingestion platform with high scalability and reliability.
  • Efficient delivery of clean, structured datasets aligned with the Medallion architecture.
  • Improved metadata management processes that enhance internal consultant productivity.
  • A well-trained, high-performing data engineering team.



Outcomes:



This role will ensure the organisation’s data ingestion platform is a robust foundation for analytics, enabling efficient and effective use of data for business insights while supporting scalability and evolving business needs.

Get the latest insights and jobs direct. Sign up for our newsletter.

By subscribing you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

Portfolio Projects That Get You Hired for Machine Learning Jobs (With Real GitHub Examples)

In today’s data-driven landscape, the field of machine learning (ML) is one of the most sought-after career paths. From startups to multinational enterprises, organisations are on the lookout for professionals who can develop and deploy ML models that drive impactful decisions. Whether you’re an aspiring data scientist, a seasoned researcher, or a machine learning engineer, one element can truly make your CV shine: a compelling portfolio. While your CV and cover letter detail your educational background and professional experiences, a portfolio reveals your practical know-how. The code you share, the projects you build, and your problem-solving process all help prospective employers ascertain if you’re the right fit for their team. But what kinds of portfolio projects stand out, and how can you showcase them effectively? This article provides the answers. We’ll look at: Why a machine learning portfolio is critical for impressing recruiters. How to select appropriate ML projects for your target roles. Inspirational GitHub examples that exemplify strong project structure and presentation. Tangible project ideas you can start immediately, from predictive modelling to computer vision. Best practices for showcasing your work on GitHub, personal websites, and beyond. Finally, we’ll share how you can leverage these projects to unlock opportunities—plus a handy link to upload your CV on Machine Learning Jobs when you’re ready to apply. Get ready to build a portfolio that underscores your skill set and positions you for the ML role you’ve been dreaming of!

Machine Learning Job Interview Warm‑Up: 30 Real Coding & System‑Design Questions

Machine learning is fuelling innovation across every industry, from healthcare to retail to financial services. As organisations look to harness large datasets and predictive algorithms to gain competitive advantages, the demand for skilled ML professionals continues to soar. Whether you’re aiming for a machine learning engineer role or a research scientist position, strong interview performance can open doors to dynamic projects and fulfilling careers. However, machine learning interviews differ from standard software engineering ones. Beyond coding proficiency, you’ll be tested on algorithms, mathematics, data manipulation, and applied problem-solving skills. Employers also expect you to discuss how to deploy models in production and maintain them effectively—touching on MLOps or advanced system design for scaling model inferences. In this guide, we’ve compiled 30 real coding & system‑design questions you might face in a machine learning job interview. From linear regression to distributed training strategies, these questions aim to test your depth of knowledge and practical know‑how. And if you’re ready to find your next ML opportunity in the UK, head to www.machinelearningjobs.co.uk—a prime location for the latest machine learning vacancies. Let’s dive in and gear up for success in your forthcoming interviews.

Negotiating Your Machine Learning Job Offer: Equity, Bonuses & Perks Explained

How to Secure a Compensation Package That Matches Your Technical Mastery and Strategic Influence in the UK’s ML Landscape Machine learning (ML) has rapidly shifted from an emerging discipline to a mission-critical function in modern enterprises. From optimising e-commerce recommendations to powering autonomous vehicles and driving innovation in healthcare, ML experts hold the keys to transformative outcomes. As a mid‑senior professional in this field, you’re not only crafting sophisticated algorithms; you’re often guiding strategic decisions about data pipelines, model deployment, and product direction. With such a powerful impact on business results, companies across the UK are going beyond standard salary structures to attract top ML talent. Negotiating a compensation package that truly reflects your value means looking beyond the numbers on your monthly payslip. In addition to a competitive base salary, you could be securing equity, performance-based bonuses, and perks that support your ongoing research, development, and growth. However, many mid‑senior ML professionals leave these additional benefits on the table—either because they’re unsure how to negotiate them or they simply underestimate their long-term worth. This guide explores every critical aspect of negotiating a machine learning job offer. Whether you’re joining an AI-focused start-up or a major tech player expanding its ML capabilities, understanding equity structures, bonus schemes, and strategic perks will help you lock in a package that matches your technical expertise and strategic influence. Let’s dive in.