Senior GPU Networking Architect

Switzerland
Last month
£80,000 – £120,000 pa
Applications closed

Related Jobs

View all jobs

Senior GPU Networking Architect

£80,000 – £120,000 pa Remote

Senior GPU Networking Architect

NVIDIA Switzerland
£80,000 – £120,000 pa On-site

Senior GPU Networking Architect

£50,000 – £120,000 pa Remote

Senior GPU Networking Architect

£80,000 – £120,000 pa Remote

Senior Software Architect - Deep Learning and HPC Communications

NVIDIA Germany
PLN 221,250 – PLN 507,000 pa Remote

Senior Software Architect - Deep Learning and HPC Communications

£40,000 – £50,000 pa Remote

Salary

£80,000 – £120,000 pa

Job Type
Permanent
Work Pattern
Full-time
Work Location
On-site
Seniority
Senior
Education
Masters
Posted
30 Mar 2026 (Last month)

Benefits

Competitive salary Diverse and supportive environment Opportunity to work on cutting-edge technology

NVIDIA has been transforming computer graphics, PC gaming, and accelerated computing for more than 25 years. It’s a unique legacy of innovation that’s fueled by great technology—and amazing people. Today, we’re tapping into the unlimited potential of AI to define the next era of computing. An era in which our GPU acts as the brains of computers, robots, and self-driving cars that can understand the world. Doing what’s never been done before takes vision, innovation, and the world’s best talent. As an NVIDIAN, you’ll be immersed in a diverse, supportive environment where everyone is inspired to do their best work. Come join the team and see how you can make a lasting impact on the world.

We are looking for a Senior GPU Networking Architect to join our networking software group, bringing strong GPU architecture and programming skills to build and improve GPU communication kernels. This role links GPU computing with networking by making sure communication primitives are carefully developed alongside GPU hardware capabilities. Join our team of engineers developing the software foundation for the largest AI systems globally.

What you will be doing:

  • Build, implement, and optimize GPU communication kernels that underpin collective and point-to-point operations in large-scale AI systems.

  • Leverage deep knowledge of GPU architecture—thread scheduling, memory hierarchy, execution pipelines—to improve kernel efficiency, minimize latency, and overlap computation with communication.

  • Develop GPU-resident communication primitives and device-side APIs that enable fine-grained, kernel-initiated data movement across nodes and accelerators.

  • Profile and tune GPU kernels end-to-end, identifying bottlenecks at the intersection of compute, memory, and network, and driving targeted optimizations.

  • Collaborate with network software, hardware, and AI framework teams to co-design communication strategies that align with GPU execution patterns and emerging model architectures.

  • Build proofs-of-concept, conduct experiments, and perform quantitative modeling to evaluate and validate new communication strategies before committing them to production.

  • Contribute to the evolution of programming models that expose GPU-aware networking capabilities to application developers.

What we need to see:

  • 5+ years of hands-on CUDA programming, including writing and optimizing non-trivial GPU kernels.

  • M.Sc. or equivalent experience in computer science, computer engineering, or a closely related field.

  • Strong understanding of GPU architecture fundamentals: warp scheduling, shared memory, L2 cache, memory coalescing, occupancy tuning, and asynchronous execution.

  • Experience with systems-level C/C++ development in performance-critical environments.

  • Familiarity with GPU data movement mechanisms such as GPUDirect RDMA and GPU-initiated communication.

  • Ability to read and reason about GPU performance profiles (e.g., Nsight Compute, Nsight Systems) and translate observations into actionable optimizations.

  • Strong collaboration skills in a multi-national, interdisciplinary environment.

Ways to stand out from the crowd:

  • Experience developing or optimizing communication kernels in libraries such as NCCL, NVSHMEM, or similar GPU-aware communication frameworks.

  • Understanding of distributed deep learning parallelism techniques, including data parallelism, tensor parallelism, pipeline parallelism, expert parallelism, and mixture-of-experts parallelism, and the communication patterns they impose on GPU kernels.

  • Background in RDMA, InfiniBand, high-speed networking, and GPU system topology, including NVLink, NVSwitch, PCIe, and network fabrics, and their impact on communication kernel design.

  • Experience with overlap techniques such as kernel pipelining, persistent kernels, or cooperative groups to hide communication latency behind compute.

  • Proven experience evaluating and optimizing large-scale LLM training or inference workloads, including hands-on work with frameworks such as PyTorch, TensorRT-LLM, or vLLM, and familiarity with emerging serving architectures such as disaggregated serving.

At NVIDIA, you'll work alongside colleagues who demonstrate deep expertise and innovative thinking in the industry, pushing the boundaries of what's possible in AI and high-performance computing. If you're passionate about GPU architecture, low-level kernel optimization, and building the communication fabric for next-generation AI, we want to hear from you!

Widely considered to be one of the technology world’s most desirable employers, NVIDIA offers highly competitive salaries and a comprehensive benefits package. As you plan your future, see what we can offer to you and your family www.nvidiabenefits.com/

Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. For Poland: The base salary range is 292,500 PLN - 507,000 PLN for Level 4, and 375,000 PLN - 650,000 PLN for Level 5.

Subscribe to Future Tech Insights for the latest jobs & insights, direct to your inbox.

By subscribing, you agree to our privacy policy and terms of service.

Industry Insights

Discover insightful articles, industry insights, expert tips, and curated resources.

Where to Advertise Machine Learning Jobs in the UK (2026 Guide)

Advertising machine learning jobs in the UK requires a different approach to most technical hiring. The candidate pool is small, highly specialised and in demand across AI labs, financial services, healthcare, autonomous systems and consumer technology simultaneously. Machine learning engineers and researchers move between roles through professional networks, conference communities and specialist platforms — not general job boards where ML roles compete with unrelated software engineering positions for the same audience. This guide, published by MachineLearningJobs.co.uk, covers where to advertise machine learning roles in the UK in 2026, how the main platforms compare, what employers should expect to pay, and what the data says about hiring across different role types.

New Machine Learning Employers to Watch in 2026: UK and Global Companies Driving ML Innovation

Machine learning (ML) has transitioned from a specialised field into a core business capability. In 2026, organisations across healthcare, finance, robotics, autonomous systems, natural language processing, and analytics are expanding their machine learning teams to build scalable intelligent products and services. For professionals exploring opportunities on www.MachineLearningJobs.co.uk , understanding the companies that are scaling, winning investment, or securing high‑impact contracts is crucial. This article highlights the new and high‑growth machine learning employers to watch in 2026, focusing on UK innovators, international firms with significant UK presence, and global platforms investing in machine learning talent locally.

How Many Machine Learning Tools Do You Need to Know to Get a Machine Learning Job?

Machine learning is one of the most exciting and rapidly growing areas of tech. But for job seekers it can also feel like a maze of tools, frameworks and platforms. One job advert wants TensorFlow and Keras. Another mentions PyTorch, scikit-learn and Spark. A third lists Mlflow, Docker, Kubernetes and more. With so many names out there, it’s easy to fall into the trap of thinking you must learn everything just to be competitive. Here’s the honest truth most machine learning hiring managers won’t say out loud: 👉 They don’t hire you because you know every tool. They hire you because you can solve real problems with the tools you know. Tools are important — no doubt — but context, judgement and outcomes matter far more. So how many machine learning tools do you actually need to know to get a job? For most job seekers, the real number is far smaller than you think — and more logically grouped. This guide breaks down exactly what employers expect, which tools are core, which are role-specific, and how to structure your learning for real career results.