Scaling Compute

Backed by nearly £50m, this programme looks to redefine our current compute paradigm

Why this programme

The digital electronics industry that has transformed our lives in immeasurable ways is defined by the simple fact that, for 60+ years, we have benefited from exponentially more computing power, at a continually lower cost. 

This fact is no longer true. For the first time in history, increased performance requires increasing costs and this coincides with an explosion of demand for more compute power driven by AI. 

Our current mechanisms for training AI systems utilise a narrow set of algorithms and hardware building blocks, which require significant capital to develop and manufacture. The combination of this significance and scarcity has far-reaching economic, geopolitical, and societal implications.

What we’re shooting for

We see an opportunity to draw inspiration from natural processing systems, which innately process complex information more efficiently (on several orders or magnitude) than today’s largest AI systems.

Our goal: to increase + open up new vectors of progress in the field of computing by bringing the cost of AI hardware down by >1000x.

Meet the R&D Creators

We’re bringing expertise across three critical technology domains (AI systems design, mixed-signal CMOS circuits, and advanced networking) and a strong institutional mix (spanning academia, non-profit R&D organisations, startups and multinational companies), as we look to pull novel ideas to prototypes and into real-world applications. 

If successful, this programme will unlock a new technological lever for next-generation AI hardware, alleviate dependence on leading-edge chip manufacturing, and open up new avenues to scale AI hardware – an industry which is worth trillions of pounds. 

We’re also delighted that the programme’s ambition has drawn ambitious international organisations to either establish or expand their UK operations – a crucial step in building the UK’s capabilities in this field.

1 | Charting the Course

We’re funding two projects to develop software simulators to help the research community map the expected performance/power/cost for any future combination of algorithm, hardware, componentry, and system scale. The goal is to quantify the bottlenecks from different components in the stack, and enable agile adaptation to a fast-paced algorithms research community.
Heterogeneous Scale-out Platform Simulator
Project lead: James Myers, Imec
Team: Nathan Laubeuf, Debjyoti Bhattacharjee, Abubakr Nada, Arjun Singh, Jonas Svedas and Diksha Moolchandani

This project imagines a future where compute infrastructure systems are designed with a heterogeneous mix of logic, memory and interconnect technology options, to mitigate the different scaling challenges. James and the team at Imec are building a software framework to estimate system efficiency and cost in these systems.

Breaking Down the Compute Graph Step by Step: A Scalable and Modular Simulation
Project leads: Aaron Zhao, Imperial College London; Luo Mai, University of Edinburgh; Robert Mullins, University of Cambridge
Team: George Constantinides, Wayne Luk, Imperial College London; Michael O’Boyle, University of Edinburgh; Timothy Jones and Rika Antonova, University of Cambridge

There is prevalent demand for performance estimation of the systems used in training frontier AI models. This project, co-led by teams from the University of Edinburgh, Imperial College London, and the University of Cambridge, focuses on the development of a scalable and modular performance simulation framework for future systems.

“The most exciting aspect is the breadth of cross-domain expertise, ensuring a thorough examination of all aspects of current machine learning systems and opening avenues for potential groundbreaking discoveries in this field of research.”

2 | Advanced Networking and Interconnect

We know the movement of data has become as critical as raw computational power, so we’re funding two projects to interrogate system-level and advanced network design opportunities.
Scalable AI Systems
Project lead: Noa Zilberman, University of Oxford
Team: Amro Awad, Martin Booth, Nick McKeown, Dominic O’Brien, Patrick Salter

Noa and the team are aiming to introduce a new interconnect for scalable AI systems that solves the communication bottleneck. By rethinking communication at multiple levels and bringing together different disciplines, the project aims to revolutionise the design of AI systems.

Connectivity Technology for Sustainable AI Scaling
Project lead: Tony Chan Carusone, Alphawave Semi
Team: Behzad Dehlaghi, Project Technical Lead

This project will develop and demonstrate the next generation of connectivity technologies for sustainable AI scaling. The Alphawave Semi team are reaching for hardware solutions that will enable 10,000s of AI accelerator chips to be interconnected across distances up to 150m with low cost and power consumption without limiting performance.

“We’re excited about Scaling Compute’s potential to impact daily life by catalysing AI innovations that will advance healthcare, education, the environment, and beyond.”

3 | New Computational Primitives

While the computing industry continues to progress an established path for improved performance, a variety of alternative ideas have emerged which harness noise, statistics, or unique physics in existing mass-manufacturable circuits to perform specific computing primitives. We’re funding seven teams to develop new technologies with the potential to open up new vectors of progress for the field of computing, with a targeted relevance for modern AI algorithms.
Low-Precision Training using Backpropagation and Backpropagation-free Algorithms
Project lead: Peter McMahon, Cornell University

This project aims to develop training methods that can leverage low-precision hardware, and to develop neural-network architectures that are better-suited to being trained with low-precision hardware. If successful, Peter’s team at Cornell will explore how well hardware designed for accelerating neural-network inference can be applied to training.

CMOS Digital Thermodynamic Hardware Accelerator for Linear Algebra
Project lead: Phillip Stanley-Marbell, Signaloid
Team: Bilgesu Bilgin, Apostolos Vailakis

Building on insights from analogue thermodynamic computing and randomised linear algebra, Phillip and the Signaloid team are looking to develop digital hardware to accelerate approximate matrix inversion.

“There is currently a missed opportunity to exploit insights from analogue computing systems to enable more efficient analogue and digital hardware for speeding up linear algebraic kernels.”
Thermodynamic Matrix Inversion
Project lead: Patrick Coles, Normal Computing UK
Team: Gavin Crooks, Maxwell Aifer, Kaelan Donatella, Denis Melanson, Zachary Belateche, Samuel Duffield, Vincent Cheung

Patrick + the team at Normal Computing will build physics-based computing chips to invert matrices and explore applications in training large-scale AI models, targeting ~1000x energy savings over GPUs.

Better Analogue in Memory Matrix-vector Multiplication
Project lead: Walter Goodwin, Fractile
Team: Fractile Team

Walter and the team at Fractile aim to demonstrate that analogue in-memory compute can drive the world’s highest density and most efficient matrix-vector multiply operations. Their goal is to leverage the developed approaches in Fractile’s inference acceleration hardware to run frontier models orders of magnitude faster than current state of the art. Whether sufficient precision can be achieved for application, in theory, to large scale training systems remains an open question.

“This is a nascent technology and we think we can reach even more energy efficiency. There’s a whole frontier to be explored.”
Neuromorphic Matrix Multiplication
Project lead: Bipin Rajendran, King’s College London
Team: Osvaldo Simeone, Kai Xu

Combining the multiplicative advantages that arise from (i) event-driven, backprop-free learning algorithms, (ii) stochastic computing, and (iii) in-memory computing based on Si CMOS technology, Bipin and the team will design and demonstrate a neuromorphic framework to reduce the cost of developing AI models.

Energy-Efficient SRAM-Based Analogue Accelerator for Second-Order Optimisation
Project lead: Jack Kendall, Rain AI UK
Team: Maxence Ernoult, Greg Kollmer, Georgios Konstadinidis, and six UK team members to be hired

Jack’s team from Rain AI are looking to develop a novel accelerator architecture for performing fast vector-matrix inverse multiplication using digitally-programmable transistor arrays with feedback control.

“The Scaling Compute programme enables researchers across many disciplines, from machine learning to photonics, analogue computing, and materials science, to come together under one roof with the unified goal of creating new paradigms of energy-efficient AI systems.”
Shortening the Salesman’s Travel: Massive Parallelism for Combinational Optimisation Problems
Project leads: Marian Verhelst, Wim Dehaene, KU Leuven
Team: Toon Bettens, Sofie De Weer

Marian and the team at KU Leuven are targeting a new class of mixed-signal processors that are specifically conceived to solve combinatorial optimisation problems.

Training Analogue Electrical Networks with Equilibrium Propagation
Project lead: Benjamin Scellier, Rain AI UK
Team: Jack Kendall, Greg Kollmer + two teams member to be hired

Ben’s team from Rain AI aims to demonstrate, through simulations, the feasibility of training analogue hardware with Equilibrium Propagation of the size of modern deep learning architectures.

Jobs

Our Scaling Compute Creators are hiring for a number of technical roles – check out the live jobs below.
KU ​​Leuven – Post-doc in Ising Processor Architectures and Chip Design
close-small

Location: Leuven, Belgium

Start date: Immediate/ASAP

Deadline: 1st December 2024

​​Signaloid – Senior Engineer, Algorithms and High-Performance Computing
close-small

Location: UK/Remote

Start date: Immediate/ASAP

Deadline: Rolling

Signaloid – Engineering Project Manager
close-small

Location: UK/Remote

Start date: Immediate/ASAP

Deadline: Rolling

Signaloid – Hardware RTL Tools and Verification Engineer
close-small

Location: UK/Remote

Start date: Immediate/ASAP

Deadline: Rolling

Normal Computing UK – Mixed Signal Design Engineer
close-small

Location: London, UK/New York City, USA

Start date: Immediate/ASAP

Deadline: Rolling

Normal Computing UK – IC Design Engineer
close-small

Location: London, UK/New York City, USA

Start date: Immediate/ASAP

Deadline: Rolling

Normal Computing UK – Digital Design Engineer
close-small

Location: London, UK

Start date: Immediate/ASAP

Deadline: Rolling

Normal Computing UK – Staff Physical Design Engineer
close-small

Location: London, UK

Start date: Immediate/ASAP

Deadline: Rolling

University of Oxford – Senior Postdoctoral Researcher / Senior Research Software Engineer in AI Systems
close-small

Location: Oxford, UK

Start date: Immediate/ASAP

Deadline: Rolling

University of Oxford – Postdoctoral Researcher in Interconnect for AI Systems
close-small

Location: Oxford, UK

Start date: Immediate/ASAP

Deadline: Rolling

University of Oxford – Project Manager
close-small

Location: Oxford, UK

Start date: Immediate/ASAP

Deadline: Rolling

University of Oxford – PhD (DPhil) studentship in Engineer Science - High performance interconnect for AI
close-small

Location: Oxford, UK

Start date: Immediate/ASAP

Deadline: 3 December 2024

University of Oxford – PhD (DPhil) studentship in Engineer Science - Advanced memory architectures for AI systems
close-small

Location: Oxford, UK

Start date: Immediate/ASAP

Deadline: 3 December 2024

University of Oxford – PhD (DPhil) studentship in Engineer Science - High performance AI Systems
close-small

Location: Oxford, UK

Start date: Immediate/ASAP

Deadline: 3 December 2024

University of Oxford – PhD (DPhil) studentship in Engineer Science - Modelling optical systems
close-small

Location: Oxford, UK

Start date: Immediate/ASAP

Deadline: 3 December 2024

Alphawave Semi – Connectivity Systems Engineer
close-small

Location: Flexible

Start date: Immediate/ASAP

Deadline: Rolling

Alphawave Semi – Mixed Signal Engineer
close-small

Location: Flexible

Start date: Immediate/ASAP

Deadline: Rolling

Programme Director: Suraj Bramhavar

Suraj aims to redefine the way computers process information. He is directing funding into how we can build more efficient computers using principles ubiquitously found in nature.

Prior to ARIA, Suraj was co-founder and CTO of Sync Computing, a VC-backed startup optimising the use of modern cloud computing resources. The company was spun-out from his research at MIT Lincoln Laboratory. Prior to that, Suraj worked at Intel Corp, helping transition silicon photonics technology from an R&D effort into a now >$1BN business.

background image