Listen "Have we finally figured out how to make efficient AI?"
Episode Synopsis
A fantastic research paper published in this month's Nature Computational Science suggests a solution may be in our midst for the incredible inefficiency in generative AI.Large Language Models' (LLMs) transformer architecture requires the next token (generally part of a word) be predicted based on all the output tokens before it.Power demands for this process are huge. Shuffling data between memory and processors isn't an easy pipeline, and when you need it to work quickly, those energy demands quickly stack up.And in an AI arms race, where everyone wants bigger and better models, requiring increasingly powerful compute is required to stretch their limits, the dependence on energy to power, and cool, those processing units grows exponentially.But what if there was a different, and better, way, to make AI work? That's the driving force behind work of Nathan Leroux and his team proposing a totally different paradigm: analog in-memory computing.And that's exactly what we're discussing today.Zip yourself in that flame retardant suit: things are about to get hot in here...Ping me at [email protected] to get on the show or talk about AI in your world.
More episodes of the podcast AI Today
When's the right time to go all-in with AI?
18/10/2025
ELephantLM: the AI that never forgets!
13/10/2025
brAIn: thinking of the future?
01/10/2025
Does AI work?
26/09/2025
China's got AI in the bag
14/09/2025
I'm working on the Zeitgeist
03/09/2025
What happens when AI fires all the hirers?
21/08/2025
ZARZA We are Zarza, the prestigious firm behind major projects in information technology.