Pinned post

As inference splits into prefill and decode, Nvidia's Groq deal could enable a "Rubin SRAM" variant optimized for ultra-low latency agentic reasoning workloads (Gavin Baker/@gavinsbaker)

Gavin Baker / @gavinsbaker : As inference splits into prefill and decode, Nvidia's Groq deal could enable a “Rubin SRAM” variant optim...

24 November 2024

An overview of macro tech trends for 2025, focusing on generative AI, LLMs, scaling challenges with training ever bigger AI models, the capex surge, and more (Benedict Evans)

Benedict Evans:
An overview of macro tech trends for 2025, focusing on generative AI, LLMs, scaling challenges with training ever bigger AI models, the capex surge, and more  —  Presentations Every year, I produce a big presentation exploring macro and strategic trends in the tech industry.  For 2025, ‘AI eats the world’.

Posted from: this blog via Microsoft Power Automate.

Daily Deals