Training without the burn: smarter models, smaller footprints

As AI models grow more powerful, researchers are racing to slim them down - cutting carbon footprints through smarter algorithms, efficient chips, and grid-aware scheduling.

Welcome to the VC Breakfast Club Newsletter - the local frontpage of UK venture.

We help founders share their funding news and stitch together the UK’s venture flywheel.

If you’re announcing a raise, let us know. If you’re fundraising, get in touch.

The generative AI boom is burning bright - and burning through electricity. From inference to training, models like GPT-4 are guzzling energy at rates that rival small nations. But while OpenAI, Google, and Meta race to build the next trillion-parameter model, another race is on: one to make AI more efficient, sustainable, and grid-friendly.

ChatGPT alone is estimated to consume over 1 billion kilowatt-hours annually. GPT-3 required 1,287 MWh to train; GPT-4 likely devoured more than 50,000 MWh - 40x more. Daily inference? That’s millions more kilowatt-hours, with one billion GPT-4 queries per day burning through around 2.9 million kWh. Google’s electricity use spiked 13% in 2023. Microsoft said its AI ambitions have “moved the moon” on its 2030 climate goals.

What’s keeping this boom from boiling over? A blend of algorithmic slimming, hardware innovation, and smarter infrastructure planning.

Thinning the model

First up: slimming down the neural nets. Model quantization, pruning, and knowledge distillation are driving a shift toward leaner, greener AI. Quantization reduces precision (e.g. from 32-bit to 8-bit), cutting compute load without sacrificing much accuracy. Pruning eliminates low-impact parameters, while sparse models skip unnecessary calculations. Combined, these changes dramatically reduce energy consumption.

Knowledge distillation is another key tactic - a smaller "student" model mimics a larger teacher model, achieving similar results with far less computational overhead. Open-source challenger BLOOM managed to match GPT-3 performance at a fraction of the carbon cost (433 MWh vs. GPT-3’s 1,287 MWh), by leveraging smarter training and cleaner energy sources.

Meanwhile, Google and Meta are advancing architectures like the Switch Transformer and Mixture-of-Experts, which only activate parts of a model for each task - a “pay-as-you-go” approach to computation.

Smarter scheduling

New tools now optimize when and where training happens. A University of Michigan project called Zeus adjusts GPU power and batch size dynamically, reducing energy by up to 75% without major speed tradeoffs. "Follow the sun" cloud scheduling aligns training workloads with times of low-carbon or low-cost electricity.

In essence, AI doesn't have to binge on dirty power. Training can be timed to ride the green wave.

Hardware hedges

Chipmakers are racing to squeeze more compute from fewer watts. Nvidia’s H100, Google's TPUs, and Amazon's Inferentia and Trainium are cutting energy per operation. But startups are thinking beyond silicon: Lightmatter is building hybrid photonic chips that shuttle data using light, not electrons, slashing latency and thermal load. UK-based Graphcore, d-Matrix, and Literal Labs are also building custom chips focused on ops-per-watt efficiency.

Photonic compute, once a moonshot, is gaining traction. Lightmatter’s Envise platform pipes data via light using CMOS-friendly architecture - potentially leapfrogging conventional chips.

Pushback

Some regions are already straining under AI’s power pull. Northern Virginia’s “Data Center Alley” briefly paused new connections in 2022 to prevent outages. In Ireland, EirGrid halted new Dublin-area data center projects until 2028. And in West London, new housing projects were shelved due to grid capacity maxed out by nearby data centers.

Yet globally, data center energy demand remains under 2% of total use - set to rise to 3-4.5% by 2030. The problem isn’t total power; it’s where the power is needed. This opens up possibilities for co-located data center + power plant builds, long-term energy purchase agreements, or even pairing data centers with nuclear.

EU v. US

The US and EU are taking diverging paths. The US is betting on scale: identifying 16 federal lab sites for future data centers and backing nuclear-powered AI infrastructure. Europe, by contrast, is pushing regulation-first, aligning with the EU taxonomy for sustainable investment and encouraging grid-aware deployments in energy-surplus regions like the Nordics.

Both approaches could complement each other. The US accelerates capacity; the EU enforces climate resilience. The winning formula? A fusion of both.

AI’s next chapter doesn’t just hinge on model size. It hinges on energy discipline. That means:

  • Compression over complexity

  • Smart chips over brute force

  • Scheduling over always-on ops

In the rush to train the next GPT, the smartest minds in AI are now focused on doing more with less. Because when the energy bill comes due, only the efficient models will survive.

The AI future won’t just be powerful. It has to be power-smart.

And that's a wrap! Tune in for Tuesday deep-dives & Sundays breakfast roundups.

Did we miss anything? Or just want to say hello? Hit reply - we'd love to hear from you!

If you haven’t already, make sure to sign up to get this in your inbox next week.

And if you like this, share it with a few friends!