We're looking for funding - see Investors page.

17× Greener: How One AI Model Could've Saved 15,225 kg CO₂e

Training AI models in cleaner regions can cut emissions by over 90%. Stable Diffusion could have saved 15,225 kg CO₂e using smarter compute and saved $150k.

The Problem – Dirty AI

Most AI training runs on dirty, fossil fueled energy and it’s expensive.

Stable Diffusion; a popular text-to-image model recently trained their v.1.4 model on:

  • Hardware: A100 PCIe 40GB
  • Usage: 150,000 hours
  • Cloud Provider: AWS
  • Region: US-East (N. Virginia)
  • Grid Intensity: 430gCO₂e/kWh
  • Total Emissions: ~16,125 kg CO₂e
  • Total Cost: $600,000 (Yikes!)

Confirmed by Emad Mostaque, former CEO of Stability AI, the next version (0.5) of Stable Diffusion cost an eye-watering $600,000 to train. We can reasonably estimate that version 0.4 had a similar cost profile, likely running at around $4/hour; highlighting just how expensive training large AI models can be.

The Solution – Cleaner compute

Imagine running the same job... in a region where the grid intensity is 17× cleaner, greener and low-carbon...

If we ran this job in Paris (24gCO2/kWh). It would have saved 15,225 kg CO₂e.

Same job. Same cloud. Smarter infrastructure.

This is the power of carbon-aware compute.

Conclusion

It’s hard to find GPUs in clean, low-carbon regions, but it’s possible, and the benefits aren’t just environmental.

Running this model in a region like Paris would have produced just 900 kg of CO₂e, compared to 16,125 kg in us-east-1. That’s over a 90% reduction in emissions, equivalent to 15 flights from New York to London.

Shopping around for different providers offering similar boxes is well worth it, using potential cost saving of over $150,000.

CarbonRunner doesn’t just bring carbon-aware infrastructure to CI/CD workflows on GitHub Actions, we also offer a powerful API that helps you train AI/ML models in the lowest-carbon regions across multiple clouds.

We arbitrage across multiple providers in real time to surface the best-value, lowest-emissions boxes, so you don’t just save carbon, you save money too.




Want to Shift AI Workloads to Cleaner Grids?

Stable Diffusion burned 15,225 kg CO₂e on US-East. The same training on a low-carbon grid would've cut emissions by 94%. Clean compute isn't slower — just smarter and cheaper!

17× Greener: How One AI Model Could’ve Saved 15,225 kg CO₂e | CarbonRunner