Get the latest tech news

Diffusion training from scratch on a micro-budget


Official repository for our work on micro-budget training of large-scale diffusion models. - SonyResearch/micro_diffusion

This repository provides a minimalistic implementation of our approach to training large-scale diffusion models from scratch on an extremely low budget. In particular, using only 37M publicly available real and synthetic images, we train a 1.16 billion parameter sparse transformer with a total cost of only $1,890 and achieve an FID of 12.7 in zero-shot generation on the COCO dataset. Using the patch-mixer significantly reduces performance degradation with masking while providing a large reduction in training time.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Budget

Budget

Photo of Scratch

Scratch

Photo of Dollar

Dollar

Related news:

News photo

India's Nifty Bulls See Stock Market Struggle Ahead of Budget

News photo

Nuclear startup Deep Fission plans to bury micro-reactors to power data centers

News photo

IBM and GlobalFoundries Settle Multibillion-Dollar Trade Secret and Contract Lawsuits