Jevon’s paradox says that efficiency will increase use. Microsoft is buying a nuclear power plant for AI shit. If they can train a trillion-parameter model for one-thousandth the cost… they will instead train a quadrillion-parameter model.
Or I guess if they’re smart they’ll train a trillion-parameter model longer. Or iterate like crazy, when training takes hours instead of months.
If bitnet takes off, that’s very good news for everyone.
The problem isn’t AI, it’s AI that’s so intensive to host that only corporations with big datacenters can do it.
The fuck is bitnet
https://www.microsoft.com/en-us/research/publication/bitnet-scaling-1-bit-transformers-for-large-language-models/ use 1 bit instead of 8 or 16, yay performance gainz
Jevon’s paradox says that efficiency will increase use. Microsoft is buying a nuclear power plant for AI shit. If they can train a trillion-parameter model for one-thousandth the cost… they will instead train a quadrillion-parameter model.
Or I guess if they’re smart they’ll train a trillion-parameter model longer. Or iterate like crazy, when training takes hours instead of months.
So will the return of the flag conclude the adventures of ressource usage in computers?
What star said, but what it also does is turn hard matrix multiplication into simple addition.
Basically, AI will be hilariously easy to run compared to now once ASICs start coming out, thought it will run on CPUs/GPUs just fine.