Researchers upend AI status quo by eliminating matrix multiplication in LLMs
Researchers upend AI status quo by eliminating matrix multiplication in LLMs
arstechnica.com Researchers upend AI status quo by eliminating matrix multiplication in LLMs
Running AI models without matrix math means far less power consumption—and fewer GPUs?
Running AI models without matrix math means far less power consumption—and fewer GPUs?
You're viewing a single thread.
All Comments
5 comments
Good
Edit: Oh shit nvm. It still requires dedicated HW (FPGA). This is no different than say, an NPU. But to be fair, they also said the researcher tested the model on traditional GPU too and reduce memory consumption.
5 1 ReplyOnly for maximum efficiency. LLMs already run tolerably well on normal CPUs and this technique would make it much more efficient there as well.
2 0 Reply