Event JSON
{
"id": "fb02c278ad078c1c32797c41508ccd6be879f9cd88d6c223a002b37de96a7bd3",
"pubkey": "a3af17104f91f7f9b5667b14717d1d434931195e4c6e075b7dc13d8ed71bc46f",
"created_at": 1719354828,
"kind": 1,
"tags": [
[
"imeta",
"url https://files.mastodon.social/media_attachments/files/112/679/638/004/208/393/original/50de5d8a22d2bfd8.jpg",
"m image/jpeg"
],
[
"proxy",
"https://mastodon.social/@arstechnica/112679638052673660",
"web"
],
[
"proxy",
"https://mastodon.social/users/arstechnica/statuses/112679638052673660",
"activitypub"
],
[
"L",
"pink.momostr"
],
[
"l",
"pink.momostr.activitypub:https://mastodon.social/users/arstechnica/statuses/112679638052673660",
"pink.momostr"
],
[
"expiration",
"1721946833"
]
],
"content": "Researchers upend AI status quo by eliminating matrix multiplication in LLMs\n\nRunning AI models without matrix math means far less power consumption—and fewer GPUs?\n\nhttps://arstechnica.com/information-technology/2024/06/researchers-upend-ai-status-quo-by-eliminating-matrix-multiplication-in-llms/?utm_brand=arstechnica\u0026utm_social-type=owned\u0026utm_source=mastodon\u0026utm_medium=social\nhttps://files.mastodon.social/media_attachments/files/112/679/638/004/208/393/original/50de5d8a22d2bfd8.jpg\n",
"sig": "39d988dcbeaeda9a18006b66a2f2a5fb56101ad0c15d93fbe79ca1267179953f190e8c9f511d4d4b0f44756bb0546dca9cf316417498c07ff70cd5cd5ed96b21"
}