Event JSON
{
"id": "7e58114f23242db0830aeacdbc125357e9b22cc97d0279c820cd1bf21358f34f",
"pubkey": "97f848adcc4c6276685fe48426de5614887c8a51ada0468cec71fba938272911",
"created_at": 1745459575,
"kind": 1,
"tags": [
[
"e",
"ada93e00c547bc6303b4ece74e3d4fadc25e5909edd8a109c868cf9af54f0e31",
"nostr-idb://cache-relay",
"root",
"0461fcbecc4c3374439932d6b8f11269ccdb7cc973ad7a50ae362db135a474dd"
],
[
"e",
"ada93e00c547bc6303b4ece74e3d4fadc25e5909edd8a109c868cf9af54f0e31",
"nostr-idb://cache-relay",
"reply",
"0461fcbecc4c3374439932d6b8f11269ccdb7cc973ad7a50ae362db135a474dd"
],
[
"p",
"0461fcbecc4c3374439932d6b8f11269ccdb7cc973ad7a50ae362db135a474dd"
]
],
"content": "Word vectors. Word2Vec - now I think more than 10 years old, but when you first start learning about large language models, and you encounter this, you're like \"oh, that is cool\"..... https://en.wikipedia.org/wiki/File:Word_vector_illustration.jpg. https://en.wikipedia.org/wiki/Word_embedding\n",
"sig": "b575a1d8bf2fb54a0922cec959dd067382a7fdeea37d104192d3182d8cd20ee62de928d615c7ebdb372f2e5538a05330a0369b60fab6f0cd6e145b030c25d466"
}