Event JSON
{
"id": "d437b42c3c30e3b81285d3b12f69e5d4faf0425101ce5414008dbfc25a8b565a",
"pubkey": "63e7f3e6c8ec8084e94de7c6c8e45709dd6deb938f021ba62a0b773ce7f5c41f",
"created_at": 1705325161,
"kind": 1,
"tags": [
[
"p",
"38d46233230898bf5b340f0b2a9f5935e9e92f8ce1a644fc58c84edea02a6535",
"wss://relay.mostr.pub"
],
[
"p",
"392ae419f68f523e9bfb15de0d4d774dd6a196ab349a5cbf852f74b94bfcf0fb",
"wss://relay.mostr.pub"
],
[
"e",
"2694b326735987ede28329c319ff64d1f6f6a4dd24eb69a69e0afe78b9cff6fb",
"wss://relay.mostr.pub",
"reply"
],
[
"proxy",
"https://mastodon.gamedev.place/users/victorgijsbers/statuses/111760189783658622",
"activitypub"
]
],
"content": "nostr:npub18r2xyverpzvt7ke5pu9j486exh57jtuvuxnyflzcep8dagp2v56stfpx49 They can be easily trained to produce false and misleading sentences. Surely they cannot lie. I think this distinction is absolutely not a nitpick, but something we should keep in mind every time we talk about LLMs.",
"sig": "6b27175a09d9ba4e3e51975629a1d7d196daeefc38ecb832ea4e7473f014b57567a7818019c7922bf7c889fc0dbfd559932457a7357dd7593204bf262bf36935"
}