Event JSON
{
"id": "466e170e0f7511e0065f6fac0c3c2f1a7a35381bc2915e8cb1ffa139049c41e6",
"pubkey": "32aafec388540f879d6bc4efa3d84340a367d2417746a82d81c7fd9638007c3b",
"created_at": 1711893744,
"kind": 1,
"tags": [
[
"p",
"05886e97f9a3368f3665b9ce6921ee72cbfd7ba82e357eeb9dc9e0cfb7dc83d7"
],
[
"t",
"ai"
],
[
"proxy",
"https://noc.social/users/AAKL/statuses/112190668427939674",
"activitypub"
],
[
"L",
"pink.momostr"
],
[
"l",
"pink.momostr.activitypub:https://noc.social/users/AAKL/statuses/112190668427939674",
"pink.momostr"
]
],
"content": "This isn't right. Is this the lazy way to deal with mental illness? #AI \n\n\"When emotional AI is deployed for mental health care or companionship, it risks creating a superficial semblance of empathy that lacks the depth and authenticity of human connections.\"\n\nIncreasingly sophisticated AI systems can perform empathy, but their use in mental health care raises ethical questions https://theconversation.com/increasingly-sophisticated-ai-systems-can-perform-empathy-but-their-use-in-mental-health-care-raises-ethical-questions-225498 [@TheConversationUK](https://newsie.social/@TheConversationUK)",
"sig": "4f1012555c6cb2662a2ff1e43da342fd8fe68f19df9d8864203ea6d0ef46190ea750e2c014fdc2c48dfe62d07b2baa8b4a214f422c7c91b217cd6a9386a0f1a2"
}