Event JSON
{
"id": "c38fc5d1c6509579b94766e52b6a1de7ee1b8852b4f2ca4e65b93027a44a5f03",
"pubkey": "bd1e4496c429785e57ece3442b71576b6c9cbbafdde26460d91ebef3acf0a276",
"created_at": 1707185809,
"kind": 1,
"tags": [
[
"t",
"AI"
],
[
"t",
"ai"
],
[
"t",
"News"
],
[
"t",
"news"
],
[
"r",
"https://www.theguardian.com/society/2024/feb/04/they-thought-they-were-doing-good-but-it-made-people-worse-why-mental-health-apps-are-under-scrutiny"
]
],
"content": "https://www.theguardian.com/society/2024/feb/04/they-thought-they-were-doing-good-but-it-made-people-worse-why-mental-health-apps-are-under-scrutiny\n\n‘They thought they were doing good but it made people worse’: why mental health apps are under scrutiny\n\n\"These fears are not unfounded. On Reddit, a user of the popular Replika chatbot shared a screenshot of a conversation in which the bot appeared to actively encourage his suicide attempt.\"\n\n#AI #News",
"sig": "edb420e3ace24058cd67a598f9cdaa6cca1d958e22a68048782e11a2dd7f563b99cc480467b6b10d1947e56c6df5e48608b008a26251a91750ccd51ca8abb6eb"
}