Event JSON
{
"id": "530566b57791343f0c31b97c89e852a9e676788a1f5e5e48995ae3a23fd25484",
"pubkey": "cd1e48c145a46c437637a19149049827a77286aeeef568566ef0399870bde650",
"created_at": 1685974826,
"kind": 1,
"tags": [
[
"p",
"1a1add9955a05ca54d7963a0babcc1d11923d64ecaffb57a5bee1035288b1368",
"wss://relay.mostr.pub"
],
[
"p",
"f70f687f680a5021100ff60d8909388eccb7afb5681ce1698f4d79bbf2a27bab",
"wss://relay.mostr.pub"
],
[
"p",
"b410b83e37b375bc8f3a6aca3dfd309d25da45909c2832cfd0a902540ac42864",
"wss://relay.mostr.pub"
],
[
"p",
"b25e18a046faff6d31453f5adbc0042688f15005f732e5f22399db9b1036a97c",
"wss://relay.mostr.pub"
],
[
"mostr",
"https://cybervillains.com/users/alex/statuses/110492046256853568"
]
],
"content": "In the course of conducting a large investigation into online child exploitation, our team at the Stanford Internet Observatory discovered serious failings with the child protection systems at Twitter.\n\nThis discovery, that Twitter’s systems for stopping the posting of known child sexual abuse material (CSAM) had failed, occurred in the context of a larger project that we will release later this week alongside the Wall Street Journal.\n\ncc nostr:npub1rgddmx245pw22ntevwst40xp6yvj84jwetlm27jmacgr22ytzd5qzxfyzy nostr:npub17u8kslmgpfgzzyq07cxcjzfc3mxt0ta4dqwwz6v0f4umhu4z0w4svyuhnv \n\nhttps://www.wsj.com/articles/twitter-missed-dozens-of-known-images-of-child-sexual-abuse-material-researchers-say-58d44f7b",
"sig": "3436e5603d38673a82ff121bc4f748e64d5318da69f7d282de6a913fa572e4f73a706cc030064aa5f3c89fd90920e383af076b8dbcc8cee4a3e4cc20f82ff86f"
}