Event JSON
{
"id": "3c8dae013c9958f0a0dc8990dae28cfc61adedf18b7a06cc5c0452837a9f2be5",
"pubkey": "98978ab8ee4b615be48667520b4a4be9987f644cb541116993a4dd0e35d3db61",
"created_at": 1719953989,
"kind": 1,
"tags": [
[
"proxy",
"https://mstdn.social/@paulknightly/112718904680463556",
"web"
],
[
"proxy",
"https://mstdn.social/users/paulknightly/statuses/112718904680463556",
"activitypub"
],
[
"L",
"pink.momostr"
],
[
"l",
"pink.momostr.activitypub:https://mstdn.social/users/paulknightly/statuses/112718904680463556",
"pink.momostr"
],
[
"expiration",
"1722547612"
]
],
"content": "I had an interaction with a TSA officer traveling home over the weekend regarding facial recognition tech (that's a thread for another day), but it zeroed in on an important point that gets glossed over as far as why people like myself are concerned with biometric tech and AI models: \n\nThe lack of consent when it's forced onto us, and the apparent inability to opt out of becoming a living dataset in any way, shape or form.\n\nCase and point:\n\nhttps://arstechnica.com/tech-policy/2024/07/ai-trains-on-kids-photos-even-when-parents-use-strict-privacy-settings/",
"sig": "6b192ce0085a6f24140d18060d58b8e6f9ead5ffd104fd62dfb9621310b1b5b3ecc8baacc731d60585849436bd6287636b202db95b0bcd38f00da7dece8753a7"
}