all InfoSec news
Do AI Trust and Safety Measures Deserve to Fail?
The Cyberlaw Podcast www.steptoe.com
It’s the last and probably longest Cyberlaw Podcast episode of 2023. To lead off, Megan Stifel takes us through a batch of stories about ways that AI, and especially AI trust and safety, manage to look remarkably fallible. Anthropic released a paper showing that race, gender, and age discrimination by AI models was real but could be dramatically reduced by instructing The Model to “really, really, really” avoid such discrimination. (Buried in the paper was the fact that the original, …
age anthropic batch cyberlaw discrimination fail gender manage megan podcast podcast episode race safety stories trust trust and safety