When is a research paper not a research paper? When a big tech company uses a preprint server as a means to dodge peer review -- in this case, of their wild speculations on the 'dangerous capabilities' of large language models. Ali Alkhatib joins Emily to explain why a recent Google DeepMind document about the hunt for evidence that LLMs might intentionally deceive us was bad science, and yet is still influencing the public conversation about AI.
Ali Alkhatib is a computer scientist and former director of the University of San Francisco’s Center for Applied Data Ethics. His research focuses on human-computer interaction, and why our technological problems are really social – and why we should apply social science lenses to data work, algorithmic justice, and even the errors and reality distortions inherent in AI models.
References:
Google DeepMind paper-like
... show moreWhen is a research paper not a research paper? When a big tech company uses a preprint server as a means to dodge peer review -- in this case, of their wild speculations on the 'dangerous capabilities' of large language models. Ali Alkhatib joins Emily to explain why a recent Google DeepMind document about the hunt for evidence that LLMs might intentionally deceive us was bad science, and yet is still influencing the public conversation about AI.
Ali Alkhatib is a computer scientist and former director of the University of San Francisco’s Center for Applied Data Ethics. His research focuses on human-computer interaction, and why our technological problems are really social – and why we should apply social science lenses to data work, algorithmic justice, and even the errors and reality distortions inherent in AI models.
References:
Google DeepMind paper-like object: Evaluating Frontier Models for Dangerous Capabilities
Fresh AI Hell:
Hacker tool extracts all the data collected by Windows' 'Recall' AI
In NYC, ShotSpotter calls are 87 percent false alarms
"AI" system to make callers sound less angry to call center workers
Anthropic's Claude Sonnet 3.5 evaluated for "graduate level reasoning"
OpenAI's Mira Murati says "AI" will have 'PhD-level' intelligence
OpenAI's Mira Murati also says AI will take some creative jobs, maybe they shouldn't have been there to start out with
*****
You can check out future streams at http://twitch.tv/dair_institute.
Follow Emily at https://twitter.com/EmilyMBender and https://dair-community.social/@EmilyMBender
Follow Alex at https://twitter.com/alexhanna and https://dair-community.social/@alex
Music: Toby Menon.
Production: Christie Taylor.
Graphic Design: Naomi Pleasure-Park.
Introducing Claude 3.5 Sonnet—our most intelligent model yet. This is the first release in our 3.5 model family. Claude 3.5 Sonnet now outperforms competitor… | 248 comments on LinkedIn
Anthropic (www.linkedin.com)