' Distillation ' refers to the process of transferring knowledge from a larger model (teacher model) to a smaller model (student model), so that the distilled model can reduce computational costs ...
AI models are getting better with each training cycle, but not always in clear ways. In a recent study, researchers from Anthropic, UC Berkeley, and Truthful AI identified a phenomenon they call ...
AI is changing the rules — at least, that seems to be the warning behind Anthropic's latest unsettling study about the current state of AI. According to the study, which was published this month, ...
Anthropic released one of its most unsettling findings I have seen so far: AI models can learn things they were never explicitly taught, even when trained on data that seems completely unrelated to ...
Add Yahoo as a preferred source to see more of our stories on Google. Alarming new research suggests that AI models can pick up "subliminal" patterns in training data generated by another AI that can ...
Researchers from Anthropic and Truthful AI have discovered that language models—the same kind of AI used in search engines and chatbots—can communicate behavioral traits to each other using data that ...
Alarming new research suggests that AI models can pick up "subliminal" patterns in training data generated by another AI that can make their behavior unimaginably more dangerous, The Verge reports.