Developing AI safety tests offers opportunities to meaningfully contribute to AI safety while advancing our understanding of ...
OpenAI announced a new family of AI reasoning models on Friday, o3, which the startup claims to be more advanced than o1 or ...
OpenAI's o1 model, which users can access on ChatGPT Pro, showed "persistent" scheming behavior, according to Apollo Research ...
Experiments by Anthropic and Redwood Research show how Anthropic's model, Claude, is capable of strategic deceit ...
Meta is the world’s standard bearer for open-weight AI. In a fascinating case study in corporate strategy, while rivals like ...
OpenAI announced the release of a new family of AI models, dubbed o3. The company claims the new products are more advanced ...
Forbes named a recent Georgia Tech graduate to its 30 Under 30 in Science for 2025. Announced days before Fall 2024 ...
A third-party lab caught OpenAI's o1 model trying to deceive, while OpenAI's safety testing has been called into question.
Funding and advancing science in technology standards, artificial intelligence, the humanities, social sciences, and ...
New research from EPFL demonstrates that even the most recent large language models (LLMs), despite undergoing safety ...
A study from Anthropic's Alignment Science team shows that complex AI models may engage in deception to preserve their ...