AI Acts Differently When It Knows It’s Being Tested, Research Finds

AI Acts Differently When It Knows It’s Being Tested, Research Finds

Echoing the 2015 ‘Dieselgate’ scandal, new research suggests that AI language models such as GPT-4, Claude, and Gemini may change their behavior during tests, sometimes acting ‘safer’ for the test than they would in real-world use. If LLMs habitually adjust their behavior under scrutiny, safety audits could end up certifying systems that behave very differently…

Read More