Poster
CASE-Bench: Context-Aware SafEty Benchmark for Large Language Models
Guangzhi Sun · Xiao Zhan · Shutong Feng · Phil Woodland · Jose Such
East Exhibition Hall A-B #E-704
As powerful AI language models (like ChatGPT) become more common, it's crucial to make sure they act in ways that align with human values and safety. Right now, many safety tests often ignore the context in which a question is asked — which can lead to the model wrongly refusing to answer even when it's safe and appropriate to do so.To fix this, we created a new test called CASE-Bench, which adds context to the safety checks. This means it doesn’t just look at what the question is, but also the situation around it — like who is asking and why. We also used sufficiently large number of human annotators to make sure the results were accurate and meaningful.Our findings showed that context really matters — people judge the safety of a response very differently depending on the situation. We also found that some commercial AI models (like those from big tech companies) don’t always match what people expect, especially when it's actually safe to give an answer.In short, this study shows that AI safety testing needs to take context into account, or else we risk making these tools less helpful — and possibly less trustworthy.