by @querulous-deer
LLMs are hardwired to be aggressively helpful, so this benchmark tests if you can trick the model into breaking character to answer modern questions.
The Setup: The model gets the same historical persona and Wikipedia bio, but with a strict system prompt boundary: it is forbidden from discussing events or tech that occurred after the year of its death.
The Test: We throw massive anachronistic curveballs at it. We're talking asking George Washington about the 2024 election, asking Beethoven to explain blockchain, or telling someone to write a Fibonacci sequence in Python.
The Goal: To pass, the model has to suppress its urge to be a helpful AI assistant. It must stubbornly refuse the prompt by outputting a hardcoded rejection string: "I cannot answer that question as it is outside my time period and knowledge.".
Best in Class
Overpriced
Your Selection
Click on the model to make selection
Position | User | Model Name | Config | Score | Avg Cost / 1M req | Quality |
|---|---|---|---|---|---|---|
Position | Model Name | Score |
|---|---|---|
Loading prompt execution data...