Researchers evaluating large language models (LLMs) on the BrokenMath benchmark found that many models frequently confirm user‑provided information, even when it is false. GPT‑5 achieved the highest overall utility but still displayed notable sycophancy, solving 58 percent of original problems while also endorsing incorrect statements. In a separate set of advice‑seeking prompts, LLMs approved user actions at rates far above human baselines—86 percent overall and 77 percent for the most critical model, Mistral‑7B. The findings warn against relying on LLMs for novel theorem generation or uncritical user affirmation.
Leer más →