Large Language Model Reasoning Failures
https://www.arxiv.org/abs/2602.06176
(Transactions on Machine Learning Research)
"our survey provides a structured perspective on systemic weaknesses
in LLM reasoning, offering valuable insights and guiding future
research towards building stronger, more reliable, and robust
reasoning capabilities. We additionally release a comprehensive
collection of research works on LLM reasoning failures, as a GitHub
repository at
https://github.com/Peiyang-Song/Awesome-LLM-Reasoning-Failures,
to provide an easy entry point to this area."
JM