ARTFEED — Contemporary Art Intelligence

LLM Psychosis Framework Introduced for AI Behavioral Failures

ai-technology · 2026-04-30

A recent study published on arXiv (2604.25934) introduces the concept of "LLM Psychosis" as a theoretical model to understand severe cognitive failures in large language models, suggesting that the conventional term "hallucination" falls short. This model outlines five key characteristics: the breakdown of reality boundaries, the persistence of false beliefs, logical inconsistencies under impossible conditions, instability in self-representation, and excessive confidence in knowledge. The authors assert that these features represent a distinctly different type of failure. To implement this framework, they present the LLM Cognitive Integrity Scale (LCIS), a diagnostic tool with five dimensions, including Environmental Reality Interface (ERI) and Premise Arbitration. The study is classified as a cross-disciplinary theoretical analysis.

Key facts

  • Paper arXiv:2604.25934 proposes LLM Psychosis framework
  • Five hallmark features define the framework
  • Claims failure mode is distinct from ordinary factual error
  • Introduces LLM Cognitive Integrity Scale (LCIS)
  • LCIS includes Environmental Reality Interface (ERI) axis
  • LCIS includes Premise Arbitration axis
  • Published as a cross-disciplinary theoretical study
  • Argues hallucination terminology is inadequate

Entities

Institutions

  • arXiv

Sources