ARTFEED — Contemporary Art Intelligence

Frictive Policy Optimization: A New Framework for LLM Alignment

ai-technology · 2026-04-30

A new framework called Frictive Policy Optimization (FPO) has been introduced by researchers to train language models in managing both epistemic and normative risks through specific control actions such as clarification, verification, challenge, redirection, and refusal. Unlike traditional alignment techniques that focus on optimizing superficial preferences or task utility, FPO approaches alignment as a risk-sensitive epistemic control challenge, prioritizing interventions based on their anticipated impact on downstream epistemic quality rather than immediate rewards. This framework presents a classification of frictive interventions, a structured friction functional to address alignment failure modes, and a cohesive set of FPO methods, including reward shaping and preference pairing. The study can be found on arXiv with the identifier 2604.25136.

Key facts

  • FPO regulates what, when, and how to intervene in LLM outputs.
  • Interventions include clarification, verification, challenge, redirection, and refusal.
  • Alignment is formalized as a risk-sensitive epistemic control problem.
  • Interventions are selected based on expected effect on downstream epistemic quality.
  • A taxonomy of frictive interventions is introduced.
  • A structured friction functional operationalizes alignment failure modes.
  • FPO methods include reward shaping and preference pairing.
  • The paper is on arXiv with ID 2604.25136.

Entities

Institutions

  • arXiv

Sources