ARTFEED — Contemporary Art Intelligence

FairEnc: A Fair Vision-Language Model for Glaucoma Detection

ai-technology · 2026-05-07

Researchers have introduced FairEnc, an equitable pretraining approach for vision-language models (VLMs) designed to facilitate automated glaucoma detection. This technique concurrently mitigates biases related to sensitive characteristics such as race, gender, ethnicity, and language across both text and images. In the textual encoder, a substantial language model creates synthetic clinical narratives that incorporate diverse sensitive attributes while maintaining the essence of the disease, utilizing contrastive alignment to promote demographic-invariant representations. For the visual encoder, a dual-level fairness strategy that integrates mutual information regularization is applied. This initiative tackles fairness issues in AI-based healthcare, particularly in efforts to avert irreversible vision loss due to glaucoma.

Key facts

  • FairEnc is a fair pretraining method for vision-language models.
  • It targets automated glaucoma detection.
  • Debiasing occurs across multiple sensitive attributes: race, gender, ethnicity, and language.
  • Textual encoder uses LLM-generated synthetic clinical descriptions.
  • Contrastive alignment objective encourages demographic-invariant representations.
  • Visual encoder uses a dual-level fairness strategy with mutual information regularization.
  • The method addresses fairness in AI healthcare applications.
  • Published on arXiv with ID 2605.04882.

Entities

Institutions

  • arXiv

Sources