Skip to main content

What is LAFT?

LAFT (Language-Assisted Feature Transformation) is a novel approach to anomaly detection that leverages language guidance to transform visual features in semantically meaningful ways. Presented at ICLR 2025, LAFT enables more accurate and interpretable anomaly detection by allowing users to specify which semantic concepts should guide the detection process.
LAFT was accepted at the Thirteenth International Conference on Learning Representations (ICLR 2025).Read the paper | OpenReview

The Core Problem

Traditional anomaly detection methods struggle with semantic anomalies where:
  • Visual features alone are insufficient - Anomalies may be semantically meaningful but visually subtle
  • Domain knowledge is hard to incorporate - Existing methods don’t allow users to specify what makes something anomalous
  • Spurious correlations interfere - Models may focus on irrelevant features instead of the true anomaly indicators
For example, in a medical imaging scenario, you might want to detect tumors (guide by pathology) while ignoring patient demographics. In quality control, you might want to focus on defects while ignoring variations in lighting or background.

How LAFT Works

LAFT transforms visual features using language-defined concept subspaces:
  1. Load a vision-language model (CLIP) that understands both images and text
  2. Define concept prompts describing normal and anomalous states in natural language
  3. Construct a concept subspace by computing pairwise differences between prompt embeddings
  4. Transform image features by projecting them onto (guide) or away from (ignore) the concept subspace
  5. Detect anomalies using k-NN or other distance-based methods on the transformed features

Key Features

Language Guidance

Use natural language to specify which semantic concepts should guide anomaly detection

Dual Transformation

Project features onto (guide) or orthogonal to (ignore) concept subspaces

Zero-Shot Detection

Detect anomalies without task-specific training, using only CLIP embeddings

Interpretable Results

Understand what the model focuses on through explicit language concepts

Use Cases

Semantic Anomaly Detection

  • ColorMNIST: Detect digits 5-9 as anomalies while ignoring color variations
  • Waterbirds: Identify bird species anomalies independent of background
  • CelebA: Detect facial attribute anomalies (e.g., blonde hair, glasses)

Industrial Anomaly Detection

  • MVTec AD: Detect manufacturing defects in industrial products
  • VisA: Identify visual anomalies in complex industrial scenarios

Custom Domains

LAFT’s flexibility allows you to:
  • Define custom prompt templates for your domain
  • Specify multiple concept subspaces
  • Combine guided and ignored transformations
  • Use auxiliary prompts to expand the semantic space

Research Background

Authors: EungGu Yun, Heonjin Ha, Yeongwoo Nam, Bryan Dongik Lee Citation:
@inproceedings{yun2025laft,
  title={Language-Assisted Feature Transformation for Anomaly Detection},
  author={EungGu Yun and Heonjin Ha and Yeongwoo Nam and Bryan Dongik Lee},
  booktitle={The Thirteenth International Conference on Learning Representations},
  year={2025},
  url={https://openreview.net/forum?id=2p03KljxE9}
}

Resources

Paper

Read the full ICLR 2025 paper on arXiv

GitHub

Access the source code and examples

Installation

Get started with setup instructions

Quick Start

Run your first LAFT example

Next Steps

1

Install LAFT

Set up your environment and install dependenciesGo to Installation →
2

Try the Quick Start

Run a complete example to understand the workflowGo to Quick Start →
3

Explore the API

Learn about all available functions and datasetsGo to API Reference →

Build docs developers (and LLMs) love