What is LAFT?
LAFT (Language-Assisted Feature Transformation) is a novel approach to anomaly detection that leverages language guidance to transform visual features in semantically meaningful ways. Presented at ICLR 2025, LAFT enables more accurate and interpretable anomaly detection by allowing users to specify which semantic concepts should guide the detection process.LAFT was accepted at the Thirteenth International Conference on Learning Representations (ICLR 2025).Read the paper | OpenReview
The Core Problem
Traditional anomaly detection methods struggle with semantic anomalies where:- Visual features alone are insufficient - Anomalies may be semantically meaningful but visually subtle
- Domain knowledge is hard to incorporate - Existing methods don’t allow users to specify what makes something anomalous
- Spurious correlations interfere - Models may focus on irrelevant features instead of the true anomaly indicators
How LAFT Works
LAFT transforms visual features using language-defined concept subspaces:- Load a vision-language model (CLIP) that understands both images and text
- Define concept prompts describing normal and anomalous states in natural language
- Construct a concept subspace by computing pairwise differences between prompt embeddings
- Transform image features by projecting them onto (guide) or away from (ignore) the concept subspace
- Detect anomalies using k-NN or other distance-based methods on the transformed features
Key Features
Language Guidance
Use natural language to specify which semantic concepts should guide anomaly detection
Dual Transformation
Project features onto (guide) or orthogonal to (ignore) concept subspaces
Zero-Shot Detection
Detect anomalies without task-specific training, using only CLIP embeddings
Interpretable Results
Understand what the model focuses on through explicit language concepts
Use Cases
Semantic Anomaly Detection
- ColorMNIST: Detect digits 5-9 as anomalies while ignoring color variations
- Waterbirds: Identify bird species anomalies independent of background
- CelebA: Detect facial attribute anomalies (e.g., blonde hair, glasses)
Industrial Anomaly Detection
- MVTec AD: Detect manufacturing defects in industrial products
- VisA: Identify visual anomalies in complex industrial scenarios
Custom Domains
LAFT’s flexibility allows you to:- Define custom prompt templates for your domain
- Specify multiple concept subspaces
- Combine guided and ignored transformations
- Use auxiliary prompts to expand the semantic space
Research Background
Authors: EungGu Yun, Heonjin Ha, Yeongwoo Nam, Bryan Dongik Lee Citation:Resources
Paper
Read the full ICLR 2025 paper on arXiv
GitHub
Access the source code and examples
Installation
Get started with setup instructions
Quick Start
Run your first LAFT example
Next Steps
Install LAFT
Set up your environment and install dependenciesGo to Installation →
Try the Quick Start
Run a complete example to understand the workflowGo to Quick Start →
Explore the API
Learn about all available functions and datasetsGo to API Reference →
