Skip to content

serre-lab/Horama

Repository files navigation

Horama logo

Horama is a compact library designed for Feature Visualization experiments, initially providing the implementation code for the research paper Maco, developed mainly to support exploratory research.

This repository includes:

While Horama builds on ideas from tools like Lucent, it focuses on flexibility and extensibility in PyTorch. It is not intended as a strict reproduction of Distill’s work -- only Maco is officially reproduced here.

Quick start notebook:
Open in Google Colab

🚀 Getting Started

Horama requires Python 3.7–3.10 and works with PyTorch models. Installation is available via PyPI:

pip install horama

Once installed, you can generate feature visualizations with minimal setup. The API is consistent across all methods and designed for easy experimentation. Example using timm:

import torch
import timm
import matplotlib.pyplot as plt
from horama import maco, fourier, plot_maco

model = timm.create_model('resnet18', pretrained=True).cuda().eval()
objective = lambda images: torch.mean(model(images)[:, 1])
# run maco
image1, alpha1 = maco(objective, device='cuda')
plot_maco(image1, alpha1)
plt.show()
# run fourier
image2, alpha2 = fourier(objective, device='cuda')
plot_maco(image2, alpha2)
plt.show()

💡 Be sure to match device='cuda' or 'cpu' depending on where your model is loaded.

Notebooks 📓

We provide a set of Colab notebooks to help you get started and explore different visualization techniques in Horama.

  • Starter Notebook (Colab) – quick overview and usage examples
  • Feature Inversion – Coming soon
  • Feature Accentuation – Coming soon

If you're new to feature visualization or just want a fast way to test Horama, the starter notebook is a great place to begin.

API Reference

Horama provides three main feature visualization functions. All three share a common structure and are designed for simple, composable use.

maco(objective_function,
     total_steps=1000,
     learning_rate=1.0,
     image_size=1000,
     model_input_size=224,
     noise=0.1,
     values_range=(-2.5, 2.5),
     crops_per_iteration=6,
     box_size=(0.20, 0.25),
     penalty=1.0,
     device='cuda')

fourier(objective_function,
        decay_power=1.5,
        total_steps=1000,
        learning_rate=1.0,
        image_size=1000,
        model_input_size=224,
        noise=0.1,
        values_range=(-2.5, 2.5),
        crops_per_iteration=6,
        box_size=(0.20, 0.25),
        penalty=1.0,
        device='cuda')

accentuation(objective_function,
             image_seed,
             decay_power=1.5,
             total_steps=1000,
             learning_rate=1.0,
             image_size=1000,
             model_input_size=224,
             noise=0.05,
             values_range=(-2.5, 2.5),
             crops_per_iteration=6,
             box_size=(0.20, 0.25),
             penalty=1.0,
             device='cuda')

Notes

  • objective_function(images) → scalar defines what you're visualizing. It typically targets a class score, a neuron or a direction.
  • values_range should match your model's expected input scale. For most timm models: (-2.5, 2.5) is usually fine.
  • box_size controls crop scale: (min_ratio, max_ratio).
  • model_input_size must match your model's expected input size (e.g. 224).

When optimizing, it's crucial to fine-tune the hyperparameters. Parameters like the decay spectrum in the Fourier method significantly impact the visual output, controlling the energy distribution across frequencies. Additionally, adjust the values_range to match your model's preprocessing requirements, and ensure model_input_size matches the expected input size of your model.

Citation

@article{fel2023maco,
      title={Unlocking Feature Visualization for Deeper Networks with MAgnitude Constrained Optimization},
      author={Thomas, Fel and Thibaut, Boissin and Victor, Boutin and Agustin, Picard and Paul, Novello and Julien, Colin and Drew, Linsley and Tom, Rousseau and Rémi, Cadène and Laurent, Gardes and Thomas, Serre},
      journal={Advances in Neural Information Processing Systems (NeurIPS)},
      year={2023},
}

Additional Resources

For a simpler and maintained implementation of the code for TensorFlow and the other feature visualization methods used in the paper, refer to the Xplique toolbox. Additionally, we have created a website called the LENS Project, which features the 1000 classes of ImageNet.

For code faithful to the original work of the Clarity team, we highly recommend Lucent.

Authors

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages