Skip to content

MadryLab/trak

Folders and files

NameName
Last commit message
Last commit date
Nov 2, 2023
Nov 11, 2023
Jan 17, 2024
Nov 18, 2024
Apr 12, 2023
Jan 17, 2024
Jan 17, 2024
Nov 2, 2023
Oct 25, 2023
Mar 28, 2023
Nov 11, 2023
Jan 27, 2023
Nov 11, 2023
Jun 1, 2023
Jan 17, 2024

Repository files navigation

arXiv PyPI version Documentation Status Code style: black

TRAK: Attributing Model Behavior at Scale

[docs & tutorials] [blog post] [website]

In our paper, we introduce a new data attribution method called TRAK (Tracing with the Randomly-Projected After Kernel). Using TRAK, you can make accurate counterfactual predictions (e.g., answers to questions of the form “what would happen to this prediction if these examples are removed from the training set?"). Computing data attribution with TRAK is 2-3 orders of magnitude cheaper than comparably effective methods, e.g., see our evaluation on:

Main figure

Usage

Check our docs for more detailed examples and tutorials on how to use TRAK. Below, we provide a brief blueprint of using TRAK's API to compute attribution scores.

Make a TRAKer instance

from trak import TRAKer

model, checkpoints = ...
train_loader = ...

traker = TRAKer(model=model, task='image_classification', train_set_size=...)

Compute TRAK features on training data

for model_id, checkpoint in enumerate(checkpoints):
  traker.load_checkpoint(checkpoint, model_id=model_id)
  for batch in loader_train:
      # batch should be a tuple of inputs and labels
      traker.featurize(batch=batch, ...)
traker.finalize_features()

Compute TRAK scores for target examples

targets_loader = ...

for model_id, checkpoint in enumerate(checkpoints):
  traker.start_scoring_checkpoint(checkpoint,
                                  model_id=model_id,
                                  exp_name='test',
                                  num_targets=...)
  for batch in targets_loader:
    traker.score(batch=batch, ...)

scores = traker.finalize_scores(exp_name='test')

Then, you can use the compute TRAK scores to analyze your model's behavior. For example, here are the most (positively and negatively) impactful examples for a ResNet18 model trained on ImageNet for three targets from the ImageNet validation set: ImageNet Figure

Check out the quickstart for a complete ready-to-run example notebook. You can also find several end-to-end examples in the examples/ directory.

Contributing

We welcome contributions to this project! Please see our contributing guidelines for more information.

Citation

If you use this code in your work, please cite using the following BibTeX entry:

@inproceedings{park2023trak,
  title = {TRAK: Attributing Model Behavior at Scale},
  author = {Sung Min Park and Kristian Georgiev and Andrew Ilyas and Guillaume Leclerc and Aleksander Madry},
  booktitle = {International Conference on Machine Learning (ICML)},
  year = {2023}
}

Installation

To install the version of our package which contains a fast, custom CUDA kernel for the JL projection step, use

pip install traker[fast]

You will need compatible versions of gcc and CUDA toolkit to install it. See the installation FAQs for tips regarding this. To install the basic version of our package that requires no compilation, use

pip install traker

Questions?

Please send an email to trak@mit.edu

Maintainers

Kristian Georgiev
Andrew Ilyas
Sung Min Park