Skip to content

Active Learning on a Budget - Opposite Strategies Suit High and Low Budgets

License

Notifications You must be signed in to change notification settings

avihu111/TypiClust

Repository files navigation

Typiclust, ProbCover & DCoM Official Code Repository

This is the official implementation for the papers Active Learning on a Budget - Opposite Strategies Suit High and Low Budgets and Active Learning Through a Covering Lens.

This code implements TypiClust, ProbCover and DCoM - Simple and Effective Low Budget Active Learning methods.

Typiclust

Arxiv link, Twitter Post link, Blog Post link

TypiClust first employs a representation learning method, then clusters the data into K clusters, and selects the most Typical (Dense) sample from every cluster. In other words, TypiClust selects samples from dense and diverse regions of the data distribution.

Selection of 30 samples on CIFAR-10:

Selection of 10 samples from a GMM:

TypiClust Results summary

Probability Cover

Arxiv link, Twitter Post link, Blog Post link

ProbCover also uses a representation learning method. Then, around every point is placed a $\delta$-radius ball, and the subset of $b$ (budget) balls which covers the most of the points is selected, with their centers chosen as the samples to be labeled.

Unfolding selection of ProbCover

ProbCover results in the Semi-Supervised training framework

DCoM

Arxiv link

DCoM employs a representation learning approach. Initially, a $\Delta_{\text{avg}}$-radius ball is placed around each point. The $\Delta$ list provides a specific radius for each labeled example individually. From these, a subset of $b$ balls is chosen based on their coverage of the most points, with the centers of these balls selected as the samples to be labeled. After training the model, the $\Delta$ list is updated according to the purity of the balls to achieve more accurate radii and coverage. DCoM utilizes this coverage to determine the competence score, which balances typicality and uncertainty.

Illustration of DCoM's $\Delta$ updating

DCoM results in the Supervised training framework

DCoM results in the Semi-Supervised training framework

Usage

Please see USAGE for brief instructions on installation and basic usage examples.

Citing this Repository

This Repository makes use of two repositories: (SCAN and Deep-AL) Please consider citing their work and ours:

@article{hacohen2022active,
  title={Active learning on a budget: Opposite strategies suit high and low budgets},
  author={Hacohen, Guy and Dekel, Avihu and Weinshall, Daphna},
  journal={arXiv preprint arXiv:2202.02794},
  year={2022}
}

@article{yehudaActiveLearningCovering2022,
  title = {Active {{Learning Through}} a {{Covering Lens}}},
  author = {Yehuda, Ofer and Dekel, Avihu and Hacohen, Guy and Weinshall, Daphna},
  journal={arXiv preprint arXiv:2205.11320},
  year={2022}
}

@article{mishal2024dcom,
      title={DCoM: Active Learning for All Learners}, 
      author={Mishal, Inbal and Weinshall, Daphna},
      journal={arXiv preprint arXiv:2407.01804},
      year={2024}
}

License

This toolkit is released under the MIT license. Please see the LICENSE file for more information.

About

Active Learning on a Budget - Opposite Strategies Suit High and Low Budgets

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages