A Comprehensive Survey of Forgetting in Deep Learning Beyond Continual Learning. TPAMI, 2024.
- Updated
![]() |
VOOZH | about |
A Comprehensive Survey of Forgetting in Deep Learning Beyond Continual Learning. TPAMI, 2024.
[IJCAI 2024] Papers about graph reduction including graph coarsening, graph condensation, graph sparsification, graph summarization, etc.
IEEE JBHI 2026 | TinyUSFM: Towards Compact and Efficient Ultrasound Foundation Models
Coresets over Multiple Tables for Feature-rich and Data-efficient Machine Learning
An Efficient Dataset Condensation Plugin and Its Application to Continual Learning. NeurIPS, 2023.
Code implementation for ICLR 2025 paper: ELFS: Label-Free Coreset Selection with Proxy Training Dynamics
DataCull is a modular, light-weight data pruning library containing many dataset pruning (coreset selection) algorithm including the official Implementation of the paper, titled, RCAP: Robust, Class-Aware, Probab ilistic Dynamic Dataset Pruning
Docs: https://erasus.readthedocs.io/en/latest/ Forget data from any foundation model without retraining. Erasus surgically removes concepts, behaviors, or training samples from LLMs, VLMs, and Diffusion models using coreset selection. 90% less compute, certified removal, multimodal support.
[ACCV 2022] The official repository of ''COLLIDER: A Robust Training Framework for Backdoor Data''.
an implementation of unsupervised core-set selection
Add a description, image, and links to the coreset-selection topic page so that developers can more easily learn about it.
To associate your repository with the coreset-selection topic, visit your repo's landing page and select "manage topics."