Skip to content
/ IDIM Public

Official implementation for "ID as a Model-Free Measure of Class Imbalance" (Neurocomputing, 2026)

License

Notifications You must be signed in to change notification settings

cagries/IDIM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

IDIM - ID as a Model-Free Measure of Class Imbalance

This repository contains the official implementation of the Neurocomputing article:

“Intrinsic Dimensionality as a Model-Free Measure of Class Imbalance”
Çağrı Eser, Zeynep Sonat Baltacı, Emre Akbaş, and Sinan Kalkan.
Neurocomputing 674 (2026) 132938

Paper: https://doi.org/10.1016/j.neucom.2026.132938
Preprint: https://arxiv.org/abs/2511.10475

Teaser figure for ID as a Model-Free Measure of Class Imbalance

ID is robust against random noise

Table of Contents

Setup

We recommend using Python 3.8+ and a virtual environment (e.g. conda). Dependencies (for ID estimation) to be installed (via pip, conda, etc.) include:

  • torch >= 1.12.0 and any appropriate version of torchvision
  • numpy
  • scikit-dimension
  • tqdm

For dependencies of individual integrations, please consult the relative README file of the respective method under methods/.

Repository Structure

.
├── docs/
├── methods/
│   ├── Bag-of-Tricks/
│   ├── BCL/
│   ├── DRO-LT/
│   ├── GLMC/
│   ├── logit_adjustment/
│   └── SURE/
├── utils/
│   └── id-estimation/
│       ├── id_cifar.py
│       ├── id_imagenet.py
│       ├── id_places.py
│       └── README.md
├── LICENSE
└── README.md

Usage

ID Estimation

We provide scripts to estimate ID on CIFAR-LT, PlacesLT and ImageNet-LT datasets in the utils/id-estimation directory.

Using ID with Long-Tailed Methods

This directory contains code for using our ID-based method with multiple integrations:

  • Bag of Tricks (Zhang et al., 2021)
  • Logit Adjustment (Menon et al., 2021)
  • DRO-LT (Samuel and Chechik, 2021)
  • BCL (Zhu et al., 2022)
  • GLMC (Du et al., 2023)
  • SURE (Li et al., 2024)

Each method has a dedicated directory under methods/ with its own instructions and an ID.md file describing how to plug in our ID-based measure.

Results

ID on CIFAR-LT against cardinality-based methods

ID on CIFAR-LT SOTA

ID on Places-LT SOTA

ID on ImageNet-LT SOTA

Pretrained Models

We release a subset of the models used in the paper.

Dataset Method Imbalance Ratio Top-1 Accuracy Checkpoints and Logs
CIFAR-10-LT GLMC + ID 100 87.9 link
50 90.5 link
CIFAR-100-LT GLMC + ID 100 58.0 link
50 62.8 link
CIFAR-10-LT SURE + RW + ID 100 87.0 link
50 90.4 link
CIFAR-100-LT SURE + RW + ID 100 57.7 link
50 62.7 link
Dataset Method Backbone Top-1 Accuracy Checkpoints and Logs
Places-LT BoT + ID ResNet-152 43.4 link
ImageNet-LT BoT + ID ResNet-10 42.9 link
ImageNet-LT GLMC + ID ResNeXt-50 56.3 link
ImageNet-LT BCL + ID ResNet-50 (90EP) 56.5 link
ImageNet-LT BCL + ID ResNeXt-50 (90EP) 57.9 link
ImageNet-LT BCL + ID ResNeXt-50 (180EP) 58.2 link

Citation

If you would like to cite this work, please use:

@article{eser2026intrinsic,
title = {Intrinsic dimensionality as a model-free measure of class imbalance},
journal = {Neurocomputing},
volume = {674},
pages = {132938},
year = {2026},
issn = {0925-2312},
doi = {https://doi.org/10.1016/j.neucom.2026.132938},
url = {https://www.sciencedirect.com/science/article/pii/S0925231226003358},
author = {Cagri Eser and Zeynep Sonat Baltaci and Emre Akbas and Sinan Kalkan},
keywords = {Intrinsic dimension, Long-tailed visual recognition, Class imbalance, Long-tailed learning},
abstract = {Imbalance in classification tasks is commonly quantified by the cardinalities of examples across classes. This, however, disregards the presence of redundant examples and inherent differences in the learning difficulties of classes. Alternatively, one can use complex measures such as training loss and uncertainty, which, however, depend on training a machine learning model. Our paper proposes using data Intrinsic Dimensionality (ID) as an easy-to-compute, model-free measure of imbalance that can be seamlessly incorporated into various imbalance mitigation methods. Our results across five different datasets with a diverse range of imbalance ratios show that ID consistently outperforms cardinality-based re-weighting and re-sampling techniques used in the literature. Moreover, we show that combining ID with cardinality can further improve performance. Our code and models are available at https://github.com/cagries/IDIM.}
}

Contact

For questions and suggestions, please contact:

About

Official implementation for "ID as a Model-Free Measure of Class Imbalance" (Neurocomputing, 2026)

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published