Abstract
In neural network models, obtaining a high-quality dataset is critical because they are generally reliant on training data. A common problem that occurs is class imbalance, in which models tend to be biased to the majority class when the training data is not balanced. To overcome this problem, we propose a novel regularization method that provides a penalty to the loss function, using two facets of the distribution of the model's output p(ŷ|x): (1) skewed mean and (2) variance divergence between p(ŷ|x∈D+) and p(ŷ|x∈D−). The experimental results demonstrate that our methods consistently improve the performance on imbalanced datasets. Moreover, the combination of two regularization methods provides a substantial performance improvement on five sentence classification datasets and also an image classification dataset; notably, state-of-the-art performances are achieved on the WikiQA and SelQA datasets.
| Original language | English |
|---|---|
| Article number | 115974 |
| Journal | Expert Systems with Applications |
| Volume | 188 |
| DOIs | |
| State | Published - Feb 2022 |
Keywords
- Class imbalance
- Image classification
- Regularization
- Sentence classification
Fingerprint
Dive into the research topics of 'Novel regularization method for the class imbalance problem'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver