Constrained Reweighting for Training Deep Neural Nets with Noisy Labels
training ML models involves minimizing a loss function
each step, the loss is approximately calculated as a weighted sum of the losses of individual instances in the mini-batch of data which is operating
each instance is treated equal for updating model parameters
noisy or mislabeled instances tend to have higher loss values than ones that are clean
assigning uniform importance weights to all instances might degrade accuracy when there are noisy instances
SOLUTION
create a family of constrained optimization problems
assign importance weights to individual instances in dataset to reduce effect of those that are likely to be noisy
controls how much the weights deviate from uniform
quantified by divergence measure
final loss —> weighted sum of individual instance losses which is used for updating the model parameters (CIW)
Schematic of the proposed Constrained Instance reWeighting (CIW) method. (src: paper linked above)
CIW method re-weights instances in each mini-batch based on their corresponding loss values
assigns importance weights over all possible class labels