Skip to content

Latest commit

 

History

History
36 lines (26 loc) · 2.08 KB

README.md

File metadata and controls

36 lines (26 loc) · 2.08 KB

CorruptedDataLoader

Chen Liu ([email protected])

Please kindly Star Github Stars this repo for better reach if you find it useful.

Contributions

We provide a simple wrapper around PyTorch DataLoader to intentionally mess up the input/label correspondence.

Motivation

In the majority of times, when we train a machine learning model, we pay extra attention to make sure the inputs and labels are correctly matched. In occasional situations, however, we may want the opposite to happen. One such possibility is, as outlined in the paper "Understanding deep learning requires rethinking generalization", we may want to corrupt the training set and intentionally overfit a model on random labels.

Despite careful search on the internet, we were unable to find existing open-source implementations to achieve this purpose. Therefore we designed our own method to achieve this purpose and provided it to those who may have a similar need.

Example

train_loader = ...  # define `train_loader` as you normally would
train_loader = CorruptedLabelDataLoader(train_loader)
for (x, y) in train_loader:
    ...

Details

This repository currently only contains a single file, which itself contains a single class called CorruptedDataLoader. CorruptedDataLoader is a wrapper around a Pytorch DataLoader. The Dataloader may hold arbitrary datasets, while in the current implementation, we only support the following datasets:

  1. torchvision.datasets.MNIST
  2. torchvision.datasets.CIFAR10
  3. torchvision.datasets.CIFAR100
  4. torchvision.datasets.STL10

Meanwhile, it can be easily adapted to any custom dataset, as long as you know under what key the labels are stored.

Usage

To use, simply copy CorruptedDataLoader to an appropriate location in your codebase and modify as you need. Don't forget to give us a star if you use it and find it helpful.

Citation

To be added