Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Experiment Relation to Lottery Ticket #6

Open
simi2525 opened this issue Jan 9, 2020 · 1 comment
Open

Experiment Relation to Lottery Ticket #6

simi2525 opened this issue Jan 9, 2020 · 1 comment
Assignees

Comments

@simi2525
Copy link
Member

simi2525 commented Jan 9, 2020

Attempt the classic Lottery Ticket method of training a large model, pruning it and than retraining the subnet from a previous state forward.

Try to train the subnet from a random initialization but with "idea" order of training (brute force it)

The current ideas around tickets are that the initialization of a given subnet is what is the defining feature of the ticket, the best way of training the subnet currently being IMP with Rewinding technique.

Perhaps by having optimal order of training it would minimize the effect of the initialization of the subnet.

Lets see if the difference in accuracy between random initialization of subnet + perfect order vs IMP with Rewinding technique is reduced.

@simi2525
Copy link
Member Author

simi2525 commented Jan 9, 2020

In the 2019 Frankle et al paper, the researchers presented a new technique called Iterative Magnitude Pruning (IMP) with Rewinding.

Instead of looking at the weights of neurons at initialization, rewinding to iteration zero, it instead looks at their weights after several training iterations aka rewinding to iteration k.

The rationale behind the technique is that in many architectures, some neurons are already at a “winning” weight at the beginning while others only reach a “winning” weight after some training.

The same rationale for using perfect ordering being that we increase a given ticket's change of reaching a "wining" state after some training

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants