-
Notifications
You must be signed in to change notification settings - Fork 45
Open
Description
Hi Tim, thanks for making this library. I am trying to test it on speech generation models and i have some questions from your code template:
- The models come with their own schedulers and optimizers. Can i simply wrap them around with
decay = CosineDecay ...andmask = Masking(optimizer, ...)? Should i change the optimizer to followoptim.SGD(...)and ignore the scheduler? It looks likemask.step()runs every epoch and replaces the scheduler, but i think i should still keep the optimizer specific to the model i have. - I understand that density/sparsity is the desired % of weights to keep, while prune/death rate is an internal parameter to determine what % weights should be redistributed at each iteration. Is this correct?
- Density looks like = sparsity in your code, although normally i would think density = 1 - sparsity.
- Code fails at
core.pyline 221-223 when there are RNNs, because for thembiasis a boolean and the bias terms are actuallybias_ihandbias_hh. I think this might count the parameters better:
for p, tensor in self.modules[0].named_parameters():
total_size += tensor.numel()
Metadata
Metadata
Assignees
Labels
No labels