You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Good catch. It's only broken for TransformerLanguageModel in 0.10.1, because that model was migrated to Hydra (dataclass-based config) before the migration was completed.
The fix is to patch this commit: b7d8b9d. I'll release 0.10.2 with the fix shortly.
The hyper parameters of transformer model are still overrided by default settings in TransformerConfig in 0.10.2.
If I set arch=transformer_tiny, there 6 layers in the model, which should be 2.
I think that's because add_args() in TransformerModelBase sets the default hyper parameters.
🐛 Bug
Using register_model_architecture to register new hyper parameter set failed. The hyper parameter is overrided by default settings.
To Reproduce
Steps to reproduce the behavior (always include the command you ran):
There are 6 layers in the model, which should be 12 as defined in
transformer_lm_big
. https://github.com/pytorch/fairseq/blob/v0.10.1/fairseq/models/transformer_lm.py#L311Code sample
Expected behavior
Environment
pip
, source):Additional context
The text was updated successfully, but these errors were encountered: