-
Notifications
You must be signed in to change notification settings - Fork 6.6k
Closed
Description
🐛 Bug
Using register_model_architecture to register new hyper parameter set failed. The hyper parameter is overrided by default settings.
To Reproduce
Steps to reproduce the behavior (always include the command you ran):
- Run cmd 'fairseq-train --task language_modeling data-bin --arch transformer_lm_big --batch-size 1 --optimizer adam'
- See error
There are 6 layers in the model, which should be 12 as defined intransformer_lm_big. https://github.com/pytorch/fairseq/blob/v0.10.1/fairseq/models/transformer_lm.py#L311
Code sample
Expected behavior
Environment
- fairseq Version (e.g., 1.0 or master): 0.10.1
- PyTorch Version (e.g., 1.0)
- OS (e.g., Linux):
- How you installed fairseq (
pip, source): - Build command you used (if compiling from source):
- Python version:
- CUDA/cuDNN version:
- GPU models and configuration:
- Any other relevant information: