Skip to content

register_model_architecture not working in version 0.10.1 #3097

@voidmagic

Description

@voidmagic

🐛 Bug

Using register_model_architecture to register new hyper parameter set failed. The hyper parameter is overrided by default settings.

To Reproduce

Steps to reproduce the behavior (always include the command you ran):

  1. Run cmd 'fairseq-train --task language_modeling data-bin --arch transformer_lm_big --batch-size 1 --optimizer adam'
  2. See error
    There are 6 layers in the model, which should be 12 as defined in transformer_lm_big . https://github.com/pytorch/fairseq/blob/v0.10.1/fairseq/models/transformer_lm.py#L311

Code sample

Expected behavior

Environment

  • fairseq Version (e.g., 1.0 or master): 0.10.1
  • PyTorch Version (e.g., 1.0)
  • OS (e.g., Linux):
  • How you installed fairseq (pip, source):
  • Build command you used (if compiling from source):
  • Python version:
  • CUDA/cuDNN version:
  • GPU models and configuration:
  • Any other relevant information:

Additional context

Metadata

Metadata

Assignees

Labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions