Skip to content

Effect of rnn layers num #1

@xingyueye

Description

@xingyueye

@BriansIDP Hi~ Firstly, appreciate your kindly sharing. I have some questions about this implementation, and hope for your replies.

  1. I notice that the default "layerlist" is "0", and I wonder if numbers of rnn layers effect a lot? Have you tried other settings, such as add rnn before each attention block.
  2. RNN layers are builded by "nn.LSTM(d_model, rnndim, 1)". I guess if we should additionally set the "batch_first" options to "True". For the first dim of tensor is "batch" in Transformer, but I have not found Reshape operations before rnn_forward.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions