The document shows the bidirectional LSTM model output hidden state and cell of memory in the formation of [2*num_layers, batch_size, num_hidden]. I am wondering what the sequence in the first dimension.
Suppose the num_layers = 2, is the sequence of hidden state in the following way:
- the forward hidden state of the first layer,
- the forward hidden state of the second layer,
- the backward hidden state of the first layer
- the backward hidden state of the second layer?