带有小批处理的Pytorch隐藏单元LSTM

问题描述 投票:0回答:1

问题

我不了解在小批量训练中如何处理LSTM隐藏细胞,因为训练数据以n序列的批次发送到网络,而在测试期间每次仅处理1个序列。

代码

具体地说,我的网络是:

class Pytorch_LSTM(nn.Module):
    def __init__(self, params):
        super(Pytorch_LSTM, self).__init__()
        self.params = params
        self.hidden_layer_size = params['hidden_layer_size']
        # Define layers
        self.lstm = nn.LSTM(input_size = params['in_features'], hidden_size = params['hidden_layer_size'])        
        self.linear1 = nn.Linear(params['hidden_layer_size'], params['hidden_layer_size'])
        self.linear2 = nn.Linear(params['hidden_layer_size'], params['out_features'])
        self.hidden_cell = (torch.zeros(1,self.params['batch_size'],self.hidden_layer_size),
                           torch.zeros(1,self.params['batch_size'],self.hidden_layer_size))

    def forward(self, input_seq):        
        lstm_out, self.hidden_cell = self.lstm(input_seq.view(self.params['time_window'],-1,self.params['in_features']), self.hidden_cell)
        linear1_out = self.linear1(lstm_out)
        predictions = self.linear2(linear1_out)
        return predictions[-1]

在我的train()方法中:

def train(self, input_sequence, params, test_idx, final, verbose=True):        

        ....
        ....

        # Model
        self.model = Pytorch_LSTM(params)
        # Let's train the model
        for epoch in range(epochs):
            for count_1,seq in enumerate(train_data_batch):      
                optimizer.zero_grad()
                self.model.hidden_cell = (torch.zeros(1, params['batch_size'], self.model.hidden_layer_size),
                                          torch.zeros(1, params['batch_size'], self.model.hidden_layer_size))   
                y_pred = self.model(seq)     # seq.shape: (n_batches, 25, 4)
                single_loss = mse_loss(y_pred, y_label)    # y_pred.shape, y_label.shape : (batch_size, 4)

我相信,这是在迷你批次中训练模型。测试时,每次只能有一个序列,而不是多个批次。在我的test()中:

for count,seq in enumerate(val_data[j]):                   
    y_pred = self.model(seq)           # seq.shape: (25,4)
    single_loss = mse_loss(y_pred, y_label)

这将返回错误:

RuntimeError: Expected hidden[0] size (1, 1, 100), got (1, 704, 100)

其中n_batches = 704。

我应该如何处理hidden_​​cell?

python machine-learning pytorch lstm recurrent-neural-network
1个回答
0
投票
您在每次调用形状(1,batch_size,100)时将(h_0, c_0)参数传递给lstm。 batch_size用于并行处理,并且是任意的,但是您每次都要在[]

self.hidden_cell = (torch.zeros(1,self.params['batch_size'],self.hidden_layer_size), torch.zeros(1,self.params['batch_size'],self.hidden_layer_size))

hidden_cellh_0c_0参数,即隐藏状态和单元状态的初始值。

尝试传递大小为(1,batch_size,100)的数组是不必要的,因为它本身具有所需大小的it defaults to zero向量。

仅除去self.hidden_cell,仅在input_seq方法中将self.lstm传递给forward。它应该工作

© www.soinside.com 2019 - 2024. All rights reserved.