class LSTMNet(torch.nn.Module): def __init__(self, num_hiddens, num_outputs): super(LSTMNet, self).__init__() #nn.Conv1d(1,16,2), #nn.Sigmoid(), # nn.MaxPool1d(2), #nn.Conv1d(1,32,2), self.hidden_size = num_hiddens # RNN 层,这里的 batch_first 指定传入的是 (批大小,序列长度,序列每个位置的大小) # 如果不指定其为 True,传入顺序应当是 (序列长度,批大小,序列每个位置的大小) input_size= num_inputs.view(len(input_x), 1, -1)//24 self.rnn = torch.nn.LSTM(input_size, hidden_size=num_hiddens,batch_first=True) # 线性层 self.dense = torch.nn.Linear(self.hidden_size*24, 256) self.dense2 = torch.nn.Linear(256,num_outputs) # dropout 层,这里的参数指 dropout 的概率 self.dropout = torch.nn.Dropout(0.3) self.dropout2 = torch.nn.Dropout(0.5) # ReLU 层 self.relu = torch.nn.ReLU() # 前向传播函数,这是一个拼接的过程,使用大量变量是为了避免混淆,不做过多讲解 def forward(self, x): # x shape: (batch_size, 24, 307) # LSTM 层会传出其参数,这里用 _ 将其舍弃 h, _ = self.rnn(x) # LSTM 层会传出 (batch_size, 24, num_hiddens) 个参数,故需要 reshape 后丢入全连接层 h_r = h.reshape(-1,self.hidden_size*24) h_d = self.dropout(h_r) y = self.dense(h_d) drop_y = self.dropout2(y) a = self.relu(drop_y) y2 = self.dense2(a) return y2
时间: 2024-04-19 16:30:13 浏览: 237
null
阅读全文
相关推荐




















