pytorch 预期输入批次大小(56180)与目标批次大小(100)匹配

hrirmatl  于 2022-12-23  发布在  其他
关注(0)|答案(1)|浏览(208)

我得到以下错误。
ValueError:预期输入batch_size(56180)与目标batch_size(100)匹配。
我的模型的输入是3通道(RGB)227x227图像和批量大小是100。

  1. torch.Size([100, 3, 227, 227])
  2. torch.Size([100, 10, 111, 111])
  3. torch.Size([100, 20, 53, 53])
  4. torch.Size([56180, 100])
  5. torch.Size([56180, 64])
  6. torch.Size([56180, 64])
  7. torch.Size([56180, 32])
  8. torch.Size([56180, 32])
  9. torch.Size([56180, 1])

这是二进制分类(真,假),所以我使最终输出为1

  1. class Net(nn.Module):
  2. def __init__(self):
  3. super(Net, self).__init__()
  4. #input image 227x227x3
  5. self.conv1 = nn.Conv2d(3, 10, kernel_size=5)
  6. self.conv2 = nn.Conv2d(10, 20, kernel_size=5)
  7. self.conv2_drop = nn.Dropout2d()
  8. self.fc1 = nn.Linear(100, 64)
  9. self.fc3 = nn.Linear(64, 32)
  10. self.fc6 = nn.Linear(32, 1)
  11. def forward(self, x):
  12. print(x.shape)
  13. x = F.relu(F.max_pool2d(self.conv1(x), 2))
  14. print(x.shape)
  15. x = F.relu(F.max_pool2d(self.conv2_drop(self.conv2(x)), 2))
  16. print(x.shape)
  17. x = x.view(-1, x.size(0))
  18. print(x.shape)
  19. x = F.relu(self.fc1(x))
  20. print(x.shape)
  21. x = F.dropout(x, training=self.training)
  22. print(x.shape)
  23. x = self.fc3(x)
  24. print(x.shape)
  25. x = F.dropout(x, training=self.training)
  26. print(x.shape)
  27. x = self.fc6(x)
  28. print(x.shape)
  29. return x
  30. def train(model, train_loader, optimizer):
  31. model.train()
  32. for batch_idx, (data, target) in enumerate(train_loader):
  33. data, target = data.to(DEVICE), target.to(DEVICE)
  34. optimizer.zero_grad()
  35. output = model(data)
  36. target = target.unsqueeze(-1)
  37. loss = F.cross_entropy(output, target)
  38. loss.backward()
  39. optimizer.step()

我的问题是,我有100批图像,所以目标(Y)是100单位。但为什么我得到56180单位的结果?

xkrw2x1b

xkrw2x1b1#

更改视图功能(在正向方法中):

  1. x = x.view(x.size(0), -1)

批量大小必须在0维中。
您的forward方法应该像这样定义:

  1. def forward(self, x):
  2. x = F.relu(F.max_pool2d(self.conv1(x), 2))
  3. x = F.relu(F.max_pool2d(self.conv2_drop(self.conv2(x)), 2))
  4. x = x.view(x.size(0), -1)
  5. x = F.relu(self.fc1(x))
  6. x = F.dropout(x, training=self.training)
  7. x = self.fc3(x)
  8. x = F.dropout(x, training=self.training)
  9. x = self.fc6(x)
  10. return x

相关问题