Open qihuanwuqi opened 4 years ago
class ConvGRUBlock(nn.Module): def init(self): super(ConvGRUBlock, self).init() self.combine = nn.Sequential(OrderedDict([ ('layer1', ConvGRU(input_size=8, hidden_sizes=[32, 32, 64, 64], kernel_sizes=[3, 5, 3, 5], n_layers=4)), ('layer2', ConvGRU(input_size=64, hidden_sizes=[32, 32, 64, 64], kernel_sizes=[3, 5, 3, 5], n_layers=4)), ('layer3', ConvGRU(input_size=64, hidden_sizes=[32, 32, 64, 64], kernel_sizes=[3, 5, 3, 5], n_layers=4)), ('layer4', ConvGRU(input_size=64, hidden_sizes=[32, 32, 64, 64], kernel_sizes=[3, 5, 3, 5], n_layers=4)) ]))
def forward(self, x): out=self.combine(x) return out
model = ConvGRUBlock() input = torch.FloatTensor(1,8,64,64) print(output.type()) AttributeError: 'list' object has no attribute 'type'
class ConvGRUBlock(nn.Module): def init(self): super(ConvGRUBlock, self).init() self.combine = nn.Sequential(OrderedDict([ ('layer1', ConvGRU(input_size=8, hidden_sizes=[32, 32, 64, 64], kernel_sizes=[3, 5, 3, 5], n_layers=4)), ('layer2', ConvGRU(input_size=64, hidden_sizes=[32, 32, 64, 64], kernel_sizes=[3, 5, 3, 5], n_layers=4)), ('layer3', ConvGRU(input_size=64, hidden_sizes=[32, 32, 64, 64], kernel_sizes=[3, 5, 3, 5], n_layers=4)), ('layer4', ConvGRU(input_size=64, hidden_sizes=[32, 32, 64, 64], kernel_sizes=[3, 5, 3, 5], n_layers=4)) ]))
model = ConvGRUBlock() input = torch.FloatTensor(1,8,64,64) print(output.type()) AttributeError: 'list' object has no attribute 'type'