Open LiuYuZzz opened 4 years ago
Hi @LiuYuZzz
For training, you do not need to make any changes inside our model. What you need to do is computing losses using training data and backward them.
Hi @LiuYuZzz
- In our current implementation, it only works for 1 sample-per-GPU. We used batch size 4 with 4-GPU machine.
- You do not need to resize them just use torch.CrossEntropyLoss function (see the documentation for usage, they take indices for GT).
- Looks like you did it right.
For training, you do not need to make any changes inside our model. What you need to do is computing losses using training data and backward them.
thanks for your reply,I got it
For multi-obj, the input frame size is
[batch_size, color channels, H, W]
, and the input objects mask size is[batch_size, num_objects + BG, H, W]
, and the questions are:batch_size =1
;[batch_size, num_objects + BG, H, W]
, then resize it to[batch_size*H*W, num_objects + BG]
, and input the new size tensor into the CrossEntropyLoss, is it right?