ShaoqingRen / faster_rcnn

Faster R-CNN
Other
2.7k stars 1.22k forks source link

Deep Residual Learning for Image Recognition paper availability #47

Closed futurely closed 8 years ago

futurely commented 8 years ago

@ShaoqingRen It is very exciting to know that your MSRA team have broken all the records on ImageNet classification, ImageNet detection, ImageNet localization, COCO detection, COCO segmentation and PASCAL VOC detection with large margins. Your achievements mark very significant milestones in the related areas. When and where will the ResNet paper [1] be released? Thanks a lot!

Sorry about that this question is not directly related to the project. I just can't wait to read the paper.

[1] "Deep Residual Learning for Image Recognition", Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun. Tech Report 2015.

ShaoqingRen commented 8 years ago

The paper can be found in http://arxiv.org/abs/1512.03385.

futurely commented 8 years ago

The new architecture is so simple to implement and the results are so strong that you might win the ICCV Best Paper Award (Marr Prize). It is certain that all types of deep neural network models will adopt residual learning and gain consistent performance improvements very soon.

Unfortunately, the paper did not mention the details of your segmentation algorithm. We will have to wait for another paper focus on that topic.

n3011 commented 8 years ago

@ShaoqingRen any hint on implementing it using caffe? @futurely

futurely commented 8 years ago

Several examples are available in MXNet, Keras and Lasagne. https://github.com/dmlc/mxnet/issues/931

futurely commented 8 years ago

Caffe: https://github.com/beijbom/beijbom_python_lib/blob/master/beijbom_caffe_tools.py

webmaven commented 8 years ago

@futurely (and anyone else): is there a TensorFlow implementation?

gcr commented 8 years ago

A warning: Quite a few of these implementations seem to be adding ReLU after every convolution layer. That doesn't seem right to me. If your building block is y = ReLU(f(x)) + x, then every layer will only increase y and never decrease y because one of the terms of the addition is constrained to be nonnegative.

The original paper only uses ReLU just after the addition y = ReLU(f(x) + x), and right after the first (but not the second) convolution layer.

erogol commented 8 years ago

@gcr good point

guiyang882 commented 8 years ago

I want to use tensorflow to complish the residual network of MSAR. Who can give me some code support ! Not the paper ! Thx!

futurely commented 8 years ago

https://github.com/tensorflow/skflow/blob/master/examples/resnet.py

ethanhe42 commented 8 years ago

The above residual network example of tensorflow has been moved here