open-mmlab / mmdetection

OpenMMLab Detection Toolbox and Benchmark
https://mmdetection.readthedocs.io
Apache License 2.0
29.46k stars 9.44k forks source link

Creating multi-modal (bi-modal) detectors / pipelines for them #8939

Open Maenansyn opened 2 years ago

Maenansyn commented 2 years ago

I adapted Faster R-CNN to be bi-modal (colour images and an aligned second 3-channel 2D sensor data). For this I duplicated the ResNet50-backbone and use a splitted forward-function. Given input x for the function is a list of two elements. x[0] is forwarded into the first ResNet50-backbone, x[1] is forwarded into the second branch. Afterwards is fusioned by an element-wise max-operation and then continued as regular Faster R-CNN.

The model structure created looks as expected, loading an adapted checkpoint works as well, but running images through the network does not work as the pipeline is not made for two images per instance and interprets it as a batch. Adapting the first-level methods (e.g. inference_detector()) does not work as the problem is deeply rooted inside the base-code of the models. So I figured out that I have to adapt the training- and test-pipelines.

But the custom pipeline tutorial doesn't explain how I can achieve that or at least I can't derive it from its content. During research in the issues here I came across issue #1861 where it was asked how to change backbone input from 3 to 6 channels. This seems to be a suitable solution for my problem to have [6, width, height]-shaped image tensors as input instead of [2, 3, width, height]. During the forward-function of my custom model I could then split it into 2 [3, 320, 240] tensors and use it as before and the batch-problem would not occur.

But neither in this issue nor in #6977 nor #4182 I can find a description about what to do and I haven't found the solution for creating such a custom pipeline by myself. That would be a great help. I also don't know how the img_norm_cfg get its values and how I would need to adopt this for 6 channels:

img_norm_cfg = dict( mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True)

The annotations-json includes only the colour images (and annotations only refer to colour images therefore), but they are working for both modals as they are aligned and a pixel in one modal refers to the same point in the real world as the pixel at the same position in the second modal image.

Summary: The new custom-backbone for 2 three-channel 2D-data is registered in registry and the adapted model is built by only adapting the backbone-dict in the config. Checkpoints are adapted accordingly. Data set is a custom data set with 2 same-sized images per instance. The reason for it not working is most likely an un-adapted pipeline config which I don't know how to customize correctly.

PeterVennerstrom commented 2 years ago

LoadMultiChannelImageFromFiles will load both image files and stack them. This will likely be compatible with some pipeline functions, but not others.

There might be some functions like resize that are compatible with a [h, w, 6] shape, but not [h, w, 3, 2] so you may want to change stack to concatenate in LoadMultiChannelImageFromFiles.

Some functions like PhotoMetricDistortion will require a [h, w, 3] input. If you are using function with this requirement a possible solution would be to decorate with a function that splits the 6 channel image into two 3 channel images and sends the appropriate one through the transform function then concatenates them after.

Maenansyn commented 2 years ago

Thank you, I'll look into it and see if it works. Is there an overview of possible selections for Load-types? I only know some because of examples or model configs and following them through references, code and imports to their actual definition/implementation to find others can be a bit tedious sometimes.. ^^

Edited: Found them now with multichannel. I should have just googled for the "LoadImageFromFile" or something before instead of searching through the repo by myself...

PeterVennerstrom commented 2 years ago

loading.py and transforms.py should contain most/all the functions you may need to modify. The best existing documentation for your use case is probably the code/comments. You may find the tests useful as a reference as well.

Maenansyn commented 2 years ago

Yes, that works mostly. But now I'm encountering an error with the FileClient where I can't find the reason for it (loading.py seems to be the place where the error roots, as explained further below). Base code for forwarding the two images:

from mmdet.apis import inference_detector, show_result_pyplot
from mmdet.datasets.pipelines import LoadMultiChannelImageFromFiles
import os.path as osp

instance_ = {'img_prefix': '/content/drive/MyDrive/test_data',
            'img_info': {'filename': ['filename1.png', 'filename2.png']}}
transform = LoadMultiChannelImageFromFiles()
instance_ = transform(instance_)
result = inference_detector(model, instance_)

The last line results in "FileNotFoundError: [Errno 2] No such file or directory: 'img_prefix'" via:

[<ipython-input-32-053cf46cbb3c>](https://localhost:8080/#) in inference_detector(model, imgs)
---> 50         data = test_pipeline(data)

[/content/mmdetection/mmdet/datasets/pipelines/compose.py](https://localhost:8080/#) in __call__(self, data)
---> 41             data = t(data)  # t being one in self.transforms

[/content/mmdetection/mmdet/datasets/pipelines/loading.py](https://localhost:8080/#) in __call__(self, results)
--> 176             img_bytes = self.file_client.get(name)

[/usr/local/lib/python3.7/dist-packages/mmcv/fileio/file_client.py](https://localhost:8080/#) in get(self, filepath)
-> 1014         return self.client.get(filepath)

[/usr/local/lib/python3.7/dist-packages/mmcv/fileio/file_client.py](https://localhost:8080/#) in get(self, filepath)
--> 535         with open(filepath, 'rb') as f:

So filepath seems to equal '_imgprefix'. But when extracting the loc right before the call in loading.py (lines 166-170) and executing it manually after the transform()-call in the first part of code above I get the right paths and files are shown as images when I open them with plt or similar:

instance_ = {'img_prefix': '/content/drive/MyDrive/test_data',
            'img_info': {'filename': ['filename1.png', 'filename2.png']}}
transform = LoadMultiChannelImageFromFiles()
instance_ = transform(instance_)
if instance_['img_prefix'] is not None:
            filename = [
                osp.join(instance_['img_prefix'], fname)
                for fname in ergebnis['img_info']['filename']
            ]
            print(filename)

--> ['/content/drive/MyDrive/test_data/filename1.png', '/content/drive/MyDrive/test_data/filename2.png']

The variable 'name' in loading.py is one element when iterating through the 'filename'-list. Do you have any clue how the dict-key '_imgprefix' becomes filepath when the joining of the paths is actually working?

And for the sake of completeness self.transform is only

test_pipeline = [
    dict(type='LoadMultiChannelImageFromFiles')
]

I'm not sure if this is actually necessary or if doing the transform by myself before calling the inference_detector is not the right way of handling the data.

Maenansyn commented 2 years ago

Origin of the error is in inference.py lines 129-131

 else:
            # add information into dict
            data = dict(img_info=dict(filename=img), img_prefix=None)

img_prefix is set to None, so setting my path earlier is undone.

This leads in loading.py lines 64 ff. to go into the else block instead of the if block:

if results['img_prefix'] is not None:
            filename = osp.join(results['img_prefix'],
                                results['img_info']['filename'])
else:
            filename = results['img_info']['filename']

The actual paths could be found at result['img_info']['filename']['img_info']['filename'], if I add the full path into img_info when assembling the instance.

Therefore the dict is somehow nested too much by the firstly mentioned else-block. On the other hand, when I only take the absolute file paths as 2D-list of data I surpass it and run into a recursion error of collate (Sequence and Mapping block):

instance_ = [['/path/to/file/filename1.png', '/path/to/file/filename2.png']]

[/usr/local/lib/python3.7/dist-packages/mmcv/parallel/collate.py](https://localhost:8080/#) in collate(batch, samples_per_gpu)
     75     elif isinstance(batch[0], Sequence):
     76         transposed = zip(*batch)
---> 77         return [collate(samples, samples_per_gpu) for samples in transposed]
     78     elif isinstance(batch[0], Mapping):
     79         return {

RecursionError: maximum recursion depth exceeded while calling a Python object
YihanWang-chengdu commented 2 years ago

I am doing the same thing as you . I have modified the Class : CustomDataset in mmdet.dataset.custom.py and transform.py 、loading.py . i relize the training and testing for dual-resnet with the input RGB and Depth image.

nsnlz commented 2 years ago

I am doing the same thing as you . I have modified the Class : CustomDataset in mmdet.dataset.custom.py and transform.py 、loading.py . i relize the training and testing for dual-resnet with the input RGB and Depth image.

Can you share your code with me ? Thx !

YihanWang-chengdu commented 2 years ago

I am doing the same thing as you . I have modified the Class : CustomDataset in mmdet.dataset.custom.py and transform.py 、loading.py . i relize the training and testing for dual-resnet with the input RGB and Depth image.

Can you share your code with me ? Thx !

what way can i share the modified code with you

YihanWang-chengdu commented 2 years ago

I am doing the same thing as you . I have modified the Class : CustomDataset in mmdet.dataset.custom.py and transform.py 、loading.py . i relize the training and testing for dual-resnet with the input RGB and Depth image.

Can you share your code with me ? Thx ! or i will share the modified code on my home page,

nsnlz commented 2 years ago

I am doing the same thing as you . I have modified the Class : CustomDataset in mmdet.dataset.custom.py and transform.py 、loading.py . i relize the training and testing for dual-resnet with the input RGB and Depth image.

Can you share your code with me ? Thx ! or i will share the modified code on my home page,

Thx!Just on your homepage.

YihanWang-chengdu commented 2 years ago

can you give me your email, i sent the code for you , git the modified project on github is complicated for me .

nsnlz commented 2 years ago

can you give me your email, i sent the code for you , git the modified project on github is complicated for me .

Thx ! My email is 2558149981@qq.com .

YihanWang-chengdu commented 2 years ago

代码已经发给你了,我做的是同时读取深度图像和RGB图像 因此在配置文件中要给定深度图的路径 同时主干网络也需要改写  给予两个输入

------------------ 原始邮件 ------------------ 发件人: "open-mmlab/mmdetection" @.>; 发送时间: 2022年10月15日(星期六) 中午12:22 @.>; @.**@.>; 主题: Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939)

can you give me your email, i sent the code for you , git the modified project on github is complicated for me .

Thx ! My email is @.*** .

— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.***>

nsnlz commented 2 years ago

代码已经发给你了,我做的是同时读取深度图像和RGB图像 因此在配置文件中要给定深度图的路径 同时主干网络也需要改写  给予两个输入 ------------------ 原始邮件 ------------------ 发件人: "open-mmlab/mmdetection" @.>; 发送时间: 2022年10月15日(星期六) 中午12:22 @.>; @.**@.>; 主题: Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939) can you give me your email, i sent the code for you , git the modified project on github is complicated for me . Thx ! My email is @. . — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.>

感谢 !

Maenansyn commented 2 years ago

I am doing the same thing as you . I have modified the Class : CustomDataset in mmdet.dataset.custom.py and transform.py 、loading.py . i relize the training and testing for dual-resnet with the input RGB and Depth image.

Right now I'm working through the calls from test.py to the forward-calls. I worked on loading and transforms as Peter has recommended. The dataset-class itself I haven't touched yet, but I'll probably need to change it for the data_loader.

I also found some more defined multi-channel transforms here. Have you done the same?

YihanWang-chengdu commented 2 years ago

I am doing the same thing as you . I have modified the Class : CustomDataset in mmdet.dataset.custom.py and transform.py 、loading.py . i relize the training and testing for dual-resnet with the input RGB and Depth image.

Right now I'm working through the calls from test.py to the forward-calls. I worked on loading and transforms as Peter has recommended. The dataset-class itself I haven't touched yet, but I'll probably need to change it for the data_loader.

I also found some more defined multi-channel transforms here. Have you done the same?

i relize the multi-modal training and testing by fixxing the mmdet.dataset.custom.py . i chose to load both depth image and RGB image, and do the same data augumentation for them. Finally, i utilize the collect function to output these images.

and the next are the part lines in my config.py , the jet_prefix indicate the path of the depth image(JET, 3-channels).

data = dict( samples_per_gpu=3, workers_per_gpu=3,
train=dict( img_prefix='D:/wyh/grap_dataset_coco/train', jet_prefix = 'D:/wyh/grap_dataset_coco/JET/', classes=classes, ann_file='D:/wyh/grap_dataset_coco/train/JPEGImages/annotations_train.json'), val=dict( img_prefix='D:/wyh/grap_dataset_coco/val/', jet_prefix = 'D:/wyh/grap_dataset_coco/JET/', classes=classes, ann_file='D:/wyh/grap_dataset_coco/val/JPEGImages/annotations_val.json'),
test=dict( img_prefix='D:/wyh/grap_dataset_coco/test/',
jet_prefix = 'D:/wyh/grap_dataset_coco/JET/', classes=classes, ann_file='D:/wyh/grap_dataset_coco/JPEGImages/annotations_test.json'))

WuTao-CS commented 2 years ago

can you give me your email, i sent the code for you , git the modified project on github is complicated for me .

Can you share the code to me too ? I face the same issue, please help me. My email is polariz@126.com.

I have sent his code to you .

Hi, can you also share your code with me ? Thanks a lot! My email is 747866429@qq.com

dagezhuang commented 1 year ago

代码已经发给你了,我做的是同时读取深度图像和RGB图像 因此在配置文件中要给定深度图的路径 同时主干网络也需要改写  给予两个输入 ------------------ 原始邮件 ------------------ 发件人: "open-mmlab/mmdetection" @.>; 发送时间: 2022年10月15日(星期六) 中午12:22 @.>; @.**@.>; 主题: Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939) can you give me your email, i sent the code for you , git the modified project on github is complicated for me . Thx ! My email is @. . — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.>

您好!可以共享下代码吗?万分感谢! e-mail: 760721935@qq.com

Troy-peng-0327 commented 1 year ago

can you give me your email, i sent the code for you , git the modified project on github is complicated for me .

Hi, can you also share your code with me ? Thanks a lot! My email is troy_peng0327@163.com

dagezhuang commented 1 year ago

can you give me your email, i sent the code for you , git the modified project on github is complicated for me .

Can you share the code to me too ? I face the same issue, please help me. My email is polariz@126.com.

I have sent his code to you .

Would you please share the code with me ? Thanks a lot. My email is 760721935@qq.com

nikky4D commented 1 year ago

can you give me your email, i sent the code for you , git the modified project on github is complicated for me .

This is an nice modification. Can you also share your data loading code with me as well? nchukaobah@gmail.com

YihanWang-chengdu commented 1 year ago

the modification code is followed in the adjunct.

------------------ 原始邮件 ------------------ 发件人: "open-mmlab/mmdetection" @.>; 发送时间: 2022年11月20日(星期天) 凌晨2:31 @.>; @.**@.>; 主题: Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939)

can you give me your email, i sent the code for you , git the modified project on github is complicated for me .

This is an nice modification. Can you also share your data loading code with me as well? @.***

— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.***>

从QQ邮箱发来的超大附件

mmdetection_deep.zip (92.69M, 2022年12月22日 15:13 到期)进入下载页面:http://mail.qq.com/cgi-bin/ftnExs_download?t=exs_ftn_download&k=2d623461df9956cb33e3117b4063501e5c5b02505605030949030402004e5654025019585002031c00540152575351555d0107556677625c09065115030016580b0c6b050306121f1e0b44615b&code=db4afcb1

Troy-peng-0327 commented 1 year ago

the modification code is followed in the adjunct.

Thanks a lot!

qqq1521902442 commented 1 year ago

I am doing the same thing as you . I have modified the Class : CustomDataset in mmdet.dataset.custom.py and transform.py 、loading.py . i relize the training and testing for dual-resnet with the input RGB and Depth image.我在做和你一样的事情。我修改了 Class: CustomDataset 的 mmdet.dataset.custom.py 和 form.py、 loading.py。我实现了训练和测试的双重重网与输入 RGB 和深度图像。

Right now I'm working through the calls from test.py to the forward-calls. I worked on loading and transforms as Peter has recommended. The dataset-class itself I haven't touched yet, but I'll probably need to change it for the dataloader.现在我正在处理 test.py 和转发呼叫之间的通话。我按照彼得的建议进行了加载和转换。我还没有涉及到数据集类本身,但是我可能需要为 data loader 更改它。 I also found some more defined multi-channel transforms here. Have you done the same?我还在这里发现了一些定义更加明确的多通道变换。您也这样做了吗?

i relize the multi-modal training and testing by fixxing the mmdet.dataset.custom.py . i chose to load both depth image and RGB image, and do the same data augumentation for them. Finally, i utilize the collect function to output these images.

通过固定 mmdet.dataset.custom.py ,我体会到了多模式训练和测试。我选择加载深度图像和 RGB 图像,并为他们做相同的数据增强。最后,我利用收集函数输出这些图像。

and the next are the part lines in my config.py , the jet_prefix indicate the path of the depth image(JET, 3-channels).

接下来是 config.py 中的部分线,JET _ 前缀表示深度图像(JET,3通道)的路径。

data = dict( samples_per_gpu=3, workers_per_gpu=3, train=dict( img_prefix='D:/wyh/grap_dataset_coco/train', jet_prefix = 'D:/wyh/grap_dataset_coco/JET/', classes=classes, ann_file='D:/wyh/grap_dataset_coco/train/JPEGImages/annotations_train.json'), val=dict( img_prefix='D:/wyh/grap_dataset_coco/val/', jet_prefix = 'D:/wyh/grap_dataset_coco/JET/', classes=classes, ann_file='D:/wyh/grap_dataset_coco/val/JPEGImages/annotations_val.json'), test=dict( img_prefix='D:/wyh/grap_dataset_coco/test/', jet_prefix = 'D:/wyh/grap_dataset_coco/JET/', classes=classes, ann_file='D:/wyh/grap_dataset_coco/JPEGImages/annotations_test.json'))

Something co/test/’,JET prefix = ‘ D:/wyh/grap datet coco/JET/’,class = class,ann file = ‘ D:/wyh/grap datet coco/JPEGImages/annotions _ test. json’))

I'm also currently having problems loading multimodal data, could you please share the code for loading data with me? My email is: juxweeru@163.com Thanks a lot!

jinuncle commented 1 year ago

can you give me your email, i sent the code for you , git the modified project on github is complicated for me .

Hi, can you also share your code with me ? Thanks a lot! My email is 962527411@qq.com

I have sent his code to you .

I'm also currently having problems loading multimodal data, could you please share the code for loading data with me? My email is: 2110456075@email.szu.edu.cn Thanks a lot!

jinuncle commented 1 year ago

thank you so much

--------------原始邮件-------------- 发件人:"zz @.>; 发送时间:2023年4月19日(星期三) 中午11:51 收件人:"open-mmlab/mmdetection" @.>; 抄送:"侯宏锦 @.>;"Comment @.>; 主题:Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939)

@.***

Original Email

Sender:"jinuncle"< @.*** &gt;;

Sent Time:2023/4/18 21:50

To:"open-mmlab/mmdetection"< @.*** &gt;;

Cc recipient:"zz"< @. &gt;;"Comment"< @. &gt;;

Subject:Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal)detectors / pipelines for them (Issue #8939)

can you give me your email, i sent the code for you , git the modified project on github is complicated for me .

Hi, can you also share your code with me ? Thanks a lot! My email is @.***

I have sent his code to you .

I'm also currently having problems loading multimodal data, could you please share the code for loading data with me? My email is: @.*** Thanks a lot!

— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.***&gt;

                    从QQ邮箱发来的超大附件     

                                                                    mmdetection_deep.zip                             (92.7MB, 2023年5月19日 11:50)                                                                             进入下载页面                          :https://wx.mail.qq.com/ftn/download?func=3&amp;k=c6c651329c2ea445f4b51a32376131647e3cff37356131641c164c46050357535c0e1904565505490b5c0d06180055570e140305010005510b010c010c582564545450574104521050565a6d5104541417435d42213895c9089ef2a53f5130d3836e15e5a991635e89&amp;key=c6c651329c2ea445f4b51a32376131647e3cff37356131641c164c46050357535c0e1904565505490b5c0d06180055570e140305010005510b010c010c582564545450574104521050565a6d5104541417435d42213895c9089ef2a53f5130d3836e15e5a991635e89&amp;code=99425a1d&amp;from= 

— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.***>

realkanchi commented 1 year ago

代码已经发给你了,我做的是同时读取深度图像和RGB图像 因此在配置文件中要给定深度图的路径 同时主干网络也需要改写  给予两个输入 ------------------ 原始邮件 ------------------ 发件人: "open-mmlab/mmdetection" @.**>; 发送时间: 2022年10月15日(星期六) 中午12:22 @.**>; @.**@.**>; 主题: Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939) can you give me your email, i sent the code for you , git the modified project on github is complicated for me . Thx ! My email is @. . — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.>

感谢 !

方便给我也发一份吗,我的邮箱realkanchi@gmail.com

jinuncle commented 1 year ago

十分感谢

--------------原始邮件-------------- 发件人:"realkanchi @.>; 发送时间:2023年7月13日(星期四) 晚上6:08 收件人:"open-mmlab/mmdetection" @.>; 抄送:"侯宏锦 @.>;"Comment @.>; 主题:Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939)

代码已经发给你了,我做的是同时读取深度图像和RGB图像 因此在配置文件中要给定深度图的路径 同时主干网络也需要改写  给予两个输入 … ------------------ 原始邮件 ------------------ 发件人: "open-mmlab/mmdetection" @.>; 发送时间: 2022年10月15日(星期六) 中午12:22 @**.>; @.*@.>; 主题: Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939) can you give me your email, i sent the code for you , git the modified project on github is complicated for me . Thx ! My email is @_. . — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.>

感谢 !

@.***

— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.***>

unnamed0216 commented 1 year ago

I also currently having problems loading multimodal data, could you please share the code for loading data with me? My email is unnamed870216@gmail.com thank you

vlaminckaxel commented 1 year ago

Can you also share this code? Or we could make a contribution to mmdetection to support multimodal/multi-frame detections out of the box?

realkanchi commented 1 year ago

I've been quite busy recently, working on other company projects. Once I have some free time and finish this project, I will share the code.

axel.vlaminck @.***> 于2023年9月5日周二 21:13写道:

Can you also share this code? Or we could make a contribution to mmdetection to support multimodal/multi-frame detections out of the box?

— Reply to this email directly, view it on GitHub https://github.com/open-mmlab/mmdetection/issues/8939#issuecomment-1706596651, or unsubscribe https://github.com/notifications/unsubscribe-auth/AC7VJLMOONVOQKPIL3XYVZDXY4QOXANCNFSM6AAAAAAQ4765KA . You are receiving this because you commented.Message ID: @.***>

-- Yaming Jin Nanjing National Laboratory of Microstructures, Physics Department, Nanjing University, Nanjing, 210093, P. R. China Tel:+86-25-8359 3940, Fax:+86-25-8359 5535 Cell Phone: +86-15601588976 E-mail: @.***

KiZemmy commented 10 months ago

代码已经发给你了,我做的是同时读取深度图像和RGB图像 因此在配置文件中要给定深度图的路径 同时主干网络也需要改写  给予两个输入 ------------------ 原始邮件 ------------------ 发件人: "open-mmlab/mmdetection" @.>; 发送时间: 2022年10月15日(星期六) 中午12:22 @.>; @.**@.>; 主题: Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939) can you give me your email, i sent the code for you , git the modified project on github is complicated for me . Thx ! My email is @. . — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.>

您好,现在还方便分享一下吗?我的邮箱是c_charyk@163.com 谢谢

zmaomia commented 8 months ago

@YihanWang-chengdu

代码已经发给你了,我做的是同时读取深度图像和RGB图像 因此在配置文件中要给定深度图的路径 同时主干网络也需要改写  给予两个输入 ------------------ 原始邮件 ------------------ 发件人: "open-mmlab/mmdetection" @.>; 发送时间: 2022年10月15日(星期六) 中午12:22 @.>; @.**@.>; 主题: Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939) can you give me your email, i sent the code for you , git the modified project on github is complicated for me . Thx ! My email is @. . — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.>

请问还方便分享代码吗?我的邮箱是 627821399@qq.com

goodday-94 commented 5 months ago

代码已经发给你了,我做的是同时读取深度图像和RGB图像 因此在配置文件中要给定深度图的路径 同时主干网络也需要改写  给予两个输入 ------------------ 原始邮件 ------------------ 发件人: "open-mmlab/mmdetection" @.>; 发送时间: 2022年10月15日(星期六) 中午12:22 @.>; @.**@.>; 主题: Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939) can you give me your email, i sent the code for you , git the modified project on github is complicated for me . Thx ! My email is @. . — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.>

Hello! Could you please share the code with me? Thank you a lot! ANd here is em email:qilinzhang94@gmail.com

JINAOLONG commented 5 months ago

代码已经发给你了,我做的是同时读取深度图像和RGB图像 因此在配置文件中要给定深度图的路径 同时主干网络也需要改写  给予两个输入 ------------------ 原始邮件 ------------------ 发件人: "open-mmlab/mmdetection" @.>; 发送时间: 2022年10月15日(星期六) 中午12:22 @.>; @.**@.>; 主题: Re: [open-mmlab/mmdetection] Creating multi-modal (bi-modal) detectors / pipelines for them (Issue #8939) can you give me your email, i sent the code for you , git the modified project on github is complicated for me . Thx ! My email is @. . — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.>

您好!我做的是融合RGB图像和深度图像,遇到了类似的问题,代码能共享一下吗?万分感谢!e-mail:17356725817@163.com

YangParky commented 5 months ago

I am doing the same thing as you . I have modified the Class : CustomDataset in mmdet.dataset.custom.py and transform.py 、loading.py . i relize the training and testing for dual-resnet with the input RGB and Depth image.

我也遇到了类似的问题,代码能共享一下吗?万分感谢!lopeture@gmail.com