Closed jing-zhao9 closed 7 months ago
To be honest, I do not understand your question.
This code implements the classic behavior of a UNet model:
How to interpolate...
Could you please provide a detailed explanation of the process of downsampling and upsampling?
If you are asking about how the "sampling" is done (ie you are looking for a point sampling, or a grid voxelization, or farthest point sampling operation somewhere), then the answer is: it is a hierarchical superpoint partition that we build at preprocessing time. This is the core idea of this whole project, please check our paper https://arxiv.org/abs/2306.08045.
Thank you very much for your patient answer. I have another question about the code. If I want to train semantic segmentation with my first GPU and panoptic segmentation with my second GPU. So how should I set up my graphics card usage?
Look into CUDA_VISIBLE_DEVICES
.
Your history of issues in this repo suggest you are not too familiar with deep learning and that you have not read the paper in details. I have to warn you, the project at hand involves some fairly advanced deep learning concepts. Besides, I cannot provide tutoring here, only support for true issues related to the code, or aspects of the project that lack clarity.
When I encountered some confusion while debugging your code, I used the following code during the decoding phase
down_outputs:{list:2}=[(32538,64),(10227,64)]
How to interpolate to obtain
up_outputs={list: 1}=[(32538,64)]
and ultimately obtainout={list: 2}=[(32538,64), (10227,64)]
Could you please provide a detailed explanation of the process of downsampling and upsampling? Thank you very much!