-
Hi, thank you for your open source codebase.
I was wondering, when might MoCa be supported in this repo? It was mentioned in the paper that this repository would be its code release.
-
Hello Adrian,
I’m using VXM for multi-modality (CT-MRI) image registration. In the beginning, I use the loss is [MI](https://github.com/voxelmorph/voxelmorph/blob/ed94aa2c7f43565ef5220a087c1fa4b07234…
-
original redmine: https://redmine.cbrain.mcgill.ca/issues/14507
-
Thanks for your excellent work!
Will the code for lidar and fusion modality be released?
-
### What is the feature?
I am looking into whether it is possible to export the BevFusion model to an external format like ONNX or torchscript.
### Any other context?
I have been able to load…
-
Thanks for this excellent codebase. I find for Waymo Open Dataset mmdet3d currently only processes the first image during data preprocessing and there is no multi-modality method implemented on Waymo.…
-
Hey, thanks a lot for your work!
I want to extend this model to more modalities(audio and video along with text and images). How difficult would that be? Also, if possible, how will that ?
-
[paper](https://arxiv.org/abs/2311.04257)
## TL;DR
- **I read this because.. :** very recent VLM model
- **task :** VLM + LLM
- **problem :** multi-modal task는 LLM freeze 시키고 사실상 V+L을 잘하려고…
-
- [ ] Landing page should be descriptive. Should give a nice overview of the project
- [ ] Claiming process should be streamlined. Discussions on this are most welcomed
- [ ] For deployment purposes…
-
only see the UniformSampleFrames class,because configs/rgbpose_conv3d/rgb_only.py used the MMUniformSampleFrames class