-
First of all congrats on the paper and thanks for providing the code!
In the paper at 'Zero-shot language-based multi-modal joint retrieval' you mention that integrating/combining multiple embeddin…
-
In the file `llava/model/llava_arch.py` under the class `LlavaMetaForCausalLM` there is a function`prepare_inputs_labels_for_multimodal` that is called when calling the `generate` and `forward` functi…
-
great job! I want to know how to get pseudo pairs when I chose one modality(for example, Image) as a starting point. I can use audio-image and image-text model to retrieve audio and text, but how ca…
-
Hello, I would like to ask, the current code seems to support only one modality and text modality at a time of inference, is it possible to input multiple modal data (such as audio, video and text) at…
-
**🚀 Feature**
Transforms that shift voxel intensity, such as intensity flipping (i.e. 1 - val, val in [0, 1]), cluster and remap, contrast jitter etc.
**Motivation**
I am working on a spine segme…
-
Hello,
In the dataset_format documentation, it is specified that "The different input channels MUST have the same geometry (same shape, spacing (if applicable) etc.) and must be co-registered (if …
-
### What version of the bids app were you using?
bids/brainsuite:v23a
### Describe your problem in detail.
I was trying run BrainSuite for resting-state analysis, and it failed to run.
My datase…
-
**Is your feature request related to a problem? Please describe.**
I trained a model with the Auto3dSeg Autorunner and was trying to run inference using the SlicerMONAIAuto3DSeg extension (https://…
che85 updated
4 months ago
-
### Model/Pipeline/Scheduler description
Lumina-T2X is a text-to-any generation model. Our model is capable of generating multiple modalities, most notably image generation. Currently, our image ge…
-
### What feature or change would you like to see made?
This ticket aims to address several reported issues concerning the rendering and display of ultrasound images within the OHIF viewer. The issu…