-
Add support to multimodal models as discussed with @haileyschoelkopf
- This PR #1832 would be a great starting point.
- List all tasks we want to support for the first iteration.
-
Hello! I hope you're doing well. I'm interested in Multi-ion radiotherapy (MIRT) and would like to know if it's possible to use Matrad for creating MIRT plans. Is there a way to implement this?
rega…
-
I dowloaded checkpoint following 3.1 and running the app.sh following section 4. However, I found all multi modality instructions are not working properly.
```python
all_gen_img_idx: []
all_g…
-
AnyGPT is quite a promising project released 2 months before GPT4o.
It is a versatile multimodal *LLaMA-based* model, which is able not only to take images as an input, but also non-transcribed spe…
-
Glad to be the first one here!
Looking forward to the final release version of epiScanpy and it's full tutorial and the download link of test data.
By the way, Could epiScanpy be used as a multi-mod…
-
Hi,
If I want to utilize one of the industrial sensors as this multi modality model, how to get started? for example, replacing audio modality by industrial (analog to digital conveted) sensor if I…
-
### Is this your first time submitting a feature request?
- [X] I have searched the existing issues, and I could not find an existing issue for this feature
- [ ] I am requesting a straightforward ex…
-
springAiVersion: 0.8.1
This is an example I saw here, but based on my testing, it seems that the `gpt-4-vision-preview` model is not yet supported.
[Spring AI - Multimodality - Orbis Sensualium …
-
**The bug**
Loading and prompting the transformer model `openbmb/MiniCPM-Llama3-V-2_5` does not work.
It tries to load the model (but according to nvtop nothing is allocated on my gpu). No error is …
-
The currently followed architecture of is still too closely bound to traditional NLU based voice interaction concepts. While it aimed at including LLM with speech, LLM with multimodality, ... it is po…