-
I would like to know the details of cross-modal training, is there any relevant training code that is responsible for performing this part?
-
Hi! Thanks for open-sourcing APE, it is fantastic! 👍
I am new to the field of open-vocabulary vision foundation models, and I have some questions on the "gated cross-modality interaction" when goi…
-
Hi, thanks for your good job.
```
# Latent Fusion
def fusion(self, audio_tokens, visual_tokens):
# shapes
BS = audio_tokens.shape[0]
# concat all the tokens
…
-
Hi!
Great work! Congratulations! Thanks for releasing the code!
However, I am not able to reproduce the results for taskrunners using any of the `allenai/uio2-large`, `allenai/uio2-xl` or `allen…
-
Hi,
I wonder know if GLUE can generate cross modality data, like you have a pre-trained model, and if you can generate scATAC-seq data using the scRNA-seq data?
-
### Describe your problem in detail.
Currently, task events are presented as modality-specific.
### Describe what you expected.
That task events are cross-modality.
### BIDS specification section
…
-
Thanks for your work. I have some questions about the Abdomen CT (ABCT) dataset. Does this dataset derive from the cross-modality dataset Abdomen MR-CT in L2R? I found that I can't figure out the spl…
-
FYI @bherr2
# User story
As a consumer of the ubkg-api (actually, the hs-ontology-api deployment), I want to know how a dataset can be associated with a code in a standard biomedical ontology, such…
-
## 一言でいうと
VQAのような、画像+言語のタスクでTransformerを適用した研究。画像は物体領域の位置ベクトルを使ってSelf-Attention(物体間の関係を学習)、言語は通常通りSelf-Attention、最後にCross(言語to画像、画像to言語のAttention計算)をした後にSelf-Attentionをとって出力を行う。事前学習を通じSOTAを達成
![…
-
Hi, I have tried navigating your model via the instructions you have provided but with no success. After completing the set-up, I tried running a simple Text-to-Text inference but with no success. The…