-
### Describe the issue
Running a model for N iterations in a single ONNX session is way faster than running the same model in 2 independent sessions, each session is run for N/2 iterations each.
¿W…
-
Hi, thank you for your nice work!
I'm reproducing the results in Table 2, using Mistral-7B model on MMLU and TydiQA and select 5% data.
I adhere to the scripts in your repo to conduct the wa…
-
### Description
we need to train wav2vec2 model for specific speaker accent and compare the performance with the base model on test data of that particular speaker.
### Completion Criteria
A model th…
-
I noticed the article's 3D backbone is pretrained model like resnet, I wonder if the model is pretrained independently on other datasets like kinetics400 and the used in YOWOv3 directly? I mean, how t…
-
## 🚀 Feature
Add early stopping functionality to DeepChem’s model training modules. This would allow models to stop training when improvements in validation loss or accuracy plateau, enhancing effi…
-
Hello,
I'm wondering if the models supported in this project have been evaluated to see if their performance replicates the reported metrics in the original papers (primarily FID and CLIP score f…
gtuzi updated
1 month ago
-
Hi guys,
I have a question regarding the performance impact and potential optimizations for distributing a large model across multiple GPUs. Specifically:
When running a 70B parameter model, how d…
-
### Your current environment
[Performance] 100% performance drop using multiple lora vs no lora(qwen-chat model)
gpu: 4 * T4
vllm version: v0.5.4
model:qwenhalf-14b-chat
### Model Input Dumps…
-
Thanks for releasing the code for the great work.
I utilized the VideoLLaMA2-7B (chat) model for offline reasoning and noted significant discrepancies compared to its online performance using the sam…
-
### Proposal to improve performance
_No response_
### Report of performance regression
_No response_
### Misc discussion on performance
Hi,
Thank you for your contribution to the LLM community…