-
## 🚀 Feature
[Documentation says](https://lightning.ai/docs/pytorch/latest/advanced/compile.html#limitations) that torch compile is not supported over distributed training right now. Since torch co…
-
Wonderful work! Following Q and looking forward ur reply.
1) I am curious about the method in your paper that copy the KV cache from cpu memory to gpu memory.
Since I have test the following…
-
Proposal to create a MaaS(Model as a Service) API family:
MaaS (Model as a Services) refers to the packaging of AI models and their associated capabilities into reusable services, enabling users to…
-
```
accelerate launch -m lm_eval \
--model hf \
--model_args pretrained=mistralai/Mistral-7B-v0.3 \
--tasks winogrande \
--num_fewshot 0 \
--batch_size 16 \
```
I tr…
-
Hello! I had a thought. To minimize constant load for tasks that occur infrequently, is there a way to keep the Docker container running with the HTTP server, but only load the model when a query is m…
-
Similar like what authors shown in official git repo, can use this efficient-kan model for continual learning settings. . For using efficient-kan for CL settings, I haven't found some attributes that …
-
This is the output of the model.
python -W ignore llava/eval/run_vila.py \
--model-path Efficient-Large-Model/Llama-3-VILA1.5-8b \
--conv-mode llama_3 \
--query "\…
-
Has the bug of the efficient-b0 model size 40M fixed?
Thanks.
clhne updated
3 years ago
-
`Distribution.sample()` evaluates all distribution parameters, and then samples from the resulting distribution, this means that if parameters are RVs, only one sample is taken. For 'full model' sampl…
-
Solvers require calculating a next sample vector from evaluating the error and/or local gradient of the forward model with respect to the measured data.
In LiberTEM, the data and computation can be…