-
decoder_base: {path to store the official weights of LLaMA-2}
which model is the official weights of LLaMA-2
-
Hi Team,
I have attempted Knowledge Distillation using Torchtune for the 8B and 1B Instruct models. However, I still need to apply KD to the Vision Instruct model. I followed the same steps and cre…
-
/kind bug
**What steps did you take and what happened:**
[A clear and concise description of what the bug is.]
I installed Kserve in k8s following steps here https://kserve.github.io/website/late…
-
Do you have plans to support other LLM models like Llama 3?
Or would it be easy to modify code implementing interface to OpenAI. I would like inerface using Ollama.
Any hints would be appreciate…
-
I would like to run Llama-3.2 11B Vision in KoboldCPP. Ollama recently added the support, so I guess it is possible to pull it here :)
Also, there is support needed for GGUF conversion!
-
Hello author.
We would like to fine-tune the method proposed in your paper on a new dataset. I now have extracted features from audio and visual. We currently only need to implement the emotion reco…
-
Dear authors,
@shuyansy @UnableToUseGit
I kindly think you need to discuss VoCo-LLaMA[1] in the "Intro" section of your paper at the very least.
As I find the citation and discussions related to …
-
Found out that **IBM ChatWatsonx** integration with **Langchain** and **IBM watsonx.ai Text Chat API** provide not consistent outputs for the same input with `decoding_method=greedy` or `temperature=0…
-
explain what a llm is in low detail and also tell me about some of the most notable llms and how they differ split your response into 2 sections
-