-
`gemma-2-27b-it` and `gemma-2-9b-it` were the highest ranked models under 70B in 2024-07-26.
However they disappeared from the latest release of 2024-08-31.
Even weirder, they're not on the [homepag…
-
### Your `minimal.lua` config
```lua
---@diagnostic disable: missing-fields
--NOTE: Set config path to enable the copilot adapter to work.
--It will search the follwoing paths for the for copilo…
-
For more info, see: https://github.com/mistralai/mistral-src and references there in.
Also: https://arxiv.org/pdf/2310.06825v1.pdf
With #3228 it should be relatively easy to support this.
-
Can I identify and analyze videos? How to input video? Do you have any examples,How much GPU is needed to run
-
### What is the issue?
When using the llm benchmark with ollama https://github.com/MinhNgyuen/llm-benchmark , I get around 80 t/s with gemma 2 2b. When asking the same questions to llama.cpp in conve…
-
Could anyone please advise if it is possible to run inference with OVIS 1.6 on a single 4090 GPU? After loading the model, it appears to consume approximately 20GB of VRAM. I attempted an inference, b…
-
## Description
I am encountering a timeout error when running the following code on macOS. The error occurs approximately 10 seconds after the request is made. I would like to know if there is a wa…
-
### What do you need?
After some experimentation, q8 llama3 and mistral **run locally** are performing **much** better after adjusting the prompt order. I propose restructuring the patterns so tha…
-
When will support for batch size > 1 be available, or where should I make modifications to enable this feature?
-
### What is the issue?
hi? I'm studying fine tuning.
I learned using the "unsloth/gemma-2-2b-it" model.
I created the dataset myself and it contains less than 100 cases.
I want to use only the fin…