Closed adampingel closed 2 months ago
Yes. Bigger models, newer versions, managed by the same process, and feeding the same BI/analytics reports.
For instance, it's not immediately clear to me that ollama isn't missing the -instruct
variants: https://ollama.com/library/granite-code
Please note the scope of this ticket is to freshen up the Ollama granite code models once laboriously. There is a follow-on ticket for automation and hardening (which will possibly grow beyond a single ticket): https://github.com/ibm-granite-cookbooks/pm/issues/28
Based on the pace of communication with the ollama
folks, we've decided to go with a staging approach. I'm using my personal account (gabegoodhart
). The 3b-128k
and 8b-128k
models are up now: https://ollama.com/gabegoodhart/granite-code:8b-128k
For posterity, here's the script I'm using to do the imports:
To do the push, you need to do the following:
ollama
accountcat ~/.ollama/id_ed25519.pub | pbcopy
ollama cp <import name> <username>/<model>:<tag>
ollama push <username>/<model>:<tag>
Ollama does show considerable metadata about the model, along with the license and template: https://ollama.com/gabegoodhart/granite-code:8b-128k/blobs/a8fe02e5a50c
This includes the context length as llama.context_length
.
Wow, great find! This metadata seems to all be parsed from the source gguf
file, so we may be able to embed additional metadata during the HF -> gguf conversion
It's already available. Do you mean newer models?