Closed dan-homebrew closed 1 month ago
model.yaml
for three models8b-gguf-q4-km
This is an example command to run model with tag
cortex pull llama3.1:8b-gguf-q4-km
cortex run llama3.1:8b-gguf-q4-km
This concise command provides sufficient information for users.
[x] Develop CI runner for building all quantization for each model:
[x] Create script to update model.yaml
for models:
This approach will streamline model management and improve user experience when working with cortex.cpp.
This PR introduces two CI pipelines to streamline the model processing workflow:
This pipeline automates the process of converting and quantizing models.
After successful processing, new tags will be added to the model repository. For example, see the llama3 repository:
This pipeline updates the model.yml file with new information.
These pipelines automate crucial steps in model processing and metadata management, streamlining the workflow for model updates and deployments.
@nguyenhoangthuan99 how do we use this pipeline? i.e. how are we adding new models
The cortexso model repo must be created before running this pipeline (e.g. llama3
must be created before running below example, the hf login token in CI doesn't have permission to create repo)
Supported quantization levels: q2-k
, q3-ks
, q3-km
, q3-kl
, q4-ks
, q4-km
, q5-ks
, q5-km
, q6-k
, q8-0
To use this pipeline:
Select the Convert model to gguf with specified quant
workflow inside the action tab
Click on Run workflow
And input all parameters
Note that the Target HuggingFace model ID to push
is cortexso model repo, in my example it is llama3
After click run, go to tab Action
and we can see the workflow is running
When the CI is finished, we can go to the cortex so repo https://huggingface.co/cortexso/llama3, to check if the model is updated
@nguyenhoangthuan99 I am refactoring the "Built-in Model Library" to a separate epic: https://github.com/janhq/models/issues/21
Infra:
janhq/models
janhq/models
I add the updated model converter pipeline to janhq/models repo. And also add a pipeline to automatically update the model.yml file in hugging face cc @gabrielle-ong, now we can run CI pipeline in this repo.
Guild for update model.yml
file
Update model.yml with specific quant
Please update with the format
"top_p=0.9" "top_k=40" "stop=['<|end_of_text|>', '<|eot_id|>']"
Note that the prompt_template
field should not update this way because this field sometimes cannot handle proper special character on string.
Marking as complete, successfully done for mistral-nemo and llama3.2 To run model converter pipeline from janhq/models repo
Goal
model.yaml
with best parametersUser Story
Decisions
Tasklist
Model Compilation Pipeline
Future Roadmap