OpenBMB / llama.cpp

Port of Facebook's LLaMA model in C/C++
MIT License
65 stars 13 forks source link

llama-cpp-python support #12

Closed cegutica closed 1 month ago

cegutica commented 3 months ago

Prerequisites

Feature Description

Hi, I have tested the MiniCPM-Llama3-V-2_5 model with my own data using the code provided in this repository and it works really nice. I saw you have also uploaded quantified versions of this model here and I would like to try the model using the library llama-cpp-python. Is there a way to do so currently? If not, it would be very useful that this model can be run with this library as well.

Motivation

This library provides a simple way to use models in GGUF format and is very popular in the community.

Possible Implementation

No response

thiner commented 3 months ago

I think that only happens after llama.cpp merge this fork code into its main branch, then we can expect llama-cpp-python add model support on their end.

Kkordik commented 2 months ago

As I see it is already supported by llama.cpp: https://github.com/ggerganov/llama.cpp/tree/master#:~:text=model.md)-,Multimodal%20models%3A,-LLaVA%201.5%20models

What about llama-cpp-python ?

tc-mb commented 1 month ago

@all Hi, I don't always pay attention to the issue area in this fork code repo.

  1. If this issue still needs to be answered, please raise an issue in the main repo with "llamacpp" label. I will respond very quickly.
  2. If this issue no longer needs to be answered, I will close it this week.