Closed vansangpfiev closed 1 month ago
When building the binary, in window I met a problem that doesn't appear on Linux and Mac:
To resolve this please check this commit if you meet this error in the future
to use commands:
./cortex-cpp engines list
It will show result like this
The onnx engine'status will be not_supported
in Linux and Mac.
and if the engine was found the status will be ready
otherwise not_initialized
@nguyenhoangthuan99 @vansangpfiev I recommend for us to simplify the cortex engines list
to a simpler table
# | Name | Supported Formats | Version | Status |
---|---|---|---|---|
1 | ONNXRuntime | ONNX | 0.0.1 | Incompatible |
2 | llama.cpp | GGUF | 0.0.1 | ready |
3 | TensorRT-LLM | TensorRT Engines | 0.0.1 | Not installed |
I personally think the description and product_name is a bit extraneous, should keep this simple
For the TensorRT-LLM supported format, I would need to ask you guys what the "Supported Format" is - .engine files? (not sure)
@namchuai Can I reassign this to you, since you are the main owner of engines
API endpoint and CLI
@dan-homebrew yes, I will work on this. However, I have question regarding the semantic naming for engine which we have a discussion here: https://github.com/janhq/cortex.cpp/issues/1168
QAed with #1073 (see screenshots there)
Linked to open followup issue on #1168
command:
cortex-cpp engines list