-
### Anything you want to discuss about vllm.
flashinfer already supports sliding window in https://github.com/flashinfer-ai/flashinfer/issues/159 , and we should update our code and pass sliding wind…
-
I like the notion of 'living documentation' as provided by http://www.relishapp.com/ and http://getgauge.io/, however I think BDDfy has the correct model (code -> documentation).
I'd like to add supp…
-
Maybe this is "solved" but I'd like to see us create a series of small codelabs that are "living", that is, are updated as the packages are updated, and are required to pass (on travis, or at least on…
-
### Your current environment
```text
The output of `python collect_env.py`
```
### How you are installing vllm
```sh
pip install -vvv vllm
```
### Before submitting a new issue...
- [X] M…
-
Jenkins and Asciidoctor integration is not straightforward and always caused issues on every jenkins or plugin update. When we updated Cukedoctor (the library used to generate documentation from cucum…
-
### Your current environment
The output of `python collect_env.py`
```text
latest vllm 0.6.1
```
### Model Input Dumps
tt
### 🐛 Describe the bug
InternVL multi image speed is slower …
-
### 📚 The doc issue
vllm serve microsoft/Phi-3.5-vision-instruct --dtype auto --trust-remote-code --max_model_len 8192 --limit-mm-per-prompt image=2
raise self._make_status_error_from_res…
-
### Your current environment
```text
The output of `python collect_env.py`
```
python3.10.12
### How would you like to use vllm
I want to run inference of a [specific model](put link here). I do…
-
### Your current environment
```text
The output of `python collect_env.py`
```
### How would you like to use vllm
I am using Qwen2VL and have deployed an online server. Does it support online …
-
### What happened?
I am currently use living documentation with cucumber by adding readme file under feature level but I am trying same approach with Junit and I don't have success question is this…