As of #769, our test units have been overhauled. This issue documents the progress to pass all those tests.
Current status:
Total test count: 183
Passed tests: 128
Partial fail: 23
Complete fail: 27
Untested: 5
To run a test:
pip install -r requirements-dev.txt
Then
pytest tests/your_test_module.py
Tests that currently pass will be marked with a ✅, tests that do not pass will be marked with ❌, and untested ones will be left blank (default).
[!NOTE]
A failed test does not mean that the associated feature does not work. A test may have many items (sometimes hundreds). The number of passed items will be logged for each test. Some feature tests may completely fail, but still work end-to-end.
The following features are known to be currently broken:
Llava-based Vision Model Loading
Out-Of-Tree model registration
General Tests
[x] test_cache_block_hashing
[x] test_config
[x] test_embedded_commit
[x] test_inputs
[x] test_logits_processor
[ ] test_regression ❌ (3/4) -- the vram release test fails
[x] test_sampling_params
[x] test_scalartype
[x] test_sequence
[x] test_sharded_state_loader
[x] test_utils
Async Aphrodite
[ ] test_api_server_async_aphrodite ❌
[x] test_async_aphrodite
[x] test_chat_template
[ ] test_openapi_server_ray ❌ (2/3)
[x] test_request_tracker
Basic Correctness
[x] test_basic_correctness
[ ] test_chunked_prefill ❌ (30/36)
[x] test_cpu_offload
[ ] test_preemption ❌ (3/5)
Compilation
[x] test_full_graph
Core
[x] test_block_manager
[x] test_chunked_prefill_scheduler
[x] test_scheduler_encoder_decoder
[x] test_scheduler
Distributed
[x] test_basic_distributed_correctness
[x] test_basic_distributed_correctness_enc_dec
[x] test_chunked_prefill_distributed
[x] test_comm_ops
[x] test_custom_all_reduce
[ ] test_distributed_oot ❌ (0/1)
[ ] test_multimodal_broadcast ❌ (0/6)
[ ] test_pipeline_parallel ❌ (1/10)
[ ] test_pipeline_partition ❌ (0/1)
[x] test_pp_cudagraph
[x] test_pynccl
[x] test_same_node (run with APHRODITE_TEST_SAME_HOST=1 torchrun --nproc-per-node=4 tests/distributed/test_same_node.py)
[x] test_shm_broadcast
Endpoints
OpenAI
[ ] test_audio ❌ (1/4)
[x] test_basic
[ ] test_chat ❌ (25/33)
[ ] test_completion ❌ (77/112)
[x] test_embedding
[x] test_encoder_decoder
[x] test_guided_processors
[x] test_metrics
[x] test_models
[ ] test_mp_api_server (takes too long to run, investigate)
[ ] test_oot_registeration ❌ (0/1)
[x] test_return_tokens_as_ids
[x] test_run_batch
[x] test_serving_chat
[ ] test_shutdown ❌ (0/1)
[x] test_tokenization
[ ] test_vision ❌ (0/16) -- seems to be issues with fetching the images
As of #769, our test units have been overhauled. This issue documents the progress to pass all those tests.
Current status:
To run a test:
Then
Tests that currently pass will be marked with a ✅, tests that do not pass will be marked with ❌, and untested ones will be left blank (default).
The following features are known to be currently broken:
General Tests
Async Aphrodite
Basic Correctness
Compilation
Core
Distributed
APHRODITE_TEST_SAME_HOST=1 torchrun --nproc-per-node=4 tests/distributed/test_same_node.py
)Endpoints
OpenAI
LLM
Engine
Output Processor
Kernels
LoRA
Metrics
Modeling
Models
Multimodal
Prefix Caching
Prompt Adapter
Quantization
Samplers
Spec Decode
End-to-end spec decode tests
Tensorizer Loader
Tokenization
Weight Loading
Worker