issues
search
ray-project
/
ray-llm
RayLLM - LLMs on Ray
https://aviary.anyscale.com
Apache License 2.0
1.2k
stars
87
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Record telemetry when RayLLM is launched using a Serve config
#102
shrekris-anyscale
closed
7 months ago
0
No available node types can fulfill resource request defaultdict - error on local deployment
#101
NikolayTV
opened
7 months ago
1
Remote address refuse queries
#100
rifkybujana
opened
7 months ago
1
Is there a way to increase the scaling up speed?
#99
rifkybujana
opened
7 months ago
2
Support function calling models
#98
richardliaw
opened
7 months ago
0
Support for multi-modal models
#97
richardliaw
opened
7 months ago
0
Langchain integration
#96
XBeg9
closed
5 months ago
1
Add AWQ and SqueezeLLM quantization configs
#95
uvikas
closed
8 months ago
0
Multiple models second models always request GPU: 1
#94
lynkz-matt-psaltis
opened
8 months ago
2
Update deploy-on-gke.md
#93
ChaosEternal
opened
8 months ago
0
fix: no attribute 'set_url'
#92
marov
closed
8 months ago
0
Request for Comment: RayLLM <-> FastChat Integration
#91
Extremys
opened
8 months ago
2
LLM Deployment Observability
#90
roelschr
opened
8 months ago
3
rayllm's frontend can't work properly via rayllm:0.4.0 image
#89
k0286
closed
8 months ago
0
VLLM Ray Workers are being killed by GCS
#88
rtwang1997
closed
6 months ago
6
Rename ray-llm docker image name in doc
#87
sihanwang41
closed
8 months ago
0
fix: Add explanation for copying files
#86
enori
opened
8 months ago
0
Doc/Config update for rayllm
#85
sihanwang41
closed
8 months ago
0
fix: docker image references
#84
roelschr
closed
8 months ago
1
Update README.md to use new docker repo name
#83
ethanyanjiali
closed
8 months ago
1
Add AWQ Quantized Llama 2 70B Model Config & Update README
#82
YQ-Wang
closed
8 months ago
4
No example for quantized model
#81
jinnig
closed
8 months ago
2
[doc] Cannot deploy an LLM model on EKS with KubeRay
#80
enori
opened
8 months ago
3
0.4.0 release
#79
avnishn
closed
8 months ago
0
Anyscale Image
#78
harsh-goglocal
opened
8 months ago
2
Update README.md with link to kuberay instructions
#77
akshay-anyscale
closed
9 months ago
0
Issues serving other models from HF
#76
kenthua
closed
8 months ago
5
[Docs] "max_total_tokens" is missing in the doc
#75
scottsun94
opened
9 months ago
0
[docs] Improve docs around configuration
#74
richardliaw
opened
9 months ago
2
Deploying RayLLM locally failed with exit code 0 even if deployment is ready
#73
lamhoangtung
opened
9 months ago
1
Ray LLM on Nvidia RTX series?
#72
shahrukhx01
opened
9 months ago
3
Update method of accessing Serve controller
#71
shrekris-anyscale
closed
9 months ago
0
fixes cluster yaml to use the correct resource attribute naming
#70
JGSweets
closed
8 months ago
1
Set `route_prefixes` in Serve configs to `/`
#69
shrekris-anyscale
closed
9 months ago
0
Revert "Fix serve config parsing"
#68
Yard1
closed
9 months ago
0
[docs] Update the application's config to be compatible with v0.3.0
#67
YQ-Wang
closed
9 months ago
1
Fix serve config parsing
#66
gvspraveen
closed
9 months ago
0
Add Serve config for LightGPT
#65
shrekris-anyscale
closed
9 months ago
0
Add back LightGPT
#64
Yard1
closed
9 months ago
0
[docs] Update changes to RayLLM
#63
richardliaw
closed
9 months ago
0
RayLLM v0.3.0
#62
Yard1
closed
9 months ago
1
issue with run locally
#61
omlomloml
opened
9 months ago
1
ray-llm support for ML Accelerators (Google's TPU, AWS Inferential & etc)
#60
sudujr
closed
9 months ago
1
Minor Update for Model Warmup
#59
FerdinandZhong
closed
9 months ago
1
Follow the doc to deploy llama2 70b throws error
#58
YQ-Wang
closed
9 months ago
1
Embedding model support in ray-llm
#57
YQ-Wang
closed
9 months ago
1
[docs] GKE / EKS updates
#56
richardliaw
closed
10 months ago
0
S3 bucket model download fails silently if the cluster doesn't have the right permissions
#55
architkulkarni
opened
10 months ago
1
Create serve configs to deploy LLMs to production
#54
shrekris-anyscale
closed
10 months ago
0
Weight caching being based on model-id creates confusion
#53
ArturNiederfahrenhorst
opened
10 months ago
0
Previous
Next