Closed Halfknow closed 1 year ago
start fastchat use new chatglm2-6b model,it wroks
python3 -m fastchat.serve.model_worker --model-names "gpt-3.5-turbo,text-davinci-003,text-embedding-ada-002" --device mps --model-path /Users/XXX/Projects/6b/chatglm2-6b
but get error when token max than 2048,"openai.error.APIError: Invalid response object from API: '{"object":"error","message":"This model\'s maximum context length is 2048 tokens. However, you requested 2895 tokens (2383 in the messages, 512 in the completion). Please reduce the length of the messages or completion.","code":40303}' (HTTP response code was 400)"" accord to chatglm2 document,chatglm2-6b support more than 2048 token
can it used as embedding model?wo need embedding when use vector store
chatglm2-6b has released with great updates bilingually, hope you guys can add this open llm in your website which will be really exciting. The github repo is here: https://github.com/THUDM/ChatGLM2-6B.