Open xyb opened 1 week ago
Thanks for sharing that. Could please tell me the current config in your computer? And do you use the vlm feature?
As far as I know some embedding models may have issue about this. For example I have such issue when using https://huggingface.co/jinaai/jina-embeddings-v3
Here's my configuration:
base_dir: ~/.memos
database_path: database.db
default_library: screenshots
screenshots_dir: screenshots
server_host: 0.0.0.0
server_port: 8839
# Enable authentication by uncommenting the following lines
# auth_username: admin
# auth_password: changeme
default_plugins:
- builtin_ocr
- builtin_vlm
# using ollama as the vlm server
vlm:
concurrency: 8
endpoint: http://localhost:11434
force_jpeg: true
modelname: minicpm-v
prompt: 请帮描述这个图片中的内容,包括画面格局、出现的视觉元素等
token: ''
# using local ocr
ocr:
concurrency: 8
# this is not used if use_local is true
endpoint: http://localhost:5555/predict
force_jpeg: false
token: ''
use_local: true
# using local embedding
embedding:
# this is not used if use_local is true
endpoint: http://localhost:11434/api/embed
model: arkohut/jina-embeddings-v2-base-zh
num_dim: 768
use_local: true
use_modelscope: false
After tracking the memory using in embedding.py
, I found that the memory usage get a peak and then quickly fall down. But the monitor will keep showing the peak. Click the Details button can show the total usage looks like this:
After running the serve process for an extended period, I noticed it was consuming an unusually high amount of memory—approximately 5.5GB on my laptop. This suggests a potential memory leak that we should investigate and resolve.