KoboldAI / KoboldAI-Client

For GGUF support, see KoboldCPP: https://github.com/LostRuins/koboldcpp
https://koboldai.com
GNU Affero General Public License v3.0
3.51k stars 758 forks source link

Is the disk cache id supposed to be N/A? #231

Open Katsu2098 opened 1 year ago

Katsu2098 commented 1 year ago
 0  |      14  |  NVIDIA GeForce RTX 3080
         N/A  |      14  |  (Disk cache)
         N/A  |       0  |  (CPU)

INIT | Loading model tensors: 100%|##########| 341/341 [01:44<00:00, 3.26it/s] Exception in thread Thread-15: Traceback (most recent call last): File "B:\python\lib\threading.py", line 932, in _bootstrap_inner self.run() File "B:\python\lib\threading.py", line 870, in run self._target(*self._args, self._kwargs) File "B:\python\lib\site-packages\socketio\server.py", line 731, in _handle_event_internal r = server._trigger_event(data[0], namespace, sid, data[1:]) File "B:\python\lib\site-packages\socketio\server.py", line 756, in _trigger_event return self.handlers[namespace][event](args) File "B:\python\lib\site-packages\flask_socketio__init.py", line 282, in _handler return self._handle_event(handler, message, namespace, sid, File "B:\python\lib\site-packages\flask_socketio\init__.py", line 826, in _handle_event ret = handler(args) File "aiserver.py", line 466, in g return f(a, k) File "aiserver.py", line 3914, in get_message load_model(use_gpu=msg['use_gpu'], gpu_layers=msg['gpu_layers'], disk_layers=msg['disk_layers'], online_model=msg['online_model']) File "aiserver.py", line 2651, in load_model move_model_to_devices(model) File "aiserver.py", line 941, in move_model_to_devices accelerate.utils.set_module_tensor_to_device(model, key, target_dtype) File "B:\python\lib\site-packages\accelerate\utils\modeling.py", line 118, in set_module_tensor_to_device raise ValueError(f"{tensor_name} is on the meta device, we need a value to put in on {device}.") ValueError: bias is on the meta device, we need a value to put in on torch.float16.