Closed shiwanlin closed 3 months ago
found the problem:
For running OSX using mlx, this change from the testing codes in the read.me is needed:
import mlx.core as mx
generation_output = model.generate(
# input_tokens['input_ids'].cuda(), --->
mx.array(input_tokens['input_ids']),
closed.
Has anyone runs into this trying to run llama 3.1 405B, following the instructions in this repo? It happens apparently after the model was loaded fine.
====================
have to fix this error earlier:
alueError: [load] Input must be a file-like object opened in binary mode, or string
by adding str() to the following line in "airllm/persist/mlx_model_persister.py"
layer_state_dict = mx.load(str(to_load_path))
==================== This is pristine Apple native Conda env on a Mac M1 Pro:
tried mlx 16.1, 15.2, and 14.1, all with the same error.
airllm version: 2.9.1 for py3
====================
The input into self.weight[x] is a tensor type, which corresponds to the input tokens:
I guess it's an API mismatching or the sort?
Any help is appreciated.