jankais3r / LLaMA_MPS

Run LLaMA (and Stanford-Alpaca) inference on Apple Silicon GPUs.
GNU General Public License v3.0
583 stars 47 forks source link

op Height/Width dimensions must be less than 16384 #20

Open itrcz opened 1 year ago

itrcz commented 1 year ago

MacOS 14.0 MacBook Pro M1 Max

autocomplete and instruction-response give same result:

$python3 chat.py --ckpt_dir models/7B-alpaca --tokenizer_path models/tokenizer.model --max_batch_size 8 --max_seq_len 256 Seed: 30112 Loading checkpoint Loaded in 94.05 seconds Running the fine-tuned 'alpaca' model in an instruction-response mode. Instruction: hello loc("mps_transpose"("(mpsFileLoc): /AppleInternal/Library/BuildRoots/d8ee83b8-11b4-11ee-a66d-46d450270006/Library/Caches/com.apple.xbs/Sources/MetalPerformanceShadersGraph/mpsgraph/MetalPerformanceShadersGraph/Core/Files/MPSGraphUtilities.mm":206:0)): error: 'anec.transpose' op Height/Width dimensions must be less than 16384 Response: hello, hello

seems like it not works as expected. I have waited couple minutes and I got just "hello, hello"