abetlen / llama-cpp-python

Python bindings for llama.cpp
https://llama-cpp-python.readthedocs.io
MIT License
8.14k stars 969 forks source link

llama_cpp installs into base instead of Conda for M1 installation #375

Open nivibilla opened 1 year ago

nivibilla commented 1 year ago

Prerequisites

Please answer the following questions for yourself before submitting an issue.

Expected Behavior

Install into the Conda library

Current Behavior

Installs into base python

Environment and Context

Please provide detailed information about your computer setup. This is important in case the issue is not reproducible except for under certain specific conditions.

M1 Mac

MacOS

python 3.9

Steps to Reproduce

Please provide detailed steps for reproducing the issue. We are not sitting in front of your screen, so the more detail the better.

follow instruction for m1 Mac but do it inside Conda env

abetlen commented 1 year ago

Can you install into the conda env with verbose output and paste that here? Should give me an idea of what's going wrong.

ianscrivener commented 1 year ago

@nivibilla check out the MacOS install instructions here: https://github.com/abetlen/llama-cpp-python/blob/main/docs/macos_install.md

did you skip the conda activate llama step?

nivibilla commented 1 year ago

Hey, no Im pretty sure I built it inside the conda env. I haven't got round to testing this again. Will try do it tomorrow

K-Mistele commented 10 months ago

+1 on this, copied/pasted the install commands from instructions and have uninstalled & reinstalled it multiple times in the llama env, but passing --n_gpu_layers -1 (-1 = all layers), or any other number of layers is SLOWER than CPU inference until I switch back to the base env, at which point it's super fast. Happy to make a loom demonstrating this :)

Macbook pro M2 Max here, latest version of llama-cpp-python, which as of today is 0.2.29