Closed horcruxen closed 9 months ago
cmd: make -j"$(grep -c ^processor /proc/cpuinfo)"
if i execute this command,Raise error :“[ 1%] Built target layernorm_kernels;make: *** [Makefile:136: all] Error 2 ” ;
Have you take this error?
cmd: make -j"$(grep -c ^processor /proc/cpuinfo)" if i execute this command,Raise error :“[ 1%] Built target layernorm_kernels;make: *** [Makefile:136: all] Error 2 ” ; Have you take this error?
No, run in docker nvcr.io/nvidia/pytorch:22.09-py3
cmd: make -j"$(grep -c ^processor /proc/cpuinfo)" if i execute this command,Raise error :“[ 1%] Built target layernorm_kernels;make: *** [Makefile:136: all] Error 2 ” ; Have you take this error?
No, run in docker nvcr.io/nvidia/pytorch:22.09-py3
can i take friend with you , this is my wechat: “hlhaxb”, i hvae some question for this project consultant.
@horcruxen
Sorry for the late reply, this repository is not for CodeLlama but mainly focuses on CodeFuse-13B.
Although we have implemented support for the Llama structure internally, it is not yet open-sourced.
Using FasterTransformer currently might not be the most efficient way, you can use other open-source methods for inference. Perhaps VLLM or TensorRT-LLM?
Branch/Tag/Commit
main
Docker Image Version
nvcr.io/nvidia/pytorch:22.09-py3
GPU name
A6000
CUDA Driver
525.60.11
Reproduced Steps