abetlen / llama-cpp-python

Python bindings for llama.cpp
https://llama-cpp-python.readthedocs.io
MIT License
7.77k stars 934 forks source link

Unable to build on windows (llava cmake issue) #898

Closed tk-master closed 10 months ago

tk-master commented 10 months ago

Hey @abetlen can you take a look at this ? https://github.com/abetlen/llama-cpp-python/commit/aab74f0b2bd25cc2a9baeb743a2057dd5cada6e4#commitcomment-132151477 Im still unable to build without the change i mention

abetlen commented 10 months ago

@tk-master certainly, can you just confirm this bug also occurs in llama.cpp if you build with the same flags ie -DLLAMA_CUBLAS=ON -DBUILD_SHARED_LIBS, if so we should make an issue / PR in llama.cpp to fix this. Unfortunately I don't run Windows and the CI environment doesn't suppot CUDA so there's no way I can really test a fix.

tk-master commented 10 months ago

@tk-master certainly, can you just confirm this bug also occurs in llama.cpp if you build with the same flags ie -DLLAMA_CUBLAS=ON -DBUILD_SHARED_LIBS, if so we should make an issue / PR in llama.cpp to fix this. Unfortunately I don't run Windows and the CI environment doesn't suppot CUDA so there's no way I can really test a fix.

@abetlen I can try.. can you please tell me what exactly I should put in -DBUILD_SHARED_LIBS= ? llava_shared?

abetlen commented 10 months ago

@tk-master my bad typo there -DBUILD_SHARED_LIBS=ON

tk-master commented 10 months ago

@abetlen right.. well llama.cpp from master seems to build fine with those params so the issue must be elsewhere?

abetlen commented 10 months ago

@tk-master yeah do you mind posting the full logs of both? Like pip install with --verbose and the cmake build logs for llama.cpp.

tk-master commented 10 months ago

Lets see.. this is the log after I run the following:

git clone https://github.com/ggerganov/llama.cpp
cd llama.cpp

mkdir build
cd build
cmake .. -DLLAMA_CUBLAS=ON -DBUILD_SHARED_LIBS=ON -DLLAMA_AVX=off -DLLAMA_AVX2=off -DLLAMA_FMA=off
cmake --build . --config Release --verbose

Log:

Cloning into 'llama.cpp'...
remote: Enumerating objects: 12057, done.
remote: Counting objects: 100% (12057/12057), done.
remote: Compressing objects: 100% (3608/3608), done.
remote: Total 12057 (delta 8418), reused 11966 (delta 8374), pack-reused 0
Receiving objects: 100% (12057/12057), 14.17 MiB | 8.78 MiB/s, done.
Resolving deltas: 100% (8418/8418), done.
-- Building for: Visual Studio 17 2022
-- The C compiler identification is MSVC 19.37.32825.0
-- The CXX compiler identification is MSVC 19.37.32825.0
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working C compiler: I:/Program Files/Microsoft Visual Studio/2022/Community/VC/Tools/MSVC/14.37.32822/bin/Hostx64/x64/cl.exe - skipped
-- Detecting C compile features
-- Detecting C compile features - done
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: I:/Program Files/Microsoft Visual Studio/2022/Community/VC/Tools/MSVC/14.37.32822/bin/Hostx64/x64/cl.exe - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed
-- Looking for pthread_create in pthreads
-- Looking for pthread_create in pthreads - not found
-- Looking for pthread_create in pthread
-- Looking for pthread_create in pthread - not found
-- Found Threads: TRUE
-- Found CUDAToolkit: C:/Program Files/NVIDIA GPU Computing Toolkit/CUDA/v12.3/include (found version "12.3.52")
-- cuBLAS found
-- The CUDA compiler identification is NVIDIA 12.3.52
-- Detecting CUDA compiler ABI info
-- Detecting CUDA compiler ABI info - done
-- Check for working CUDA compiler: C:/Program Files/NVIDIA GPU Computing Toolkit/CUDA/v12.3/bin/nvcc.exe - skipped
-- Detecting CUDA compile features
-- Detecting CUDA compile features - done
-- Using CUDA architectures: 52;61;70
-- CMAKE_SYSTEM_PROCESSOR: AMD64
-- CMAKE_GENERATOR_PLATFORM:
-- x86 detected
-- Performing Test HAS_AVX_1
-- Performing Test HAS_AVX_1 - Failed
-- Performing Test HAS_AVX_2
-- Performing Test HAS_AVX_2 - Failed
-- Performing Test HAS_AVX2_1
-- Performing Test HAS_AVX2_1 - Failed
-- Performing Test HAS_AVX2_2
-- Performing Test HAS_AVX2_2 - Failed
-- Performing Test HAS_FMA_1
-- Performing Test HAS_FMA_1 - Failed
-- Performing Test HAS_FMA_2
-- Performing Test HAS_FMA_2 - Failed
-- Performing Test HAS_AVX512_1
-- Performing Test HAS_AVX512_1 - Failed
-- Performing Test HAS_AVX512_2
-- Performing Test HAS_AVX512_2 - Failed
-- Configuring done (24.5s)
-- Generating done (0.7s)
-- Build files have been written to: I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build
Change Dir: 'I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build'

Run Build Command(s): "I:/Program Files/Microsoft Visual Studio/2022/Community/MSBuild/Current/Bin/amd64/MSBuild.exe" ALL_BUILD.vcxproj /p:Configuration=Release /p:Platform=x64 /p:VisualStudioVersion=17.0 /v:n
MSBuild version 17.7.2+d6990bcfa for .NET Framework
Build started 10/11/2023 21:28:51.

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" on node 1 (default targets).
Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\ZERO_CHECK.vcxproj" (2) on node 1 (default targets).
PrepareForBuild:
  Creating directory "x64\Release\ZERO_CHECK\".
  Creating directory "x64\Release\ZERO_CHECK\ZERO_CHECK.tlog\".
InitializeBuildStatus:
  Creating "x64\Release\ZERO_CHECK\ZERO_CHECK.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "x64\Release\ZERO_CHECK\ZERO_CHECK.tlog\unsuccessfulbuild".
CustomBuild:
  1>Checking Build System
FinalizeBuildStatus:
  Deleting file "x64\Release\ZERO_CHECK\ZERO_CHECK.tlog\unsuccessfulbuild".
  Touching "x64\Release\ZERO_CHECK\ZERO_CHECK.tlog\ZERO_CHECK.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ZERO_CHECK.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\common\build_info.vcxproj" (3) on node 1 (default targets).
PrepareForBuild:
  Creating directory "build_info.dir\Release\".
  Creating directory "build_info.dir\Release\build_info.tlog\".
InitializeBuildStatus:
  Creating "build_info.dir\Release\build_info.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "build_info.dir\Release\build_info.tlog\unsuccessfulbuild".
CustomBuild:
  Generating build details from Git
  -- Found Git: C:/Program Files/Git/cmd/git.exe (found version "2.38.1.windows.1")
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/common/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"C:\Pro
  gram Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D
   WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION
  =2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\""
   /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"build_info.dir\Release\\" /Fd"build_info.
  dir\Release\build_info.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\comm
  on\build-info.cpp"
  build-info.cpp
Lib:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\Lib.exe /OUT:"build
  _info.dir\Release\build_info.lib" /NOLOGO /MACHINE:X64  /machine:x64 "build_info.dir\Release\build-info.obj"
  build_info.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\build_info.dir\Release\build_info.lib
FinalizeBuildStatus:
  Deleting file "build_info.dir\Release\build_info.tlog\unsuccessfulbuild".
  Touching "build_info.dir\Release\build_info.tlog\build_info.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\build_info.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\ggml.vcxproj" (4) on node 1 (default targets).
PrepareForBuild:
  Creating directory "ggml.dir\Release\".
  Creating directory "ggml.dir\Release\ggml.tlog\".
InitializeBuildStatus:
  Creating "ggml.dir\Release\ggml.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "ggml.dir\Release\ggml.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/CMakeLists.txt
AddCudaCompileDeps:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\cl.exe /E /nologo /
  showIncludes /TP /D__CUDACC__ /D__CUDACC_VER_MAJOR__=12 /D__CUDACC_VER_MINOR__=3 /D_WINDOWS /DNDEBUG /DGGML_USE_CUBLA
  S /DGGML_CUDA_DMMV_X=32 /DGGML_CUDA_MMV_Y=1 /DK_QUANTS_PER_ITERATION=2 /DGGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D_CRT_SEC
  URE_NO_WARNINGS /D_XOPEN_SOURCE=600 /DCMAKE_INTDIR="Release" /D_MBCS /DWIN32 /D_WINDOWS /DNDEBUG /DGGML_USE_CUBLAS /D
  GGML_CUDA_DMMV_X=32 /DGGML_CUDA_MMV_Y=1 /DK_QUANTS_PER_ITERATION=2 /DGGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D_CRT_SECURE_
  NO_WARNINGS /D_XOPEN_SOURCE=600 /DCMAKE_INTDIR="Release" /II:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\. /I"C:\Progra
  m Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\
  bin" /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /I. /FIcuda_runtime.h /c I:\ai-chat_v1\llam
  a.cpp\llama-cpp\llama.cpp\ggml-cuda.cu
Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.vcxproj" (4) is building "I:\ai-chat_v1\llama.cpp\llama
-cpp\llama.cpp\build\ggml.vcxproj" (4:2) on node 1 (CudaBuildCore target(s)).
CudaBuildCore:
  Compiling CUDA source file ..\ggml-cuda.cu...
  cmd.exe /C "C:\Users\Tkay\AppData\Local\Temp\tmpd121bf8a9dcc4b49acdd93e5e0d15775.cmd"
  "C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\bin\nvcc.exe"  --use-local-env -ccbin "I:\Program Files\Mic
  rosoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64" -x cu   -I"I:\ai-chat_v1\llama.cpp\lla
  ma-cpp\llama.cpp\." -I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" -I"C:\Program Files\NVIDIA G
  PU Computing Toolkit\CUDA\v12.3\include"     --keep-dir x64\Release -use_fast_math -maxrregcount=0   --machine 64 --c
  ompile -cudart static --generate-code=arch=compute_52,code=[compute_52,sm_52] --generate-code=arch=compute_61,code=[c
  ompute_61,sm_61] --generate-code=arch=compute_70,code=[compute_70,sm_70] -Xcompiler="/EHsc -Ob2"   -D_WINDOWS -DNDEBU
  G -DGGML_USE_CUBLAS -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMV_Y=1 -DK_QUANTS_PER_ITERATION=2 -DGGML_CUDA_PEER_MAX_BATCH_S
  IZE=128 -D_CRT_SECURE_NO_WARNINGS -D_XOPEN_SOURCE=600 -D"CMAKE_INTDIR=\"Release\"" -D_MBCS -DWIN32 -D_WINDOWS -DNDEBU
  G -DGGML_USE_CUBLAS -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMV_Y=1 -DK_QUANTS_PER_ITERATION=2 -DGGML_CUDA_PEER_MAX_BATCH_S
  IZE=128 -D_CRT_SECURE_NO_WARNINGS -D_XOPEN_SOURCE=600 -D"CMAKE_INTDIR=\"Release\"" -Xcompiler "/EHsc /W3 /nologo /O2
  /FS   /MD /GR" -Xcompiler "/Fdggml.dir\Release\ggml.pdb" -o ggml.dir\Release\ggml-cuda.obj "I:\ai-chat_v1\llama.cpp\l
  lama-cpp\llama.cpp\ggml-cuda.cu"

  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build>"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\bin\nvcc.
  exe"  --use-local-env -ccbin "I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\H
  ostX64\x64" -x cu   -I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\." -I"C:\Program Files\NVIDIA GPU Computing Toolki
  t\CUDA\v12.3\include" -I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include"     --keep-dir x64\Release
   -use_fast_math -maxrregcount=0   --machine 64 --compile -cudart static --generate-code=arch=compute_52,code=[compute
  _52,sm_52] --generate-code=arch=compute_61,code=[compute_61,sm_61] --generate-code=arch=compute_70,code=[compute_70,s
  m_70] -Xcompiler="/EHsc -Ob2"   -D_WINDOWS -DNDEBUG -DGGML_USE_CUBLAS -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMV_Y=1 -DK_Q
  UANTS_PER_ITERATION=2 -DGGML_CUDA_PEER_MAX_BATCH_SIZE=128 -D_CRT_SECURE_NO_WARNINGS -D_XOPEN_SOURCE=600 -D"CMAKE_INTD
  IR=\"Release\"" -D_MBCS -DWIN32 -D_WINDOWS -DNDEBUG -DGGML_USE_CUBLAS -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMV_Y=1 -DK_Q
  UANTS_PER_ITERATION=2 -DGGML_CUDA_PEER_MAX_BATCH_SIZE=128 -D_CRT_SECURE_NO_WARNINGS -D_XOPEN_SOURCE=600 -D"CMAKE_INTD
  IR=\"Release\"" -Xcompiler "/EHsc /W3 /nologo /O2 /FS   /MD /GR" -Xcompiler "/Fdggml.dir\Release\ggml.pdb" -o ggml.di
  r\Release\ggml-cuda.obj "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\ggml-cuda.cu"
  ggml-cuda.cu
  tmpxft_00001e08_00000000-7_ggml-cuda.compute_70.cudafe1.cpp
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.vcxproj" (CudaBuildCore target(s)).

ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo
   /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X
  =32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS
  /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /std:
  c11 /Fo"ggml.dir\Release\\" /Fd"ggml.dir\Release\ggml.pdb" /external:W0 /Gd /TC /errorReport:queue  /external:I "C:/P
  rogram Files/NVIDIA GPU Computing Toolkit/CUDA/v12.3/include" "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\ggml.c" "I
  :\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\ggml-alloc.c" "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\ggml-backend.c"
   "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\ggml-quants.c"
  ggml.c
  ggml-alloc.c
  ggml-backend.c
  ggml-quants.c
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\ggml-quants.c(532,26): warning C4244: '=': conversion from 'float' to 'int8
_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\ggml-quants.c(750,36): warning C4244: '=': conversion from 'float' to 'int8
_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\ggml-quants.c(751,36): warning C4244: '=': conversion from 'float' to 'int8
_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.vcxproj]
  Generating Code...
Lib:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\Lib.exe /OUT:"ggml.
  dir\Release\ggml.lib" /NOLOGO /MACHINE:X64  /machine:x64 "ggml.dir\Release\ggml-cuda.obj"
  ggml.dir\Release\ggml.obj
  "ggml.dir\Release\ggml-alloc.obj"
  "ggml.dir\Release\ggml-backend.obj"
  "ggml.dir\Release\ggml-quants.obj"
  ggml.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml.lib
FinalizeBuildStatus:
  Deleting file "ggml.dir\Release\ggml.tlog\unsuccessfulbuild".
  Touching "ggml.dir\Release\ggml.tlog\ggml.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\llava\llava.vcxproj" (5) on node 1 (default targets).
Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxproj" (5) is building "I:\ai-chat_v1
\llama.cpp\llama-cpp\llama.cpp\build\llama.vcxproj" (6) on node 1 (default targets).
PrepareForBuild:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\".
  Creating directory "llama.dir\Release\llama.tlog\".
InitializeBuildStatus:
  Creating "llama.dir\Release\llama.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "llama.dir\Release\llama.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo
   /W3 /WX- /diagnostics:column /O2 /Ob2 /D _WINDLL /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D LLAMA_SHARED /D LLAMA_BU
  ILD /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_
  BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /D llama_EXPORTS /Gm- /E
  Hsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"llama.dir\Release\\" /Fd"llama.dir\Release\vc143.
  pdb" /external:W0 /Gd /TP /errorReport:queue  /external:I "C:/Program Files/NVIDIA GPU Computing Toolkit/CUDA/v12.3/i
  nclude" "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\llama.cpp"
  llama.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\llama.cpp(1103,31): warning C4305: 'initializing': truncation from 'double'
 to 'float' [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\llama.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\llama.cpp(2331,69): warning C4566: character represented by universal-chara
cter-name '\u010A' cannot be represented in the current code page (1252) [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\b
uild\llama.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\llama.cpp(9054,28): warning C4146: unary minus operator applied to unsigned
 type, result still unsigned [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\llama.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\llama.cpp(9084,28): warning C4146: unary minus operator applied to unsigned
 type, result still unsigned [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\llama.vcxproj]
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\Release\".
PreLinkEvent:
  Auto build dll exports
  setlocal
  cd I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build
  if %errorlevel% neq 0 goto :cmEnd
  I:
  if %errorlevel% neq 0 goto :cmEnd
  "C:\Program Files\CMake\bin\cmake.exe" -E __create_def I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/llama.dir/Re
  lease/exports.def I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/llama.dir/Release//objects.txt
  if %errorlevel% neq 0 goto :cmEnd
  :cmEnd
  endlocal & call :cmErrorLevel %errorlevel% & goto :cmDone
  :cmErrorLevel
  exit /b %1
  :cmDone
  if %errorlevel% neq 0 goto :VCEnd
  :VCEnd
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\llama.dll" /INCREMENTAL:NO /NOLOGO /LIBP
  ATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" "C:\Program Files\NVIDIA GPU Computing Toolkit
  \CUDA\v12.3\lib\x64\cudart.lib" "C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64\cublas.lib" "C:\Pro
  gram Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64\cublasLt.lib" cudadevrt.lib cudart_static.lib kernel32.lib
   user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib /DEF:"I:/ai-
  chat_v1/llama.cpp/llama-cpp/llama.cpp/build/llama.dir/Release/exports.def" /MANIFEST /MANIFESTUAC:"level='asInvoker'
  uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/bin/Release/llama.pdb" /SUB
  SYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/Release/lla
  ma.lib" /MACHINE:X64  /machine:x64 /DLL llama.dir\Release\llama.obj
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-alloc.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-backend.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-quants.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-cuda.obj"
     Creating library I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/Release/llama.lib and object I:/ai-chat_v1/llam
  a.cpp/llama-cpp/llama.cpp/build/Release/llama.exp
  llama.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\llama.dll
FinalizeBuildStatus:
  Deleting file "llama.dir\Release\llama.tlog\unsuccessfulbuild".
  Touching "llama.dir\Release\llama.tlog\llama.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\llama.vcxproj" (default targets).

PrepareForBuild:
  Creating directory "llava.dir\Release\".
  Creating directory "llava.dir\Release\llava.tlog\".
InitializeBuildStatus:
  Creating "llava.dir\Release\llava.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "llava.dir\Release\llava.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/llava/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\." /I"I
  :\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\..\.." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examp
  les\llava\..\..\common" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing To
  olkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D LL
  AMA_SHARED /D LLAMA_BUILD /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2
  /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /G
  m- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"llava.dir\Release\\" /Fd"llava.dir\Release\l
  lava.pdb" /external:W0 /Gd /TP /errorReport:queue  /external:I "C:/Program Files/NVIDIA GPU Computing Toolkit/CUDA/v1
  2.3/include" "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\llava.cpp" "I:\ai-chat_v1\llama.cpp\llama-cp
  p\llama.cpp\examples\llava\clip.cpp"
  llava.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\llava.cpp(32,24): warning C4244: 'initializing': conversion
from 'double' to 'float', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava
.vcxproj]
  clip.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(251,20): warning C4267: 'initializing': conversion
from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.v
cxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(465,9): warning C4297: 'clip_model_load': function
assumed not to throw an exception but does [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxp
roj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(465,9): message : __declspec(nothrow), throw(), noe
xcept(true), or noexcept was specified on the function [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llav
a\llava.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(714,46): warning C4267: 'argument': conversion from
 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxpr
oj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(817,44): warning C4244: 'initializing': conversion
from 'const _Ty' to 'uint8_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\
llava.vcxproj]
          with
          [
              _Ty=float
          ]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(817,34): warning C4244: 'initializing': conversion
from 'const _Ty' to 'const uint8_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\
llava\llava.vcxproj]
          with
          [
              _Ty=float
          ]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(855,20): warning C4267: 'initializing': conversion
from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.v
cxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(999,88): warning C4244: 'argument': conversion from
 'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxp
roj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(999,71): warning C4267: 'argument': conversion from
 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxpr
oj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1002,88): warning C4244: 'argument': conversion fro
m 'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcx
proj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1002,71): warning C4267: 'argument': conversion fro
m 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxp
roj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1005,88): warning C4244: 'argument': conversion fro
m 'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcx
proj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1005,71): warning C4267: 'argument': conversion fro
m 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxp
roj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1008,88): warning C4244: 'argument': conversion fro
m 'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcx
proj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1008,71): warning C4267: 'argument': conversion fro
m 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxp
roj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1011,88): warning C4244: 'argument': conversion fro
m 'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcx
proj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1011,71): warning C4267: 'argument': conversion fro
m 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxp
roj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1073,42): warning C4244: 'return': conversion from
'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxpr
oj]
  Generating Code...
Lib:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\Lib.exe /OUT:"llava
  .dir\Release\llava.lib" /NOLOGO /MACHINE:X64  /machine:x64 llava.dir\Release\llava.obj
  llava.dir\Release\clip.obj
  llava.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.dir\Release\llava.lib
FinalizeBuildStatus:
  Deleting file "llava.dir\Release\llava.tlog\unsuccessfulbuild".
  Touching "llava.dir\Release\llava.tlog\llava.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxproj" (default targets
).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\baby-llama\baby-llama.vcxproj" (7) on node 1 (default targets).
Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\baby-llama\baby-llama.vcxproj" (7) is building "I:\
ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj" (8) on node 1 (default targets).
PrepareForBuild:
  Creating directory "common.dir\Release\".
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\Release\".
  Creating directory "common.dir\Release\common.tlog\".
InitializeBuildStatus:
  Creating "common.dir\Release\common.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "common.dir\Release\common.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/common/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"common.dir\Release\\" /Fd"I:\ai-chat_v1\llama.cpp\l
  lama-cpp\llama.cpp\build\common\Release\common.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\
  llama-cpp\llama.cpp\common\common.cpp" "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\sampling.cpp" "I:\ai-chat_
  v1\llama.cpp\llama-cpp\llama.cpp\common\console.cpp" "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\grammar-pars
  er.cpp" "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\train.cpp"
  common.cpp
  sampling.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\sampling.cpp(75,45): warning C4267: 'initializing': conversion from
'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\sampling.cpp(75,20): warning C4267: 'initializing': conversion from
'size_t' to 'const int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj
]
  console.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\console.cpp(253,30): warning C4267: 'initializing': conversion from
'size_t' to 'DWORD', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\console.cpp(407,28): warning C4267: 'initializing': conversion from
'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj]
  grammar-parser.cpp
  train.cpp
  Generating Code...
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\common.cpp(879): warning C4715: 'gpt_random_prompt': not all control
 paths return a value [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj]
Lib:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\Lib.exe /OUT:"I:\ai
  -chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\Release\common.lib" /NOLOGO /MACHINE:X64  /machine:x64 common.dir
  \Release\common.obj
  common.dir\Release\sampling.obj
  common.dir\Release\console.obj
  "common.dir\Release\grammar-parser.obj"
  common.dir\Release\train.obj
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\build_info.dir\Release\build-info.obj"
  common.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\Release\common.lib
FinalizeBuildStatus:
  Deleting file "common.dir\Release\common.tlog\unsuccessfulbuild".
  Touching "common.dir\Release\common.tlog\common.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj" (default targets).

PrepareForBuild:
  Creating directory "baby-llama.dir\Release\".
  Creating directory "baby-llama.dir\Release\baby-llama.tlog\".
InitializeBuildStatus:
  Creating "baby-llama.dir\Release\baby-llama.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "baby-llama.dir\Release\baby-llama.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/baby-llama/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"baby-llama.dir\Release\\" /Fd"baby-llama.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\baby-llama\baby-llama.cpp"
  baby-llama.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\baby-llama\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\baby-llama.exe" /INCREMENTAL:NO /NOLOGO
  /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Rel
  ease\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.li
  b advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.
  cpp/llama-cpp/llama.cpp/build/bin/Release/baby-llama.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:
  "I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/baby-llama/Release/baby-llama.lib" /MACHINE:X64  /machine
  :x64 "baby-llama.dir\Release\baby-llama.obj"
  baby-llama.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\baby-llama.exe
FinalizeBuildStatus:
  Deleting file "baby-llama.dir\Release\baby-llama.tlog\unsuccessfulbuild".
  Touching "baby-llama.dir\Release\baby-llama.tlog\baby-llama.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\baby-llama\baby-llama.vcxproj" (defau
lt targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\batched\batched.vcxproj" (9) on node 1 (default targets).
PrepareForBuild:
  Creating directory "batched.dir\Release\".
  Creating directory "batched.dir\Release\batched.tlog\".
InitializeBuildStatus:
  Creating "batched.dir\Release\batched.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "batched.dir\Release\batched.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/batched/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"batched.dir\Release\\" /Fd"batched.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v
  1\llama.cpp\llama-cpp\llama.cpp\examples\batched\batched.cpp"
  batched.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\batched\batched.cpp(72,45): warning C4267: 'initializing': convers
ion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\batched\b
atched.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\batched\batched.cpp(72,24): warning C4267: 'initializing': convers
ion from 'size_t' to 'const int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\bat
ched\batched.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\batched\batched.cpp(114,50): warning C4267: 'argument': conversion
 from 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\batched\
batched.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\batched\batched.cpp(118,48): warning C4267: 'argument': conversion
 from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\batche
d\batched.vcxproj]
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\batched\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\batched.exe" /INCREMENTAL:NO /NOLOGO /LI
  BPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Releas
  e\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib a
  dvapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp
  /llama-cpp/llama.cpp/build/bin/Release/batched.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai
  -chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/batched/Release/batched.lib" /MACHINE:X64  /machine:x64 batched
  .dir\Release\batched.obj
  batched.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\batched.exe
FinalizeBuildStatus:
  Deleting file "batched.dir\Release\batched.tlog\unsuccessfulbuild".
  Touching "batched.dir\Release\batched.tlog\batched.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\batched\batched.vcxproj" (default tar
gets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\batched-bench\batched-bench.vcxproj" (10) on node 1 (default targets).
PrepareForBuild:
  Creating directory "batched-bench.dir\Release\".
  Creating directory "batched-bench.dir\Release\batched-bench.tlog\".
InitializeBuildStatus:
  Creating "batched-bench.dir\Release\batched-bench.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "batched-bench.dir\Release\batched-bench.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/batched-bench/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"batched-bench.dir\Release\\" /Fd"batched-bench.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\batched-bench\batched-bench.cpp"
  batched-bench.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\batched-bench\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\batched-bench.exe" /INCREMENTAL:NO /NOLO
  GO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\
  Release\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32
  .lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/lla
  ma.cpp/llama-cpp/llama.cpp/build/bin/Release/batched-bench.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /I
  MPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/batched-bench/Release/batched-bench.lib" /MACHINE:X
  64  /machine:x64 "batched-bench.dir\Release\batched-bench.obj"
  batched-bench.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\batched-bench.exe
FinalizeBuildStatus:
  Deleting file "batched-bench.dir\Release\batched-bench.tlog\unsuccessfulbuild".
  Touching "batched-bench.dir\Release\batched-bench.tlog\batched-bench.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\batched-bench\batched-bench.vcxproj"
(default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\beam-search\beam-search.vcxproj" (11) on node 1 (default targets).
PrepareForBuild:
  Creating directory "beam-search.dir\Release\".
  Creating directory "beam-search.dir\Release\beam-search.tlog\".
InitializeBuildStatus:
  Creating "beam-search.dir\Release\beam-search.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "beam-search.dir\Release\beam-search.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/beam-search/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"beam-search.dir\Release\\" /Fd"beam-search.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\a
  i-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\beam-search\beam-search.cpp"
  beam-search.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\beam-search\beam-search.cpp(163,83): warning C4267: 'argument': co
nversion from 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\
beam-search\beam-search.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\beam-search\beam-search.cpp(168,31): warning C4267: '+=': conversi
on from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\beam-searc
h\beam-search.vcxproj]
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\beam-search\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\beam-search.exe" /INCREMENTAL:NO /NOLOGO
   /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Re
  lease\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.l
  ib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama
  .cpp/llama-cpp/llama.cpp/build/bin/Release/beam-search.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLI
  B:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/beam-search/Release/beam-search.lib" /MACHINE:X64  /mac
  hine:x64 "beam-search.dir\Release\beam-search.obj"
  beam-search.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\beam-search.exe
FinalizeBuildStatus:
  Deleting file "beam-search.dir\Release\beam-search.tlog\unsuccessfulbuild".
  Touching "beam-search.dir\Release\beam-search.tlog\beam-search.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\beam-search\beam-search.vcxproj" (def
ault targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\benchmark\benchmark.vcxproj" (12) on node 1 (default targets).
PrepareForBuild:
  Creating directory "benchmark.dir\Release\".
  Creating directory "benchmark.dir\Release\benchmark.tlog\".
InitializeBuildStatus:
  Creating "benchmark.dir\Release\benchmark.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "benchmark.dir\Release\benchmark.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/benchmark/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\benchmark\..\
  ..\common" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v1
  2.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS
   /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT
  _SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc
  :forScope /Zc:inline /GR /Fo"benchmark.dir\Release\\" /Fd"benchmark.dir\Release\vc143.pdb" /external:W3 /Gd /TP /erro
  rReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\benchmark\benchmark-matmult.cpp"
  benchmark-matmult.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\benchmark\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\benchmark.exe" /INCREMENTAL:NO /NOLOGO /
  LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\Release\llama.lib kernel32.lib user3
  2.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib /MANIFEST /MANIFES
  TUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/bin
  /Release/benchmark.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp
  /llama.cpp/build/examples/benchmark/Release/benchmark.lib" /MACHINE:X64  /machine:x64 "benchmark.dir\Release\benchmar
  k-matmult.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\build_info.dir\Release\build-info.obj"
  benchmark.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\benchmark.exe
FinalizeBuildStatus:
  Deleting file "benchmark.dir\Release\benchmark.tlog\unsuccessfulbuild".
  Touching "benchmark.dir\Release\benchmark.tlog\benchmark.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\benchmark\benchmark.vcxproj" (default
 targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\convert-llama2c-to-ggml\convert-llama2c-to-ggml.vcxproj" (13) on node 1 (default tar
gets).
PrepareForBuild:
  Creating directory "convert-llama2c-to-ggml.dir\Release\".
  Creating directory "convert-llama2c-to-ggml.dir\Release\convert-.161408BC.tlog\".
InitializeBuildStatus:
  Creating "convert-llama2c-to-ggml.dir\Release\convert-.161408BC.tlog\unsuccessfulbuild" because "AlwaysCreate" was sp
  ecified.
  Touching "convert-llama2c-to-ggml.dir\Release\convert-.161408BC.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/convert-llama2c-to-ggml/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"convert-llama2c-to-ggml.dir\Release\\" /Fd"convert-llama2c-to-ggml.dir\Release\vc143.pdb" /external:W3 /Gd /TP
  /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\convert-llama2c-to-ggml\convert-llama2c-to-g
  gml.cpp"
  convert-llama2c-to-ggml.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\convert-llama2c-to-ggml\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\convert-llama2c-to-ggml.exe" /INCREMENTA
  L:NO /NOLOGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.
  lib ..\..\Release\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.li
  b comdlg32.lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-c
  hat_v1/llama.cpp/llama-cpp/llama.cpp/build/bin/Release/convert-llama2c-to-ggml.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNA
  MICBASE /NXCOMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/convert-llama2c-to-ggml/Release
  /convert-llama2c-to-ggml.lib" /MACHINE:X64  /machine:x64 "convert-llama2c-to-ggml.dir\Release\convert-llama2c-to-ggml
  .obj"
  convert-llama2c-to-ggml.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\convert-llama2c-to-g
  gml.exe
FinalizeBuildStatus:
  Deleting file "convert-llama2c-to-ggml.dir\Release\convert-.161408BC.tlog\unsuccessfulbuild".
  Touching "convert-llama2c-to-ggml.dir\Release\convert-.161408BC.tlog\convert-llama2c-to-ggml.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\convert-llama2c-to-ggml\convert-llama
2c-to-ggml.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\embedding\embedding.vcxproj" (14) on node 1 (default targets).
PrepareForBuild:
  Creating directory "embedding.dir\Release\".
  Creating directory "embedding.dir\Release\embedding.tlog\".
InitializeBuildStatus:
  Creating "embedding.dir\Release\embedding.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "embedding.dir\Release\embedding.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/embedding/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"embedding.dir\Release\\" /Fd"embedding.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-ch
  at_v1\llama.cpp\llama-cpp\llama.cpp\examples\embedding\embedding.cpp"
  embedding.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\embedding\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\embedding.exe" /INCREMENTAL:NO /NOLOGO /
  LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Rele
  ase\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib
   advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.c
  pp/llama-cpp/llama.cpp/build/bin/Release/embedding.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I
  :/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/embedding/Release/embedding.lib" /MACHINE:X64  /machine:x64
   embedding.dir\Release\embedding.obj
  embedding.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\embedding.exe
FinalizeBuildStatus:
  Deleting file "embedding.dir\Release\embedding.tlog\unsuccessfulbuild".
  Touching "embedding.dir\Release\embedding.tlog\embedding.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\embedding\embedding.vcxproj" (default
 targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\export-lora\export-lora.vcxproj" (15) on node 1 (default targets).
PrepareForBuild:
  Creating directory "export-lora.dir\Release\".
  Creating directory "export-lora.dir\Release\export-lora.tlog\".
InitializeBuildStatus:
  Creating "export-lora.dir\Release\export-lora.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "export-lora.dir\Release\export-lora.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/export-lora/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"export-lora.dir\Release\\" /Fd"export-lora.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\a
  i-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\export-lora\export-lora.cpp"
  export-lora.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\export-lora\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\export-lora.exe" /INCREMENTAL:NO /NOLOGO
   /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Re
  lease\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.l
  ib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama
  .cpp/llama-cpp/llama.cpp/build/bin/Release/export-lora.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLI
  B:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/export-lora/Release/export-lora.lib" /MACHINE:X64  /mac
  hine:x64 "export-lora.dir\Release\export-lora.obj"
  export-lora.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\export-lora.exe
FinalizeBuildStatus:
  Deleting file "export-lora.dir\Release\export-lora.tlog\unsuccessfulbuild".
  Touching "export-lora.dir\Release\export-lora.tlog\export-lora.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\export-lora\export-lora.vcxproj" (def
ault targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\finetune\finetune.vcxproj" (16) on node 1 (default targets).
PrepareForBuild:
  Creating directory "finetune.dir\Release\".
  Creating directory "finetune.dir\Release\finetune.tlog\".
InitializeBuildStatus:
  Creating "finetune.dir\Release\finetune.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "finetune.dir\Release\finetune.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/finetune/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"finetune.dir\Release\\" /Fd"finetune.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat
  _v1\llama.cpp\llama-cpp\llama.cpp\examples\finetune\finetune.cpp"
  finetune.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\finetune\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\finetune.exe" /INCREMENTAL:NO /NOLOGO /L
  IBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Relea
  se\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib
  advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cp
  p/llama-cpp/llama.cpp/build/bin/Release/finetune.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/
  ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/finetune/Release/finetune.lib" /MACHINE:X64  /machine:x64 fin
  etune.dir\Release\finetune.obj
  finetune.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\finetune.exe
FinalizeBuildStatus:
  Deleting file "finetune.dir\Release\finetune.tlog\unsuccessfulbuild".
  Touching "finetune.dir\Release\finetune.tlog\finetune.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\finetune\finetune.vcxproj" (default t
argets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\ggml_shared.vcxproj" (17) on node 1 (default targets).
PrepareForBuild:
  Creating directory "ggml_shared.dir\Release\ggml_shared.tlog\".
InitializeBuildStatus:
  Creating "ggml_shared.dir\Release\ggml_shared.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "ggml_shared.dir\Release\ggml_shared.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/CMakeLists.txt
PreLinkEvent:
  Auto build dll exports
  setlocal
  cd I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build
  if %errorlevel% neq 0 goto :cmEnd
  I:
  if %errorlevel% neq 0 goto :cmEnd
  "C:\Program Files\CMake\bin\cmake.exe" -E __create_def I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/ggml_shared.
  dir/Release/exports.def I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/ggml_shared.dir/Release//objects.txt
  if %errorlevel% neq 0 goto :cmEnd
  :cmEnd
  endlocal & call :cmErrorLevel %errorlevel% & goto :cmDone
  :cmErrorLevel
  exit /b %1
  :cmDone
  if %errorlevel% neq 0 goto :VCEnd
  :VCEnd
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\ggml_shared.dll" /INCREMENTAL:NO /NOLOGO
   /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" "C:\Program Files\NVIDIA GPU Computing T
  oolkit\CUDA\v12.3\lib\x64\cudart.lib" "C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64\cublas.lib" "
  C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64\cublasLt.lib" cudadevrt.lib cudart_static.lib kernel
  32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib /DEF:"
  I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/ggml_shared.dir/Release/exports.def" /MANIFEST /MANIFESTUAC:"level=
  'asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/bin/Release/ggm
  l_shared.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp
  /build/Release/ggml_shared.lib" /MACHINE:X64  /machine:x64 /DLL "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\gg
  ml.dir\Release\ggml.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-alloc.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-backend.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-quants.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-cuda.obj"
     Creating library I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/Release/ggml_shared.lib and object I:/ai-chat_v
  1/llama.cpp/llama-cpp/llama.cpp/build/Release/ggml_shared.exp
  ggml_shared.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\ggml_shared.dll
FinalizeBuildStatus:
  Deleting file "ggml_shared.dir\Release\ggml_shared.tlog\unsuccessfulbuild".
  Touching "ggml_shared.dir\Release\ggml_shared.tlog\ggml_shared.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml_shared.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\ggml_static.vcxproj" (18) on node 1 (default targets).
PrepareForBuild:
  Creating directory "ggml_static.dir\Release\".
  Creating directory "ggml_static.dir\Release\ggml_static.tlog\".
InitializeBuildStatus:
  Creating "ggml_static.dir\Release\ggml_static.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "ggml_static.dir\Release\ggml_static.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/CMakeLists.txt
Lib:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\Lib.exe /OUT:"I:\ai
  -chat_v1\llama.cpp\llama-cpp\llama.cpp\build\Release\ggml_static.lib" /NOLOGO /MACHINE:X64  /machine:x64 "I:\ai-chat_
  v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-alloc.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-backend.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-quants.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-cuda.obj"
  ggml_static.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\Release\ggml_static.lib
FinalizeBuildStatus:
  Deleting file "ggml_static.dir\Release\ggml_static.tlog\unsuccessfulbuild".
  Touching "ggml_static.dir\Release\ggml_static.tlog\ggml_static.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml_static.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\infill\infill.vcxproj" (19) on node 1 (default targets).
PrepareForBuild:
  Creating directory "infill.dir\Release\".
  Creating directory "infill.dir\Release\infill.tlog\".
InitializeBuildStatus:
  Creating "infill.dir\Release\infill.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "infill.dir\Release\infill.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/infill/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"infill.dir\Release\\" /Fd"infill.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\
  llama.cpp\llama-cpp\llama.cpp\examples\infill\infill.cpp"
  infill.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\infill\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\infill.exe" /INCREMENTAL:NO /NOLOGO /LIB
  PATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Release
  \llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib ad
  vapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/
  llama-cpp/llama.cpp/build/bin/Release/infill.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-c
  hat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/infill/Release/infill.lib" /MACHINE:X64  /machine:x64 infill.dir\
  Release\infill.obj
  infill.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\infill.exe
FinalizeBuildStatus:
  Deleting file "infill.dir\Release\infill.tlog\unsuccessfulbuild".
  Touching "infill.dir\Release\infill.tlog\infill.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\infill\infill.vcxproj" (default targe
ts).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\llama-bench\llama-bench.vcxproj" (20) on node 1 (default targets).
PrepareForBuild:
  Creating directory "llama-bench.dir\Release\".
  Creating directory "llama-bench.dir\Release\llama-bench.tlog\".
InitializeBuildStatus:
  Creating "llama-bench.dir\Release\llama-bench.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "llama-bench.dir\Release\llama-bench.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/llama-bench/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"llama-bench.dir\Release\\" /Fd"llama-bench.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\a
  i-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llama-bench\llama-bench.cpp"
  llama-bench.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llama-bench\llama-bench.cpp(72,13): warning C4244: 'initializing':
 conversion from 'double' to 'T', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\lla
ma-bench\llama-bench.vcxproj]
          with
          [
              T=uint64_t
          ]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llama-bench\llama-bench.cpp(531,18): message : see reference to fu
nction template instantiation 'T stdev<uint64_t>(const std::vector<uint64_t,std::allocator<uint64_t>> &)' being compile
d [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llama-bench\llama-bench.vcxproj]
          with
          [
              T=uint64_t
          ]
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llama-bench\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\llama-bench.exe" /INCREMENTAL:NO /NOLOGO
   /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Re
  lease\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.l
  ib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama
  .cpp/llama-cpp/llama.cpp/build/bin/Release/llama-bench.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLI
  B:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/llama-bench/Release/llama-bench.lib" /MACHINE:X64  /mac
  hine:x64 "llama-bench.dir\Release\llama-bench.obj"
  llama-bench.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\llama-bench.exe
FinalizeBuildStatus:
  Deleting file "llama-bench.dir\Release\llama-bench.tlog\unsuccessfulbuild".
  Touching "llama-bench.dir\Release\llama-bench.tlog\llama-bench.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llama-bench\llama-bench.vcxproj" (def
ault targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\llava\llava-cli.vcxproj" (21) on node 1 (default targets).
PrepareForBuild:
  Creating directory "llava-cli.dir\Release\".
  Creating directory "llava-cli.dir\Release\llava-cli.tlog\".
InitializeBuildStatus:
  Creating "llava-cli.dir\Release\llava-cli.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "llava-cli.dir\Release\llava-cli.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/llava/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\." /I"I:\ai-chat_
  v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\..\.." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\
  ..\..\common" /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:colu
  mn /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D
   K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CM
  AKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"llava-cli.dir\Rele
  ase\\" /Fd"llava-cli.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cp
  p\llama.cpp\examples\llava\llava-cli.cpp"
  llava-cli.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\llava-cli.cpp(150,105): warning C4267: 'argument': conversio
n from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava
-cli.vcxproj]
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\llava-cli.exe" /INCREMENTAL:NO /NOLOGO /
  LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Rele
  ase\llama.lib "C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64\cudart.lib" "C:\Program Files\NVIDIA
  GPU Computing Toolkit\CUDA\v12.3\lib\x64\cublas.lib" "C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x6
  4\cublasLt.lib" kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.l
  ib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama
  .cpp/llama-cpp/llama.cpp/build/bin/Release/llava-cli.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:
  "I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/llava/Release/llava-cli.lib" /MACHINE:X64  /machine:x64 "
  llava-cli.dir\Release\llava-cli.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.dir\Release\llava.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.dir\Release\clip.obj"
     Creating library I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/llava/Release/llava-cli.lib and object
   I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/llava/Release/llava-cli.exp
  llava-cli.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\llava-cli.exe
FinalizeBuildStatus:
  Deleting file "llava-cli.dir\Release\llava-cli.tlog\unsuccessfulbuild".
  Touching "llava-cli.dir\Release\llava-cli.tlog\llava-cli.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava-cli.vcxproj" (default tar
gets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\llava\llava_shared.vcxproj" (22) on node 1 (default targets).
PrepareForBuild:
  Creating directory "llava_shared.dir\Release\llava_shared.tlog\".
InitializeBuildStatus:
  Creating "llava_shared.dir\Release\llava_shared.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "llava_shared.dir\Release\llava_shared.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/llava/CMakeLists.txt
PreLinkEvent:
  Auto build dll exports
  setlocal
  cd I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava
  if %errorlevel% neq 0 goto :cmEnd
  I:
  if %errorlevel% neq 0 goto :cmEnd
  "C:\Program Files\CMake\bin\cmake.exe" -E __create_def I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/lla
  va/llava_shared.dir/Release/exports.def I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/llava/llava_shared
  .dir/Release//objects.txt
  if %errorlevel% neq 0 goto :cmEnd
  :cmEnd
  endlocal & call :cmErrorLevel %errorlevel% & goto :cmDone
  :cmErrorLevel
  exit /b %1
  :cmDone
  if %errorlevel% neq 0 goto :VCEnd
  :VCEnd
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\llava_shared.dll" /INCREMENTAL:NO /NOLOG
  O /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\Release\llama.lib "C:\Program Fil
  es\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64\cudart.lib" "C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v1
  2.3\lib\x64\cublas.lib" "C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64\cublasLt.lib" cudadevrt.lib
   cudart_static.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg3
  2.lib advapi32.lib /DEF:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/llava/llava_shared.dir/Release/ex
  ports.def" /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/
  llama-cpp/llama.cpp/build/bin/Release/llava_shared.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I
  :/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/llava/Release/llava_shared.lib" /MACHINE:X64  /machine:x64
  /DLL "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.dir\Release\llava.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.dir\Release\clip.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-alloc.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-backend.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-quants.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.dir\Release\ggml-cuda.obj"
     Creating library I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/llava/Release/llava_shared.lib and obj
  ect I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/llava/Release/llava_shared.exp
  llava_shared.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\llava_shared.dll
FinalizeBuildStatus:
  Deleting file "llava_shared.dir\Release\llava_shared.tlog\unsuccessfulbuild".
  Touching "llava_shared.dir\Release\llava_shared.tlog\llava_shared.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava_shared.vcxproj" (default
targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\llava\llava_static.vcxproj" (23) on node 1 (default targets).
PrepareForBuild:
  Creating directory "llava_static.dir\Release\".
  Creating directory "llava_static.dir\Release\llava_static.tlog\".
InitializeBuildStatus:
  Creating "llava_static.dir\Release\llava_static.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "llava_static.dir\Release\llava_static.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/llava/CMakeLists.txt
Lib:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\Lib.exe /OUT:"I:\ai
  -chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\Release\llava_static.lib" /NOLOGO /MACHINE:X64  /machine:
  x64 "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.dir\Release\llava.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.dir\Release\clip.obj"
  llava_static.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\Release\llava_static.lib
FinalizeBuildStatus:
  Deleting file "llava_static.dir\Release\llava_static.tlog\unsuccessfulbuild".
  Touching "llava_static.dir\Release\llava_static.tlog\llava_static.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava_static.vcxproj" (default
targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\main\main.vcxproj" (24) on node 1 (default targets).
PrepareForBuild:
  Creating directory "main.dir\Release\".
  Creating directory "main.dir\Release\main.tlog\".
InitializeBuildStatus:
  Creating "main.dir\Release\main.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "main.dir\Release\main.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/main/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"main.dir\Release\\" /Fd"main.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llam
  a.cpp\llama-cpp\llama.cpp\examples\main\main.cpp"
  main.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\main\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\main.exe" /INCREMENTAL:NO /NOLOGO /LIBPA
  TH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Release\l
  lama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib adva
  pi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/ll
  ama-cpp/llama.cpp/build/bin/Release/main.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-chat_
  v1/llama.cpp/llama-cpp/llama.cpp/build/examples/main/Release/main.lib" /MACHINE:X64  /machine:x64 main.dir\Release\ma
  in.obj
  main.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\main.exe
FinalizeBuildStatus:
  Deleting file "main.dir\Release\main.tlog\unsuccessfulbuild".
  Touching "main.dir\Release\main.tlog\main.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\main\main.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\parallel\parallel.vcxproj" (25) on node 1 (default targets).
PrepareForBuild:
  Creating directory "parallel.dir\Release\".
  Creating directory "parallel.dir\Release\parallel.tlog\".
InitializeBuildStatus:
  Creating "parallel.dir\Release\parallel.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "parallel.dir\Release\parallel.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/parallel/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"parallel.dir\Release\\" /Fd"parallel.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat
  _v1\llama.cpp\llama-cpp\llama.cpp\examples\parallel\parallel.cpp"
  parallel.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\parallel\parallel.cpp(157,21): warning C4267: '=': conversion from
 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\parallel\para
llel.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\parallel\parallel.cpp(163,55): warning C4267: 'initializing': conv
ersion from 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\pa
rallel\parallel.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\parallel\parallel.cpp(163,35): warning C4267: 'initializing': conv
ersion from 'size_t' to 'const int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examp
les\parallel\parallel.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\parallel\parallel.cpp(248,68): warning C4267: 'argument': conversi
on from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\para
llel\parallel.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\parallel\parallel.cpp(256,58): warning C4267: '=': conversion from
 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\parallel\para
llel.vcxproj]
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\parallel\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\parallel.exe" /INCREMENTAL:NO /NOLOGO /L
  IBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Relea
  se\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib
  advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cp
  p/llama-cpp/llama.cpp/build/bin/Release/parallel.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/
  ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/parallel/Release/parallel.lib" /MACHINE:X64  /machine:x64 par
  allel.dir\Release\parallel.obj
  parallel.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\parallel.exe
FinalizeBuildStatus:
  Deleting file "parallel.dir\Release\parallel.tlog\unsuccessfulbuild".
  Touching "parallel.dir\Release\parallel.tlog\parallel.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\parallel\parallel.vcxproj" (default t
argets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\perplexity\perplexity.vcxproj" (26) on node 1 (default targets).
PrepareForBuild:
  Creating directory "perplexity.dir\Release\".
  Creating directory "perplexity.dir\Release\perplexity.tlog\".
InitializeBuildStatus:
  Creating "perplexity.dir\Release\perplexity.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "perplexity.dir\Release\perplexity.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/perplexity/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"perplexity.dir\Release\\" /Fd"perplexity.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\perplexity\perplexity.cpp"
  perplexity.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\perplexity\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\perplexity.exe" /INCREMENTAL:NO /NOLOGO
  /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Rel
  ease\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.li
  b advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.
  cpp/llama-cpp/llama.cpp/build/bin/Release/perplexity.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:
  "I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/perplexity/Release/perplexity.lib" /MACHINE:X64  /machine
  :x64 perplexity.dir\Release\perplexity.obj
  perplexity.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\perplexity.exe
FinalizeBuildStatus:
  Deleting file "perplexity.dir\Release\perplexity.tlog\unsuccessfulbuild".
  Touching "perplexity.dir\Release\perplexity.tlog\perplexity.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\perplexity\perplexity.vcxproj" (defau
lt targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\pocs\vdot\q8dot.vcxproj" (27) on node 1 (default targets).
PrepareForBuild:
  Creating directory "q8dot.dir\Release\".
  Creating directory "q8dot.dir\Release\q8dot.tlog\".
InitializeBuildStatus:
  Creating "q8dot.dir\Release\q8dot.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "q8dot.dir\Release\q8dot.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/pocs/vdot/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\pocs" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-chat_v1
  \llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /W
  X- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D
  GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOP
  EN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /F
  o"q8dot.dir\Release\\" /Fd"q8dot.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.
  cpp\llama-cpp\llama.cpp\pocs\vdot\q8dot.cpp"
  q8dot.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\pocs\vdot\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\q8dot.exe" /INCREMENTAL:NO /NOLOGO /LIBP
  ATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Release\
  llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib adv
  api32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/l
  lama-cpp/llama.cpp/build/bin/Release/q8dot.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-cha
  t_v1/llama.cpp/llama-cpp/llama.cpp/build/pocs/vdot/Release/q8dot.lib" /MACHINE:X64  /machine:x64 q8dot.dir\Release\q8
  dot.obj
  q8dot.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\q8dot.exe
FinalizeBuildStatus:
  Deleting file "q8dot.dir\Release\q8dot.tlog\unsuccessfulbuild".
  Touching "q8dot.dir\Release\q8dot.tlog\q8dot.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\pocs\vdot\q8dot.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\quantize\quantize.vcxproj" (28) on node 1 (default targets).
PrepareForBuild:
  Creating directory "quantize.dir\Release\".
  Creating directory "quantize.dir\Release\quantize.tlog\".
InitializeBuildStatus:
  Creating "quantize.dir\Release\quantize.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "quantize.dir\Release\quantize.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/quantize/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\quantize\..\.
  .\common" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12
  .3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS
  /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_
  SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:
  forScope /Zc:inline /GR /Fo"quantize.dir\Release\\" /Fd"quantize.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorRe
  port:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\quantize\quantize.cpp"
  quantize.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\quantize\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\quantize.exe" /INCREMENTAL:NO /NOLOGO /L
  IBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\Release\llama.lib kernel32.lib user32
  .lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib /MANIFEST /MANIFEST
  UAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/bin/
  Release/quantize.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/l
  lama.cpp/build/examples/quantize/Release/quantize.lib" /MACHINE:X64  /machine:x64 quantize.dir\Release\quantize.obj
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\build_info.dir\Release\build-info.obj"
  quantize.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\quantize.exe
FinalizeBuildStatus:
  Deleting file "quantize.dir\Release\quantize.tlog\unsuccessfulbuild".
  Touching "quantize.dir\Release\quantize.tlog\quantize.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\quantize\quantize.vcxproj" (default t
argets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\quantize-stats\quantize-stats.vcxproj" (29) on node 1 (default targets).
PrepareForBuild:
  Creating directory "quantize-stats.dir\Release\".
  Creating directory "quantize-stats.dir\Release\quantize-stats.tlog\".
InitializeBuildStatus:
  Creating "quantize-stats.dir\Release\quantize-stats.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "quantize-stats.dir\Release\quantize-stats.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/quantize-stats/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\quantize-stat
  s\..\..\common" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CU
  DA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_C
  UBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D
   _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_
  t /Zc:forScope /Zc:inline /GR /Fo"quantize-stats.dir\Release\\" /Fd"quantize-stats.dir\Release\vc143.pdb" /external:W
  3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\quantize-stats\quantize-stats.cpp"
  quantize-stats.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\quantize-stats\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\quantize-stats.exe" /INCREMENTAL:NO /NOL
  OGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\Release\llama.lib kernel32.lib
  user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib /MANIFEST /MA
  NIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/buil
  d/bin/Release/quantize-stats.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp
  /llama-cpp/llama.cpp/build/examples/quantize-stats/Release/quantize-stats.lib" /MACHINE:X64  /machine:x64 "quantize-s
  tats.dir\Release\quantize-stats.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\build_info.dir\Release\build-info.obj"
  quantize-stats.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\quantize-stats.exe
FinalizeBuildStatus:
  Deleting file "quantize-stats.dir\Release\quantize-stats.tlog\unsuccessfulbuild".
  Touching "quantize-stats.dir\Release\quantize-stats.tlog\quantize-stats.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\quantize-stats\quantize-stats.vcxproj
" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\save-load-state\save-load-state.vcxproj" (30) on node 1 (default targets).
PrepareForBuild:
  Creating directory "save-load-state.dir\Release\".
  Creating directory "save-load-state.dir\Release\save-load-state.tlog\".
InitializeBuildStatus:
  Creating "save-load-state.dir\Release\save-load-state.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "save-load-state.dir\Release\save-load-state.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/save-load-state/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"save-load-state.dir\Release\\" /Fd"save-load-state.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:que
  ue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\save-load-state\save-load-state.cpp"
  save-load-state.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\save-load-state\save-load-state.cpp(42,69): warning C4267: 'argume
nt': conversion from 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ex
amples\save-load-state\save-load-state.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\save-load-state\save-load-state.cpp(43,26): warning C4267: '+=': c
onversion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\sav
e-load-state\save-load-state.vcxproj]
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\save-load-state\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\save-load-state.exe" /INCREMENTAL:NO /NO
  LOGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\.
  .\Release\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg
  32.lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/l
  lama.cpp/llama-cpp/llama.cpp/build/bin/Release/save-load-state.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPA
  T /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/save-load-state/Release/save-load-state.lib" /M
  ACHINE:X64  /machine:x64 "save-load-state.dir\Release\save-load-state.obj"
  save-load-state.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\save-load-state.exe
FinalizeBuildStatus:
  Deleting file "save-load-state.dir\Release\save-load-state.tlog\unsuccessfulbuild".
  Touching "save-load-state.dir\Release\save-load-state.tlog\save-load-state.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\save-load-state\save-load-state.vcxpr
oj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\server\server.vcxproj" (31) on node 1 (default targets).
PrepareForBuild:
  Creating directory "server.dir\Release\".
  Creating directory "server.dir\Release\server.tlog\".
InitializeBuildStatus:
  Creating "server.dir\Release\server.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "server.dir\Release\server.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/server/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server" /I"I:
  \ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v
  1\llama.cpp\llama-cpp\llama.cpp\examples\llava\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\..\..
  " /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\..\..\common" /I"C:\Program Files\NVIDIA GPU Computin
  g Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /
  D SERVER_VERBOSE=1 /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"server.dir\Release\\" /Fd"server.dir\Release\vc143.
  pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp"
  server.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(82,16): warning C4267: 'initializing': conversio
n from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\serv
er.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(96,52): warning C4267: '=': conversion from 'siz
e_t' to 'uint8_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vcxp
roj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(120,48): warning C4267: '=': conversion from 'si
ze_t' to 'uint8_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vcx
proj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(781,49): warning C4267: '=': conversion from 'si
ze_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vcx
proj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(785,93): warning C4267: 'argument': conversion f
rom 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.
vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(832,67): warning C4101: 'e': unreferenced local
variable [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(887,67): warning C4267: 'argument': conversion f
rom 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\s
erver.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(960,53): warning C4267: '-=': conversion from 's
ize_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vc
xproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1496,71): warning C4267: 'argument': conversion
from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\
server.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1558,64): warning C4267: '=': conversion from 's
ize_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vc
xproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1594,72): warning C4267: '=': conversion from 's
ize_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vc
xproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1604,50): warning C4267: '=': conversion from 's
ize_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vc
xproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1612,78): warning C4267: 'argument': conversion
from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\
server.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1635,96): warning C4267: 'argument': conversion
from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\
server.vcxproj]
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\server.exe" /INCREMENTAL:NO /NOLOGO /LIB
  PATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ws2_32.lib ..
  \..\Release\llama.lib "C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64\cudart.lib" "C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64\cublas.lib" "C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.
  3\lib\x64\cublasLt.lib" kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib co
  mdlg32.lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_
  v1/llama.cpp/llama-cpp/llama.cpp/build/bin/Release/server.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IM
  PLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/server/Release/server.lib" /MACHINE:X64  /machine:x6
  4 server.dir\Release\server.obj
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.dir\Release\llava.obj"
  "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.dir\Release\clip.obj"
     Creating library I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/server/Release/server.lib and object I
  :/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/server/Release/server.exp
  server.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\server.exe
FinalizeBuildStatus:
  Deleting file "server.dir\Release\server.tlog\unsuccessfulbuild".
  Touching "server.dir\Release\server.tlog\server.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vcxproj" (default targe
ts).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\simple\simple.vcxproj" (32) on node 1 (default targets).
PrepareForBuild:
  Creating directory "simple.dir\Release\".
  Creating directory "simple.dir\Release\simple.tlog\".
InitializeBuildStatus:
  Creating "simple.dir\Release\simple.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "simple.dir\Release\simple.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/simple/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"simple.dir\Release\\" /Fd"simple.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\
  llama.cpp\llama-cpp\llama.cpp\examples\simple\simple.cpp"
  simple.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\simple\simple.cpp(71,45): warning C4267: 'initializing': conversio
n from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\simple\simp
le.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\simple\simple.cpp(71,24): warning C4267: 'initializing': conversio
n from 'size_t' to 'const int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\simpl
e\simple.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\simple\simple.cpp(99,48): warning C4267: 'argument': conversion fr
om 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\simple\si
mple.vcxproj]
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\simple\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\simple.exe" /INCREMENTAL:NO /NOLOGO /LIB
  PATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Release
  \llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib ad
  vapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/
  llama-cpp/llama.cpp/build/bin/Release/simple.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-c
  hat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/simple/Release/simple.lib" /MACHINE:X64  /machine:x64 simple.dir\
  Release\simple.obj
  simple.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\simple.exe
FinalizeBuildStatus:
  Deleting file "simple.dir\Release\simple.tlog\unsuccessfulbuild".
  Touching "simple.dir\Release\simple.tlog\simple.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\simple\simple.vcxproj" (default targe
ts).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\speculative\speculative.vcxproj" (33) on node 1 (default targets).
PrepareForBuild:
  Creating directory "speculative.dir\Release\".
  Creating directory "speculative.dir\Release\speculative.tlog\".
InitializeBuildStatus:
  Creating "speculative.dir\Release\speculative.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "speculative.dir\Release\speculative.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/speculative/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"speculative.dir\Release\\" /Fd"speculative.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\a
  i-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\speculative\speculative.cpp"
  speculative.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\speculative\speculative.cpp(117,33): warning C4267: 'initializing'
: conversion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\
speculative\speculative.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\speculative\speculative.cpp(117,23): warning C4267: 'initializing'
: conversion from 'size_t' to 'const int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\exa
mples\speculative\speculative.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\speculative\speculative.cpp(138,20): warning C4267: 'initializing'
: conversion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\
speculative\speculative.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\speculative\speculative.cpp(139,20): warning C4267: 'initializing'
: conversion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\
speculative\speculative.vcxproj]
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\speculative\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\speculative.exe" /INCREMENTAL:NO /NOLOGO
   /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Re
  lease\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.l
  ib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama
  .cpp/llama-cpp/llama.cpp/build/bin/Release/speculative.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLI
  B:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/speculative/Release/speculative.lib" /MACHINE:X64  /mac
  hine:x64 speculative.dir\Release\speculative.obj
  speculative.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\speculative.exe
FinalizeBuildStatus:
  Deleting file "speculative.dir\Release\speculative.tlog\unsuccessfulbuild".
  Touching "speculative.dir\Release\speculative.tlog\speculative.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\speculative\speculative.vcxproj" (def
ault targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-c.vcxproj" (34) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-c.dir\Release\".
  Creating directory "test-c.dir\Release\test-c.tlog\".
InitializeBuildStatus:
  Creating "test-c.dir\Release\test-c.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "test-c.dir\Release\test-c.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo
   /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X
  =32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS
  /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /std:
  c11 /Fo"test-c.dir\Release\\" /Fd"test-c.dir\Release\vc143.pdb" /external:W3 /Gd /TC /errorReport:queue "I:\ai-chat_v
  1\llama.cpp\llama-cpp\llama.cpp\tests\test-c.c"
  test-c.c
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-c.exe" /INCREMENTAL:NO /NOLOGO /LIB
  PATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\Release\llama.lib kernel32.lib user32.lib
  gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib /MANIFEST /MANIFESTUAC:"
  level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/bin/Relea
  se/test-c.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cp
  p/build/tests/Release/test-c.lib" /MACHINE:X64  /machine:x64 "test-c.dir\Release\test-c.obj"
  test-c.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-c.exe
FinalizeBuildStatus:
  Deleting file "test-c.dir\Release\test-c.tlog\unsuccessfulbuild".
  Touching "test-c.dir\Release\test-c.tlog\test-c.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-c.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-grad0.vcxproj" (35) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-grad0.dir\Release\".
  Creating directory "test-grad0.dir\Release\test-grad0.tlog\".
InitializeBuildStatus:
  Creating "test-grad0.dir\Release\test-grad0.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "test-grad0.dir\Release\test-grad0.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"test-grad0.dir\Release\\" /Fd"test-grad0.dir\Releas
  e\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\tests\test-grad0.cp
  p"
  test-grad0.cpp
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-grad0.exe" /INCREMENTAL:NO /NOLOGO
  /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\common\Release\common.lib ..\Release\l
  lama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib adva
  pi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/ll
  ama-cpp/llama.cpp/build/bin/Release/test-grad0.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai
  -chat_v1/llama.cpp/llama-cpp/llama.cpp/build/tests/Release/test-grad0.lib" /MACHINE:X64  /machine:x64 "test-grad0.dir
  \Release\test-grad0.obj"
  test-grad0.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-grad0.exe
FinalizeBuildStatus:
  Deleting file "test-grad0.dir\Release\test-grad0.tlog\unsuccessfulbuild".
  Touching "test-grad0.dir\Release\test-grad0.tlog\test-grad0.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-grad0.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-grammar-parser.vcxproj" (36) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-grammar-parser.dir\Release\".
  Creating directory "test-grammar-parser.dir\Release\test-gra.05F1B7A5.tlog\".
InitializeBuildStatus:
  Creating "test-grammar-parser.dir\Release\test-gra.05F1B7A5.tlog\unsuccessfulbuild" because "AlwaysCreate" was specif
  ied.
  Touching "test-grammar-parser.dir\Release\test-gra.05F1B7A5.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"test-grammar-parser.dir\Release\\" /Fd"test-grammar
  -parser.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\t
  ests\test-grammar-parser.cpp"
  test-grammar-parser.cpp
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-grammar-parser.exe" /INCREMENTAL:NO
   /NOLOGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\common\Release\common.lib ..\
  Release\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32
  .lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/lla
  ma.cpp/llama-cpp/llama.cpp/build/bin/Release/test-grammar-parser.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOM
  PAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/tests/Release/test-grammar-parser.lib" /MACHINE:X64  /
  machine:x64 "test-grammar-parser.dir\Release\test-grammar-parser.obj"
  test-grammar-parser.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-grammar-parser.exe
FinalizeBuildStatus:
  Deleting file "test-grammar-parser.dir\Release\test-gra.05F1B7A5.tlog\unsuccessfulbuild".
  Touching "test-grammar-parser.dir\Release\test-gra.05F1B7A5.tlog\test-grammar-parser.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-grammar-parser.vcxproj" (default ta
rgets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-llama-grammar.vcxproj" (37) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-llama-grammar.dir\Release\".
  Creating directory "test-llama-grammar.dir\Release\test-lla.F9AAAF21.tlog\".
InitializeBuildStatus:
  Creating "test-llama-grammar.dir\Release\test-lla.F9AAAF21.tlog\unsuccessfulbuild" because "AlwaysCreate" was specifi
  ed.
  Touching "test-llama-grammar.dir\Release\test-lla.F9AAAF21.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"test-llama-grammar.dir\Release\\" /Fd"test-llama-gr
  ammar.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\tes
  ts\test-llama-grammar.cpp"
  test-llama-grammar.cpp
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\.\llama.cpp(1103,31): warning C4305: 'initializing': truncation from 'doubl
e' to 'float' [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-llama-grammar.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\.\llama.cpp(2331,69): warning C4566: character represented by universal-cha
racter-name '\u010A' cannot be represented in the current code page (1252) [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp
\build\tests\test-llama-grammar.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\.\llama.cpp(9054,28): warning C4146: unary minus operator applied to unsign
ed type, result still unsigned [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-llama-grammar.vcxproj]
I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\.\llama.cpp(9084,28): warning C4146: unary minus operator applied to unsign
ed type, result still unsigned [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-llama-grammar.vcxproj]
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-llama-grammar.exe" /INCREMENTAL:NO
  /NOLOGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\common\Release\common.lib ..\R
  elease\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.
  lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llam
  a.cpp/llama-cpp/llama.cpp/build/bin/Release/test-llama-grammar.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPA
  T /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/tests/Release/test-llama-grammar.lib" /MACHINE:X64  /mac
  hine:x64 "test-llama-grammar.dir\Release\test-llama-grammar.obj"
  test-llama-grammar.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-llama-grammar.exe
FinalizeBuildStatus:
  Deleting file "test-llama-grammar.dir\Release\test-lla.F9AAAF21.tlog\unsuccessfulbuild".
  Touching "test-llama-grammar.dir\Release\test-lla.F9AAAF21.tlog\test-llama-grammar.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-llama-grammar.vcxproj" (default tar
gets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-quantize-fns.vcxproj" (38) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-quantize-fns.dir\Release\".
  Creating directory "test-quantize-fns.dir\Release\test-qua.4C84FB51.tlog\".
InitializeBuildStatus:
  Creating "test-quantize-fns.dir\Release\test-qua.4C84FB51.tlog\unsuccessfulbuild" because "AlwaysCreate" was specifie
  d.
  Touching "test-quantize-fns.dir\Release\test-qua.4C84FB51.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"test-quantize-fns.dir\Release\\" /Fd"test-quantize-
  fns.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\tests
  \test-quantize-fns.cpp"
  test-quantize-fns.cpp
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-quantize-fns.exe" /INCREMENTAL:NO /
  NOLOGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\common\Release\common.lib ..\Re
  lease\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.l
  ib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama
  .cpp/llama-cpp/llama.cpp/build/bin/Release/test-quantize-fns.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT
  /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/tests/Release/test-quantize-fns.lib" /MACHINE:X64  /machin
  e:x64 "test-quantize-fns.dir\Release\test-quantize-fns.obj"
  test-quantize-fns.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-quantize-fns.exe
FinalizeBuildStatus:
  Deleting file "test-quantize-fns.dir\Release\test-qua.4C84FB51.tlog\unsuccessfulbuild".
  Touching "test-quantize-fns.dir\Release\test-qua.4C84FB51.tlog\test-quantize-fns.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-quantize-fns.vcxproj" (default targ
ets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-quantize-perf.vcxproj" (39) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-quantize-perf.dir\Release\".
  Creating directory "test-quantize-perf.dir\Release\test-qua.AAA6E98D.tlog\".
InitializeBuildStatus:
  Creating "test-quantize-perf.dir\Release\test-qua.AAA6E98D.tlog\unsuccessfulbuild" because "AlwaysCreate" was specifi
  ed.
  Touching "test-quantize-perf.dir\Release\test-qua.AAA6E98D.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"test-quantize-perf.dir\Release\\" /Fd"test-quantize
  -perf.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\tes
  ts\test-quantize-perf.cpp"
  test-quantize-perf.cpp
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-quantize-perf.exe" /INCREMENTAL:NO
  /NOLOGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\common\Release\common.lib ..\R
  elease\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.
  lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llam
  a.cpp/llama-cpp/llama.cpp/build/bin/Release/test-quantize-perf.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPA
  T /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/tests/Release/test-quantize-perf.lib" /MACHINE:X64  /mac
  hine:x64 "test-quantize-perf.dir\Release\test-quantize-perf.obj"
  test-quantize-perf.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-quantize-perf.exe
FinalizeBuildStatus:
  Deleting file "test-quantize-perf.dir\Release\test-qua.AAA6E98D.tlog\unsuccessfulbuild".
  Touching "test-quantize-perf.dir\Release\test-qua.AAA6E98D.tlog\test-quantize-perf.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-quantize-perf.vcxproj" (default tar
gets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-rope.vcxproj" (40) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-rope.dir\Release\".
  Creating directory "test-rope.dir\Release\test-rope.tlog\".
InitializeBuildStatus:
  Creating "test-rope.dir\Release\test-rope.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "test-rope.dir\Release\test-rope.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"test-rope.dir\Release\\" /Fd"test-rope.dir\Release\
  vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\tests\test-rope.cpp"
  test-rope.cpp
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-rope.exe" /INCREMENTAL:NO /NOLOGO /
  LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\common\Release\common.lib ..\Release\ll
  ama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advap
  i32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/lla
  ma-cpp/llama.cpp/build/bin/Release/test-rope.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-c
  hat_v1/llama.cpp/llama-cpp/llama.cpp/build/tests/Release/test-rope.lib" /MACHINE:X64  /machine:x64 "test-rope.dir\Rel
  ease\test-rope.obj"
  test-rope.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-rope.exe
FinalizeBuildStatus:
  Deleting file "test-rope.dir\Release\test-rope.tlog\unsuccessfulbuild".
  Touching "test-rope.dir\Release\test-rope.tlog\test-rope.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-rope.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-sampling.vcxproj" (41) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-sampling.dir\Release\".
  Creating directory "test-sampling.dir\Release\test-sampling.tlog\".
InitializeBuildStatus:
  Creating "test-sampling.dir\Release\test-sampling.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "test-sampling.dir\Release\test-sampling.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"test-sampling.dir\Release\\" /Fd"test-sampling.dir\
  Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\tests\test-sa
  mpling.cpp"
  test-sampling.cpp
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-sampling.exe" /INCREMENTAL:NO /NOLO
  GO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\common\Release\common.lib ..\Releas
  e\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib a
  dvapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp
  /llama-cpp/llama.cpp/build/bin/Release/test-sampling.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:
  "I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/tests/Release/test-sampling.lib" /MACHINE:X64  /machine:x64 "test-
  sampling.dir\Release\test-sampling.obj"
  test-sampling.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-sampling.exe
FinalizeBuildStatus:
  Deleting file "test-sampling.dir\Release\test-sampling.tlog\unsuccessfulbuild".
  Touching "test-sampling.dir\Release\test-sampling.tlog\test-sampling.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-sampling.vcxproj" (default targets)
.

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-tokenizer-0-falcon.vcxproj" (42) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-tokenizer-0-falcon.dir\Release\".
  Creating directory "test-tokenizer-0-falcon.dir\Release\test-tok.3836F38F.tlog\".
InitializeBuildStatus:
  Creating "test-tokenizer-0-falcon.dir\Release\test-tok.3836F38F.tlog\unsuccessfulbuild" because "AlwaysCreate" was sp
  ecified.
  Touching "test-tokenizer-0-falcon.dir\Release\test-tok.3836F38F.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"test-tokenizer-0-falcon.dir\Release\\" /Fd"test-tok
  enizer-0-falcon.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\lla
  ma.cpp\tests\test-tokenizer-0-falcon.cpp"
  test-tokenizer-0-falcon.cpp
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-tokenizer-0-falcon.exe" /INCREMENTA
  L:NO /NOLOGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\common\Release\common.lib
   ..\Release\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comd
  lg32.lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1
  /llama.cpp/llama-cpp/llama.cpp/build/bin/Release/test-tokenizer-0-falcon.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBAS
  E /NXCOMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/tests/Release/test-tokenizer-0-falcon.lib" /MA
  CHINE:X64  /machine:x64 "test-tokenizer-0-falcon.dir\Release\test-tokenizer-0-falcon.obj"
  test-tokenizer-0-falcon.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-tokenizer-0-fal
  con.exe
FinalizeBuildStatus:
  Deleting file "test-tokenizer-0-falcon.dir\Release\test-tok.3836F38F.tlog\unsuccessfulbuild".
  Touching "test-tokenizer-0-falcon.dir\Release\test-tok.3836F38F.tlog\test-tokenizer-0-falcon.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-tokenizer-0-falcon.vcxproj" (defaul
t targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-tokenizer-0-llama.vcxproj" (43) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-tokenizer-0-llama.dir\Release\".
  Creating directory "test-tokenizer-0-llama.dir\Release\test-tok.AC5C71AD.tlog\".
InitializeBuildStatus:
  Creating "test-tokenizer-0-llama.dir\Release\test-tok.AC5C71AD.tlog\unsuccessfulbuild" because "AlwaysCreate" was spe
  cified.
  Touching "test-tokenizer-0-llama.dir\Release\test-tok.AC5C71AD.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"test-tokenizer-0-llama.dir\Release\\" /Fd"test-toke
  nizer-0-llama.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama
  .cpp\tests\test-tokenizer-0-llama.cpp"
  test-tokenizer-0-llama.cpp
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-tokenizer-0-llama.exe" /INCREMENTAL
  :NO /NOLOGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\common\Release\common.lib
  ..\Release\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdl
  g32.lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/
  llama.cpp/llama-cpp/llama.cpp/build/bin/Release/test-tokenizer-0-llama.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE
  /NXCOMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/tests/Release/test-tokenizer-0-llama.lib" /MACHI
  NE:X64  /machine:x64 "test-tokenizer-0-llama.dir\Release\test-tokenizer-0-llama.obj"
  test-tokenizer-0-llama.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-tokenizer-0-llam
  a.exe
FinalizeBuildStatus:
  Deleting file "test-tokenizer-0-llama.dir\Release\test-tok.AC5C71AD.tlog\unsuccessfulbuild".
  Touching "test-tokenizer-0-llama.dir\Release\test-tok.AC5C71AD.tlog\test-tokenizer-0-llama.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-tokenizer-0-llama.vcxproj" (default
 targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-tokenizer-1-bpe.vcxproj" (44) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-tokenizer-1-bpe.dir\Release\".
  Creating directory "test-tokenizer-1-bpe.dir\Release\test-tok.944D62A8.tlog\".
InitializeBuildStatus:
  Creating "test-tokenizer-1-bpe.dir\Release\test-tok.944D62A8.tlog\unsuccessfulbuild" because "AlwaysCreate" was speci
  fied.
  Touching "test-tokenizer-1-bpe.dir\Release\test-tok.944D62A8.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"test-tokenizer-1-bpe.dir\Release\\" /Fd"test-tokeni
  zer-1-bpe.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp
  \tests\test-tokenizer-1-bpe.cpp"
  test-tokenizer-1-bpe.cpp
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-tokenizer-1-bpe.exe" /INCREMENTAL:N
  O /NOLOGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\common\Release\common.lib ..
  \Release\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg3
  2.lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/ll
  ama.cpp/llama-cpp/llama.cpp/build/bin/Release/test-tokenizer-1-bpe.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXC
  OMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/tests/Release/test-tokenizer-1-bpe.lib" /MACHINE:X64
    /machine:x64 "test-tokenizer-1-bpe.dir\Release\test-tokenizer-1-bpe.obj"
  test-tokenizer-1-bpe.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-tokenizer-1-bpe.ex
  e
FinalizeBuildStatus:
  Deleting file "test-tokenizer-1-bpe.dir\Release\test-tok.944D62A8.tlog\unsuccessfulbuild".
  Touching "test-tokenizer-1-bpe.dir\Release\test-tok.944D62A8.tlog\test-tokenizer-1-bpe.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-tokenizer-1-bpe.vcxproj" (default t
argets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\tests\test-tokenizer-1-llama.vcxproj" (45) on node 1 (default targets).
PrepareForBuild:
  Creating directory "test-tokenizer-1-llama.dir\Release\".
  Creating directory "test-tokenizer-1-llama.dir\Release\test-tok.BB225721.tlog\".
InitializeBuildStatus:
  Creating "test-tokenizer-1-llama.dir\Release\test-tok.BB225721.tlog\unsuccessfulbuild" because "AlwaysCreate" was spe
  cified.
  Touching "test-tokenizer-1-llama.dir\Release\test-tok.BB225721.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/tests/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\." /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"C:\Program Files
  \NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D
  _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML
  _CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHs
  c /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /Fo"test-tokenizer-1-llama.dir\Release\\" /Fd"test-toke
  nizer-1-llama.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama
  .cpp\tests\test-tokenizer-1-llama.cpp"
  test-tokenizer-1-llama.cpp
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-tokenizer-1-llama.exe" /INCREMENTAL
  :NO /NOLOGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\common\Release\common.lib
  ..\Release\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdl
  g32.lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/
  llama.cpp/llama-cpp/llama.cpp/build/bin/Release/test-tokenizer-1-llama.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE
  /NXCOMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/tests/Release/test-tokenizer-1-llama.lib" /MACHI
  NE:X64  /machine:x64 "test-tokenizer-1-llama.dir\Release\test-tokenizer-1-llama.obj"
  test-tokenizer-1-llama.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\test-tokenizer-1-llam
  a.exe
FinalizeBuildStatus:
  Deleting file "test-tokenizer-1-llama.dir\Release\test-tok.BB225721.tlog\unsuccessfulbuild".
  Touching "test-tokenizer-1-llama.dir\Release\test-tok.BB225721.tlog\test-tokenizer-1-llama.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-tokenizer-1-llama.vcxproj" (default
 targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\examples\train-text-from-scratch\train-text-from-scratch.vcxproj" (46) on node 1 (default tar
gets).
PrepareForBuild:
  Creating directory "train-text-from-scratch.dir\Release\".
  Creating directory "train-text-from-scratch.dir\Release\train-te.17E4B3D6.tlog\".
InitializeBuildStatus:
  Creating "train-text-from-scratch.dir\Release\train-te.17E4B3D6.tlog\unsuccessfulbuild" because "AlwaysCreate" was sp
  ecified.
  Touching "train-text-from-scratch.dir\Release\train-te.17E4B3D6.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/examples/train-text-from-scratch/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\examples" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-cha
  t_v1\llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W
  3 /WX- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32
   /D GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D
  _XOPEN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /G
  R /Fo"train-text-from-scratch.dir\Release\\" /Fd"train-text-from-scratch.dir\Release\vc143.pdb" /external:W3 /Gd /TP
  /errorReport:queue "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\train-text-from-scratch\train-text-from-scra
  tch.cpp"
  train-text-from-scratch.cpp
MakeDirsForLink:
  Creating directory "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\train-text-from-scratch\Release\".
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\train-text-from-scratch.exe" /INCREMENTA
  L:NO /NOLOGO /LIBPATH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.
  lib ..\..\Release\llama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.li
  b comdlg32.lib advapi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-c
  hat_v1/llama.cpp/llama-cpp/llama.cpp/build/bin/Release/train-text-from-scratch.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNA
  MICBASE /NXCOMPAT /IMPLIB:"I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/build/examples/train-text-from-scratch/Release
  /train-text-from-scratch.lib" /MACHINE:X64  /machine:x64 "train-text-from-scratch.dir\Release\train-text-from-scratch
  .obj"
  train-text-from-scratch.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\train-text-from-scra
  tch.exe
FinalizeBuildStatus:
  Deleting file "train-text-from-scratch.dir\Release\train-te.17E4B3D6.tlog\unsuccessfulbuild".
  Touching "train-text-from-scratch.dir\Release\train-te.17E4B3D6.tlog\train-text-from-scratch.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\train-text-from-scratch\train-text-fr
om-scratch.vcxproj" (default targets).

Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (1) is building "I:\ai-chat_v1\llama.cpp\
llama-cpp\llama.cpp\build\pocs\vdot\vdot.vcxproj" (47) on node 1 (default targets).
PrepareForBuild:
  Creating directory "vdot.dir\Release\".
  Creating directory "vdot.dir\Release\vdot.tlog\".
InitializeBuildStatus:
  Creating "vdot.dir\Release\vdot.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "vdot.dir\Release\vdot.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/pocs/vdot/CMakeLists.txt
ClCompile:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\CL.exe /c /I"I:\ai-
  chat_v1\llama.cpp\llama-cpp\llama.cpp\pocs" /I"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\." /I"I:\ai-chat_v1
  \llama.cpp\llama-cpp\llama.cpp\." /I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\include" /nologo /W3 /W
  X- /diagnostics:column /O2 /Ob2 /D _MBCS /D WIN32 /D _WINDOWS /D NDEBUG /D GGML_USE_CUBLAS /D GGML_CUDA_DMMV_X=32 /D
  GGML_CUDA_MMV_Y=1 /D K_QUANTS_PER_ITERATION=2 /D GGML_CUDA_PEER_MAX_BATCH_SIZE=128 /D _CRT_SECURE_NO_WARNINGS /D _XOP
  EN_SOURCE=600 /D "CMAKE_INTDIR=\"Release\"" /Gm- /EHsc /MD /GS /fp:precise /Zc:wchar_t /Zc:forScope /Zc:inline /GR /F
  o"vdot.dir\Release\\" /Fd"vdot.dir\Release\vc143.pdb" /external:W3 /Gd /TP /errorReport:queue "I:\ai-chat_v1\llama.cp
  p\llama-cpp\llama.cpp\pocs\vdot\vdot.cpp"
  vdot.cpp
Link:
  I:\Program Files\Microsoft Visual Studio\2022\Community\VC\Tools\MSVC\14.37.32822\bin\HostX64\x64\link.exe /ERRORREPO
  RT:QUEUE /OUT:"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\vdot.exe" /INCREMENTAL:NO /NOLOGO /LIBPA
  TH:"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.3\lib\x64" ..\..\common\Release\common.lib ..\..\Release\l
  lama.lib kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib adva
  pi32.lib /MANIFEST /MANIFESTUAC:"level='asInvoker' uiAccess='false'" /manifest:embed /PDB:"I:/ai-chat_v1/llama.cpp/ll
  ama-cpp/llama.cpp/build/bin/Release/vdot.pdb" /SUBSYSTEM:CONSOLE /TLBID:1 /DYNAMICBASE /NXCOMPAT /IMPLIB:"I:/ai-chat_
  v1/llama.cpp/llama-cpp/llama.cpp/build/pocs/vdot/Release/vdot.lib" /MACHINE:X64  /machine:x64 vdot.dir\Release\vdot.o
  bj
  vdot.vcxproj -> I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\bin\Release\vdot.exe
FinalizeBuildStatus:
  Deleting file "vdot.dir\Release\vdot.tlog\unsuccessfulbuild".
  Touching "vdot.dir\Release\vdot.tlog\vdot.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\pocs\vdot\vdot.vcxproj" (default targets).

PrepareForBuild:
  Creating directory "x64\Release\ALL_BUILD\".
  Creating directory "x64\Release\ALL_BUILD\ALL_BUILD.tlog\".
InitializeBuildStatus:
  Creating "x64\Release\ALL_BUILD\ALL_BUILD.tlog\unsuccessfulbuild" because "AlwaysCreate" was specified.
  Touching "x64\Release\ALL_BUILD\ALL_BUILD.tlog\unsuccessfulbuild".
CustomBuild:
  Building Custom Rule I:/ai-chat_v1/llama.cpp/llama-cpp/llama.cpp/CMakeLists.txt
FinalizeBuildStatus:
  Deleting file "x64\Release\ALL_BUILD\ALL_BUILD.tlog\unsuccessfulbuild".
  Touching "x64\Release\ALL_BUILD\ALL_BUILD.tlog\ALL_BUILD.lastbuildstate".
Done Building Project "I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default targets).

Build succeeded.

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.vcxproj" (default target) (4) ->
(ClCompile target) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\ggml-quants.c(532,26): warning C4244: '=': conversion from 'float' to 'in
t8_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\ggml-quants.c(750,36): warning C4244: '=': conversion from 'float' to 'in
t8_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\ggml-quants.c(751,36): warning C4244: '=': conversion from 'float' to 'in
t8_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ggml.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxproj" (default target) (5) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\llama.vcxproj" (default target) (6) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\llama.cpp(1103,31): warning C4305: 'initializing': truncation from 'doubl
e' to 'float' [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\llama.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\llama.cpp(2331,69): warning C4566: character represented by universal-cha
racter-name '\u010A' cannot be represented in the current code page (1252) [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp
\build\llama.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\llama.cpp(9054,28): warning C4146: unary minus operator applied to unsign
ed type, result still unsigned [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\llama.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\llama.cpp(9084,28): warning C4146: unary minus operator applied to unsign
ed type, result still unsigned [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\llama.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcxproj" (default target) (5) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\llava.cpp(32,24): warning C4244: 'initializing': conversio
n from 'double' to 'float', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\lla
va.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(251,20): warning C4267: 'initializing': conversio
n from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava
.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(465,9): warning C4297: 'clip_model_load': functio
n assumed not to throw an exception but does [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vc
xproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(714,46): warning C4267: 'argument': conversion fr
om 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcx
proj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(817,44): warning C4244: 'initializing': conversio
n from 'const _Ty' to 'uint8_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llav
a\llava.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(817,34): warning C4244: 'initializing': conversio
n from 'const _Ty' to 'const uint8_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\example
s\llava\llava.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(855,20): warning C4267: 'initializing': conversio
n from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava
.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(999,88): warning C4244: 'argument': conversion fr
om 'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vc
xproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(999,71): warning C4267: 'argument': conversion fr
om 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcx
proj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1002,88): warning C4244: 'argument': conversion f
rom 'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.v
cxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1002,71): warning C4267: 'argument': conversion f
rom 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vc
xproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1005,88): warning C4244: 'argument': conversion f
rom 'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.v
cxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1005,71): warning C4267: 'argument': conversion f
rom 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vc
xproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1008,88): warning C4244: 'argument': conversion f
rom 'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.v
cxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1008,71): warning C4267: 'argument': conversion f
rom 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vc
xproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1011,88): warning C4244: 'argument': conversion f
rom 'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.v
cxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1011,71): warning C4267: 'argument': conversion f
rom 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vc
xproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\clip.cpp(1073,42): warning C4244: 'return': conversion fro
m 'int64_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava.vcx
proj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\baby-llama\baby-llama.vcxproj" (default target) (7) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj" (default target) (8) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\sampling.cpp(75,45): warning C4267: 'initializing': conversion fro
m 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\sampling.cpp(75,20): warning C4267: 'initializing': conversion fro
m 'size_t' to 'const int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxpr
oj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\console.cpp(253,30): warning C4267: 'initializing': conversion fro
m 'size_t' to 'DWORD', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\console.cpp(407,28): warning C4267: 'initializing': conversion fro
m 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\common\common.cpp(879): warning C4715: 'gpt_random_prompt': not all contr
ol paths return a value [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\common\common.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\batched\batched.vcxproj" (default target) (9) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\batched\batched.cpp(72,45): warning C4267: 'initializing': conve
rsion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\batched
\batched.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\batched\batched.cpp(72,24): warning C4267: 'initializing': conve
rsion from 'size_t' to 'const int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\b
atched\batched.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\batched\batched.cpp(114,50): warning C4267: 'argument': conversi
on from 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\batche
d\batched.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\batched\batched.cpp(118,48): warning C4267: 'argument': conversi
on from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\batc
hed\batched.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\beam-search\beam-search.vcxproj" (default target) (11) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\beam-search\beam-search.cpp(163,83): warning C4267: 'argument':
conversion from 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\example
s\beam-search\beam-search.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\beam-search\beam-search.cpp(168,31): warning C4267: '+=': conver
sion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\beam-sea
rch\beam-search.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llama-bench\llama-bench.vcxproj" (default target) (20) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llama-bench\llama-bench.cpp(72,13): warning C4244: 'initializing
': conversion from 'double' to 'T', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\l
lama-bench\llama-bench.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\llava-cli.vcxproj" (default target) (21) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\llava\llava-cli.cpp(150,105): warning C4267: 'argument': convers
ion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\llava\lla
va-cli.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\parallel\parallel.vcxproj" (default target) (25) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\parallel\parallel.cpp(157,21): warning C4267: '=': conversion fr
om 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\parallel\pa
rallel.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\parallel\parallel.cpp(163,55): warning C4267: 'initializing': co
nversion from 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\
parallel\parallel.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\parallel\parallel.cpp(163,35): warning C4267: 'initializing': co
nversion from 'size_t' to 'const int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\exa
mples\parallel\parallel.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\parallel\parallel.cpp(248,68): warning C4267: 'argument': conver
sion from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\pa
rallel\parallel.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\parallel\parallel.cpp(256,58): warning C4267: '=': conversion fr
om 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\parallel\pa
rallel.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\save-load-state\save-load-state.vcxproj" (default target) (
30) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\save-load-state\save-load-state.cpp(42,69): warning C4267: 'argu
ment': conversion from 'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\
examples\save-load-state\save-load-state.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\save-load-state\save-load-state.cpp(43,26): warning C4267: '+=':
 conversion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\s
ave-load-state\save-load-state.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vcxproj" (default target) (31) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(82,16): warning C4267: 'initializing': convers
ion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\se
rver.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(96,52): warning C4267: '=': conversion from 's
ize_t' to 'uint8_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vc
xproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(120,48): warning C4267: '=': conversion from '
size_t' to 'uint8_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.v
cxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(781,49): warning C4267: '=': conversion from '
size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.v
cxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(785,93): warning C4267: 'argument': conversion
 from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\serve
r.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(832,67): warning C4101: 'e': unreferenced loca
l variable [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(887,67): warning C4267: 'argument': conversion
 from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server
\server.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(960,53): warning C4267: '-=': conversion from
'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.
vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1496,71): warning C4267: 'argument': conversio
n from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\serve
r\server.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1558,64): warning C4267: '=': conversion from
'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.
vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1594,72): warning C4267: '=': conversion from
'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.
vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1604,50): warning C4267: '=': conversion from
'size_t' to 'int32_t', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\server\server.
vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1612,78): warning C4267: 'argument': conversio
n from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\serve
r\server.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\server\server.cpp(1635,96): warning C4267: 'argument': conversio
n from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\serve
r\server.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\simple\simple.vcxproj" (default target) (32) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\simple\simple.cpp(71,45): warning C4267: 'initializing': convers
ion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\simple\si
mple.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\simple\simple.cpp(71,24): warning C4267: 'initializing': convers
ion from 'size_t' to 'const int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\sim
ple\simple.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\simple\simple.cpp(99,48): warning C4267: 'argument': conversion
from 'size_t' to 'llama_pos', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\simple\
simple.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\examples\speculative\speculative.vcxproj" (default target) (33) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\speculative\speculative.cpp(117,33): warning C4267: 'initializin
g': conversion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\example
s\speculative\speculative.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\speculative\speculative.cpp(117,23): warning C4267: 'initializin
g': conversion from 'size_t' to 'const int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\e
xamples\speculative\speculative.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\speculative\speculative.cpp(138,20): warning C4267: 'initializin
g': conversion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\example
s\speculative\speculative.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\examples\speculative\speculative.cpp(139,20): warning C4267: 'initializin
g': conversion from 'size_t' to 'int', possible loss of data [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\example
s\speculative\speculative.vcxproj]

"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\ALL_BUILD.vcxproj" (default target) (1) ->
"I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-llama-grammar.vcxproj" (default target) (37) ->
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\.\llama.cpp(1103,31): warning C4305: 'initializing': truncation from 'dou
ble' to 'float' [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-llama-grammar.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\.\llama.cpp(2331,69): warning C4566: character represented by universal-c
haracter-name '\u010A' cannot be represented in the current code page (1252) [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.c
pp\build\tests\test-llama-grammar.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\.\llama.cpp(9054,28): warning C4146: unary minus operator applied to unsi
gned type, result still unsigned [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-llama-grammar.vcxproj]
  I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\.\llama.cpp(9084,28): warning C4146: unary minus operator applied to unsi
gned type, result still unsigned [I:\ai-chat_v1\llama.cpp\llama-cpp\llama.cpp\build\tests\test-llama-grammar.vcxproj]

    70 Warning(s)
    0 Error(s)

Time Elapsed 00:02:44.05

tk-master commented 10 months ago

And I just tried a clean install llama-cpp-python to virtual env..

git clone --recurse-submodules https://github.com/abetlen/llama-cpp-python.git llama-cpp-python-main
cd llama-cpp-python-main

set FORCE_CMAKE=1 && set CMAKE_ARGS=-DLLAMA_CUBLAS=on -DBUILD_SHARED_LIBS=ON -DLLAMA_AVX=off -DLLAMA_AVX2=off -DLLAMA_FMA=off
python -m pip install -e .[all] --verbose --no-cache-dir

Log:

Cloning into 'llama-cpp-python-main'...
remote: Enumerating objects: 5280, done.
remote: Counting objects: 100% (1872/1872), done.
remote: Compressing objects: 100% (411/411), done.
remote: Total 5280 (delta 1638), reused 1583 (delta 1443), pack-reused 3408
Receiving objects: 100% (5280/5280), 2.63 MiB | 5.68 MiB/s, done.
Resolving deltas: 100% (3298/3298), done.
Submodule 'vendor/llama.cpp' (https://github.com/ggerganov/llama.cpp.git) registered for path 'vendor/llama.cpp'
Cloning into 'I:/ai-chat_v1/llama-cpp-python/llama-cpp-python-main/vendor/llama.cpp'...
remote: Enumerating objects: 12057, done.
remote: Counting objects: 100% (12057/12057), done.
remote: Compressing objects: 100% (3607/3607), done.
remote: Total 12057 (delta 8418), reused 11967 (delta 8375), pack-reused 0
Receiving objects: 100% (12057/12057), 14.21 MiB | 1.79 MiB/s, done.
Resolving deltas: 100% (8418/8418), done.
Submodule path 'vendor/llama.cpp': checked out 'a75fa576abba9d37f463580c379e4bbf1e1ad03c'
Using pip 23.3 from I:\ai-chat_v1\llama-cpp-python\venv_mamba\env\lib\site-packages\pip (python 3.10)
Obtaining file:///I:/ai-chat_v1/llama-cpp-python/llama-cpp-python-main
  Running command pip subprocess to install build dependencies
  Collecting scikit-build-core>=0.5.1 (from scikit-build-core[pyproject]>=0.5.1)
    Downloading scikit_build_core-0.6.1-py3-none-any.whl.metadata (17 kB)
  Collecting exceptiongroup (from scikit-build-core>=0.5.1->scikit-build-core[pyproject]>=0.5.1)
    Downloading exceptiongroup-1.1.3-py3-none-any.whl.metadata (6.1 kB)
  Collecting packaging>=20.9 (from scikit-build-core>=0.5.1->scikit-build-core[pyproject]>=0.5.1)
    Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)
  Collecting tomli>=1.1 (from scikit-build-core>=0.5.1->scikit-build-core[pyproject]>=0.5.1)
    Downloading tomli-2.0.1-py3-none-any.whl (12 kB)
  Collecting pathspec>=0.10.1 (from scikit-build-core[pyproject]>=0.5.1)
    Downloading pathspec-0.11.2-py3-none-any.whl.metadata (19 kB)
  Collecting pyproject-metadata>=0.5 (from scikit-build-core[pyproject]>=0.5.1)
    Downloading pyproject_metadata-0.7.1-py3-none-any.whl (7.4 kB)
  Downloading scikit_build_core-0.6.1-py3-none-any.whl (134 kB)
     -------------------------------------- 134.5/134.5 kB 176.8 kB/s eta 0:00:00
  Downloading packaging-23.2-py3-none-any.whl (53 kB)
     ---------------------------------------- 53.0/53.0 kB 195.3 kB/s eta 0:00:00
  Downloading pathspec-0.11.2-py3-none-any.whl (29 kB)
  Downloading exceptiongroup-1.1.3-py3-none-any.whl (14 kB)
  Installing collected packages: tomli, pathspec, packaging, exceptiongroup, scikit-build-core, pyproject-metadata
  Successfully installed exceptiongroup-1.1.3 packaging-23.2 pathspec-0.11.2 pyproject-metadata-0.7.1 scikit-build-core-0.6.1 tomli-2.0.1
  Installing build dependencies ... done
  Running command Checking if build backend supports build_editable
  Checking if build backend supports build_editable ... done
  Running command Getting requirements to build editable
  Getting requirements to build editable ... done
  Running command Preparing editable metadata (pyproject.toml)
  *** scikit-build-core 0.6.1 using CMake 3.27.7 (metadata_editable)
  Preparing editable metadata (pyproject.toml) ... done
Collecting typing-extensions>=4.5.0
  Obtaining dependency information for typing-extensions>=4.5.0 from https://files.pythonhosted.org/packages/24/21/7d397a4b7934ff4028987914ac1044d3b7d52712f30e2ac7a2ae5bc86dd0/typing_extensions-4.8.0-py3-none-any.whl.metadata
  Downloading typing_extensions-4.8.0-py3-none-any.whl.metadata (3.0 kB)
Collecting numpy>=1.20.0
  Obtaining dependency information for numpy>=1.20.0 from https://files.pythonhosted.org/packages/57/09/fe9282ffb0217176b0185900945189b6beaec4f94ff46afb76bcd9b68e30/numpy-1.26.1-cp310-cp310-win_amd64.whl.metadata
  Downloading numpy-1.26.1-cp310-cp310-win_amd64.whl.metadata (61 kB)
     ---------------------------------------- 61.2/61.2 kB 653.3 kB/s eta 0:00:00
Collecting diskcache>=5.6.1
  Obtaining dependency information for diskcache>=5.6.1 from https://files.pythonhosted.org/packages/3f/27/4570e78fc0bf5ea0ca45eb1de3818a23787af9b390c0b0a0033a1b8236f9/diskcache-5.6.3-py3-none-any.whl.metadata
  Downloading diskcache-5.6.3-py3-none-any.whl.metadata (20 kB)
Collecting black>=23.3.0 (from llama_cpp_python==0.2.16)
  Obtaining dependency information for black>=23.3.0 from https://files.pythonhosted.org/packages/ce/59/b0c04cd9ddccb19660ba92c6ebb559face8533b12f04003b912c5c38f5db/black-23.11.0-cp310-cp310-win_amd64.whl.metadata
  Downloading black-23.11.0-cp310-cp310-win_amd64.whl.metadata (66 kB)
     ---------------------------------------- 66.9/66.9 kB 3.5 MB/s eta 0:00:00
Collecting twine>=4.0.2 (from llama_cpp_python==0.2.16)
  Downloading twine-4.0.2-py3-none-any.whl (36 kB)
Collecting mkdocs>=1.4.3 (from llama_cpp_python==0.2.16)
  Obtaining dependency information for mkdocs>=1.4.3 from https://files.pythonhosted.org/packages/89/58/aa3301b23966a71d7f8e55233f467b3cec94a651434e9cd9053811342539/mkdocs-1.5.3-py3-none-any.whl.metadata
  Downloading mkdocs-1.5.3-py3-none-any.whl.metadata (6.2 kB)
Collecting mkdocstrings>=0.22.0 (from mkdocstrings[python]>=0.22.0; extra == "dev"->llama_cpp_python==0.2.16)
  Obtaining dependency information for mkdocstrings>=0.22.0 from https://files.pythonhosted.org/packages/64/2f/6b72f8f8bf168a5820c6c38bffe54d25cfdafd9b4be6fbb335a9a57dd7c9/mkdocstrings-0.23.0-py3-none-any.whl.metadata
  Downloading mkdocstrings-0.23.0-py3-none-any.whl.metadata (7.6 kB)
Collecting mkdocs-material>=9.1.18 (from llama_cpp_python==0.2.16)
  Obtaining dependency information for mkdocs-material>=9.1.18 from https://files.pythonhosted.org/packages/b6/cd/588b4e12296edd6c554475d2bc90237e304d479cd9ba971ae6fa80d0051f/mkdocs_material-9.4.8-py3-none-any.whl.metadata
  Downloading mkdocs_material-9.4.8-py3-none-any.whl.metadata (15 kB)
Collecting pytest>=7.4.0 (from llama_cpp_python==0.2.16)
  Obtaining dependency information for pytest>=7.4.0 from https://files.pythonhosted.org/packages/f3/8c/f16efd81ca8e293b2cc78f111190a79ee539d0d5d36ccd49975cb3beac60/pytest-7.4.3-py3-none-any.whl.metadata
  Downloading pytest-7.4.3-py3-none-any.whl.metadata (7.9 kB)
Collecting httpx>=0.24.1 (from llama_cpp_python==0.2.16)
  Obtaining dependency information for httpx>=0.24.1 from https://files.pythonhosted.org/packages/82/61/a5fca4a1e88e40969bbd0cf0d981f3aa76d5057db160b94f49603fc18740/httpx-0.25.1-py3-none-any.whl.metadata
  Downloading httpx-0.25.1-py3-none-any.whl.metadata (7.1 kB)
Collecting uvicorn>=0.22.0 (from llama_cpp_python==0.2.16)
  Obtaining dependency information for uvicorn>=0.22.0 from https://files.pythonhosted.org/packages/7e/17/4b7a76fffa7babf397481040d8aef2725b2b81ae19f1a31b5ca0c17d49e6/uvicorn-0.24.0.post1-py3-none-any.whl.metadata
  Downloading uvicorn-0.24.0.post1-py3-none-any.whl.metadata (6.4 kB)
Collecting fastapi>=0.100.0 (from llama_cpp_python==0.2.16)
  Obtaining dependency information for fastapi>=0.100.0 from https://files.pythonhosted.org/packages/f3/4f/0ce34195b63240b6693086496c9bab4ef23999112184399a3e88854c7674/fastapi-0.104.1-py3-none-any.whl.metadata
  Downloading fastapi-0.104.1-py3-none-any.whl.metadata (24 kB)
Collecting pydantic-settings>=2.0.1 (from llama_cpp_python==0.2.16)
  Obtaining dependency information for pydantic-settings>=2.0.1 from https://files.pythonhosted.org/packages/46/92/918ef6b14d54c6a4fccdecd65b3ee15360ca2b4aa52d5c9c4f39f99b4c56/pydantic_settings-2.0.3-py3-none-any.whl.metadata
  Downloading pydantic_settings-2.0.3-py3-none-any.whl.metadata (2.9 kB)
Collecting sse-starlette>=1.6.1 (from llama_cpp_python==0.2.16)
  Obtaining dependency information for sse-starlette>=1.6.1 from https://files.pythonhosted.org/packages/e9/83/5825dd7be255bbd3587a45651804092817eadf7070db757331d0e9c2ee47/sse_starlette-1.6.5-py3-none-any.whl.metadata
  Downloading sse_starlette-1.6.5-py3-none-any.whl.metadata (6.7 kB)
Collecting starlette-context<0.4,>=0.3.6 (from llama_cpp_python==0.2.16)
  Downloading starlette_context-0.3.6-py3-none-any.whl (12 kB)
Collecting click>=8.0.0 (from black>=23.3.0->llama_cpp_python==0.2.16)
  Obtaining dependency information for click>=8.0.0 from https://files.pythonhosted.org/packages/00/2e/d53fa4befbf2cfa713304affc7ca780ce4fc1fd8710527771b58311a3229/click-8.1.7-py3-none-any.whl.metadata
  Downloading click-8.1.7-py3-none-any.whl.metadata (3.0 kB)
Collecting mypy-extensions>=0.4.3 (from black>=23.3.0->llama_cpp_python==0.2.16)
  Downloading mypy_extensions-1.0.0-py3-none-any.whl (4.7 kB)
Collecting packaging>=22.0 (from black>=23.3.0->llama_cpp_python==0.2.16)
  Obtaining dependency information for packaging>=22.0 from https://files.pythonhosted.org/packages/ec/1a/610693ac4ee14fcdf2d9bf3c493370e4f2ef7ae2e19217d7a237ff42367d/packaging-23.2-py3-none-any.whl.metadata
  Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)
Collecting pathspec>=0.9.0 (from black>=23.3.0->llama_cpp_python==0.2.16)
  Obtaining dependency information for pathspec>=0.9.0 from https://files.pythonhosted.org/packages/b4/2a/9b1be29146139ef459188f5e420a66e835dda921208db600b7037093891f/pathspec-0.11.2-py3-none-any.whl.metadata
  Downloading pathspec-0.11.2-py3-none-any.whl.metadata (19 kB)
Collecting platformdirs>=2 (from black>=23.3.0->llama_cpp_python==0.2.16)
  Obtaining dependency information for platformdirs>=2 from https://files.pythonhosted.org/packages/31/16/70be3b725073035aa5fc3229321d06e22e73e3e09f6af78dcfdf16c7636c/platformdirs-4.0.0-py3-none-any.whl.metadata
  Downloading platformdirs-4.0.0-py3-none-any.whl.metadata (11 kB)
Collecting tomli>=1.1.0 (from black>=23.3.0->llama_cpp_python==0.2.16)
  Downloading tomli-2.0.1-py3-none-any.whl (12 kB)
Collecting anyio<4.0.0,>=3.7.1 (from fastapi>=0.100.0->llama_cpp_python==0.2.16)
  Obtaining dependency information for anyio<4.0.0,>=3.7.1 from https://files.pythonhosted.org/packages/19/24/44299477fe7dcc9cb58d0a57d5a7588d6af2ff403fdd2d47a246c91a3246/anyio-3.7.1-py3-none-any.whl.metadata
  Downloading anyio-3.7.1-py3-none-any.whl.metadata (4.7 kB)
Collecting pydantic!=1.8,!=1.8.1,!=2.0.0,!=2.0.1,!=2.1.0,<3.0.0,>=1.7.4 (from fastapi>=0.100.0->llama_cpp_python==0.2.16)
  Obtaining dependency information for pydantic!=1.8,!=1.8.1,!=2.0.0,!=2.0.1,!=2.1.0,<3.0.0,>=1.7.4 from https://files.pythonhosted.org/packages/73/66/0a72c9fcde42e5650c8d8d5c5c1873b9a3893018020c77ca8eb62708b923/pydantic-2.4.2-py3-none-any.whl.metadata
  Downloading pydantic-2.4.2-py3-none-any.whl.metadata (158 kB)
     ---------------------------------------- 158.6/158.6 kB 1.9 MB/s eta 0:00:00
Collecting starlette<0.28.0,>=0.27.0 (from fastapi>=0.100.0->llama_cpp_python==0.2.16)
  Obtaining dependency information for starlette<0.28.0,>=0.27.0 from https://files.pythonhosted.org/packages/58/f8/e2cca22387965584a409795913b774235752be4176d276714e15e1a58884/starlette-0.27.0-py3-none-any.whl.metadata
  Downloading starlette-0.27.0-py3-none-any.whl.metadata (5.8 kB)
Collecting certifi (from httpx>=0.24.1->llama_cpp_python==0.2.16)
  Obtaining dependency information for certifi from https://files.pythonhosted.org/packages/4c/dd/2234eab22353ffc7d94e8d13177aaa050113286e93e7b40eae01fbf7c3d9/certifi-2023.7.22-py3-none-any.whl.metadata
  Downloading certifi-2023.7.22-py3-none-any.whl.metadata (2.2 kB)
Collecting httpcore (from httpx>=0.24.1->llama_cpp_python==0.2.16)
  Obtaining dependency information for httpcore from https://files.pythonhosted.org/packages/56/ba/78b0a99c4da0ff8b0f59defa2f13ca4668189b134bd9840b6202a93d9a0f/httpcore-1.0.2-py3-none-any.whl.metadata
  Downloading httpcore-1.0.2-py3-none-any.whl.metadata (20 kB)
Collecting idna (from httpx>=0.24.1->llama_cpp_python==0.2.16)
  Downloading idna-3.4-py3-none-any.whl (61 kB)
     ---------------------------------------- 61.5/61.5 kB ? eta 0:00:00
Collecting sniffio (from httpx>=0.24.1->llama_cpp_python==0.2.16)
  Downloading sniffio-1.3.0-py3-none-any.whl (10 kB)
Collecting colorama>=0.4 (from mkdocs>=1.4.3->llama_cpp_python==0.2.16)
  Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB)
Collecting ghp-import>=1.0 (from mkdocs>=1.4.3->llama_cpp_python==0.2.16)
  Downloading ghp_import-2.1.0-py3-none-any.whl (11 kB)
Collecting jinja2>=2.11.1 (from mkdocs>=1.4.3->llama_cpp_python==0.2.16)
  Downloading Jinja2-3.1.2-py3-none-any.whl (133 kB)
     ---------------------------------------- 133.1/133.1 kB 7.7 MB/s eta 0:00:00
Collecting markdown>=3.2.1 (from mkdocs>=1.4.3->llama_cpp_python==0.2.16)
  Obtaining dependency information for markdown>=3.2.1 from https://files.pythonhosted.org/packages/70/58/2c5a654173937d9f540a4971c569b44dcd55e5424a484d954cdaeebcf79c/Markdown-3.5.1-py3-none-any.whl.metadata
  Downloading Markdown-3.5.1-py3-none-any.whl.metadata (7.1 kB)
Collecting markupsafe>=2.0.1 (from mkdocs>=1.4.3->llama_cpp_python==0.2.16)
  Obtaining dependency information for markupsafe>=2.0.1 from https://files.pythonhosted.org/packages/84/a8/c4aebb8a14a1d39d5135eb8233a0b95831cdc42c4088358449c3ed657044/MarkupSafe-2.1.3-cp310-cp310-win_amd64.whl.metadata
  Downloading MarkupSafe-2.1.3-cp310-cp310-win_amd64.whl.metadata (3.1 kB)
Collecting mergedeep>=1.3.4 (from mkdocs>=1.4.3->llama_cpp_python==0.2.16)
  Downloading mergedeep-1.3.4-py3-none-any.whl (6.4 kB)
Collecting pyyaml-env-tag>=0.1 (from mkdocs>=1.4.3->llama_cpp_python==0.2.16)
  Downloading pyyaml_env_tag-0.1-py3-none-any.whl (3.9 kB)
Collecting pyyaml>=5.1 (from mkdocs>=1.4.3->llama_cpp_python==0.2.16)
  Obtaining dependency information for pyyaml>=5.1 from https://files.pythonhosted.org/packages/24/97/9b59b43431f98d01806b288532da38099cc6f2fea0f3d712e21e269c0279/PyYAML-6.0.1-cp310-cp310-win_amd64.whl.metadata
  Downloading PyYAML-6.0.1-cp310-cp310-win_amd64.whl.metadata (2.1 kB)
  Link requires a different Python (3.10.9 not in: '>=2.7, <3.6'): https://files.pythonhosted.org/packages/8e/68/261843fe0fed52c7f4364325bde666c858c582547c6a32e90ce9c184a5d7/watchdog-0.10.6.tar.gz (from https://pypi.org/simple/watchdog/) (requires-python:>=2.7, <3.6)
  Link requires a different Python (3.10.9 not in: '>=2.7, <3.6'): https://files.pythonhosted.org/packages/9d/f8/99b803268b9cc809df0934c54266777db3ab7698ff6c82021a00e4fdfc03/watchdog-0.10.7.tar.gz (from https://pypi.org/simple/watchdog/) (requires-python:>=2.7, <3.6)
Collecting watchdog>=2.0 (from mkdocs>=1.4.3->llama_cpp_python==0.2.16)
  Downloading watchdog-3.0.0-py3-none-win_amd64.whl (82 kB)
     ---------------------------------------- 82.0/82.0 kB ? eta 0:00:00
Collecting babel~=2.10 (from mkdocs-material>=9.1.18->llama_cpp_python==0.2.16)
  Obtaining dependency information for babel~=2.10 from https://files.pythonhosted.org/packages/86/14/5dc2eb02b7cc87b2f95930310a2cc5229198414919a116b564832c747bc1/Babel-2.13.1-py3-none-any.whl.metadata
  Downloading Babel-2.13.1-py3-none-any.whl.metadata (1.6 kB)
Collecting mkdocs-material-extensions~=1.3 (from mkdocs-material>=9.1.18->llama_cpp_python==0.2.16)
  Obtaining dependency information for mkdocs-material-extensions~=1.3 from https://files.pythonhosted.org/packages/e9/f1/a4d3c449b7fa9e9694236a621b12e994653bd45ad96dc041cbb58a630005/mkdocs_material_extensions-1.3-py3-none-any.whl.metadata
  Downloading mkdocs_material_extensions-1.3-py3-none-any.whl.metadata (6.9 kB)
Collecting paginate~=0.5 (from mkdocs-material>=9.1.18->llama_cpp_python==0.2.16)
  Downloading paginate-0.5.6.tar.gz (12 kB)
  Running command python setup.py egg_info
  running egg_info
  creating C:\Users\Tkay\AppData\Local\Temp\pip-pip-egg-info-60yh78xq\paginate.egg-info
  writing C:\Users\Tkay\AppData\Local\Temp\pip-pip-egg-info-60yh78xq\paginate.egg-info\PKG-INFO
  writing dependency_links to C:\Users\Tkay\AppData\Local\Temp\pip-pip-egg-info-60yh78xq\paginate.egg-info\dependency_links.txt
  writing top-level names to C:\Users\Tkay\AppData\Local\Temp\pip-pip-egg-info-60yh78xq\paginate.egg-info\top_level.txt
  writing manifest file 'C:\Users\Tkay\AppData\Local\Temp\pip-pip-egg-info-60yh78xq\paginate.egg-info\SOURCES.txt'
  reading manifest file 'C:\Users\Tkay\AppData\Local\Temp\pip-pip-egg-info-60yh78xq\paginate.egg-info\SOURCES.txt'
  reading manifest template 'MANIFEST.in'
  warning: no files found matching '*.ini'
  warning: no files found matching '*.rst'
  warning: no files found matching '*.png' under directory '*.ico'
  warning: no files found matching '*.css' under directory '*.ico'
  warning: no files found matching '*.gif' under directory '*.ico'
  warning: no files found matching '*.jpg' under directory '*.ico'
  warning: no files found matching '*.pt' under directory '*.ico'
  warning: no files found matching '*.txt' under directory '*.ico'
  warning: no files found matching '*.mak' under directory '*.ico'
  warning: no files found matching '*.mako' under directory '*.ico'
  warning: no files found matching '*.js' under directory '*.ico'
  warning: no files found matching '*.html' under directory '*.ico'
  warning: no files found matching '*.xml' under directory '*.ico'
  writing manifest file 'C:\Users\Tkay\AppData\Local\Temp\pip-pip-egg-info-60yh78xq\paginate.egg-info\SOURCES.txt'
  Preparing metadata (setup.py) ... done
Collecting pygments~=2.16 (from mkdocs-material>=9.1.18->llama_cpp_python==0.2.16)
  Obtaining dependency information for pygments~=2.16 from https://files.pythonhosted.org/packages/43/88/29adf0b44ba6ac85045e63734ae0997d3c58d8b1a91c914d240828d0d73d/Pygments-2.16.1-py3-none-any.whl.metadata
  Downloading Pygments-2.16.1-py3-none-any.whl.metadata (2.5 kB)
Collecting pymdown-extensions~=10.2 (from mkdocs-material>=9.1.18->llama_cpp_python==0.2.16)
  Obtaining dependency information for pymdown-extensions~=10.2 from https://files.pythonhosted.org/packages/0e/f3/31b0d3472a65eabe2eeeb682d7ab142b6a29045d997bf1b73aabcf59a3e2/pymdown_extensions-10.3.1-py3-none-any.whl.metadata
  Downloading pymdown_extensions-10.3.1-py3-none-any.whl.metadata (3.2 kB)
Collecting regex>=2022.4 (from mkdocs-material>=9.1.18->llama_cpp_python==0.2.16)
  Obtaining dependency information for regex>=2022.4 from https://files.pythonhosted.org/packages/33/03/91c9509b43154795fb848a4cf8cef5b37302b3b3ccf8a9763046ea528c6b/regex-2023.10.3-cp310-cp310-win_amd64.whl.metadata
  Downloading regex-2023.10.3-cp310-cp310-win_amd64.whl.metadata (41 kB)
     ---------------------------------------- 42.0/42.0 kB ? eta 0:00:00
Collecting requests~=2.26 (from mkdocs-material>=9.1.18->llama_cpp_python==0.2.16)
  Obtaining dependency information for requests~=2.26 from https://files.pythonhosted.org/packages/70/8e/0e2d847013cb52cd35b38c009bb167a1a26b2ce6cd6965bf26b47bc0bf44/requests-2.31.0-py3-none-any.whl.metadata
  Downloading requests-2.31.0-py3-none-any.whl.metadata (4.6 kB)
Collecting mkdocs-autorefs>=0.3.1 (from mkdocstrings>=0.22.0->mkdocstrings[python]>=0.22.0; extra == "dev"->llama_cpp_python==0.2.16)
  Obtaining dependency information for mkdocs-autorefs>=0.3.1 from https://files.pythonhosted.org/packages/21/5f/fe501daf6f06b93d5d9dff4319c04ad6e74965348dff22465bdd53e5e2d9/mkdocs_autorefs-0.5.0-py3-none-any.whl.metadata
  Downloading mkdocs_autorefs-0.5.0-py3-none-any.whl.metadata (4.0 kB)
Collecting mkdocstrings-python>=0.5.2 (from mkdocstrings[python]>=0.22.0; extra == "dev"->llama_cpp_python==0.2.16)
  Obtaining dependency information for mkdocstrings-python>=0.5.2 from https://files.pythonhosted.org/packages/35/ec/a9d5e3b98eb5d7da229481049f8031c2dad2a72d48192d3471c72ae0eda6/mkdocstrings_python-1.7.3-py3-none-any.whl.metadata
  Downloading mkdocstrings_python-1.7.3-py3-none-any.whl.metadata (5.7 kB)
Collecting python-dotenv>=0.21.0 (from pydantic-settings>=2.0.1->llama_cpp_python==0.2.16)
  Downloading python_dotenv-1.0.0-py3-none-any.whl (19 kB)
Collecting iniconfig (from pytest>=7.4.0->llama_cpp_python==0.2.16)
  Downloading iniconfig-2.0.0-py3-none-any.whl (5.9 kB)
Collecting pluggy<2.0,>=0.12 (from pytest>=7.4.0->llama_cpp_python==0.2.16)
  Obtaining dependency information for pluggy<2.0,>=0.12 from https://files.pythonhosted.org/packages/05/b8/42ed91898d4784546c5f06c60506400548db3f7a4b3fb441cba4e5c17952/pluggy-1.3.0-py3-none-any.whl.metadata
  Downloading pluggy-1.3.0-py3-none-any.whl.metadata (4.3 kB)
Collecting exceptiongroup>=1.0.0rc8 (from pytest>=7.4.0->llama_cpp_python==0.2.16)
  Obtaining dependency information for exceptiongroup>=1.0.0rc8 from https://files.pythonhosted.org/packages/ad/83/b71e58666f156a39fb29417e4c8ca4bc7400c0dd4ed9e8842ab54dc8c344/exceptiongroup-1.1.3-py3-none-any.whl.metadata
  Downloading exceptiongroup-1.1.3-py3-none-any.whl.metadata (6.1 kB)
Collecting pkginfo>=1.8.1 (from twine>=4.0.2->llama_cpp_python==0.2.16)
  Downloading pkginfo-1.9.6-py3-none-any.whl (30 kB)
Collecting readme-renderer>=35.0 (from twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for readme-renderer>=35.0 from https://files.pythonhosted.org/packages/b5/7e/992e0e21b37cadd668226f75fef0aa81bf21c2426c98bc06a55e514cb323/readme_renderer-42.0-py3-none-any.whl.metadata
  Downloading readme_renderer-42.0-py3-none-any.whl.metadata (2.8 kB)
Collecting requests-toolbelt!=0.9.0,>=0.8.0 (from twine>=4.0.2->llama_cpp_python==0.2.16)
  Downloading requests_toolbelt-1.0.0-py2.py3-none-any.whl (54 kB)
     ---------------------------------------- 54.5/54.5 kB ? eta 0:00:00
Collecting urllib3>=1.26.0 (from twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for urllib3>=1.26.0 from https://files.pythonhosted.org/packages/d2/b2/b157855192a68541a91ba7b2bbcb91f1b4faa51f8bae38d8005c034be524/urllib3-2.0.7-py3-none-any.whl.metadata
  Downloading urllib3-2.0.7-py3-none-any.whl.metadata (6.6 kB)
Collecting importlib-metadata>=3.6 (from twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for importlib-metadata>=3.6 from https://files.pythonhosted.org/packages/cc/37/db7ba97e676af155f5fcb1a35466f446eadc9104e25b83366e8088c9c926/importlib_metadata-6.8.0-py3-none-any.whl.metadata
  Downloading importlib_metadata-6.8.0-py3-none-any.whl.metadata (5.1 kB)
Collecting keyring>=15.1 (from twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for keyring>=15.1 from https://files.pythonhosted.org/packages/0e/8f/5772801169cf62e8232721034f91f81e33b0cfa6e51d3bf6ff65c503af2a/keyring-24.2.0-py3-none-any.whl.metadata
  Downloading keyring-24.2.0-py3-none-any.whl.metadata (20 kB)
Collecting rfc3986>=1.4.0 (from twine>=4.0.2->llama_cpp_python==0.2.16)
  Downloading rfc3986-2.0.0-py2.py3-none-any.whl (31 kB)
Collecting rich>=12.0.0 (from twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for rich>=12.0.0 from https://files.pythonhosted.org/packages/be/2a/4e62ff633612f746f88618852a626bbe24226eba5e7ac90e91dcfd6a414e/rich-13.6.0-py3-none-any.whl.metadata
  Downloading rich-13.6.0-py3-none-any.whl.metadata (18 kB)
Collecting h11>=0.8 (from uvicorn>=0.22.0->llama_cpp_python==0.2.16)
  Downloading h11-0.14.0-py3-none-any.whl (58 kB)
     ---------------------------------------- 58.3/58.3 kB 3.2 MB/s eta 0:00:00
Collecting python-dateutil>=2.8.1 (from ghp-import>=1.0->mkdocs>=1.4.3->llama_cpp_python==0.2.16)
  Downloading python_dateutil-2.8.2-py2.py3-none-any.whl (247 kB)
     ---------------------------------------- 247.7/247.7 kB 7.7 MB/s eta 0:00:00
Collecting zipp>=0.5 (from importlib-metadata>=3.6->twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for zipp>=0.5 from https://files.pythonhosted.org/packages/d9/66/48866fc6b158c81cc2bfecc04c480f105c6040e8b077bc54c634b4a67926/zipp-3.17.0-py3-none-any.whl.metadata
  Downloading zipp-3.17.0-py3-none-any.whl.metadata (3.7 kB)
Collecting jaraco.classes (from keyring>=15.1->twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for jaraco.classes from https://files.pythonhosted.org/packages/c7/6b/1bc8fa93ea85146e08f0e0883bc579b7c7328364ed7df90b1628dcb36e10/jaraco.classes-3.3.0-py3-none-any.whl.metadata
  Downloading jaraco.classes-3.3.0-py3-none-any.whl.metadata (2.9 kB)
Collecting pywin32-ctypes>=0.2.0 (from keyring>=15.1->twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for pywin32-ctypes>=0.2.0 from https://files.pythonhosted.org/packages/a4/bc/78b2c00cc64c31dbb3be42a0e8600bcebc123ad338c3b714754d668c7c2d/pywin32_ctypes-0.2.2-py3-none-any.whl.metadata
  Downloading pywin32_ctypes-0.2.2-py3-none-any.whl.metadata (3.8 kB)
Collecting griffe>=0.35 (from mkdocstrings-python>=0.5.2->mkdocstrings[python]>=0.22.0; extra == "dev"->llama_cpp_python==0.2.16)
  Obtaining dependency information for griffe>=0.35 from https://files.pythonhosted.org/packages/11/99/b8ab0cdd202af44b45fcec04d53ef1a1c5d4964b035b5fb4df2761e3c815/griffe-0.36.9-py3-none-any.whl.metadata
  Downloading griffe-0.36.9-py3-none-any.whl.metadata (6.1 kB)
Collecting annotated-types>=0.4.0 (from pydantic!=1.8,!=1.8.1,!=2.0.0,!=2.0.1,!=2.1.0,<3.0.0,>=1.7.4->fastapi>=0.100.0->llama_cpp_python==0.2.16)
  Obtaining dependency information for annotated-types>=0.4.0 from https://files.pythonhosted.org/packages/28/78/d31230046e58c207284c6b2c4e8d96e6d3cb4e52354721b944d3e1ee4aa5/annotated_types-0.6.0-py3-none-any.whl.metadata
  Downloading annotated_types-0.6.0-py3-none-any.whl.metadata (12 kB)
Collecting pydantic-core==2.10.1 (from pydantic!=1.8,!=1.8.1,!=2.0.0,!=2.0.1,!=2.1.0,<3.0.0,>=1.7.4->fastapi>=0.100.0->llama_cpp_python==0.2.16)
  Obtaining dependency information for pydantic-core==2.10.1 from https://files.pythonhosted.org/packages/45/49/1a11586e2e4bc052581cc7f9d3090c23fb49c2572536dd26004764603959/pydantic_core-2.10.1-cp310-none-win_amd64.whl.metadata
  Downloading pydantic_core-2.10.1-cp310-none-win_amd64.whl.metadata (6.6 kB)
Collecting nh3>=0.2.14 (from readme-renderer>=35.0->twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for nh3>=0.2.14 from https://files.pythonhosted.org/packages/53/4c/c34a0a48e81f3328e3300038d52195a2387d82570b0f17df53033c40bc23/nh3-0.2.14-cp37-abi3-win_amd64.whl.metadata
  Downloading nh3-0.2.14-cp37-abi3-win_amd64.whl.metadata (1.6 kB)
Collecting docutils>=0.13.1 (from readme-renderer>=35.0->twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for docutils>=0.13.1 from https://files.pythonhosted.org/packages/26/87/f238c0670b94533ac0353a4e2a1a771a0cc73277b88bff23d3ae35a256c1/docutils-0.20.1-py3-none-any.whl.metadata
  Downloading docutils-0.20.1-py3-none-any.whl.metadata (2.8 kB)
Collecting charset-normalizer<4,>=2 (from requests~=2.26->mkdocs-material>=9.1.18->llama_cpp_python==0.2.16)
  Obtaining dependency information for charset-normalizer<4,>=2 from https://files.pythonhosted.org/packages/a2/a0/4af29e22cb5942488cf45630cbdd7cefd908768e69bdd90280842e4e8529/charset_normalizer-3.3.2-cp310-cp310-win_amd64.whl.metadata
  Downloading charset_normalizer-3.3.2-cp310-cp310-win_amd64.whl.metadata (34 kB)
Collecting markdown-it-py>=2.2.0 (from rich>=12.0.0->twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for markdown-it-py>=2.2.0 from https://files.pythonhosted.org/packages/42/d7/1ec15b46af6af88f19b8e5ffea08fa375d433c998b8a7639e76935c14f1f/markdown_it_py-3.0.0-py3-none-any.whl.metadata
  Downloading markdown_it_py-3.0.0-py3-none-any.whl.metadata (6.9 kB)
Collecting mdurl~=0.1 (from markdown-it-py>=2.2.0->rich>=12.0.0->twine>=4.0.2->llama_cpp_python==0.2.16)
  Downloading mdurl-0.1.2-py3-none-any.whl (10.0 kB)
Collecting six>=1.5 (from python-dateutil>=2.8.1->ghp-import>=1.0->mkdocs>=1.4.3->llama_cpp_python==0.2.16)
  Downloading six-1.16.0-py2.py3-none-any.whl (11 kB)
Collecting more-itertools (from jaraco.classes->keyring>=15.1->twine>=4.0.2->llama_cpp_python==0.2.16)
  Obtaining dependency information for more-itertools from https://files.pythonhosted.org/packages/5a/cb/6dce742ea14e47d6f565589e859ad225f2a5de576d7696e0623b784e226b/more_itertools-10.1.0-py3-none-any.whl.metadata
  Downloading more_itertools-10.1.0-py3-none-any.whl.metadata (33 kB)
Downloading diskcache-5.6.3-py3-none-any.whl (45 kB)
   ---------------------------------------- 45.5/45.5 kB ? eta 0:00:00
Downloading numpy-1.26.1-cp310-cp310-win_amd64.whl (15.8 MB)
   ---------------------------------------- 15.8/15.8 MB 13.6 MB/s eta 0:00:00
Downloading typing_extensions-4.8.0-py3-none-any.whl (31 kB)
Downloading black-23.11.0-cp310-cp310-win_amd64.whl (1.3 MB)
   ---------------------------------------- 1.3/1.3 MB 14.1 MB/s eta 0:00:00
Downloading fastapi-0.104.1-py3-none-any.whl (92 kB)
   ---------------------------------------- 92.9/92.9 kB ? eta 0:00:00
Downloading httpx-0.25.1-py3-none-any.whl (75 kB)
   ---------------------------------------- 75.0/75.0 kB ? eta 0:00:00
Downloading mkdocs-1.5.3-py3-none-any.whl (3.7 MB)
   ---------------------------------------- 3.7/3.7 MB 13.1 MB/s eta 0:00:00
Downloading mkdocs_material-9.4.8-py3-none-any.whl (8.4 MB)
   ---------------------------------------- 8.4/8.4 MB 9.8 MB/s eta 0:00:00
Downloading mkdocstrings-0.23.0-py3-none-any.whl (25 kB)
Downloading pydantic_settings-2.0.3-py3-none-any.whl (11 kB)
Downloading pytest-7.4.3-py3-none-any.whl (325 kB)
   ---------------------------------------- 325.1/325.1 kB 10.2 MB/s eta 0:00:00
Downloading sse_starlette-1.6.5-py3-none-any.whl (9.6 kB)
Downloading uvicorn-0.24.0.post1-py3-none-any.whl (59 kB)
   ---------------------------------------- 59.7/59.7 kB ? eta 0:00:00
Downloading anyio-3.7.1-py3-none-any.whl (80 kB)
   ---------------------------------------- 80.9/80.9 kB ? eta 0:00:00
Downloading Babel-2.13.1-py3-none-any.whl (10.1 MB)
   ---------------------------------------- 10.1/10.1 MB 12.6 MB/s eta 0:00:00
Downloading click-8.1.7-py3-none-any.whl (97 kB)
   ---------------------------------------- 97.9/97.9 kB ? eta 0:00:00
Downloading exceptiongroup-1.1.3-py3-none-any.whl (14 kB)
Downloading importlib_metadata-6.8.0-py3-none-any.whl (22 kB)
Downloading keyring-24.2.0-py3-none-any.whl (37 kB)
Downloading Markdown-3.5.1-py3-none-any.whl (102 kB)
   ---------------------------------------- 102.2/102.2 kB ? eta 0:00:00
Downloading MarkupSafe-2.1.3-cp310-cp310-win_amd64.whl (17 kB)
Downloading mkdocs_autorefs-0.5.0-py3-none-any.whl (9.6 kB)
Downloading mkdocs_material_extensions-1.3-py3-none-any.whl (8.6 kB)
Downloading mkdocstrings_python-1.7.3-py3-none-any.whl (51 kB)
   ---------------------------------------- 51.7/51.7 kB ? eta 0:00:00
Downloading packaging-23.2-py3-none-any.whl (53 kB)
   ---------------------------------------- 53.0/53.0 kB ? eta 0:00:00
Downloading pathspec-0.11.2-py3-none-any.whl (29 kB)
Downloading platformdirs-4.0.0-py3-none-any.whl (17 kB)
Downloading pluggy-1.3.0-py3-none-any.whl (18 kB)
Downloading pydantic-2.4.2-py3-none-any.whl (395 kB)
   ---------------------------------------- 395.8/395.8 kB 12.4 MB/s eta 0:00:00
Downloading pydantic_core-2.10.1-cp310-none-win_amd64.whl (2.0 MB)
   ---------------------------------------- 2.0/2.0 MB 14.1 MB/s eta 0:00:00
Downloading Pygments-2.16.1-py3-none-any.whl (1.2 MB)
   ---------------------------------------- 1.2/1.2 MB 14.8 MB/s eta 0:00:00
Downloading pymdown_extensions-10.3.1-py3-none-any.whl (241 kB)
   ---------------------------------------- 241.0/241.0 kB 14.4 MB/s eta 0:00:00
Downloading PyYAML-6.0.1-cp310-cp310-win_amd64.whl (145 kB)
   ---------------------------------------- 145.3/145.3 kB ? eta 0:00:00
Downloading readme_renderer-42.0-py3-none-any.whl (13 kB)
Downloading regex-2023.10.3-cp310-cp310-win_amd64.whl (269 kB)
   ---------------------------------------- 269.6/269.6 kB 16.2 MB/s eta 0:00:00
Downloading requests-2.31.0-py3-none-any.whl (62 kB)
   ---------------------------------------- 62.6/62.6 kB ? eta 0:00:00
Downloading certifi-2023.7.22-py3-none-any.whl (158 kB)
   ---------------------------------------- 158.3/158.3 kB ? eta 0:00:00
Downloading rich-13.6.0-py3-none-any.whl (239 kB)
   ---------------------------------------- 239.8/239.8 kB ? eta 0:00:00
Downloading starlette-0.27.0-py3-none-any.whl (66 kB)
   ---------------------------------------- 67.0/67.0 kB ? eta 0:00:00
Downloading urllib3-2.0.7-py3-none-any.whl (124 kB)
   ---------------------------------------- 124.2/124.2 kB ? eta 0:00:00
Downloading httpcore-1.0.2-py3-none-any.whl (76 kB)
   ---------------------------------------- 76.9/76.9 kB ? eta 0:00:00
Downloading annotated_types-0.6.0-py3-none-any.whl (12 kB)
Downloading charset_normalizer-3.3.2-cp310-cp310-win_amd64.whl (100 kB)
   ---------------------------------------- 100.3/100.3 kB ? eta 0:00:00
Downloading docutils-0.20.1-py3-none-any.whl (572 kB)
   ---------------------------------------- 572.7/572.7 kB 12.0 MB/s eta 0:00:00
Downloading griffe-0.36.9-py3-none-any.whl (111 kB)
   ---------------------------------------- 111.7/111.7 kB ? eta 0:00:00
Downloading markdown_it_py-3.0.0-py3-none-any.whl (87 kB)
   ---------------------------------------- 87.5/87.5 kB ? eta 0:00:00
Downloading nh3-0.2.14-cp37-abi3-win_amd64.whl (574 kB)
   ---------------------------------------- 574.0/574.0 kB 17.6 MB/s eta 0:00:00
Downloading pywin32_ctypes-0.2.2-py3-none-any.whl (30 kB)
Downloading zipp-3.17.0-py3-none-any.whl (7.4 kB)
Downloading jaraco.classes-3.3.0-py3-none-any.whl (5.9 kB)
Downloading more_itertools-10.1.0-py3-none-any.whl (55 kB)
   ---------------------------------------- 55.8/55.8 kB ? eta 0:00:00
Running command Checking if build backend supports build_editable
Checking if build backend supports build_editable ... done
Building wheels for collected packages: llama_cpp_python, paginate
  Running command Building editable for llama_cpp_python (pyproject.toml)
  *** scikit-build-core 0.6.1 using CMake 3.27.7 (editable)
  *** Configuring CMake...
  2023-11-10 21:56:31,964 - scikit_build_core - WARNING - Can't find a Python library, got libdir=None, ldlibrary=None, multiarch=None, masd=None
  loading initial cache file C:\Users\Tkay\AppData\Local\Temp\tmpiubxovwy\build\CMakeInit.txt
  -- Building for: Visual Studio 17 2022
  -- The C compiler identification is MSVC 19.37.32825.0
  -- The CXX compiler identification is MSVC 19.37.32825.0
  -- Detecting C compiler ABI info
  -- Detecting C compiler ABI info - done
  -- Check for working C compiler: I:/Program Files/Microsoft Visual Studio/2022/Community/VC/Tools/MSVC/14.37.32822/bin/Hostx64/x64/cl.exe - skipped
  -- Detecting C compile features
  -- Detecting C compile features - done
  -- Detecting CXX compiler ABI info
  -- Detecting CXX compiler ABI info - done
  -- Check for working CXX compiler: I:/Program Files/Microsoft Visual Studio/2022/Community/VC/Tools/MSVC/14.37.32822/bin/Hostx64/x64/cl.exe - skipped
  -- Detecting CXX compile features
  -- Detecting CXX compile features - done
  -- Performing Test CMAKE_HAVE_LIBC_PTHREAD
  -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed
  -- Looking for pthread_create in pthreads
  -- Looking for pthread_create in pthreads - not found
  -- Looking for pthread_create in pthread
  -- Looking for pthread_create in pthread - not found
  -- Found Threads: TRUE
  -- Found CUDAToolkit: C:/Program Files/NVIDIA GPU Computing Toolkit/CUDA/v12.3/include (found version "12.3.52")
  -- cuBLAS found
  -- The CUDA compiler identification is NVIDIA 12.3.52
  -- Detecting CUDA compiler ABI info
  -- Detecting CUDA compiler ABI info - done
  -- Check for working CUDA compiler: C:/Program Files/NVIDIA GPU Computing Toolkit/CUDA/v12.3/bin/nvcc.exe - skipped
  -- Detecting CUDA compile features
  -- Detecting CUDA compile features - done
  -- Using CUDA architectures: 52;61;70
  -- CMAKE_SYSTEM_PROCESSOR: AMD64
  -- CMAKE_GENERATOR_PLATFORM: x64
  -- x86 detected
  -- Performing Test HAS_AVX_1
  -- Performing Test HAS_AVX_1 - Failed
  -- Performing Test HAS_AVX_2
  -- Performing Test HAS_AVX_2 - Failed
  -- Performing Test HAS_AVX2_1
  -- Performing Test HAS_AVX2_1 - Failed
  -- Performing Test HAS_AVX2_2
  -- Performing Test HAS_AVX2_2 - Failed
  -- Performing Test HAS_FMA_1
  -- Performing Test HAS_FMA_1 - Failed
  -- Performing Test HAS_FMA_2
  -- Performing Test HAS_FMA_2 - Failed
  -- Performing Test HAS_AVX512_1
  -- Performing Test HAS_AVX512_1 - Failed
  -- Performing Test HAS_AVX512_2
  -- Performing Test HAS_AVX512_2 - Failed
  CMake Warning (dev) at CMakeLists.txt:18 (install):
    Target llama has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION.
  This warning is for project developers.  Use -Wno-dev to suppress it.

  CMake Warning (dev) at CMakeLists.txt:27 (install):
    Target llama has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION.
  This warning is for project developers.  Use -Wno-dev to suppress it.

  -- Configuring done (27.4s)
  CMake Error in vendor/llama.cpp/examples/llava/CMakeLists.txt:
    CUDA_ARCHITECTURES is empty for target "llava_shared".

  CMake Error in vendor/llama.cpp/examples/llava/CMakeLists.txt:
    CUDA_ARCHITECTURES is empty for target "llava_shared".

  CMake Error in vendor/llama.cpp/examples/llava/CMakeLists.txt:
    CUDA_ARCHITECTURES is empty for target "llava_shared".

  CMake Error in vendor/llama.cpp/examples/llava/CMakeLists.txt:
    CUDA_ARCHITECTURES is empty for target "llava_shared".

  -- Generating done (0.2s)
  CMake Generate step failed.  Build files cannot be regenerated correctly.

  *** CMake configuration failed
  error: subprocess-exited-with-error

  × Building editable for llama_cpp_python (pyproject.toml) did not run successfully.
  │ exit code: 1
  ╰─> See above for output.

  note: This error originates from a subprocess, and is likely not a problem with pip.
  full command: 'I:\ai-chat_v1\llama-cpp-python\venv_mamba\env\python.exe' 'I:\ai-chat_v1\llama-cpp-python\venv_mamba\env\lib\site-packages\pip\_vendor\pyproject_hooks\_in_process\_in_process.py' build_editable 'C:\Users\Tkay\AppData\Local\Temp\tmp6hjivvue'
  cwd: I:\ai-chat_v1\llama-cpp-python\llama-cpp-python-main
  Building editable for llama_cpp_python (pyproject.toml) ... error
  ERROR: Failed building editable for llama_cpp_python
  Running command python setup.py bdist_wheel
  running bdist_wheel
  running build
  running build_py
  creating build
  creating build\lib
  creating build\lib\paginate
  copying paginate\ext_reverse.py -> build\lib\paginate
  copying paginate\__init__.py -> build\lib\paginate
  running egg_info
  writing paginate.egg-info\PKG-INFO
  writing dependency_links to paginate.egg-info\dependency_links.txt
  deleting paginate.egg-info\entry_points.txt
  writing top-level names to paginate.egg-info\top_level.txt
  reading manifest file 'paginate.egg-info\SOURCES.txt'
  reading manifest template 'MANIFEST.in'
  warning: no files found matching '*.ini'
  warning: no files found matching '*.rst'
  warning: no files found matching '*.png' under directory '*.ico'
  warning: no files found matching '*.css' under directory '*.ico'
  warning: no files found matching '*.gif' under directory '*.ico'
  warning: no files found matching '*.jpg' under directory '*.ico'
  warning: no files found matching '*.pt' under directory '*.ico'
  warning: no files found matching '*.txt' under directory '*.ico'
  warning: no files found matching '*.mak' under directory '*.ico'
  warning: no files found matching '*.mako' under directory '*.ico'
  warning: no files found matching '*.js' under directory '*.ico'
  warning: no files found matching '*.html' under directory '*.ico'
  warning: no files found matching '*.xml' under directory '*.ico'
  writing manifest file 'paginate.egg-info\SOURCES.txt'
  I:\ai-chat_v1\llama-cpp-python\venv_mamba\env\lib\site-packages\setuptools\_distutils\cmd.py:66: SetuptoolsDeprecationWarning: setup.py install is deprecated.
  !!

          ********************************************************************************
          Please avoid running ``setup.py`` directly.
          Instead, use pypa/build, pypa/installer or other
          standards-based tools.

          See https://blog.ganssle.io/articles/2021/10/setup-py-deprecated.html for details.
          ********************************************************************************

  !!
    self.initialize_options()
  installing to build\bdist.win-amd64\wheel
  running install
  running install_lib
  creating build\bdist.win-amd64
  creating build\bdist.win-amd64\wheel
  creating build\bdist.win-amd64\wheel\paginate
  copying build\lib\paginate\ext_reverse.py -> build\bdist.win-amd64\wheel\.\paginate
  copying build\lib\paginate\__init__.py -> build\bdist.win-amd64\wheel\.\paginate
  running install_egg_info
  Copying paginate.egg-info to build\bdist.win-amd64\wheel\.\paginate-0.5.6-py3.10.egg-info
  running install_scripts
  creating build\bdist.win-amd64\wheel\paginate-0.5.6.dist-info\WHEEL
  creating 'C:\Users\Tkay\AppData\Local\Temp\pip-wheel-tsoyy_0d\paginate-0.5.6-py3-none-any.whl' and adding 'build\bdist.win-amd64\wheel' to it
  adding 'paginate/__init__.py'
  adding 'paginate/ext_reverse.py'
  adding 'paginate-0.5.6.dist-info/METADATA'
  adding 'paginate-0.5.6.dist-info/WHEEL'
  adding 'paginate-0.5.6.dist-info/top_level.txt'
  adding 'paginate-0.5.6.dist-info/RECORD'
  removing build\bdist.win-amd64\wheel
  Building wheel for paginate (setup.py) ... done
  Created wheel for paginate: filename=paginate-0.5.6-py3-none-any.whl size=12708 sha256=45578bc694c75b5b8b8e36b3c9ff35b4dc7339a7c06cf44a986202b1b33fa49b
  Stored in directory: C:\Users\Tkay\AppData\Local\Temp\pip-ephem-wheel-cache-5r4shpvk\wheels\47\d3\18\0b5bebc873f29bea61fedece1e92cbcbef416839dfe5bd0eef
Successfully built paginate
Failed to build llama_cpp_python
ERROR: Could not build wheels for llama_cpp_python, which is required to install pyproject.toml-based projects
devosonder commented 10 months ago

The privateGPT project cannot be compiled because of the llama-cpp-python dependency. https://github.com/imartinez/privateGPT/issues/1197

abetlen commented 10 months ago

@tk-master just pushed v0.2.17 with the CUDA_ARCHITECTURES fix you suggested, hopefully that helps. In theory it shouldn't impact llava performance as the clip implementation in llava is cpu-only at the moment but we'll see down the road.

Anyways let me know if everything works correctly now.

tk-master commented 10 months ago

works for me

devosonder commented 10 months ago

It's working. Thanks