Open MartialTerran opened 2 months ago
lol. yea absolutely. there is a small fix you need to compile it to run models over 4gig, like llama 7b. Just compile it with ming64 and run it first.
Hi. According to Google Gemini Pro, compiling a Llama2.c model on a Windows 10 machine is "difficult" also because of high RAM memory requirements. And "Dependency Hell" due to "libraries optimized for Linux". I apprehend that the windows-compile task is beyond my experience and practical ability. Can Windows run the Linux version of Llama2.c within a Windows builtin Linux environment "MingW64 ~" prompt? Is that the same method as what you originally proposed? Can you point to simple instructions printed online?
On Sat, May 11, 2024, 12:03 AM James Delancey @.***> wrote:
lol. yea absolutely. there is a small fix you need to compile it to run models over 4gig, like llama 7b. Just compile it with ming64 and run it first.
— Reply to this email directly, view it on GitHub https://github.com/karpathy/llama2.c/issues/516#issuecomment-2105527029, or unsubscribe https://github.com/notifications/unsubscribe-auth/BHT6KT2R3CQSZGJRQKGVUEDZBWKAPAVCNFSM6AAAAABHJXVADWVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDCMBVGUZDOMBSHE . You are receiving this because you authored the thread.Message ID: @.***>
Assuming that a Windows 10 machine can run this code, can this run.c lllama2 be modified to be complied to run inference mode on a Windows 10 machine? I asked Google Gemini 1.5 Pro this question and this is Gemini's response:
Compiling and Running the Code on Windows 10 While the provided code appears to be primarily written in C, it includes some POSIX-specific elements like mmap and munmap which are not directly available on Windows 10. Therefore, running this code directly on Windows 10 would require some modifications and additional libraries. Here's a potential approach: