Closed redskate closed 1 year ago
And here my comment (since I observe, no reaction has come so far). Thank you again for all the llama models! I managed to launch a docker version from a good connected server and could download + quantize all the (heavier) llama models.
Of course a llama 65B model takes too long to load (5-10min on an average machine at each use) and an alpaca 65B model would be faster and very appreciated but is not available.
I close therefore this issue, hoping at some point to read something on the alpaca 65B.
Good work.
Dear community!
Thanks a lot for his very interesting system!!! However: The only interesting model here is the 65B and this happens to be huge and the only one mentioned and shown in the article https://arxiv.org/abs/2302.13971 (practically the only one of interest).
I managed to download 7B, 13B and 30B and they work "nice" but not really useful in answering (see your prompts) - (thank you).
Installing the 65B model resulted for me since 4 days in a "try" to download the consolidated.xx.pth files (each day a new try). This npx install blocks after having downloaded 3 "consolidated" files ... and I have no idea what might have gone wrong here and why (and I stop the download, but then all the downloaded files are erased to start again from scratch). This should not be (or yes?)
Is there a URL on a repository where one can directly manually download what is shown in https://github.com/cocktailpeanut/dalai/blob/main/docs/65b.png ???
Or else what do you suggest please?
Thanks a lot! Kind regards