-
Hi, thank you for providing 1.58bit implementation. Nice work! I looked through many bitnet1.58 implementations and noticed that they all use the method suggested in "The Era from 1-bit LLMs: Training…
-
Hello Professor, thank you for your contribution. As a new student, I am very interested in your research. When I read the paper and wanted to find more details from the running of the code, I encount…
-
scrypt cpu is much more useful :)
-
- Start LLM
- Close laptop
- Sleep 8 hours
- Open laptop
- Issue command to LLM
-
### Feature request
Is there any chance we coukd get this 4bit adam optimizer added to tranformers?
It has nearly the same performance as 32bit adam with significant drop in vram overhead.
[repo…
-
## AAAI-24
Benchmarking Large Language Models in Retrieval-Augmented Generation
https://arxiv.org/abs/2309.01431
Hot or Cold? Adaptive Temperature Sampling for Code Generation with Large Langua…
-
### Motivation and description
See here:
https://github.com/ggerganov/ggml/pull/254/files
I think we may need QuickGELU, for compatibility, if not same as GELU, more than just optimization.
…
-
Subject: Proposal for Implementing a 1-bit Optimizer
Dear [sd-scripts] Team,
I hope this message finds you well. I am reaching out to propose the addition of a 1-bit optimizer to your project, s…
-
I'm currently trying to replicate the extraction from an EMNIST network to MLCert. This is as far as I've gotten trying to get the code in NNCert to work, patching things up as I go:
1. Install [Mi…
-
I tried running the following code, with just having the ```ridger/MMfreeLM-1.3B```` model initialized:
```
root@r4-0:~/matmulfreellm# python
>>> import os
>>> os.environ["TOKENIZERS_PARALLELISM…