issues
search
jzhang38
/
TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
Apache License 2.0
7.3k
stars
425
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
No module named 'xentropy_cuda_lib'
#142
ajdotexecute
closed
4 months ago
1
Does it support spanish language?
#141
wilfoderek
opened
5 months ago
7
How to evaluate checkpoints during pretraining?
#140
DarthMurse
closed
3 months ago
2
How do I use these data sets to train new models?
#139
win10ogod
closed
4 months ago
2
how to determine reasonable max steps?
#138
ScottishFold007
opened
5 months ago
1
Chat v1.0 training recipe
#137
normster
closed
4 months ago
1
Update README.md
#136
ChaosCodes
closed
5 months ago
0
Update README.md
#135
hunter-lee1
closed
5 months ago
0
Update README.md
#134
hunter-lee1
closed
5 months ago
0
Unexpected behavior - Incomplete responses and nonsense outputs
#133
AmadejTratnik
closed
4 months ago
1
Saturation / epoch-accuracy plot
#132
rasbt
closed
5 months ago
6
update gradio
#131
hunter-lee1
closed
5 months ago
0
update chat_gradio
#130
hunter-lee1
closed
5 months ago
0
Usage documentation
#129
simensandhaug
closed
5 months ago
1
add_gradio
#128
hunter-lee1
closed
6 months ago
0
All 'rotary_emb.inv_freq' are not able to load with Automodel or Llamamodel.
#127
becxer
closed
3 months ago
2
Loss logs
#126
borgr
closed
6 months ago
2
Update README.md
#125
ChaosCodes
closed
6 months ago
0
Add support for sliding window context similar to Mistral?
#124
Harleen8118
opened
6 months ago
4
Results vs Chinchilla
#123
SpaceCowboy850
closed
6 months ago
5
TinyMix (Not Issue)
#122
nivibilla
closed
6 months ago
1
是否可以公布多一些checkpoint?
#121
yangsp5
closed
6 months ago
1
Full fine-tuning with llama.cpp
#120
RonanKMcGovern
closed
6 months ago
1
No dropout?
#119
sacharbit
closed
6 months ago
1
Modify readme.md and eval.md
#118
TianduoWang
closed
6 months ago
0
Why change the `_init_weights` ?
#117
larrylawl
closed
6 months ago
5
Should we add `<eos>` token during preprocessing?
#116
larrylawl
closed
6 months ago
1
3T checkpoint?
#115
coder543
closed
6 months ago
5
The 4bit-quantized TinyLlama-1.1B's weight only takes up 550MB RAM ?
#114
TapendraBaduwal
closed
6 months ago
6
How can we enable continuous learning with the Tiny Llama model ?
#113
TapendraBaduwal
opened
6 months ago
0
Update README.md
#112
ChaosCodes
closed
6 months ago
0
the dataset selection sft on OpenAssisant
#111
littleSunlxy
closed
6 months ago
1
Update EVAL.md
#110
ChaosCodes
closed
6 months ago
0
Update README.md
#109
ChaosCodes
closed
6 months ago
0
Estimation of 2.5T Tokens Checkpoint
#108
mounta11n
closed
6 months ago
1
TinyLlama-1.1B-Chat-v0.6 Tokenization
#107
phaylon
closed
6 months ago
1
如果我按中文语料CLM+SFT对TinyLlama进行重训,是否可行?
#106
skykiseki
closed
6 months ago
5
The lighting app was updated, and does not support run model !
#105
imanu20
closed
6 months ago
2
chore: correct bibtex syntax
#104
joennlae
closed
7 months ago
0
Settings for training/fine-tune of TinyLlama/TinyLlama-1.1B-Chat-v0.6
#103
sebastianschramm
closed
7 months ago
1
Is it supported to convert to Apple CoreML? I tried to use coremltools to convert this model to CoreML format, but encountered an error.
#102
HoodL
opened
7 months ago
1
Assessing performance of TinyLlama
#101
galleon
closed
7 months ago
3
ncclRemoteError
#100
JerryDaHeLian
closed
6 months ago
2
Can anyone pre train tinyllama. py on v100s?
#99
JerryDaHeLian
closed
6 months ago
1
Update EVAL.md
#98
jzhang38
closed
7 months ago
0
Consider providing safetensor files
#97
Calandiel
closed
7 months ago
2
Update EVAL.md
#96
TianduoWang
closed
7 months ago
0
请问支持中文对话么?
#95
xman1991
closed
7 months ago
1
Status of chat model
#94
galleon
closed
7 months ago
2
data mixture
#93
NonvolatileMemory
closed
7 months ago
1
Previous
Next