issues
search
mit-han-lab
/
TinyChatEngine
TinyChatEngine: On-Device LLM Inference Library
https://mit-han-lab.github.io/TinyChatEngine/
MIT License
756
stars
73
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Assistant spitting out non-readable characters on RTX 4060
#71
zhefciad
opened
1 year ago
5
Support CodeLLaMA
#70
RaymondWang0
closed
1 year ago
0
Create httpchat.cc
#69
omjee
opened
1 year ago
2
Revised CUDA support
#68
RaymondWang0
closed
1 year ago
0
Support new features
#67
RaymondWang0
closed
1 year ago
0
Unable to maintain chat history and continuous chat
#66
Rkyzzy
opened
1 year ago
1
Cleaned up output and support more models for voicechat
#65
Jiminator
closed
1 year ago
0
LLaMA2_7B_chat_awq_int4.zip Empty File
#64
tuobulatuo
closed
1 year ago
2
GPU A100 Output Random Code
#63
tuobulatuo
opened
1 year ago
2
Increase CPU usage cap from 400% to higher number, better if configurable
#62
tuobulatuo
opened
1 year ago
1
Made dictation concurrent with token generation
#61
Jiminator
closed
1 year ago
0
fix RPi support and update readme
#60
RaymondWang0
closed
1 year ago
0
minor fix
#59
RaymondWang0
closed
1 year ago
0
Support to Tesla P100 GPU inference
#58
songkq
opened
1 year ago
5
Update download_model.py
#57
meenchen
closed
1 year ago
0
Update initialization and README
#56
RaymondWang0
closed
1 year ago
0
No module named 'requests'
#55
andyluo7
closed
1 year ago
1
Support windows as experimental feature
#54
meenchen
closed
1 year ago
0
add more demos to readme
#53
RaymondWang0
closed
1 year ago
0
Cannot compile on Raspberry Pi
#52
nmeln
closed
1 year ago
4
Speed comparasion with exllama and llama.cpp?
#51
lucasjinreal
opened
1 year ago
1
Update README.md
#50
RaymondWang0
closed
1 year ago
0
customized define
#49
meenchen
closed
1 year ago
0
minor fix
#48
meenchen
closed
1 year ago
0
fix path
#47
meenchen
closed
1 year ago
0
Fix path of docs and restore patch
#46
meenchen
closed
1 year ago
0
logo
#45
meenchen
closed
1 year ago
0
update readme demo
#44
RaymondWang0
closed
1 year ago
0
Minor fix to ARM support and README
#43
meenchen
closed
1 year ago
0
update README
#42
meenchen
closed
1 year ago
0
Cleanup and fix bugs for CUDA
#41
RaymondWang0
closed
1 year ago
0
cleanup, model naming, docs
#40
meenchen
closed
1 year ago
0
Dev/fix gpu demo UI
#39
meenchen
closed
1 year ago
0
Fix cuda bug
#38
meenchen
closed
1 year ago
0
CUDA bug fix
#37
RaymondWang0
closed
1 year ago
0
minor ui fix
#36
meenchen
closed
1 year ago
0
fix gpu and update demo ui with shortcut
#35
meenchen
closed
1 year ago
0
docs generation
#34
meenchen
closed
1 year ago
0
Support opt fp32/int4
#33
meenchen
closed
1 year ago
0
support llama13B
#32
meenchen
closed
1 year ago
0
Added edge GPU voicechat instructions
#31
Jiminator
closed
1 year ago
0
minor fix for Makefile
#30
RaymondWang0
closed
1 year ago
0
Update README.md
#29
RaymondWang0
closed
1 year ago
0
Support w4a16 LLaMA on CUDA GPUs
#28
RaymondWang0
closed
1 year ago
0
Speech to text demo
#27
Jiminator
closed
1 year ago
0
Speech to text demo
#26
Jiminator
closed
1 year ago
0
Update README.md
#25
RaymondWang0
closed
1 year ago
0
Update README.md
#24
meenchen
closed
1 year ago
0
Upload quantized models
#23
meenchen
closed
1 year ago
0
Dev/minor UI update
#22
meenchen
closed
1 year ago
0
Previous
Next