An open source chat bot architecture for voice/vision (and multimodal) assistants, local and remote to run; if u run achatbot by yourself, u can learn more, star and fork to contribute~
BSD 3-Clause "New" or "Revised" License
15
stars
2
forks
source link
feat: add local llm -> transformers core llm to support local llm inference, next step for local vision llm #59
fix:
- log support LOG_LEVEL env to set log level
- check llm stream text have special char for sentences (simple process with re, maybe use [NLTK](https://www.nltk.org/) )
feat:
TQDM_DISABLE=True TOKENIZERS_PARALLELISM=false \ AUDIO_IN_STREAM_TAG=pyaudio_in_stream \ AUDIO_OUT_STREAM_TAG=pyaudio_out_stream \ VAD_DETECTOR_TAG=webrtc_silero_vad \ RECORDER_TAG=vad_recorder \ ASR_TAG=sense_voice_asr \ ASR_LANG=zn \ ASR_MODEL_NAME_OR_PATH=./models/FunAudioLLM/SenseVoiceSmall \ LLM_TAG=llm_transformers_pipeline \ LLM_MODEL_NAME_OR_PATH=./models/Qwen/Qwen2-0.5B-Instruct \ TTS_TAG=tts_edge \ python -m src.cmd.local-terminal-chat.generate_audio2audio > ./log/std_out.log