v3.5 Now it also supports for Twitch Streamer
v3.0 Now not only supports Japanese TTS using VoiceVox. But also supports TTS for RU (Russian), EN (English), DE (German), ES (Spanish), FR (French), TT (Tatar), UA (Ukrainian), UZ (Uzbek), XAL (Kalmyk), Indic (Hindi), using Seliro TTS. Change voicevox_tts
on run.py
to seliro_tts
, for detailed information of how to use Seliro TTS
This project is inspired by shioridotdev and utilizes various technologies such as VoiceVox Engine, DeepL, Whisper OpenAI, Seliro TTS and VtubeStudio to create an AI waifu virtual YouTuber.
pip install -r requirements.txt
api_key = 'yourapikey'
owner_name = "Ardha"
if you want to use it for livestream, create a list of users that you want to blacklist on run.py
blacklist = ["Nightbot", "streamelements"]
Change the lore or identity of your assistant. Change the txt file at characterConfig\Pina\identity.txt
If you want to stream on Twitch you need to change the config file at utils/twitch_config.py
. Get your token from Here. Your token should look something like oauth:43rip6j6fgio8n5xly1oum1lph8ikl1 (fake for this tutorial). After you change the config file, you can start the program using Mode - 3
server = 'irc.chat.twitch.tv'
port = 6667
nickname = 'testing' # You don't need to change this
token = 'oauth:43rip6j6fgio8n5xly1oum1lph8ikl1' # get it from https://twitchapps.com/tmi/.
user = 'ardha27' # Your Twitch username
channel = '#aikohound' # The channel you want to retrieve messages from
Choose which TTS you want to use, VoiceVox
or Silero
. Uncomment and Comment to switch between them
# Choose between the available TTS engines
# Japanese TTS
voicevox_tts(tts)
# Silero TTS, Silero TTS can generate English, Russian, French, Hindi, Spanish, German, etc. Uncomment the line below. Make sure the input is in that language
# silero_tts(tts_en, "en", "v3_en", "en_21")
If you want to use VoiceVox, you need to run VoiceVox Engine first. You can run them on local using VoiceVox Docker or on Google Colab using VoiceVox Colab. If you use the Colab one, change voicevox_url
on utils\TTS.py
using the link you get from Colab.
voicevox_url = 'http://localhost:50021'
if you want to see the voice list of VoiceVox you can check this VoiceVox and see the speaker id on speaker.json
then change it on utils/TTS.py
. For Seliro Voice sample you can check this Seliro Samples
VoiceVox
, because VoiceVox only accepts input in Japanese. The language answer from OpenAI will depens on your assistant lore language characterConfig\Pina\identity.txt
and the input languagetts = translate_deeplx(text, f"{detect}", "JA")
tts = translate_google(text, f"{detect}", "JA")
DeepLx
is free version of DeepL
(No API Key Required). You can run Deeplx on docker, or if you want to use the normal version of deepl, you can make the function on utils\translate.py
. I use DeepLx
because i can't register on DeepL
from my country. The translate result from DeepL
is more accurate and casual than Google Translate. But if you want the simple way, just use Google Translate.
If you want to use the audio output from the program as an input for your Vtubestudio
. You will need to capture your desktop audio using Virtual Cable
and use it as input on VtubeStudio microphone.
If you planning to use this program for live streaming Use chat.txt
and output.txt
as an input on OBS Text for Realtime Caption/Subtitles
def transcribe_audio(file):
global chat_now
try:
audio_file= open(file, "rb")
# Translating the audio to English
# transcript = openai.Audio.translate("whisper-1", audio_file)
# Transcribe the audio to detected language
transcript = openai.Audio.transcribe("whisper-1", audio_file)
chat_now = transcript.text
print ("Question: " + chat_now)
except:
print("Error transcribing audio")
return
result = owner_name + " said " + chat_now
conversation.append({'role': 'user', 'content': result})
openai_answer()
Change this Line of Code to this. This will help you to get more information about the error
def transcribe_audio(file):
global chat_now
audio_file= open(file, "rb")
# Translating the audio to English
# transcript = openai.Audio.translate("whisper-1", audio_file)
# Transcribe the audio to detected language
transcript = openai.Audio.transcribe("whisper-1", audio_file)
chat_now = transcript.text
print ("Question: " + chat_now)
result = owner_name + " said " + chat_now
conversation.append({'role': 'user', 'content': result})
openai_answer()
Another option to solve this problem, you can upgrade the OpenAI library to the latest version. Make sure the program capture your voice/sentence, try to hear the input.wav
this library is a little bit tricky to install. If you facing this problem, you can just delete and don't use the katakana_converter
on utils/TTS.py
. That function is optional, you can run the program without it. Delete this two line on utils/TTS.py
from utils.katakana import *
katakana_text = katakana_converter(tts)
and just pass the tts
to next line of the code
params_encoded = urllib.parse.urlencode({'text': tts, 'speaker': 46})
This project is inspired by the work of shioridotdev. Special thanks to the creators of the technologies used in this project including VoiceVox Engine, DeepL, Whisper OpenAI, and VtubeStudio.