Open dzianisv opened 1 month ago
Yeah some books do that...
And I'm not sure why....
I think there's something wrong with the logic in my code for that or something... cause it's super weird some books do that and some don't
Even when the file given is an EPUB which is the one that has best compatibility
I'll see when I can get to hunting that down
In this case word
for some reason is not array, but float, at nltk/tag/perceptron.py
I also have the same issue with my book (EPUB exported using Amazon Kindle tool) I was trying to convert. Here are the few logs before the crash.
Cannot resolve quotation
--- attribution: 2.777 seconds ---
--- name coref: 0.078 seconds ---
--- coref: 43.001 seconds ---
--- TOTAL (excl. startup): 191.703 seconds ---, 244422 words
deleted file: /home/user1/code/voxnovel/AudioBook.txt because its not needed anymore after the ebook convertsion to txt
Success, File processed successfully!
Processed 323 lines.
Removed 0 problematic lines.
Wrote 323 lines back to the file.
Saved nonquotes.csv to Working_files/Book/non_quotes.csv
All processing complete!
Traceback (most recent call last):
File "/home/user1/VoxNovel/gui_run.py", line 696, in <module>
main()
File "/home/user1/VoxNovel/gui_run.py", line 691, in main
process_files(q_file, matching_entities_files[0])
File "/home/user1/VoxNovel/gui_run.py", line 619, in process_files
if is_pronoun(mention):
File "/home/user1/VoxNovel/gui_run.py", line 596, in is_pronoun
tagged_word = nltk.pos_tag([word])
File "/home/user1/miniconda3/lib/python3.10/site-packages/nltk/tag/__init__.py", line 169, in pos_tag
return _pos_tag(tokens, tagset, tagger, lang)
File "/home/user1/miniconda3/lib/python3.10/site-packages/nltk/tag/__init__.py", line 126, in _pos_tag
tagged_tokens = tagger.tag(tokens)
File "/home/user1/miniconda3/lib/python3.10/site-packages/nltk/tag/perceptron.py", line 194, in tag
context = self.START + [self.normalize(w) for w in tokens] + self.END
File "/home/user1/miniconda3/lib/python3.10/site-packages/nltk/tag/perceptron.py", line 194, in <listcomp>
context = self.START + [self.normalize(w) for w in tokens] + self.END
File "/home/user1/miniconda3/lib/python3.10/site-packages/nltk/tag/perceptron.py", line 303, in normalize
if "-" in word and word[0] != "-":
Note that I had to do manually:
pip install pandas
Because it was complaining when running the python gui_run.py
command.
At the end I started to work on my own implementation https://github.com/dzianisv/VocalPages For me it was quicker to build a simple tool to vocalize a few epubs, rather than digging in a lot of 3rd party librarties.
Oh dang!
Nice!
So your not having issues with bark tts and "adding extra words" type hallucinations?
Also yeah sorry been busy lol
Plus my other ebook2audiobookxtts repo has really blew up so that's been taking up my tiny amount of avalible time
So your not having issues with bark tts and "adding extra words" type hallucinations?
Sometimes I see them. But quality is great. + A lot of uhhh. But it is acceptable, I am working on adding https://github.com/coqui-ai/TTS as well.
Nice nice
I tried to vocalize the book https://library.lol/main/7AB96C0626A64D7AFA33F68DC6D97CE3.