dmort27 / epitran

A tool for transcribing orthographic text as IPA (International Phonetic Alphabet)
MIT License
630 stars 121 forks source link

Speeding up mapping with HuggingFace datasets #82

Open MaveriQ opened 3 years ago

MaveriQ commented 3 years ago

Hi. I am trying to convert corpora from HF to their IPA form with the following snippet. But I am getting really slow speeds.. only a couple of examples per second. Do you know how it can be sped up? Thanks

import epitran
from datasets import load_dataset

bookscorpus = load_dataset('bookcorpus',split='train')
epi = epitran.Epitran('eng-Latn')

def transliterate(x):
    return {'trans': epi.transliterate(x['text'])}

tokenized = bookscorpus.map(lambda x: transliterate(x),num_proc=32)
dmort27 commented 2 years ago

Hi @MaveriQ , I seem to have missed this message when you originally sent it. It is true that the Python implementation of Epitran is very slow. If you can take advantage of concurrency, you can do better. I am in the process of rewriting parts of Epitran in Rust, which will increase performance—based on initial tests—by about 100 times.

dmort27 commented 2 years ago

However, for English this will not help as much since English support is provided by Flite (written in C) and the only python part converts the ARPAbet representation from Flite to IPA.

MaveriQ commented 2 years ago

Hi. If by concurrency you meant multiprocessing, I already tried that, but it's still pretty slow. Can you recommend anything else for English? Thanks

dmort27 commented 2 years ago

As I look at this issue, the real problem is that Epitran spawns a shell every time it calls lex_lookup to convert a word to IPA. This is expensive (although lex_lookup itself is quite efficient). The solutions would be to:

The second solution would be easier, but ultimately less satisfying.

juice500ml commented 6 months ago

FYI for future reference, extremely dirty quickfix would be:

import re

words = [...]  # words to be transliterated
with open("eng_words.sh", "w") as f:
    f.write("\n".join([f"lex_lookup {w} | head -n 1" for w in words]))

!bash eng_words.sh > eng_lex.txt

lexs = open("eng_lex.txt").readlines()
arpa_to_ipa = {'ey': 'ej', 'ae': 'æ', 'iy': 'i', 'eh': 'ɛ', 'ay': 'aj', 'ih': 'ɪ', 'ow': 'ow', 'aa': 'ɑ', 'ao': 'ɔ', 'aw': 'aw', 'oy': 'oj', 'ah': 'ʌ', 'ax': 'ə', 'uw': 'u', 'uh': 'ʊ', 'er': 'ɹ̩', 'b': 'b', 'ch': 't͡ʃ', 'd': 'd', 'dx': 'ɾ', 'f': 'f', 'g': 'ɡ', 'hh': 'h', 'jh': 'd͡ʒ', 'k': 'k', 'l': 'l', 'em': 'm̩', 'm': 'm', 'en': 'n̩', 'n': 'n', 'ng': 'ŋ', 'p': 'p', 'q': 'ʔ', 'r': 'ɹ', 's': 's', 'sh': 'ʃ', 't': 't', 'dh': 'ð', 'th': 'θ', 'v': 'v', 'w': 'w', 'y': 'j', 'z': 'z', 'zh': 'ʒ'}

ipas = []
for lex in lexs:
    lex = lex.strip()[1:-1].split()
    lex = map(lambda d: re.sub(r'\d', '', d), lex)
    ipa = map(lambda d: arpa_to_ipa[d], lex)
    ipas.append("".join(list(ipa)))

word_to_ipa = dict(zip(words, ipas))  # now this dict has key: word, value: transliteration result