|
import pickle
|
|
import os
|
|
import re
|
|
from g2p_en import G2p
|
|
from transformers import DebertaV2Tokenizer
|
|
|
|
from text import symbols
|
|
from text.symbols import punctuation
|
|
|
|
current_file_path = os.path.dirname(__file__)
|
|
CMU_DICT_PATH = os.path.join(current_file_path, "cmudict.rep")
|
|
CACHE_PATH = os.path.join(current_file_path, "cmudict_cache.pickle")
|
|
_g2p = G2p()
|
|
LOCAL_PATH = "./bert/deberta-v3-large"
|
|
tokenizer = DebertaV2Tokenizer.from_pretrained(LOCAL_PATH)
|
|
|
|
arpa = {
|
|
"AH0",
|
|
"S",
|
|
"AH1",
|
|
"EY2",
|
|
"AE2",
|
|
"EH0",
|
|
"OW2",
|
|
"UH0",
|
|
"NG",
|
|
"B",
|
|
"G",
|
|
"AY0",
|
|
"M",
|
|
"AA0",
|
|
"F",
|
|
"AO0",
|
|
"ER2",
|
|
"UH1",
|
|
"IY1",
|
|
"AH2",
|
|
"DH",
|
|
"IY0",
|
|
"EY1",
|
|
"IH0",
|
|
"K",
|
|
"N",
|
|
"W",
|
|
"IY2",
|
|
"T",
|
|
"AA1",
|
|
"ER1",
|
|
"EH2",
|
|
"OY0",
|
|
"UH2",
|
|
"UW1",
|
|
"Z",
|
|
"AW2",
|
|
"AW1",
|
|
"V",
|
|
"UW2",
|
|
"AA2",
|
|
"ER",
|
|
"AW0",
|
|
"UW0",
|
|
"R",
|
|
"OW1",
|
|
"EH1",
|
|
"ZH",
|
|
"AE0",
|
|
"IH2",
|
|
"IH",
|
|
"Y",
|
|
"JH",
|
|
"P",
|
|
"AY1",
|
|
"EY0",
|
|
"OY2",
|
|
"TH",
|
|
"HH",
|
|
"D",
|
|
"ER0",
|
|
"CH",
|
|
"AO1",
|
|
"AE1",
|
|
"AO2",
|
|
"OY1",
|
|
"AY2",
|
|
"IH1",
|
|
"OW0",
|
|
"L",
|
|
"SH",
|
|
}
|
|
|
|
|
|
def post_replace_ph(ph):
|
|
rep_map = {
|
|
":": ",",
|
|
";": ",",
|
|
",": ",",
|
|
"。": ".",
|
|
"!": "!",
|
|
"?": "?",
|
|
"\n": ".",
|
|
"·": ",",
|
|
"、": ",",
|
|
"…": "...",
|
|
"···": "...",
|
|
"・・・": "...",
|
|
"v": "V",
|
|
}
|
|
if ph in rep_map.keys():
|
|
ph = rep_map[ph]
|
|
if ph in symbols:
|
|
return ph
|
|
if ph not in symbols:
|
|
ph = "UNK"
|
|
return ph
|
|
|
|
|
|
rep_map = {
|
|
":": ",",
|
|
";": ",",
|
|
",": ",",
|
|
"。": ".",
|
|
"!": "!",
|
|
"?": "?",
|
|
"\n": ".",
|
|
".": ".",
|
|
"…": "...",
|
|
"···": "...",
|
|
"・・・": "...",
|
|
"·": ",",
|
|
"・": ",",
|
|
"、": ",",
|
|
"$": ".",
|
|
"“": "'",
|
|
"”": "'",
|
|
'"': "'",
|
|
"‘": "'",
|
|
"’": "'",
|
|
"(": "'",
|
|
")": "'",
|
|
"(": "'",
|
|
")": "'",
|
|
"《": "'",
|
|
"》": "'",
|
|
"【": "'",
|
|
"】": "'",
|
|
"[": "'",
|
|
"]": "'",
|
|
"—": "-",
|
|
"−": "-",
|
|
"~": "-",
|
|
"~": "-",
|
|
"「": "'",
|
|
"」": "'",
|
|
}
|
|
|
|
|
|
def replace_punctuation(text):
|
|
pattern = re.compile("|".join(re.escape(p) for p in rep_map.keys()))
|
|
|
|
replaced_text = pattern.sub(lambda x: rep_map[x.group()], text)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
return replaced_text
|
|
|
|
|
|
def read_dict():
|
|
g2p_dict = {}
|
|
start_line = 49
|
|
with open(CMU_DICT_PATH) as f:
|
|
line = f.readline()
|
|
line_index = 1
|
|
while line:
|
|
if line_index >= start_line:
|
|
line = line.strip()
|
|
word_split = line.split(" ")
|
|
word = word_split[0]
|
|
|
|
syllable_split = word_split[1].split(" - ")
|
|
g2p_dict[word] = []
|
|
for syllable in syllable_split:
|
|
phone_split = syllable.split(" ")
|
|
g2p_dict[word].append(phone_split)
|
|
|
|
line_index = line_index + 1
|
|
line = f.readline()
|
|
|
|
return g2p_dict
|
|
|
|
|
|
def cache_dict(g2p_dict, file_path):
|
|
with open(file_path, "wb") as pickle_file:
|
|
pickle.dump(g2p_dict, pickle_file)
|
|
|
|
|
|
def get_dict():
|
|
if os.path.exists(CACHE_PATH):
|
|
with open(CACHE_PATH, "rb") as pickle_file:
|
|
g2p_dict = pickle.load(pickle_file)
|
|
else:
|
|
g2p_dict = read_dict()
|
|
cache_dict(g2p_dict, CACHE_PATH)
|
|
|
|
return g2p_dict
|
|
|
|
|
|
eng_dict = get_dict()
|
|
|
|
|
|
def refine_ph(phn):
|
|
tone = 0
|
|
if re.search(r"\d$", phn):
|
|
tone = int(phn[-1]) + 1
|
|
phn = phn[:-1]
|
|
else:
|
|
tone = 3
|
|
return phn.lower(), tone
|
|
|
|
|
|
def refine_syllables(syllables):
|
|
tones = []
|
|
phonemes = []
|
|
for phn_list in syllables:
|
|
for i in range(len(phn_list)):
|
|
phn = phn_list[i]
|
|
phn, tone = refine_ph(phn)
|
|
phonemes.append(phn)
|
|
tones.append(tone)
|
|
return phonemes, tones
|
|
|
|
|
|
import re
|
|
import inflect
|
|
|
|
_inflect = inflect.engine()
|
|
_comma_number_re = re.compile(r"([0-9][0-9\,]+[0-9])")
|
|
_decimal_number_re = re.compile(r"([0-9]+\.[0-9]+)")
|
|
_pounds_re = re.compile(r"£([0-9\,]*[0-9]+)")
|
|
_dollars_re = re.compile(r"\$([0-9\.\,]*[0-9]+)")
|
|
_ordinal_re = re.compile(r"[0-9]+(st|nd|rd|th)")
|
|
_number_re = re.compile(r"[0-9]+")
|
|
|
|
|
|
_abbreviations = [
|
|
(re.compile("\\b%s\\." % x[0], re.IGNORECASE), x[1])
|
|
for x in [
|
|
("mrs", "misess"),
|
|
("mr", "mister"),
|
|
("dr", "doctor"),
|
|
("st", "saint"),
|
|
("co", "company"),
|
|
("jr", "junior"),
|
|
("maj", "major"),
|
|
("gen", "general"),
|
|
("drs", "doctors"),
|
|
("rev", "reverend"),
|
|
("lt", "lieutenant"),
|
|
("hon", "honorable"),
|
|
("sgt", "sergeant"),
|
|
("capt", "captain"),
|
|
("esq", "esquire"),
|
|
("ltd", "limited"),
|
|
("col", "colonel"),
|
|
("ft", "fort"),
|
|
]
|
|
]
|
|
|
|
|
|
|
|
_lazy_ipa = [
|
|
(re.compile("%s" % x[0]), x[1])
|
|
for x in [
|
|
("r", "ɹ"),
|
|
("æ", "e"),
|
|
("ɑ", "a"),
|
|
("ɔ", "o"),
|
|
("ð", "z"),
|
|
("θ", "s"),
|
|
("ɛ", "e"),
|
|
("ɪ", "i"),
|
|
("ʊ", "u"),
|
|
("ʒ", "ʥ"),
|
|
("ʤ", "ʥ"),
|
|
("ˈ", "↓"),
|
|
]
|
|
]
|
|
|
|
|
|
_lazy_ipa2 = [
|
|
(re.compile("%s" % x[0]), x[1])
|
|
for x in [
|
|
("r", "ɹ"),
|
|
("ð", "z"),
|
|
("θ", "s"),
|
|
("ʒ", "ʑ"),
|
|
("ʤ", "dʑ"),
|
|
("ˈ", "↓"),
|
|
]
|
|
]
|
|
|
|
|
|
_ipa_to_ipa2 = [
|
|
(re.compile("%s" % x[0]), x[1]) for x in [("r", "ɹ"), ("ʤ", "dʒ"), ("ʧ", "tʃ")]
|
|
]
|
|
|
|
|
|
def _expand_dollars(m):
|
|
match = m.group(1)
|
|
parts = match.split(".")
|
|
if len(parts) > 2:
|
|
return match + " dollars"
|
|
dollars = int(parts[0]) if parts[0] else 0
|
|
cents = int(parts[1]) if len(parts) > 1 and parts[1] else 0
|
|
if dollars and cents:
|
|
dollar_unit = "dollar" if dollars == 1 else "dollars"
|
|
cent_unit = "cent" if cents == 1 else "cents"
|
|
return "%s %s, %s %s" % (dollars, dollar_unit, cents, cent_unit)
|
|
elif dollars:
|
|
dollar_unit = "dollar" if dollars == 1 else "dollars"
|
|
return "%s %s" % (dollars, dollar_unit)
|
|
elif cents:
|
|
cent_unit = "cent" if cents == 1 else "cents"
|
|
return "%s %s" % (cents, cent_unit)
|
|
else:
|
|
return "zero dollars"
|
|
|
|
|
|
def _remove_commas(m):
|
|
return m.group(1).replace(",", "")
|
|
|
|
|
|
def _expand_ordinal(m):
|
|
return _inflect.number_to_words(m.group(0))
|
|
|
|
|
|
def _expand_number(m):
|
|
num = int(m.group(0))
|
|
if num > 1000 and num < 3000:
|
|
if num == 2000:
|
|
return "two thousand"
|
|
elif num > 2000 and num < 2010:
|
|
return "two thousand " + _inflect.number_to_words(num % 100)
|
|
elif num % 100 == 0:
|
|
return _inflect.number_to_words(num // 100) + " hundred"
|
|
else:
|
|
return _inflect.number_to_words(
|
|
num, andword="", zero="oh", group=2
|
|
).replace(", ", " ")
|
|
else:
|
|
return _inflect.number_to_words(num, andword="")
|
|
|
|
|
|
def _expand_decimal_point(m):
|
|
return m.group(1).replace(".", " point ")
|
|
|
|
|
|
def normalize_numbers(text):
|
|
text = re.sub(_comma_number_re, _remove_commas, text)
|
|
text = re.sub(_pounds_re, r"\1 pounds", text)
|
|
text = re.sub(_dollars_re, _expand_dollars, text)
|
|
text = re.sub(_decimal_number_re, _expand_decimal_point, text)
|
|
text = re.sub(_ordinal_re, _expand_ordinal, text)
|
|
text = re.sub(_number_re, _expand_number, text)
|
|
return text
|
|
|
|
|
|
def text_normalize(text):
|
|
text = normalize_numbers(text)
|
|
text = replace_punctuation(text)
|
|
text = re.sub(r"([,;.\?\!])([\w])", r"\1 \2", text)
|
|
return text
|
|
|
|
|
|
def distribute_phone(n_phone, n_word):
|
|
phones_per_word = [0] * n_word
|
|
for task in range(n_phone):
|
|
min_tasks = min(phones_per_word)
|
|
min_index = phones_per_word.index(min_tasks)
|
|
phones_per_word[min_index] += 1
|
|
return phones_per_word
|
|
|
|
|
|
def sep_text(text):
|
|
words = re.split(r"([,;.\?\!\s+])", text)
|
|
words = [word for word in words if word.strip() != ""]
|
|
return words
|
|
|
|
|
|
def text_to_words(text):
|
|
tokens = tokenizer.tokenize(text)
|
|
words = []
|
|
for idx, t in enumerate(tokens):
|
|
if t.startswith("▁"):
|
|
words.append([t[1:]])
|
|
else:
|
|
if t in punctuation:
|
|
if idx == len(tokens) - 1:
|
|
words.append([f"{t}"])
|
|
else:
|
|
if (
|
|
not tokens[idx + 1].startswith("▁")
|
|
and tokens[idx + 1] not in punctuation
|
|
):
|
|
if idx == 0:
|
|
words.append([])
|
|
words[-1].append(f"{t}")
|
|
else:
|
|
words.append([f"{t}"])
|
|
else:
|
|
if idx == 0:
|
|
words.append([])
|
|
words[-1].append(f"{t}")
|
|
return words
|
|
|
|
|
|
def g2p(text):
|
|
phones = []
|
|
tones = []
|
|
phone_len = []
|
|
|
|
|
|
words = text_to_words(text)
|
|
|
|
for word in words:
|
|
temp_phones, temp_tones = [], []
|
|
if len(word) > 1:
|
|
if "'" in word:
|
|
word = ["".join(word)]
|
|
for w in word:
|
|
if w in punctuation:
|
|
temp_phones.append(w)
|
|
temp_tones.append(0)
|
|
continue
|
|
if w.upper() in eng_dict:
|
|
phns, tns = refine_syllables(eng_dict[w.upper()])
|
|
temp_phones += [post_replace_ph(i) for i in phns]
|
|
temp_tones += tns
|
|
|
|
else:
|
|
phone_list = list(filter(lambda p: p != " ", _g2p(w)))
|
|
phns = []
|
|
tns = []
|
|
for ph in phone_list:
|
|
if ph in arpa:
|
|
ph, tn = refine_ph(ph)
|
|
phns.append(ph)
|
|
tns.append(tn)
|
|
else:
|
|
phns.append(ph)
|
|
tns.append(0)
|
|
temp_phones += [post_replace_ph(i) for i in phns]
|
|
temp_tones += tns
|
|
phones += temp_phones
|
|
tones += temp_tones
|
|
phone_len.append(len(temp_phones))
|
|
|
|
|
|
word2ph = []
|
|
for token, pl in zip(words, phone_len):
|
|
word_len = len(token)
|
|
|
|
aaa = distribute_phone(pl, word_len)
|
|
word2ph += aaa
|
|
|
|
phones = ["_"] + phones + ["_"]
|
|
tones = [0] + tones + [0]
|
|
word2ph = [1] + word2ph + [1]
|
|
assert len(phones) == len(tones), text
|
|
assert len(phones) == sum(word2ph), text
|
|
|
|
return phones, tones, word2ph
|
|
|
|
|
|
def get_bert_feature(text, word2ph):
|
|
from text import english_bert_mock
|
|
|
|
return english_bert_mock.get_bert_feature(text, word2ph)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
|
|
|
|
print(g2p("In this paper, we propose 1 DSPGAN, a GAN-based universal vocoder."))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|