| import re | |
| from .g2p import * | |
| import json | |
| import os | |
| class Tokenizer(): | |
| def __init__(self, data_path: str) -> None: | |
| self.dic = {} | |
| for line in open(os.path.join(data_path, "dictionary.txt")): | |
| items = line.split() | |
| self.dic[items[0]] = " ".join(items[1:]) | |
| self.config = json.load(open(os.path.join(data_path, "config.json"))) | |
| def g2p(self, text): | |
| text = re.sub("—", "-", text) | |
| text = re.sub("([!'(),-.:;?])", r' \1 ', text) | |
| phonemes = [] | |
| for word in text.split(): | |
| if re.match("[!'(),-.:;?]", word): | |
| phonemes.append(word) | |
| continue | |
| word = word.lower() | |
| if len(phonemes) > 0: phonemes.append(' ') | |
| if word in self.dic: | |
| phonemes.extend(self.dic[word].split()) | |
| else: | |
| phonemes.extend(convert(word).split()) | |
| phoneme_id_map = self.config["phoneme_id_map"] | |
| phoneme_ids = [] | |
| phoneme_ids.extend(phoneme_id_map["^"]) | |
| phoneme_ids.extend(phoneme_id_map["_"]) | |
| for p in phonemes: | |
| if p in phoneme_id_map: | |
| phoneme_ids.extend(phoneme_id_map[p]) | |
| phoneme_ids.extend(phoneme_id_map["_"]) | |
| phoneme_ids.extend(phoneme_id_map["$"]) | |
| return phoneme_ids, phonemes | |
| def _get_seq(self, text: str) -> list[int]: | |
| seq = self.g2p(text)[0] | |
| return seq |