flexthink
commited on
Commit
·
b6b12e5
1
Parent(s):
fead939
Clean-up and fixes
Browse files- librig2p-nostress.py +4 -48
librig2p-nostress.py
CHANGED
|
@@ -7,56 +7,12 @@ import json
|
|
| 7 |
|
| 8 |
import datasets
|
| 9 |
|
| 10 |
-
_DESCRIPTION = """
|
| 11 |
Grapheme-to-Phoneme training, validation and test sets
|
| 12 |
"""
|
| 13 |
|
| 14 |
-
_BASE_URL = "https://
|
| 15 |
-
|
| 16 |
-
_HOMEPAGE_URL = "https://github.com/flexthink/librig2p-nostress/tree/develop"
|
| 17 |
-
|
| 18 |
-
_PHONEMES = [
|
| 19 |
-
"AA",
|
| 20 |
-
"AE",
|
| 21 |
-
"AH",
|
| 22 |
-
"AO",
|
| 23 |
-
"AW",
|
| 24 |
-
"AY",
|
| 25 |
-
"B",
|
| 26 |
-
"CH",
|
| 27 |
-
"D",
|
| 28 |
-
"DH",
|
| 29 |
-
"EH",
|
| 30 |
-
"ER",
|
| 31 |
-
"EY",
|
| 32 |
-
"F",
|
| 33 |
-
"G",
|
| 34 |
-
"HH",
|
| 35 |
-
"IH",
|
| 36 |
-
"IY",
|
| 37 |
-
"JH",
|
| 38 |
-
"K",
|
| 39 |
-
"L",
|
| 40 |
-
"M",
|
| 41 |
-
"N",
|
| 42 |
-
"NG",
|
| 43 |
-
"OW",
|
| 44 |
-
"OY",
|
| 45 |
-
"P",
|
| 46 |
-
"R",
|
| 47 |
-
"S",
|
| 48 |
-
"SH",
|
| 49 |
-
"T",
|
| 50 |
-
"TH",
|
| 51 |
-
"UH",
|
| 52 |
-
"UW",
|
| 53 |
-
"V",
|
| 54 |
-
"W",
|
| 55 |
-
"Y",
|
| 56 |
-
"Z",
|
| 57 |
-
"ZH",
|
| 58 |
-
]
|
| 59 |
-
_ORIGINS = ["librispeech", "librispeech-lex"]
|
| 60 |
_NA = "N/A"
|
| 61 |
_SPLIT_TYPES = ["train", "valid", "test"]
|
| 62 |
_DATA_TYPES = ["lexicon", "sentence"]
|
|
@@ -78,7 +34,7 @@ class GraphemeToPhoneme(datasets.GeneratorBasedBuilder):
|
|
| 78 |
{
|
| 79 |
"id": datasets.Value("string"),
|
| 80 |
"speaker_id": datasets.Value("string"),
|
| 81 |
-
"origin": datasets.
|
| 82 |
"char": datasets.Value("string"),
|
| 83 |
"phn": datasets.Sequence(datasets.Value("string")),
|
| 84 |
},
|
|
|
|
| 7 |
|
| 8 |
import datasets
|
| 9 |
|
| 10 |
+
_DESCRIPTION = """
|
| 11 |
Grapheme-to-Phoneme training, validation and test sets
|
| 12 |
"""
|
| 13 |
|
| 14 |
+
_BASE_URL = "https://huggingface.co/datasets/flexthink/librig2p-nostress/resolve/main/dataset"
|
| 15 |
+
_HOMEPAGE_URL = "https://huggingface.co/datasets/flexthink/librig2p-nostress"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 16 |
_NA = "N/A"
|
| 17 |
_SPLIT_TYPES = ["train", "valid", "test"]
|
| 18 |
_DATA_TYPES = ["lexicon", "sentence"]
|
|
|
|
| 34 |
{
|
| 35 |
"id": datasets.Value("string"),
|
| 36 |
"speaker_id": datasets.Value("string"),
|
| 37 |
+
"origin": datasets.Value("string"),
|
| 38 |
"char": datasets.Value("string"),
|
| 39 |
"phn": datasets.Sequence(datasets.Value("string")),
|
| 40 |
},
|