xls-r_cv_ur / vocab.json
hadiqa123's picture
add tokenizer
d387f35
raw
history blame
202 Bytes
{
"[PAD]": 16,
"[UNK]": 15,
"|": 0,
"ا": 2,
"ب": 14,
"ت": 13,
"د": 1,
"ز": 7,
"س": 9,
"ل": 11,
"م": 12,
"ڑ": 8,
"ک": 5,
"ھ": 6,
"ہ": 4,
"ی": 3,
"ے": 10
}