polyglot-ko-1.3b / tokenizer_config.json
jason9693's picture
Upload tokenizer
40d2d2e
raw
history blame
312 Bytes
{
"eos_token": "<eos>",
"name_or_path": "EleutherAI/gpt-neox-ko-1.3b",
"pad_token": "<pad>",
"special_tokens_map_file": "/fsx/kevinai/.cache/models--EleutherAI--gpt-neox-ko-1.3b/snapshots/1736ae3c92f641c994d1361fb5d2c680fd483ab0/special_tokens_map.json",
"tokenizer_class": "PreTrainedTokenizerFast"
}