{ "added_tokens_decoder": { "0": { "content": "", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": false } }, "additional_special_tokens": [], "clean_up_tokenization_spaces": true, "model_max_length": 512, "tokenizer_class": "OpenAIGPTTokenizer", "tokenizer_file": "/root/.cache/huggingface/hub/models--openai-gpt/snapshots/adacbec1426c790163037629ab537b20464a8f71/tokenizer.json", "unk_token": "" }