88 lines
1.9 KiB
JSON
88 lines
1.9 KiB
JSON
|
|
{
|
||
|
|
"added_tokens_decoder": {
|
||
|
|
"0": {
|
||
|
|
"content": "<unk>",
|
||
|
|
"lstrip": false,
|
||
|
|
"normalized": false,
|
||
|
|
"rstrip": false,
|
||
|
|
"single_word": false,
|
||
|
|
"special": true
|
||
|
|
},
|
||
|
|
"1": {
|
||
|
|
"content": "<s>",
|
||
|
|
"lstrip": false,
|
||
|
|
"normalized": false,
|
||
|
|
"rstrip": false,
|
||
|
|
"single_word": false,
|
||
|
|
"special": true
|
||
|
|
},
|
||
|
|
"2": {
|
||
|
|
"content": "</s>",
|
||
|
|
"lstrip": false,
|
||
|
|
"normalized": false,
|
||
|
|
"rstrip": false,
|
||
|
|
"single_word": false,
|
||
|
|
"special": true
|
||
|
|
},
|
||
|
|
"32000": {
|
||
|
|
"content": "<|bos|>",
|
||
|
|
"lstrip": false,
|
||
|
|
"normalized": false,
|
||
|
|
"rstrip": false,
|
||
|
|
"single_word": false,
|
||
|
|
"special": true
|
||
|
|
},
|
||
|
|
"32001": {
|
||
|
|
"content": "<|endoftext|>",
|
||
|
|
"lstrip": false,
|
||
|
|
"normalized": false,
|
||
|
|
"rstrip": false,
|
||
|
|
"single_word": false,
|
||
|
|
"special": true
|
||
|
|
},
|
||
|
|
"32002": {
|
||
|
|
"content": "[PAD]",
|
||
|
|
"lstrip": false,
|
||
|
|
"normalized": false,
|
||
|
|
"rstrip": false,
|
||
|
|
"single_word": false,
|
||
|
|
"special": true
|
||
|
|
},
|
||
|
|
"32003": {
|
||
|
|
"content": "<|ASSISTANT|>",
|
||
|
|
"lstrip": false,
|
||
|
|
"normalized": false,
|
||
|
|
"rstrip": false,
|
||
|
|
"single_word": false,
|
||
|
|
"special": true
|
||
|
|
},
|
||
|
|
"32004": {
|
||
|
|
"content": "<|USER|>",
|
||
|
|
"lstrip": false,
|
||
|
|
"normalized": false,
|
||
|
|
"rstrip": false,
|
||
|
|
"single_word": false,
|
||
|
|
"special": true
|
||
|
|
}
|
||
|
|
},
|
||
|
|
"additional_special_tokens": [
|
||
|
|
"<|ASSISTANT|>",
|
||
|
|
"<|USER|>"
|
||
|
|
],
|
||
|
|
"bos_token": "<s>",
|
||
|
|
"clean_up_tokenization_spaces": false,
|
||
|
|
"eos_token": "</s>",
|
||
|
|
"legacy": true,
|
||
|
|
"max_length": 1536,
|
||
|
|
"model_max_length": 4096,
|
||
|
|
"pad_token": "[PAD]",
|
||
|
|
"sp_model_kwargs": {},
|
||
|
|
"spaces_between_special_tokens": false,
|
||
|
|
"stride": 0,
|
||
|
|
"tokenizer_class": "LlamaTokenizer",
|
||
|
|
"truncation_side": "right",
|
||
|
|
"truncation_strategy": "longest_first",
|
||
|
|
"unk_token": "<unk>",
|
||
|
|
"use_default_system_prompt": true
|
||
|
|
}
|