Upload tokenizer
This commit is contained in:
committed by
huggingface-web
parent
d268cdbb0d
commit
482263cf87
6
special_tokens_map.json
Normal file
6
special_tokens_map.json
Normal file
@@ -0,0 +1,6 @@
|
||||
{
|
||||
"bos_token": "</s>",
|
||||
"eos_token": "</s>",
|
||||
"pad_token": "[PAD]",
|
||||
"unk_token": "</s>"
|
||||
}
|
||||
3
tokenizer.model
Normal file
3
tokenizer.model
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
|
||||
size 499723
|
||||
9
tokenizer_config.json
Normal file
9
tokenizer_config.json
Normal file
@@ -0,0 +1,9 @@
|
||||
{
|
||||
"bos_token": "",
|
||||
"eos_token": "",
|
||||
"model_max_length": 2048,
|
||||
"padding_side": "right",
|
||||
"special_tokens_map_file": "models/decapoda-research_llama-7b-hf/special_tokens_map.json",
|
||||
"tokenizer_class": "LlamaTokenizer",
|
||||
"unk_token": ""
|
||||
}
|
||||
Reference in New Issue
Block a user