Upload folder using huggingface_hub
This commit is contained in:
6
chat_template.jinja
Normal file
6
chat_template.jinja
Normal file
@@ -0,0 +1,6 @@
|
||||
{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>
|
||||
' + message['content'] + '<|end|>
|
||||
'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>
|
||||
' + message['content'] + '<|end|>
|
||||
'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>
|
||||
' }}{% else %}{{ eos_token }}{% endif %}
|
||||
@@ -1,5 +1,4 @@
|
||||
{
|
||||
"_name_or_path": "unsloth/Phi-3-medium-4k-instruct",
|
||||
"architectures": [
|
||||
"MistralForCausalLM"
|
||||
],
|
||||
@@ -22,7 +21,8 @@
|
||||
"sliding_window": 2048,
|
||||
"tie_word_embeddings": false,
|
||||
"torch_dtype": "bfloat16",
|
||||
"transformers_version": "4.44.2",
|
||||
"transformers_version": "4.52.4",
|
||||
"unsloth_fixed": true,
|
||||
"unsloth_version": "2024.9",
|
||||
"use_cache": true,
|
||||
"vocab_size": 32064
|
||||
|
||||
@@ -8,5 +8,5 @@
|
||||
],
|
||||
"max_length": 4096,
|
||||
"pad_token_id": 32009,
|
||||
"transformers_version": "4.44.2"
|
||||
"transformers_version": "4.52.4"
|
||||
}
|
||||
|
||||
@@ -13,7 +13,13 @@
|
||||
"rstrip": false,
|
||||
"single_word": false
|
||||
},
|
||||
"pad_token": "<|placeholder6|>",
|
||||
"pad_token": {
|
||||
"content": "<|placeholder6|>",
|
||||
"lstrip": false,
|
||||
"normalized": false,
|
||||
"rstrip": false,
|
||||
"single_word": false
|
||||
},
|
||||
"unk_token": {
|
||||
"content": "<unk>",
|
||||
"lstrip": false,
|
||||
|
||||
306245
tokenizer.json
306245
tokenizer.json
File diff suppressed because it is too large
Load Diff
@@ -117,9 +117,9 @@
|
||||
}
|
||||
},
|
||||
"bos_token": "<s>",
|
||||
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}",
|
||||
"clean_up_tokenization_spaces": false,
|
||||
"eos_token": "<|endoftext|>",
|
||||
"extra_special_tokens": {},
|
||||
"legacy": false,
|
||||
"model_max_length": 4096,
|
||||
"pad_token": "<|placeholder6|>",
|
||||
@@ -127,5 +127,6 @@
|
||||
"sp_model_kwargs": {},
|
||||
"tokenizer_class": "LlamaTokenizer",
|
||||
"unk_token": "<unk>",
|
||||
"use_default_system_prompt": false
|
||||
}
|
||||
"use_default_system_prompt": false,
|
||||
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}"
|
||||
}
|
||||
Reference in New Issue
Block a user