Upload folder using huggingface_hub
This commit is contained in:
6
chat_template.jinja
Normal file
6
chat_template.jinja
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>
|
||||||
|
' + message['content'] + '<|end|>
|
||||||
|
'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>
|
||||||
|
' + message['content'] + '<|end|>
|
||||||
|
'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>
|
||||||
|
' }}{% else %}{{ eos_token }}{% endif %}
|
||||||
@@ -1,5 +1,4 @@
|
|||||||
{
|
{
|
||||||
"_name_or_path": "unsloth/Phi-3-medium-4k-instruct",
|
|
||||||
"architectures": [
|
"architectures": [
|
||||||
"MistralForCausalLM"
|
"MistralForCausalLM"
|
||||||
],
|
],
|
||||||
@@ -22,7 +21,8 @@
|
|||||||
"sliding_window": 2048,
|
"sliding_window": 2048,
|
||||||
"tie_word_embeddings": false,
|
"tie_word_embeddings": false,
|
||||||
"torch_dtype": "bfloat16",
|
"torch_dtype": "bfloat16",
|
||||||
"transformers_version": "4.44.2",
|
"transformers_version": "4.52.4",
|
||||||
|
"unsloth_fixed": true,
|
||||||
"unsloth_version": "2024.9",
|
"unsloth_version": "2024.9",
|
||||||
"use_cache": true,
|
"use_cache": true,
|
||||||
"vocab_size": 32064
|
"vocab_size": 32064
|
||||||
|
|||||||
@@ -8,5 +8,5 @@
|
|||||||
],
|
],
|
||||||
"max_length": 4096,
|
"max_length": 4096,
|
||||||
"pad_token_id": 32009,
|
"pad_token_id": 32009,
|
||||||
"transformers_version": "4.44.2"
|
"transformers_version": "4.52.4"
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -13,7 +13,13 @@
|
|||||||
"rstrip": false,
|
"rstrip": false,
|
||||||
"single_word": false
|
"single_word": false
|
||||||
},
|
},
|
||||||
"pad_token": "<|placeholder6|>",
|
"pad_token": {
|
||||||
|
"content": "<|placeholder6|>",
|
||||||
|
"lstrip": false,
|
||||||
|
"normalized": false,
|
||||||
|
"rstrip": false,
|
||||||
|
"single_word": false
|
||||||
|
},
|
||||||
"unk_token": {
|
"unk_token": {
|
||||||
"content": "<unk>",
|
"content": "<unk>",
|
||||||
"lstrip": false,
|
"lstrip": false,
|
||||||
|
|||||||
306245
tokenizer.json
306245
tokenizer.json
File diff suppressed because it is too large
Load Diff
@@ -117,9 +117,9 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"bos_token": "<s>",
|
"bos_token": "<s>",
|
||||||
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}",
|
|
||||||
"clean_up_tokenization_spaces": false,
|
"clean_up_tokenization_spaces": false,
|
||||||
"eos_token": "<|endoftext|>",
|
"eos_token": "<|endoftext|>",
|
||||||
|
"extra_special_tokens": {},
|
||||||
"legacy": false,
|
"legacy": false,
|
||||||
"model_max_length": 4096,
|
"model_max_length": 4096,
|
||||||
"pad_token": "<|placeholder6|>",
|
"pad_token": "<|placeholder6|>",
|
||||||
@@ -127,5 +127,6 @@
|
|||||||
"sp_model_kwargs": {},
|
"sp_model_kwargs": {},
|
||||||
"tokenizer_class": "LlamaTokenizer",
|
"tokenizer_class": "LlamaTokenizer",
|
||||||
"unk_token": "<unk>",
|
"unk_token": "<unk>",
|
||||||
"use_default_system_prompt": false
|
"use_default_system_prompt": false,
|
||||||
}
|
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}"
|
||||||
|
}
|
||||||
Reference in New Issue
Block a user