Upload folder using huggingface_hub

This commit is contained in:
ai-modelscope
2025-06-15 11:13:09 +08:00
parent 1c58f79c89
commit 277a84f3b9
6 changed files with 245016 additions and 61256 deletions

6
chat_template.jinja Normal file
View File

@@ -0,0 +1,6 @@
{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>
' + message['content'] + '<|end|>
'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>
' + message['content'] + '<|end|>
'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>
' }}{% else %}{{ eos_token }}{% endif %}

View File

@@ -1,5 +1,4 @@
{ {
"_name_or_path": "unsloth/Phi-3-medium-4k-instruct",
"architectures": [ "architectures": [
"MistralForCausalLM" "MistralForCausalLM"
], ],
@@ -22,7 +21,8 @@
"sliding_window": 2048, "sliding_window": 2048,
"tie_word_embeddings": false, "tie_word_embeddings": false,
"torch_dtype": "bfloat16", "torch_dtype": "bfloat16",
"transformers_version": "4.44.2", "transformers_version": "4.52.4",
"unsloth_fixed": true,
"unsloth_version": "2024.9", "unsloth_version": "2024.9",
"use_cache": true, "use_cache": true,
"vocab_size": 32064 "vocab_size": 32064

View File

@@ -8,5 +8,5 @@
], ],
"max_length": 4096, "max_length": 4096,
"pad_token_id": 32009, "pad_token_id": 32009,
"transformers_version": "4.44.2" "transformers_version": "4.52.4"
} }

View File

@@ -13,7 +13,13 @@
"rstrip": false, "rstrip": false,
"single_word": false "single_word": false
}, },
"pad_token": "<|placeholder6|>", "pad_token": {
"content": "<|placeholder6|>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false
},
"unk_token": { "unk_token": {
"content": "<unk>", "content": "<unk>",
"lstrip": false, "lstrip": false,

File diff suppressed because it is too large Load Diff

View File

@@ -117,9 +117,9 @@
} }
}, },
"bos_token": "<s>", "bos_token": "<s>",
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}",
"clean_up_tokenization_spaces": false, "clean_up_tokenization_spaces": false,
"eos_token": "<|endoftext|>", "eos_token": "<|endoftext|>",
"extra_special_tokens": {},
"legacy": false, "legacy": false,
"model_max_length": 4096, "model_max_length": 4096,
"pad_token": "<|placeholder6|>", "pad_token": "<|placeholder6|>",
@@ -127,5 +127,6 @@
"sp_model_kwargs": {}, "sp_model_kwargs": {},
"tokenizer_class": "LlamaTokenizer", "tokenizer_class": "LlamaTokenizer",
"unk_token": "<unk>", "unk_token": "<unk>",
"use_default_system_prompt": false "use_default_system_prompt": false,
} "chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}"
}