Upload folder using huggingface_hub

This commit is contained in:
ai-modelscope
2025-06-15 11:13:09 +08:00
parent 1c58f79c89
commit 277a84f3b9
6 changed files with 245016 additions and 61256 deletions

6
chat_template.jinja Normal file
View File

@@ -0,0 +1,6 @@
{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>
' + message['content'] + '<|end|>
'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>
' + message['content'] + '<|end|>
'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>
' }}{% else %}{{ eos_token }}{% endif %}

View File

@@ -1,5 +1,4 @@
{
"_name_or_path": "unsloth/Phi-3-medium-4k-instruct",
"architectures": [
"MistralForCausalLM"
],
@@ -22,7 +21,8 @@
"sliding_window": 2048,
"tie_word_embeddings": false,
"torch_dtype": "bfloat16",
"transformers_version": "4.44.2",
"transformers_version": "4.52.4",
"unsloth_fixed": true,
"unsloth_version": "2024.9",
"use_cache": true,
"vocab_size": 32064

View File

@@ -8,5 +8,5 @@
],
"max_length": 4096,
"pad_token_id": 32009,
"transformers_version": "4.44.2"
"transformers_version": "4.52.4"
}

View File

@@ -13,7 +13,13 @@
"rstrip": false,
"single_word": false
},
"pad_token": "<|placeholder6|>",
"pad_token": {
"content": "<|placeholder6|>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false
},
"unk_token": {
"content": "<unk>",
"lstrip": false,

File diff suppressed because it is too large Load Diff

View File

@@ -117,9 +117,9 @@
}
},
"bos_token": "<s>",
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}",
"clean_up_tokenization_spaces": false,
"eos_token": "<|endoftext|>",
"extra_special_tokens": {},
"legacy": false,
"model_max_length": 4096,
"pad_token": "<|placeholder6|>",
@@ -127,5 +127,6 @@
"sp_model_kwargs": {},
"tokenizer_class": "LlamaTokenizer",
"unk_token": "<unk>",
"use_default_system_prompt": false
"use_default_system_prompt": false,
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}"
}