Upload folder using huggingface_hub
This commit is contained in:
@@ -1,5 +1,4 @@
|
|||||||
{
|
{
|
||||||
"_name_or_path": "meta-llama/Llama-3.2-1B",
|
|
||||||
"architectures": [
|
"architectures": [
|
||||||
"LlamaForCausalLM"
|
"LlamaForCausalLM"
|
||||||
],
|
],
|
||||||
@@ -31,7 +30,7 @@
|
|||||||
"rope_theta": 500000.0,
|
"rope_theta": 500000.0,
|
||||||
"tie_word_embeddings": true,
|
"tie_word_embeddings": true,
|
||||||
"torch_dtype": "bfloat16",
|
"torch_dtype": "bfloat16",
|
||||||
"transformers_version": "4.48.1",
|
"transformers_version": "4.52.4",
|
||||||
"unsloth_fixed": true,
|
"unsloth_fixed": true,
|
||||||
"use_cache": true,
|
"use_cache": true,
|
||||||
"vocab_size": 128256
|
"vocab_size": 128256
|
||||||
|
|||||||
@@ -7,5 +7,5 @@
|
|||||||
"pad_token_id": 128004,
|
"pad_token_id": 128004,
|
||||||
"temperature": 0.6,
|
"temperature": 0.6,
|
||||||
"top_p": 0.9,
|
"top_p": 0.9,
|
||||||
"transformers_version": "4.48.1"
|
"transformers_version": "4.52.4"
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2061,6 +2061,6 @@
|
|||||||
"model_max_length": 131072,
|
"model_max_length": 131072,
|
||||||
"pad_token": "<|finetune_right_pad_id|>",
|
"pad_token": "<|finetune_right_pad_id|>",
|
||||||
"padding_side": "left",
|
"padding_side": "left",
|
||||||
"tokenizer_class": "PreTrainedTokenizerFast",
|
"tokenizer_class": "PreTrainedTokenizer",
|
||||||
"unk_token": null
|
"unk_token": null
|
||||||
}
|
}
|
||||||
Reference in New Issue
Block a user