Upload folder using huggingface_hub
This commit is contained in:
@@ -1,5 +1,4 @@
|
||||
{
|
||||
"_name_or_path": "meta-llama/Llama-3.2-1B",
|
||||
"architectures": [
|
||||
"LlamaForCausalLM"
|
||||
],
|
||||
@@ -31,7 +30,7 @@
|
||||
"rope_theta": 500000.0,
|
||||
"tie_word_embeddings": true,
|
||||
"torch_dtype": "bfloat16",
|
||||
"transformers_version": "4.48.1",
|
||||
"transformers_version": "4.52.4",
|
||||
"unsloth_fixed": true,
|
||||
"use_cache": true,
|
||||
"vocab_size": 128256
|
||||
|
||||
@@ -7,5 +7,5 @@
|
||||
"pad_token_id": 128004,
|
||||
"temperature": 0.6,
|
||||
"top_p": 0.9,
|
||||
"transformers_version": "4.48.1"
|
||||
"transformers_version": "4.52.4"
|
||||
}
|
||||
|
||||
@@ -2061,6 +2061,6 @@
|
||||
"model_max_length": 131072,
|
||||
"pad_token": "<|finetune_right_pad_id|>",
|
||||
"padding_side": "left",
|
||||
"tokenizer_class": "PreTrainedTokenizerFast",
|
||||
"tokenizer_class": "PreTrainedTokenizer",
|
||||
"unk_token": null
|
||||
}
|
||||
}
|
||||
Reference in New Issue
Block a user