diff --git a/config.json b/config.json index 2df560b..bd7dab8 100644 --- a/config.json +++ b/config.json @@ -1,5 +1,4 @@ { - "_name_or_path": "meta-llama/Llama-3.2-1B", "architectures": [ "LlamaForCausalLM" ], @@ -31,7 +30,7 @@ "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", - "transformers_version": "4.48.1", + "transformers_version": "4.52.4", "unsloth_fixed": true, "use_cache": true, "vocab_size": 128256 diff --git a/generation_config.json b/generation_config.json index 83b77ea..e9d7feb 100644 --- a/generation_config.json +++ b/generation_config.json @@ -7,5 +7,5 @@ "pad_token_id": 128004, "temperature": 0.6, "top_p": 0.9, - "transformers_version": "4.48.1" + "transformers_version": "4.52.4" } diff --git a/tokenizer_config.json b/tokenizer_config.json index 0269fe2..8a07177 100644 --- a/tokenizer_config.json +++ b/tokenizer_config.json @@ -2061,6 +2061,6 @@ "model_max_length": 131072, "pad_token": "<|finetune_right_pad_id|>", "padding_side": "left", - "tokenizer_class": "PreTrainedTokenizerFast", + "tokenizer_class": "PreTrainedTokenizer", "unk_token": null -} +} \ No newline at end of file