From 95ffade41c57c5570feb255cbdb491c60abc6616 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Sayhan=20Yalva=C3=A7er?= Date: Sun, 18 Feb 2024 09:03:52 +0000 Subject: [PATCH] Upload config.json with huggingface_hub --- config.json | 29 ++--------------------------- 1 file changed, 2 insertions(+), 27 deletions(-) diff --git a/config.json b/config.json index 7b78e4b..9f0f76f 100644 --- a/config.json +++ b/config.json @@ -1,28 +1,3 @@ { - "_name_or_path": "teknium/OpenHermes-2.5-Mistral-7B", - "architectures": [ - "MistralForCausalLM" - ], - "attention_dropout": 0.0, - "bos_token_id": 1, - "eos_token_id": 32000, - "hidden_act": "silu", - "hidden_size": 4096, - "initializer_range": 0.02, - "intermediate_size": 14336, - "max_position_embeddings": 32768, - "model_type": "mistral", - "num_attention_heads": 32, - "num_hidden_layers": 32, - "num_key_value_heads": 8, - "pad_token_id": 32000, - "rms_norm_eps": 1e-05, - "rope_theta": 10000.0, - "sliding_window": 4096, - "tie_word_embeddings": false, - "torch_dtype": "bfloat16", - "transformers_version": "4.37.2", - "unsloth_version": "2024.2", - "use_cache": false, - "vocab_size": 32002 -} + "model_type": "mistral" +} \ No newline at end of file