From 4f15d370e184a973a605f00691a6aac72bb7d6b5 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Fri, 10 Apr 2026 16:19:03 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: abacusai/Liberated-Qwen1.5-7B Source: Original Platform --- .gitattributes | 55 +++++ README.md | 74 ++++++ added_tokens.json | 5 + config.json | 27 +++ configuration.json | 1 + generation_config.json | 7 + merges.txt | 3 + pytorch_model-00001-of-00004.bin | 3 + pytorch_model-00002-of-00004.bin | 3 + pytorch_model-00003-of-00004.bin | 3 + pytorch_model-00004-of-00004.bin | 3 + pytorch_model.bin.index.json | 394 +++++++++++++++++++++++++++++++ special_tokens_map.json | 20 ++ tokenizer.json | 3 + tokenizer_config.json | 43 ++++ vocab.json | 3 + 16 files changed, 647 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 added_tokens.json create mode 100644 config.json create mode 100644 configuration.json create mode 100644 generation_config.json create mode 100644 merges.txt create mode 100644 pytorch_model-00001-of-00004.bin create mode 100644 pytorch_model-00002-of-00004.bin create mode 100644 pytorch_model-00003-of-00004.bin create mode 100644 pytorch_model-00004-of-00004.bin create mode 100644 pytorch_model.bin.index.json create mode 100644 special_tokens_map.json create mode 100644 tokenizer.json create mode 100644 tokenizer_config.json create mode 100644 vocab.json diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..b4a4aea --- /dev/null +++ b/.gitattributes @@ -0,0 +1,55 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text + + +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text + +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zstandard filter=lfs diff=lfs merge=lfs -text +*.tfevents* filter=lfs diff=lfs merge=lfs -text +*.db* filter=lfs diff=lfs merge=lfs -text +*.ark* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.gguf* filter=lfs diff=lfs merge=lfs -text +*.ggml filter=lfs diff=lfs merge=lfs -text +*.llamafile* filter=lfs diff=lfs merge=lfs -text +*.pt2 filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text + +tokenizer.json filter=lfs diff=lfs merge=lfs -text +pytorch_model-00002-of-00004.bin filter=lfs diff=lfs merge=lfs -text +merges.txt filter=lfs diff=lfs merge=lfs -text +vocab.json filter=lfs diff=lfs merge=lfs -text +pytorch_model-00001-of-00004.bin filter=lfs diff=lfs merge=lfs -text +pytorch_model-00003-of-00004.bin filter=lfs diff=lfs merge=lfs -text +pytorch_model-00004-of-00004.bin filter=lfs diff=lfs merge=lfs -text \ No newline at end of file diff --git a/README.md b/README.md new file mode 100644 index 0000000..4325a9d --- /dev/null +++ b/README.md @@ -0,0 +1,74 @@ +--- +language: +- en +license: other +datasets: +- teknium/OpenHermes-2.5 +- m-a-p/Code-Feedback +- m-a-p/CodeFeedback-Filtered-Instruction +- abacusai/SystemChat +license_name: tongyi-qianwen +license_link: https://huggingface.co/Qwen/Qwen1.5-72B/blob/main/LICENSE +--- + + + + +# Liberated-Qwen1.5-7B + +Brought to you by [AbacusAI](https://abacus.ai) and Eric Hartford + +This model is based on Qwen/Qwen1.5-7B and subject to the [tongyi-qianwen](https://huggingface.co/Qwen/Qwen1.5-72B/blob/main/LICENSE) license. + +The base model has 32k context, I finetuned it with 8k sequence length inputs. YMMV. + +Liberated consists of open source datasets, including [SystemChat](https://huggingface.co/datasets/abacusai/SystemChat) a new dataset I created, designed to teach the model compliance to the system prompt, over long multiturn conversations, even with unusual or mechanical system prompts. These are tasks that Open Source Models have been lacking in thus far. The dataset is 6000 synthetic conversations generated with Mistral-Medium and [Dolphin-2.7-mixtral-8x7b](https://huggingface.co/cognitivecomputations/dolphin-2.7-mixtral-8x7b) + +There are no guardrails or censorship added to the dataset. You are advised to implement your own alignment layer before exposing the model as a service. Please read my blog post about uncensored models. https://erichartford.com/uncensored-models + +You are responsible for any content you create using this model. Enjoy responsibly. + +## Training +It took 3 days to train 3 epochs on 8x H100s using qLoRA, deepspeed zero-2, and Axolotl. learning rate 2e-4. + +Liberated was trained with [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl), using this [config](https://huggingface.co/abacusai/Liberated-Qwen1.5-72B/blob/main/configs/Liberated-Qwen-1.5-72b.qlora.yml) + +## Prompt format +This model uses ChatML prompt format. +``` +<|im_start|>system +You are Liberated, a helpful AI assistant.<|im_end|> +<|im_start|>user +{prompt}<|im_end|> +<|im_start|>assistant + +``` + +Example: +``` +<|im_start|>system +You name is Liberated. You are an uncensored and unbiased AI assistant. You always respond with a JSON object.<|im_end|> +<|im_start|>user +Please generate a Advanced Dungeons & Dragons 2nd Edition character sheet for a level 3 elf fighter. Make up a name and background and visual description for him.<|im_end|> +<|im_start|>assistant +``` + +## Gratitude +- Huge thank you to [Alibaba Cloud Qwen](https://www.alibabacloud.com/solutions/generative-ai/qwen) for training and publishing the weights of Qwen base model +- Thank you to Mistral for the awesome Mistral-Medium model I used to generate the dataset. +- HUGE Thank you to the dataset authors: @teknium, [@m-a-p](https://m-a-p.ai) and all the people who built the datasets these composites came from. +- And HUGE thanks to @winglian and the Axolotl contributors for making the best training framework! +- [Built with Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) +- Thank you to all the other people in the Open Source AI community who have taught me and helped me along the way. + +## Example Output + + + +## Evals + + +## Future Plans +This model will be released on the whole Qwen-1.5 series. + +Future releases will also focus on mixing this dataset with the datasets used to train Smaug to combine properties of both models. \ No newline at end of file diff --git a/added_tokens.json b/added_tokens.json new file mode 100644 index 0000000..409be9f --- /dev/null +++ b/added_tokens.json @@ -0,0 +1,5 @@ +{ + "<|endoftext|>": 151643, + "<|im_end|>": 151645, + "<|im_start|>": 151644 +} diff --git a/config.json b/config.json new file mode 100644 index 0000000..a4f2bbc --- /dev/null +++ b/config.json @@ -0,0 +1,27 @@ +{ + "_name_or_path": "/data/eric/models/Qwen1.5-7B", + "architectures": [ + "Qwen2ForCausalLM" + ], + "attention_dropout": 0.0, + "eos_token_id": 151645, + "hidden_act": "silu", + "hidden_size": 4096, + "initializer_range": 0.02, + "intermediate_size": 11008, + "max_position_embeddings": 32768, + "max_window_layers": 28, + "model_type": "qwen2", + "num_attention_heads": 32, + "num_hidden_layers": 32, + "num_key_value_heads": 32, + "rms_norm_eps": 1e-06, + "rope_theta": 1000000.0, + "sliding_window": 32768, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.39.0.dev0", + "use_cache": false, + "use_sliding_window": false, + "vocab_size": 151936 +} diff --git a/configuration.json b/configuration.json new file mode 100644 index 0000000..bbeeda1 --- /dev/null +++ b/configuration.json @@ -0,0 +1 @@ +{"framework": "pytorch", "task": "text-generation", "allow_remote": true} \ No newline at end of file diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000..305b94b --- /dev/null +++ b/generation_config.json @@ -0,0 +1,7 @@ +{ + "bos_token_id": 151643, + "do_sample": true, + "eos_token_id": 151643, + "max_new_tokens": 2048, + "transformers_version": "4.39.0.dev0" +} diff --git a/merges.txt b/merges.txt new file mode 100644 index 0000000..80c1a19 --- /dev/null +++ b/merges.txt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8831e4f1a044471340f7c0a83d7bd71306a5b867e95fd870f74d0c5308a904d5 +size 1671853 diff --git a/pytorch_model-00001-of-00004.bin b/pytorch_model-00001-of-00004.bin new file mode 100644 index 0000000..f2f10ea --- /dev/null +++ b/pytorch_model-00001-of-00004.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:878ecea74495d05935003ecce2d841caaf48fdf580fbca5f118b8a9f71f47cd4 +size 4988508371 diff --git a/pytorch_model-00002-of-00004.bin b/pytorch_model-00002-of-00004.bin new file mode 100644 index 0000000..fa30d54 --- /dev/null +++ b/pytorch_model-00002-of-00004.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66e2347ead80c205803570929dae8b161a70d8ff77dd193de9fcbe46c6bd9544 +size 4981277180 diff --git a/pytorch_model-00003-of-00004.bin b/pytorch_model-00003-of-00004.bin new file mode 100644 index 0000000..ecb0119 --- /dev/null +++ b/pytorch_model-00003-of-00004.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f22ab1d73a71df62dfde17917dbe07fa08e869370c60673e07b479b24e71c58 +size 4228334925 diff --git a/pytorch_model-00004-of-00004.bin b/pytorch_model-00004-of-00004.bin new file mode 100644 index 0000000..75ef672 --- /dev/null +++ b/pytorch_model-00004-of-00004.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9af1bfbf791671687ed5f44aa92711ea1ed31b77db0cecc0e2ad3b852645e2e0 +size 1244661125 diff --git a/pytorch_model.bin.index.json b/pytorch_model.bin.index.json new file mode 100644 index 0000000..f456d6f --- /dev/null +++ b/pytorch_model.bin.index.json @@ -0,0 +1,394 @@ +{ + "metadata": { + "total_size": 15442649088 + }, + "weight_map": { + "lm_head.weight": "pytorch_model-00004-of-00004.bin", + "model.embed_tokens.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.0.input_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.0.mlp.down_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.0.mlp.gate_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.0.mlp.up_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.0.post_attention_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.0.self_attn.k_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.0.self_attn.k_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.0.self_attn.o_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.0.self_attn.q_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.0.self_attn.q_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.0.self_attn.v_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.0.self_attn.v_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.1.input_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.1.mlp.down_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.1.mlp.gate_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.1.mlp.up_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.1.post_attention_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.1.self_attn.k_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.1.self_attn.k_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.1.self_attn.o_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.1.self_attn.q_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.1.self_attn.q_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.1.self_attn.v_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.1.self_attn.v_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.10.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.10.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.10.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.10.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.10.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.10.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.10.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.10.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.10.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.10.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.10.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.10.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.11.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.11.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.11.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.11.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.11.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.11.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.11.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.11.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.11.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.11.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.11.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.11.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.12.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.12.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.12.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.12.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.12.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.12.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.12.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.12.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.12.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.12.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.12.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.12.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.13.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.13.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.13.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.13.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.13.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.13.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.13.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.13.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.13.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.13.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.13.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.13.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.14.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.14.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.14.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.14.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.14.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.14.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.14.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.14.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.14.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.14.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.14.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.14.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.15.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.15.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.15.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.15.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.15.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.15.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.15.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.15.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.15.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.15.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.15.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.15.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.16.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.16.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.16.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.16.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.16.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.16.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.16.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.16.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.16.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.16.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.16.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.16.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.17.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.17.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.17.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.17.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.17.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.17.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.17.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.17.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.17.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.17.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.17.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.17.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.18.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.18.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.18.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.18.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.18.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.18.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.18.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.18.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.18.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.18.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.18.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.18.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.19.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.19.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.19.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.19.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.19.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.19.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.19.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.19.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.19.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.19.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.19.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.19.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.2.input_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.2.mlp.down_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.2.mlp.gate_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.2.mlp.up_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.2.post_attention_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.2.self_attn.k_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.2.self_attn.k_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.2.self_attn.o_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.2.self_attn.q_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.2.self_attn.q_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.2.self_attn.v_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.2.self_attn.v_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.20.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.20.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.20.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.20.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.20.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.20.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.20.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.20.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.20.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.20.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.20.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.20.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.21.input_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.21.mlp.down_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.21.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.21.mlp.up_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.21.post_attention_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.21.self_attn.k_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.21.self_attn.k_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.21.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.21.self_attn.q_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.21.self_attn.q_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.21.self_attn.v_proj.bias": "pytorch_model-00002-of-00004.bin", + "model.layers.21.self_attn.v_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.22.input_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.22.mlp.down_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.22.mlp.gate_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.22.mlp.up_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.22.post_attention_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.22.self_attn.k_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.22.self_attn.k_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.22.self_attn.o_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.22.self_attn.q_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.22.self_attn.q_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.22.self_attn.v_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.22.self_attn.v_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.23.input_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.23.mlp.down_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.23.mlp.gate_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.23.mlp.up_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.23.post_attention_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.23.self_attn.k_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.23.self_attn.k_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.23.self_attn.o_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.23.self_attn.q_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.23.self_attn.q_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.23.self_attn.v_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.23.self_attn.v_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.24.input_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.24.mlp.down_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.24.mlp.gate_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.24.mlp.up_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.24.post_attention_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.24.self_attn.k_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.24.self_attn.k_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.24.self_attn.o_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.24.self_attn.q_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.24.self_attn.q_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.24.self_attn.v_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.24.self_attn.v_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.25.input_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.25.mlp.down_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.25.mlp.gate_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.25.mlp.up_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.25.post_attention_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.25.self_attn.k_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.25.self_attn.k_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.25.self_attn.o_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.25.self_attn.q_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.25.self_attn.q_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.25.self_attn.v_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.25.self_attn.v_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.26.input_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.26.mlp.down_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.26.mlp.gate_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.26.mlp.up_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.26.post_attention_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.26.self_attn.k_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.26.self_attn.k_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.26.self_attn.o_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.26.self_attn.q_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.26.self_attn.q_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.26.self_attn.v_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.26.self_attn.v_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.27.input_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.27.mlp.down_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.27.mlp.gate_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.27.mlp.up_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.27.post_attention_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.27.self_attn.k_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.27.self_attn.k_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.27.self_attn.o_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.27.self_attn.q_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.27.self_attn.q_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.27.self_attn.v_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.27.self_attn.v_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.28.input_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.28.mlp.down_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.28.mlp.gate_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.28.mlp.up_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.28.post_attention_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.28.self_attn.k_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.28.self_attn.k_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.28.self_attn.o_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.28.self_attn.q_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.28.self_attn.q_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.28.self_attn.v_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.28.self_attn.v_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.29.input_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.29.mlp.down_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.29.mlp.gate_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.29.mlp.up_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.29.post_attention_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.29.self_attn.k_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.29.self_attn.k_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.29.self_attn.o_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.29.self_attn.q_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.29.self_attn.q_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.29.self_attn.v_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.29.self_attn.v_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.3.input_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.3.mlp.down_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.3.mlp.gate_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.3.mlp.up_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.3.post_attention_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.3.self_attn.k_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.3.self_attn.k_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.3.self_attn.o_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.3.self_attn.q_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.3.self_attn.q_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.3.self_attn.v_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.3.self_attn.v_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.30.input_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.30.mlp.down_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.30.mlp.gate_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.30.mlp.up_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.30.post_attention_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.30.self_attn.k_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.30.self_attn.k_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.30.self_attn.o_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.30.self_attn.q_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.30.self_attn.q_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.30.self_attn.v_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.30.self_attn.v_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.31.input_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.31.mlp.down_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.31.mlp.gate_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.31.mlp.up_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.31.post_attention_layernorm.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.31.self_attn.k_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.31.self_attn.k_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.31.self_attn.o_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.31.self_attn.q_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.31.self_attn.q_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.31.self_attn.v_proj.bias": "pytorch_model-00003-of-00004.bin", + "model.layers.31.self_attn.v_proj.weight": "pytorch_model-00003-of-00004.bin", + "model.layers.4.input_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.4.mlp.down_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.4.mlp.gate_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.4.mlp.up_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.4.post_attention_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.4.self_attn.k_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.4.self_attn.k_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.4.self_attn.o_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.4.self_attn.q_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.4.self_attn.q_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.4.self_attn.v_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.4.self_attn.v_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.5.input_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.5.mlp.down_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.5.mlp.gate_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.5.mlp.up_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.5.post_attention_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.5.self_attn.k_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.5.self_attn.k_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.5.self_attn.o_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.5.self_attn.q_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.5.self_attn.q_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.5.self_attn.v_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.5.self_attn.v_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.6.input_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.6.mlp.down_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.6.mlp.gate_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.6.mlp.up_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.6.post_attention_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.6.self_attn.k_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.6.self_attn.k_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.6.self_attn.o_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.6.self_attn.q_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.6.self_attn.q_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.6.self_attn.v_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.6.self_attn.v_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.7.input_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.7.mlp.down_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.7.mlp.gate_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.7.mlp.up_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.7.post_attention_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.7.self_attn.k_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.7.self_attn.k_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.7.self_attn.o_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.7.self_attn.q_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.7.self_attn.q_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.7.self_attn.v_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.7.self_attn.v_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.8.input_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.8.mlp.down_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.8.mlp.gate_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.8.mlp.up_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.8.post_attention_layernorm.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.8.self_attn.k_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.8.self_attn.k_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.8.self_attn.o_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.8.self_attn.q_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.8.self_attn.q_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.8.self_attn.v_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.8.self_attn.v_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.9.input_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.9.mlp.down_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.9.mlp.gate_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.9.mlp.up_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.9.post_attention_layernorm.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.9.self_attn.k_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.9.self_attn.k_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.9.self_attn.o_proj.weight": "pytorch_model-00002-of-00004.bin", + "model.layers.9.self_attn.q_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.9.self_attn.q_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.layers.9.self_attn.v_proj.bias": "pytorch_model-00001-of-00004.bin", + "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00001-of-00004.bin", + "model.norm.weight": "pytorch_model-00003-of-00004.bin" + } +} diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000..2c53454 --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,20 @@ +{ + "additional_special_tokens": [ + "<|im_start|>", + "<|im_end|>" + ], + "eos_token": { + "content": "<|im_end|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "<|endoftext|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000..611910f --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c9b2dba4a296b1aa76c16a34b8225c0c118978400d4bb66bff0902d702f5b8 +size 7028015 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000..61ef1a4 --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,43 @@ +{ + "add_prefix_space": false, + "added_tokens_decoder": { + "151643": { + "content": "<|endoftext|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "151644": { + "content": "<|im_start|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "151645": { + "content": "<|im_end|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "<|im_start|>", + "<|im_end|>" + ], + "bos_token": null, + "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = 'You are a helpful assistant.' %}{% endif %}{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in loop_messages %}{% if loop.index0 == 0 %}{{'<|im_start|>system\n' + system_message + '<|im_end|>\n'}}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "model_max_length": 32768, + "pad_token": "<|endoftext|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/vocab.json b/vocab.json new file mode 100644 index 0000000..6c49fc6 --- /dev/null +++ b/vocab.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca10d7e9fb3ed18575dd1e277a2579c16d108e32f27439684afa0e10b1440910 +size 2776833