From b65e7fcda873691944438954a232d4af8bd25f08 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sat, 2 May 2026 16:05:23 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: Nanbeige/ToolMind-Web-3B Source: Original Platform --- .gitattributes | 36 ++ README.md | 68 ++++ added_tokens.json | 9 + config.json | 27 ++ generation_config.json | 11 + model.safetensors.index.json | 730 +++++++++++++++++++++++++++++++++++ nbg_search_performance.png | Bin 0 -> 36029 bytes pytorch_model.bin | 3 + special_tokens_map.json | 27 ++ tokenizer.json | 3 + tokenizer.model | 3 + tokenizer_config.json | 102 +++++ 12 files changed, 1019 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 added_tokens.json create mode 100644 config.json create mode 100644 generation_config.json create mode 100644 model.safetensors.index.json create mode 100644 nbg_search_performance.png create mode 100644 pytorch_model.bin create mode 100644 special_tokens_map.json create mode 100644 tokenizer.json create mode 100644 tokenizer.model create mode 100644 tokenizer_config.json diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..52373fe --- /dev/null +++ b/.gitattributes @@ -0,0 +1,36 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +tokenizer.json filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..0de3b9a --- /dev/null +++ b/README.md @@ -0,0 +1,68 @@ +--- +language: +- en +- zh +license: apache-2.0 +tags: +- search-agent +- rag +- deep-research +- 3b +- nlp +pipeline_tag: text-generation +--- + +# Overview + +ToolMind-Web-3B is a specialized lightweight agent built on top of the [**Nanbeige4-3B-Thinking-2511**](https://huggingface.co/Nanbeige/Nanbeige4-3B-Thinking-2511) foundation model. +Following extensive **SFT (Supervised Fine-Tuning)** and **RL (Reinforcement Learning)** focused on search behaviors, +our model attains leading performance among small-scale models on multiple long-horizon leaderboards like **Xbench-Deepsearch, HLE, and GAIA**, enabling reliable execution of up to hundreds of consecutive tool invocations. + +
+ + +
+ + +# Highlights + +1. **Strong Performance at Compact Scale** + +ToolMind-Web-3B delivers high-quality long-horizon reasoning and tool-augmented search capabilities while maintaining a lightweight 3B parameter footprint. Despite its compact size, it achieves competitive performance across multiple benchmarks like **Xbench-Deepsearch, GAIA, and HLE**. The model is evaluated under the **[MiroThinkers workflow](https://github.com/MiroMindAI/MiroThinker)**, ensuring standardized and reproducible assessment. + +2. **An Open-Source Complex QA Dataset Synthesized from Wikipedia Entity–Relation Knowledge Graphs** + +We provide a rich, structured QA dataset derived from Wikipedia knowledge graphs, designed to support supervised fine-tuning and reinforcement learning of search-augmented agents. + +3. **Turn-Level Judge Guided SFT and Reinforcement Learning** + +In the supervised fine-tuning (SFT) stage, a **turn-level judge** identifies which interaction turns should be used for training. +In the reinforcement learning (RL) stage, a **turn-level reward** provides feedback to refine the model's multi-turn search and tool invocation behavior. + + + +# Benchmark Results + +| Model | GAIA | BrowseComp | BrowseComp-zh | HLE | Seal-0 | Xbench-Deepsearch | Xbench-Deepsearch-10 | DSQA | +|------|------|------------|---------------|-----|--------|-------------------|----------------------|------| +| DeepSeek-V3.2 | 0.635 | 0.676 | 0.65 | 0.408 | 0.385 | 0.71 | | / | +| MiniMax-M2 | 0.757 | 0.44 | 0.485 | 0.318 | / | 0.72 | | / | +| GLM-4.6 | 0.719 | 0.451 | 0.495 | 0.304 | / | 0.7 | | / | +| MiroThinker 8B | 0.664 | 0.311 | 0.402 | 0.215 | 0.404 | 0.606 | | / | +| AgentCPM-Explore 4B | 0.639 | 0.25 | 0.29 | 0.191 | 0.4 | 0.7 | / | / | +| **Ours**| +| **ToolMind-Web-3B(w Synthetic QA only)** | 0.583 | 0.144 | 0.301 | 0.224 | 0.36 | 0.76 | 0.3 | 0.308 | +| **ToolMind-Web-3B** | 0.670 | 0.174 | 0.308 | 0.248 | 0.477 | 0.751 | 0.37 | 0.458 | + +# Limitations + +While we place great emphasis on the safety of the model during the training process, striving to ensure that its outputs align with ethical and legal requirements, it may not completely avoid generating unexpected outputs due to the model's size and probabilistic nature. These outputs may include harmful content such as bias or discrimination. Please don't propagate such content. We do not assume any responsibility for the consequences resulting from the dissemination of inappropriate information. +
+ +# Citation +If you find our model useful or want to use it in your projects, please cite this project. +
+ +# Contact +If you have any questions, please raise an issue or contact us at nanbeige@126.com. +
\ No newline at end of file diff --git a/added_tokens.json b/added_tokens.json new file mode 100644 index 0000000..8805f6d --- /dev/null +++ b/added_tokens.json @@ -0,0 +1,9 @@ +{ + "": 166104, + "": 166106, + "": 166103, + "": 166105, + "<|endoftext|>": 166102, + "<|im_end|>": 166101, + "<|im_start|>": 166100 +} diff --git a/config.json b/config.json new file mode 100644 index 0000000..33f7bd0 --- /dev/null +++ b/config.json @@ -0,0 +1,27 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 1, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 2560, + "initializer_range": 0.02, + "intermediate_size": 10496, + "max_position_embeddings": 196608, + "model_type": "llama", + "num_attention_heads": 20, + "num_hidden_layers": 32, + "num_key_value_heads": 4, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 70000000, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.38.2", + "use_cache": true, + "vocab_size": 166144 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000..9fa0eeb --- /dev/null +++ b/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "bos_token_id": 1, + "do_sample": true, + "eos_token_id": 2, + "max_length": 4096, + "pad_token_id": 0, + "temperature": 0.9, + "top_p": 0.6, + "transformers_version": "4.37.0" +} diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000..c65778a --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,730 @@ +{ + "metadata": { + "total_size": 228983265280 + }, + "weight_map": { + "lm_head.weight": "model-00048-of-00048.safetensors", + "model.embed_tokens.weight": "model-00001-of-00048.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00048.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00002-of-00048.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00048.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00002-of-00048.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00048.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00048.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00048.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00048.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00048.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00048.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00048.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00048.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00048.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00048.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00048.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00048.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00048.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00048.safetensors", + "model.layers.10.input_layernorm.weight": "model-00008-of-00048.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00008-of-00048.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.11.input_layernorm.weight": "model-00008-of-00048.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00008-of-00048.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.12.input_layernorm.weight": "model-00009-of-00048.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00009-of-00048.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00008-of-00048.safetensors", + "model.layers.13.input_layernorm.weight": "model-00009-of-00048.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00009-of-00048.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.14.input_layernorm.weight": "model-00010-of-00048.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00010-of-00048.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00010-of-00048.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00010-of-00048.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00010-of-00048.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00009-of-00048.safetensors", + "model.layers.15.input_layernorm.weight": "model-00010-of-00048.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00010-of-00048.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00010-of-00048.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00010-of-00048.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00010-of-00048.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00010-of-00048.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00010-of-00048.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00010-of-00048.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00010-of-00048.safetensors", + "model.layers.16.input_layernorm.weight": "model-00011-of-00048.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00011-of-00048.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.17.input_layernorm.weight": "model-00012-of-00048.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00012-of-00048.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00011-of-00048.safetensors", + "model.layers.18.input_layernorm.weight": "model-00012-of-00048.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00012-of-00048.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.19.input_layernorm.weight": "model-00013-of-00048.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00013-of-00048.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00012-of-00048.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00048.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00048.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.20.input_layernorm.weight": "model-00013-of-00048.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00013-of-00048.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.21.input_layernorm.weight": "model-00014-of-00048.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00014-of-00048.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00014-of-00048.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00014-of-00048.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00014-of-00048.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00013-of-00048.safetensors", + "model.layers.22.input_layernorm.weight": "model-00014-of-00048.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00014-of-00048.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00014-of-00048.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00014-of-00048.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00014-of-00048.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00014-of-00048.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00014-of-00048.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00014-of-00048.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00014-of-00048.safetensors", + "model.layers.23.input_layernorm.weight": "model-00015-of-00048.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00015-of-00048.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.24.input_layernorm.weight": "model-00016-of-00048.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00016-of-00048.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00015-of-00048.safetensors", + "model.layers.25.input_layernorm.weight": "model-00016-of-00048.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00016-of-00048.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.26.input_layernorm.weight": "model-00017-of-00048.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00017-of-00048.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00016-of-00048.safetensors", + "model.layers.27.input_layernorm.weight": "model-00017-of-00048.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00017-of-00048.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.28.input_layernorm.weight": "model-00018-of-00048.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00018-of-00048.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00018-of-00048.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00018-of-00048.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00018-of-00048.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00017-of-00048.safetensors", + "model.layers.29.input_layernorm.weight": "model-00018-of-00048.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00018-of-00048.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00018-of-00048.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00018-of-00048.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00018-of-00048.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00018-of-00048.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00018-of-00048.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00018-of-00048.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00018-of-00048.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00048.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00048.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00048.safetensors", + "model.layers.30.input_layernorm.weight": "model-00019-of-00048.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00019-of-00048.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.31.input_layernorm.weight": "model-00020-of-00048.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00020-of-00048.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00019-of-00048.safetensors", + "model.layers.32.input_layernorm.weight": "model-00020-of-00048.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00020-of-00048.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.33.input_layernorm.weight": "model-00021-of-00048.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00021-of-00048.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00020-of-00048.safetensors", + "model.layers.34.input_layernorm.weight": "model-00021-of-00048.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00021-of-00048.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.35.input_layernorm.weight": "model-00022-of-00048.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00022-of-00048.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00022-of-00048.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00022-of-00048.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00022-of-00048.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00021-of-00048.safetensors", + "model.layers.36.input_layernorm.weight": "model-00022-of-00048.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00022-of-00048.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00022-of-00048.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00022-of-00048.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00022-of-00048.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00022-of-00048.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00022-of-00048.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00022-of-00048.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00022-of-00048.safetensors", + "model.layers.37.input_layernorm.weight": "model-00023-of-00048.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00023-of-00048.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.38.input_layernorm.weight": "model-00024-of-00048.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00024-of-00048.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00023-of-00048.safetensors", + "model.layers.39.input_layernorm.weight": "model-00024-of-00048.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00024-of-00048.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.4.input_layernorm.weight": "model-00004-of-00048.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00048.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.40.input_layernorm.weight": "model-00025-of-00048.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00025-of-00048.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00024-of-00048.safetensors", + "model.layers.41.input_layernorm.weight": "model-00025-of-00048.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00025-of-00048.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.42.input_layernorm.weight": "model-00026-of-00048.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00026-of-00048.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00026-of-00048.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00026-of-00048.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00026-of-00048.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00025-of-00048.safetensors", + "model.layers.43.input_layernorm.weight": "model-00026-of-00048.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00026-of-00048.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00026-of-00048.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00026-of-00048.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00026-of-00048.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00026-of-00048.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00026-of-00048.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00026-of-00048.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00026-of-00048.safetensors", + "model.layers.44.input_layernorm.weight": "model-00027-of-00048.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00027-of-00048.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.45.input_layernorm.weight": "model-00028-of-00048.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00028-of-00048.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00027-of-00048.safetensors", + "model.layers.46.input_layernorm.weight": "model-00028-of-00048.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00028-of-00048.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.47.input_layernorm.weight": "model-00029-of-00048.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00029-of-00048.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00028-of-00048.safetensors", + "model.layers.48.input_layernorm.weight": "model-00029-of-00048.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00029-of-00048.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.49.input_layernorm.weight": "model-00030-of-00048.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00030-of-00048.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00030-of-00048.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00030-of-00048.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00030-of-00048.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00029-of-00048.safetensors", + "model.layers.5.input_layernorm.weight": "model-00005-of-00048.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00048.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00048.safetensors", + "model.layers.50.input_layernorm.weight": "model-00030-of-00048.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00030-of-00048.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00030-of-00048.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00030-of-00048.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00030-of-00048.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00030-of-00048.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00030-of-00048.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00030-of-00048.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00030-of-00048.safetensors", + "model.layers.51.input_layernorm.weight": "model-00031-of-00048.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00031-of-00048.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.52.input_layernorm.weight": "model-00032-of-00048.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00032-of-00048.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00031-of-00048.safetensors", + "model.layers.53.input_layernorm.weight": "model-00032-of-00048.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00032-of-00048.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.54.input_layernorm.weight": "model-00033-of-00048.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00033-of-00048.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00032-of-00048.safetensors", + "model.layers.55.input_layernorm.weight": "model-00033-of-00048.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00033-of-00048.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.56.input_layernorm.weight": "model-00034-of-00048.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00034-of-00048.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00034-of-00048.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00034-of-00048.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00034-of-00048.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00033-of-00048.safetensors", + "model.layers.57.input_layernorm.weight": "model-00034-of-00048.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00034-of-00048.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00034-of-00048.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00034-of-00048.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00034-of-00048.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00034-of-00048.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00034-of-00048.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00034-of-00048.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00034-of-00048.safetensors", + "model.layers.58.input_layernorm.weight": "model-00035-of-00048.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00035-of-00048.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.59.input_layernorm.weight": "model-00036-of-00048.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00036-of-00048.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00035-of-00048.safetensors", + "model.layers.6.input_layernorm.weight": "model-00005-of-00048.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00005-of-00048.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.60.input_layernorm.weight": "model-00036-of-00048.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00036-of-00048.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.61.input_layernorm.weight": "model-00037-of-00048.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00037-of-00048.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00036-of-00048.safetensors", + "model.layers.62.input_layernorm.weight": "model-00037-of-00048.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00037-of-00048.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.63.input_layernorm.weight": "model-00038-of-00048.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00038-of-00048.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00038-of-00048.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00038-of-00048.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00038-of-00048.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00037-of-00048.safetensors", + "model.layers.64.input_layernorm.weight": "model-00038-of-00048.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00038-of-00048.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00038-of-00048.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00038-of-00048.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00038-of-00048.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00038-of-00048.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00038-of-00048.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00038-of-00048.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00038-of-00048.safetensors", + "model.layers.65.input_layernorm.weight": "model-00039-of-00048.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00039-of-00048.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.66.input_layernorm.weight": "model-00040-of-00048.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00040-of-00048.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00039-of-00048.safetensors", + "model.layers.67.input_layernorm.weight": "model-00040-of-00048.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00040-of-00048.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.68.input_layernorm.weight": "model-00041-of-00048.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00041-of-00048.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00040-of-00048.safetensors", + "model.layers.69.input_layernorm.weight": "model-00041-of-00048.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00041-of-00048.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.7.input_layernorm.weight": "model-00006-of-00048.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00006-of-00048.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00048.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00006-of-00048.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00006-of-00048.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00005-of-00048.safetensors", + "model.layers.70.input_layernorm.weight": "model-00042-of-00048.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00042-of-00048.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00042-of-00048.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00042-of-00048.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00042-of-00048.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00041-of-00048.safetensors", + "model.layers.71.input_layernorm.weight": "model-00042-of-00048.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00042-of-00048.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00042-of-00048.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00042-of-00048.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00042-of-00048.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00042-of-00048.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00042-of-00048.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00042-of-00048.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00042-of-00048.safetensors", + "model.layers.72.input_layernorm.weight": "model-00043-of-00048.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00043-of-00048.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.73.input_layernorm.weight": "model-00044-of-00048.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00044-of-00048.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00043-of-00048.safetensors", + "model.layers.74.input_layernorm.weight": "model-00044-of-00048.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00044-of-00048.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.75.input_layernorm.weight": "model-00045-of-00048.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00045-of-00048.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00044-of-00048.safetensors", + "model.layers.76.input_layernorm.weight": "model-00045-of-00048.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00045-of-00048.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.77.input_layernorm.weight": "model-00046-of-00048.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00046-of-00048.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00046-of-00048.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00046-of-00048.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00046-of-00048.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00045-of-00048.safetensors", + "model.layers.78.input_layernorm.weight": "model-00046-of-00048.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00046-of-00048.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00046-of-00048.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00046-of-00048.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00046-of-00048.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00046-of-00048.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00046-of-00048.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00046-of-00048.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00046-of-00048.safetensors", + "model.layers.79.input_layernorm.weight": "model-00047-of-00048.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00047-of-00048.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00047-of-00048.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00047-of-00048.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00047-of-00048.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00047-of-00048.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00047-of-00048.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00047-of-00048.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00047-of-00048.safetensors", + "model.layers.8.input_layernorm.weight": "model-00006-of-00048.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00006-of-00048.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00006-of-00048.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00006-of-00048.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00006-of-00048.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00006-of-00048.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00006-of-00048.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00006-of-00048.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00006-of-00048.safetensors", + "model.layers.9.input_layernorm.weight": "model-00007-of-00048.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00007-of-00048.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00048.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00048.safetensors", + "model.norm.weight": "model-00047-of-00048.safetensors" + } +} diff --git a/nbg_search_performance.png b/nbg_search_performance.png new file mode 100644 index 0000000000000000000000000000000000000000..fbef14dfd421cce944b720267f2e1c01c0f7eae0 GIT binary patch literal 36029 zcmeFZc{rDCyEc4Dr9tK)DMT4cid0gD3Plo{G{_th5;A2dl^-Q3MT5u?WlE+>revsO z44KDFg=C)J{p5L`Z>{yNwcdBzzCXV2-L|*w*`E8ph2M3Z=W!m#zVF9=oZgxm`&TSm zwTz;u6{-i6v?+?di=yc6E?JDP$Q@EMz&~W|ly&TmSf8_VFt}tw9Wt=Hc){B4g1Mo< zWs^&`=GInQB_y{<$cPJ=+1XvRmEFAA@?YN}VSUMTbIxh|V*Hk+7Z04Yr6^|nMrH%L zXoV_4muns>K6ZT=`ze-ykX?*7D1iY`umNl&0Z@RegAPVOn6ng z;SMdM{tSKYGBj!E&_yPRkM_U+z*Z6 z|7rzzX*Efv`?NTQ{nFE2oE^lJcV8LP?0nl(75Ai3S5U5Cl#%S5$x-pr;R6lt?d_K> z?+uC9O@H+(_H2f1vAy$f&yCPffvJ%?>uv|?rGQ@Ml_krTeH`hnJ94Gz&Ye3p&khO+ z2^p6>TeDs*Kr#Jt@9PUSQrBEU+W<&J-v>8@S7 zrf+WU{lRIlH6|fbEuk-J8Q!e@+Y|1cwjJsNu9IV90-~ZzDq~JCP-HpCGUR$OxBU8b z>+^k)+#f$yx=sx3xw>G{Yq#kMj(+;c58tnpu<&f??JS9MS~)n#KrQ1Er(rFh%DEi3 z+~qPNj!)DKS6T5rT1%*E(Hr~?c~t$JYazY4pS;=aC8SRzoYigE9b|+*$3-E~3%g@$JcvBvWQthrZibo^bq_sAect6?b~HW_pue z=4$FZXO=QBFkn3vdiCnF40j%MMV^)=Q_;! zx8vJ#fr-;UF8%!S+;wtfsGJRL?ywr?c)k!(enSW=^E;L7AF{aq!tzz7k5|Yspr+y@a%HvyYB71w7#KXV6@3S z${~sRAVavyKA$aIvy7k<9-)kRxd zdWw->es{6G##5QA$Ky`rR8?`EJ9kd>Xf(_F%e_9lJFGsXH+>D<9Bx!_=XFw&=c7kF zzkmNOO|?87T(|$~4FI4_18rvYZzZu|-r^_UWS5n(A_v^sdKYQPThb`M`By^L z@vP^s&b*Vt*4@HOf2v7qNOuS~$h}$kXsd|0xVYvAM;peKTPQz1IiHk1#LsK{U&oWJ zoA@FR1|Dxx*k)06d$6ek=PmJ>?Ft;}!wS>auZ;iA`SC-hxKnRyxywzOw%5qXB1B1c zdAU?Z`aY$_?PcNnbv`&sQu@u=SM|;`AqbFku7!siJ^i|E)r#W@XMK@a+PwJPJZ{`r z)R^Jo_3g=yD4md|OxIhMb*azy`|)*^g)?|`4eY&g<%;(2ZKA$QyW{H8tPdin7kNlq zH=en(dZ!>47uO{Op<^1)4y&isU4)_9_}`x<1$Wk5>Ti_W={yvDH0Jmrst~8w$da3b zW8KKeh@qjO;C8aG+ zOxMYt4;p4>(E;KowJ%s&nzZHnUK;E;v~ERP}0!>#2 zeOJplA4T;OJ{Ty$X<=cJn3D1U|Johv-t%3IPGI;~bs!e!Ylh2cJI(?5c<(UL#0d^o z*0p#XDVLE8!_z<8+t;BUJ;b8;`1$#@7YCnyZ`U1W6@Iv%M{wVR(%m0iM&loQIrKOB zrkt4^3=GiQnjB(e-}Bwub|-Z*<@~X*(Su4u>ybb>C_^hN34ee8&y7uF3+n67JX3I& z5EMiXV577%oQF4C#e6Hp6S-f`z}l-eesHRsdH?q95^8*Avj2XAyBa^rdOc3)cYe){ z8#g}M?ey2Qx_ZWtUf?+E@We2AKWv!$tXr^&@2YKVl$2#{vd-j%hW9DdU7oGyKAn2u z8uv`jr6)PdabPc!bzXV7_Q%{@`-hf|A5KySZY<+mCn&hrvPa7(5b` z_1nqOu}i!z!zGPkIdP);?cp!n>xQ_?8JPK^U3HSq`PRB&k0hm}*iN55y>#i)vYPy+ zCOr;ie>`3xmhJlpj>4_T$jG2ak3LUNj@qX#UAF8V&OZ%bTex^}kN9MLmNoX3LsV0d zq6G#9)@CBJ(g~m}_}Qx?{C$pvTGzkTq7J;f^wUn_3yt=>Hsv5i^YrvIlU|iF-Q+&Y z>EhzDbougI!>i<+hghgfef95jz9zd)CYJ~bhVS=#sOL7WtTfawx6e2lFjnhaRpdc# zIsqIou{R1kcI@abeeO9i8dT#r+*4U%^`vw`1}dJ^m9dM%{k2VTNIUI6zXpEpQHW%2 zY>n@n80}AJ?BQFr%7^igb+7yd$;?)KxQMdQpDis)19_qs2fX`FVDrXwSRc%vGh`^G|gHs=<(wo@vr_u z46=4z9!&N?mFStB1t9S&JQn+e|e4gq=-2_hyfn7-`DsfVIk3$oqA<3$R{N6t(+ z+4VQ-T3S@cuZfL|W98#xVrKR;&RQHP$JOPmp~dl|u8y8kQrh$|D+1NgCKb=dLy;#P zKKT=&7gP7?SUWNiqvFt@_`tW}=oL?blEcUHuWXkA77X01x0pKmQhR-|>o>sEMn$os z(VDRfHpN$~u3&4cu~OEeC@n258M`hPkJ;khsN~Mg9m%U|ZJz~M8cvjFB_t$lmUo+4 zLPeiQ-~%+C8Xqi6G}^y<*JVbOWJ>exMT4oaHou+`%Y}hV>qyTW_prd@o7oCg{#%s(%Vzd)_#dbai({7pLLv#$kTuHVB?89gDKv$ z@`@Z!26nrFi-1Z5BqD_;RK#!kQBA;N;$t>LY?iFD0;O~6R2ULDrQIP&MapR;z361( z+f7@+yIsHM`^%M*h6|Zi%XpZ4efZq8XY|W>R?7SpSr~`^;(s0|P27%Xhn!kL$OMf`SV?tggP2V;0(f=eF%CMoayI z6W`ssn|AEpy?XQ64~oxrJ88bVWD={J&gHpy#YcdtMAVocwY6b@Q(ZMlQN2lp6SE($pF0;^jfxVvxq|n_Gp~~2T@vQw&(?;SfePFn z{223a#8h;f$>)3Pi{(Fk&{0zgxR;!zm*F-o+h~9Od-U_O;=A#E&JI0MK%G|s1qvo~ zUm*ln$~j9vlXqQaZEc-w)gY_tI{l}Ge)sO(TdW&D=zMh^>bjbDi#xZeiJx2IjK}@^ zNA8{VEN575CpXgZjo0p1)rzgA#Y&@n4K{z4 zHpaikI&}SF?jBs>?Gzju8d`EE^qG#En_JH;c!6BiU#!7%?MSr&EB#@6MWCuT8#k{ccF-Ms_a%agw~)tQdT{HDzYJGo++A-Q3bL zZpJ;!ZTip5w2OMM?6xY=(E(j`n%K4`1@qxX6o#GgvL4^HWGN}O3q zv2byX)L5%)?=9+#3z2hq7;%enX^QjkdBSk>MxuiK8gp{!)*pPhkY)PJ`^d=kA3uK7 zH!$#T9z1~)SpU|BURYRIRkl-aVl?ITYufO)#CPv@l0=JS-rQT4M&PuHAD=H`BlKj_ zITAd@S$+>59J@9#(wev2{#B}V=D2WK!Qb_BF_g^7!?TptdoZwhGYSyWaEooXQh?#9 z)C=BwJQuG;aCn$j9&yHKQVLd2LOIED#wN1yb&8e*1R+F}{uBPUt=Mz`i>a?l| z)%ygZXT0|vK$`C9{d*)z<5-=quWw6F&oY6Tws$V0606s&Ni?qzPC4XK+*T(nEzJS4 z)RQ$Dl{wV^tF<+(@!EUa&SISt8gX%Pk{5n#0sme^U2OTdh*~;13dngQJUm?XKz(2% zdx*=(?zK`qfZ&<2$TO;$uK^;nbUY7&3hqZ?EJNv|vTyTjy_f6O4#8L zc?bkde>+POl43!@%)UpnQq91=kSq7whW$e$| z;?w#wqgk^CKw%dzUGf>^`Sl6lLfWHi*MemFN9KzbEi%Zz%~Ke>E3L8q@f85dE29k# zG#{UjAUS2p=0}2m3aQ-o@b_m0^XBQ?dJtz5scXmgy?oSSX1?1f>xO{-Z(T;MVlB^= zJX@ilpdgLqE0)h2?w{;y*tMwmnybjyw%1y6QZm&TmfjPGe1Xe(I4#B{vc!Nz9K zlkJwTEvhAe_F4%CERR@9$~6?H4 zyn&Ot`10k;JxBFI`gQ@@cFHJJ`)zg_XuYAKkTUHyH6}?u4P-qONh1$$2{U zLJok24dS11jVuRFaV`Sd@KEH7N}Myh-#=Kvz?Ag$1xUU;_w8{D*cVE_bu9A@KxI&vi8%&=({W9Qmb}Y-A1Vt*s_HM0kjYK{vz-jiKKb_DyB3sg%44*@NeBzr-;}lg=uwVqo}L3JMZqSL zvzv^{1Btj-K6B>EW3So}txdfpRz-WZriAE1PIMgYTeD!n0y~M_ix)2ZG%3=d{H?iJ zf8oM~(G{!BM<3j^^rmV1Ad3=Ap%E-=fLMq=62-Lr!ViyM)d_8Q#Xk@Nk@ZVM_s|2t zt?B6K=&FnfoJcRi#*+j00q@mFcM0kI1y@O$Rl_x~fQj_aQCZC9gX36Fq%dVoO%@U= zT$@e_VCPM@bOhlA$PZK5&-Z!#>Cc+&G2YRgIP};f*l^>Sx38(v@vT<%GQsxXYit&4 zsC(FFow_yhS9H$k>aGV&C3Fm$=Nn*mM?D2;>7&AjpNTUuG2Mgc^7)xuI)oRkb45XG z`x`SVPNw0-5+^jHG}f@Pf(M^@>N@ou#A6eV^g{#%VV}@1Re&&r0a{8=ZjN^~FZtRz zF|eL=TlTR28C5&U!siXDv^~PYoY^zpHIrrS9_e<~463ksFT`efyS| zlS8n$;3{7g(&Z(>O0x3u^1jrL^eGD7bpmCv^-A2~Sy#0H(X|Na z``D_2bho%uDW5JN)7B?keskRrDL#2m zjQ~1ehgDdETb0G8W!~g37y;CfJINop?9=sgU!&26sDJa-1;O+ElmpDK6-DO1wVYnO z^#7;ds$Gx$Eqi%HMAYY+%V=MugTwsOOD()V)JLs9?C2;1$urTYfGOz7lU$Sq&;@0v za|-FYSeSu7E&HwvSDujV1#~A`b#aJ-9}e^aiim#@tv{VggLM4o&!1BBx-54ELRJnQ z{K_FCJGU@lMV^bC7gwW(BWYw;S93#v{PXwk-d|O5i6})-$flC&Bp_|u|Ej7qf3nk7 z7)A3XS7*rS9TMc1qo5_CeBcw-CwlFhYe)r{4{(h#{)X`OBh``@k8WUOL{EWEEm|dQ z=5dvtVXsna!)1u^qDaZHBVdmN45;>1ARD!`w+FQM&To{|My@WN9}hR3rl|P%c-db3 z-&@;Grm3l^)`nD~JKh)Z)2PD_ktHDfO||d7_fqRG0&_2>whTWVl$E6bHd<+2lH$X? znTdMqIA9nyl4kswjpTd8{H>It++2ESLOIoX?!H(yL((#VbXC6V@O*bocFf*%@d%2m z&gs(y=c`kpev4~O)+QP;AnkBkJbLhelSBDxU?4X>$7Y~a899Or-I1uNyZG2ZJ z)Ks6S>XRduv6n5M#Kl=&8xa!Jd2c5+G1B{_xf&%>mcpj($f*PJ#~-aJ@*~Kc_>oO# zKR72F)&bYsTW)$VCa{q_be5+AwHTm)0+7x;g zUHyXgwl+`0JfG*#1E{cDze({<4*_})1(YpBCXO{>=G(2<^6>^MCud|u{F=3yY7aJi zmR&r5;C%8I$kE4~PLn`VZSL#4(^(QiPhlNBA${uOpPC$tsL}*Ig&#FFE$yH0CyorX ze`f#Pm@%1T8rxs|Qh}k;TA3Pd(mlxeRHaq#vxOm_ze?I4o9SWmWChm|t#Ilgl6S z{e}N;K(H(_GB642J? zII=(9*@`^gVLel02`Stcv8!JluS+@uXiJo4xw{i~3wYh*4A~+d9YecR!jw_f{6J6F z=1oAAlkss`oHL(M@AcqNM`KR`d9IK2&Q3PO*tF%7q=T(kq^#9iUsY1V3=$&a(6<`W zb403w%LpTi-G0LhgEb2trlREibTWmi=l-&9N|fZv(uV@v70FIXE^P6(y@#IJRQGfuFY#@c{TjFsQz% zDTA_BX_)eIyRK5YTU?ty0#~&)Sm7WN0J>^4NVLt6yitI9OG}IB@z|9N(t?5+CPqe{ zXdC4tDb>{0=BLLY6;bHiXhjk2Pw*PJU*5Y0v=jX|o!%~e{=29$S!3cA@Xt~0sed0R zp}Ti>X4=5o+7DPS9Hb7tB%&Y6r#FPS&v*m%=BG~q?~Flk)O?}NBO@adibPMu6{zHa zT}DEJYRk|-{@^+Zf#5@_c^1+$6$&wPz{{X~U1lX-;)(hBYN?W0lL zJ_oVr9VH=ti&w1r^ll2@60Mgd&nM@69k1L{9(nKw(9lB(U7rxQ39*_2s75R#r%_{L zV_nJUE{EzQ8b+g&MG0)!uneWA@?;dyJ;n0N&_q#@LSy^EhGT{qjW`m`ot@m)#p{z5 zlpd`h_(Dn)$0!f|A!M}pJ52T1GHJ6?3Mp->m6i*Ju8hE-qaaS?VPF5WMOm0P_=)-D@)c>vntIzBO0t&f#)r@5`HtdQIa z20KgWh;o~QpWj0I9<0G=d*>7vGa<Lj*gC@crDRW z$!4+Dr*SeeC;QSJt&bfxbEir15`ToT=Dn?Hec0;e_V$R1<3O)eHj#+p+{gYHJcT~K z?M@=TR%EO+_o9&(&7-)UI;B;^vDaw_x8VTLSiw1oRE076APdNHo?u{b?|^YO*@yy zDikWZ3@jqJQP*49f3xcYMd#=LAd}9E}Dn_oT2MJeooFCU<0lAH3aIQCXiJF#b*KL z|M>C7^7{_i$o9^nIHG{EkV+XyeK~jTCSEoSTI|wVlC{Y}MQ?2Y?H!$yC!d2EuavnM zTI}q{uaN)Nrj7hg+7f6K^dcf6X#6qUSk8SXr3v5m9ES)HFB@Q>G{vGiK&GA}*P5J$ z(1@Y>_vl9{p;(pT`04Z_+vPzMnXKt0%FoZAix?OfNZ)vTuygugFK?LU# zBtrb94Kio4dBtKBw%lurmXF1*2P5O?b=(4x!N+!Y5Lk8IWKCyhr>t8xz=oWBG>UKHDLSU?baZq8T+d|ho_Qo?vc=>Jhxn=F zC1{PSYr11Aq#Qc@;*-yKIInvUXz;ObC^Wg%^dADOPZ&R56# zN|}7gMh6^8u9BxS4_ICBT0-W;GQ`0?Gzc}{7^?!+N>=UYrM6g9i9;qL-FY#)fR~=c zHO3RrUNpR%I(znP&yG$&8p;DknBYB@q*kGypN-DobhvK(v$-ByKw{E-*C}>a=B)L-7F1k4G_tmRcDGzKzE}jy~dNz7i z#w@EH?m(1rx)+65XbqFQEB3ExbDs zZ9QgSF=7XK{`@)R0Vrzt`$u9S0s{f~ut-Q4sd+VOBauU(3()P_wF_hldPFGV8?v## zbIvXtW=a!XU%TE~P8AiELNxKX#C1G~LWWLDk30g*VW_8aIfTZ8fktYK(;4H>L^;>n z*LvYIT0t|$dSVjnYI8q+ya;|kk|adqqjdlv-uTNlG!#&sl(2Yz3i#axa42N!ETmUq zap;j*kB4~k$Xpb7@6i8Dia|Z#GZarIq?U^|Hdmp}s2@7SE?sRMBJYY*cNOYfL9OYT zBPmoi4%2H0yu)glNMz!hHdTJi0>+kcosbqj^!Pe3WbXIx9Ozp|pGw(4G#xl}BAWdH z<|7pVc?c;x?y@)EGx5!vkKmY<=|7U6K7C5mv8THnn2?Wd63FoBR04vGEN1aDIk;j* z2Y@|4In|}h19Uvn?ko;wWnrQ8^z?rG_>ukf>vHrN@_`D<(Csd3iJi6H%5O zMUQmeU5eDOaYO2P`~JNlnC8kY`V`cRQbC(Bzi-Xa?&qE(t6Ofr1)4cK(w!=ADTtI;G35)hjt9o^8`OZu2w3J3D$^(CY z9+cU&i=LJ#;0{(7Y-99rvtGOe%3k%8*440sKGMAO!j*JIWSj3-VSWC|i_h0Vi@t~= zsgr;TzzBcBg~q9-W0s&gO3KQ4H6cKehRB@WyA%m`3$qJ&I-rj;=Lhg=2%f%V=~6lh zW;!lDKE5vW%Au<+6PU1iheSt@7RpBi104n1%cr|*_uU27yp3g9;V1L+wPKXI$j%D+ zPo#m@^&?T}jmg&+sk-m&mI{XT9r;>VxCDz;=Cs1y-95PauoX$fG+KEbj$7#)lY?*H zzRgCv2x7;fQpYU>QlYs?v~-f#)E`N8RmJg=Q-D(V>w`+kGr39|Q-Xxp)~hvxXpuvL z|Ky1|3>=S6hY;uioxlW_0{|8nfA|< zR1Q#nso-|?NtXLGeega2XS=5fthON`nHojaw0BnU9^I?w4=XCv+%L9RMq3gD{u~@} zVih)acAtR!Q3{p^k6b6e;9urw^>~;Ke-22N9 zozNpM?mso>H#`OOM|2=~IzF9xC~}w&bZ1wU1v&m}*RDO!Pc^6aXL=?k+B}_3OifjR zD9-NSG5_>y6aTS&W^zt{7Xppk**5B=PT&PvKZzGx_EL&EIxERh#MuQyaMP^bc=m$~ z>7ZoIOe}}(W)VfQq(xP14)j-jh#`gr^ZMV!tr9$7MDU#)*a@LHZ`ro(P-!yrDU=~4 z6kW|%M_)SXg+q7vGcb@ta)o}*0woodFtE%qpk{*$7h;q8aJ=Y{P4f5Tu1TU9kJK zCrxhKy45oy!!>q!SCb16IKM-~l2A{|af*fH@i=`NJyW^q~$o zl?kUc2kFya&gC+!RZM_a*TI@$6uASgAN>niYVm1OK(km!rIi_U*;-cKy)h^=@ zr6?GP9%!UPA9P+P_=R!aF~8pbKhXzalD3bTK=Xj207(yD3LymyLKQ(V5NY=2-dvFg zWo2DIWa@IJIn_CA9XjYd5CDx*Eb1UTJ&c~-bmlD!`EI1u`}HDldrDi?uSCxH2v$k@ zx}>8AiA(jbyf;nM@{gMs30b+C$E!})&IKj?RE^E$$I%xjM&48at5f< z)sUX%%qxcl;CJ&R2P#16KZ=C5z=5Rq3OP9sU@JJxb)g6P4MO0%ro~TH2VB`%wZ0h4 zi6j9^|C3y(2sD}b)(;=o3@CFwEiJA1`*~GXxT;D^OLYjLpHmHEr}0~Q=7b46!hg%K zeN!?>+vwXnjlHU7wl;zdHx>hZ2q@`VtO&HpP0{;losau@$*}R07F%cPol~XeB03!?cN080bpM% z?$kE%vmbOU`^zJjqj5>v=j3<-yuMCN&BrU>0M(CbH^2gJ&{R??PiG%8sl05d(JvA4 zxFW7Lsy0($$Ls}_z`LERycyRd`O`$LZa(%aQkg~(4HTR!ToWTu(i$q%l1a+ z9`Xy+sC~gIOAds*yiTKuZl4b-B_$5OKYuR5S-cuL!A(z3`VDY8Bf*L|x-c^{e~&%M z4x+RdC90*Xiv~+ZS8NiBg0w{yBb3#P7cVLj5*xms8DkJzs%A7oGdOTz`5jfy{zH9O6W zE|e$2s@NISwt(pOfbiTg+FiM#%^W0J2aoM$Ony0PUhCu zo|+yx6XQ70N_pJk-khJ_i#Z3)WHT+G>NUvuDqzdaPz}Jk-A0}`u;xH3kaXxf1GD>C zQ}*0e`>(?Rvg1^$8<*MJ!cM zFzmYHze|~1{i$G&=+6ieH5JNC8(NSnx1M9X$+X4?YQU#=-%!|y2M^xk#z zg6Oru3jlv#kQuR&sh)KmI=9|O@0(5UxI`36Grr>Alw zBGnM|8+*Zp_llB0wC%%MT?`R`-D-@^}_0~NNVx7Qb(x@D-l;s=^_Q1DoPJuJbV7{aQBOn7i4k(G7*4PW0w zPn*;%>fp6*YpB8u2ZgRubVXl#`$E{V++6&(CzLth?=+TuE4my$ue6PMbTz!m3% z&mjyQ0I(<2wl=5hAL_Ew=T$F%;CDgY*^_9PCqKCg|CR&c#TRNY53y}gL>|Eb$;SaB z+IjFWiZi2#NR%-j535%BPFdMjh_MK!8yb_ryBsKrG}s+Z!vRcrsDzi9SIcjXx11dt z5R$(##?om)Pdipv9j!$V@blJna{J8ms2(|6kS=e4m#$vDn$5;8!f!sD-l%NYE+uu% zzbw!_QU)zJDwOzodu#PBqC4=iB0ks(sVgHHD1~ZnYYVUF`!~%ikLYO8-L4ZyO@lEV zK}SJnj;P=RrB&9_j8ubUjyh2WYNC@6=V~n4LD?e^6aVTckz!0NQvj&ma!Td%KuPqZ z5T{7M- zk1p$`{kK=rvF5-NcBEfP^Knb0IRZ|t>OkJopWt@X7-nHAtsg>Eqo?-s3kyyY|0A9) z_g_f+2Hcj2tLO0`g?yMN(-qkdpRo_gdt-lE;2qiz?G_OGHpJSNKs`6DYKz)rF|@`! zVCDbV(6C1E8Z2ONAS}wv%#xD+J52(E$Ebycs z6Q+FYPlH2~&5!}iS3~n45cS#y9RskCVyJG!)*ojFHhV2UTI?RQ^j{c|3PeIrz3x9ocepYk61s;XyAkYya zXQsjY$1tQM34so#dVz5?!<7QA4sbbO7Z=3YHm3VlP{L9M`0Oj~wVe}}v8y2Y<%mWVN=`gzyptxW&K3rcf?s9M)6oN@d~H->!gdpmOiIH2U&^x0x%FWTEHlg22dO|X6=^yXFNRj^WmVM8C$ zl4OeoY59?2!F|Qm)dqH(+q^rNa8jpYR{vT-!Fpzq%$oLU7tYElZ26==S{OKsnbx7K@DghGc>K;}iga-&kGq~z=e*R+e zBsjW$kUQpwC^10Qiqn?^HAFN1%uaiLPX=ZctcCW^)hs;A_POS>HQPQ!>?9I0tH#os zR{lSkBrZd90IRf){aAyMFf;-tL~u+c~~$ekbiMyVkYc=oF*PomiV<2`%TA zLtS>ZBE)8*B9aCSnRkSnoA~8`wJ>&lwOGE@B{(R7b>&JL)bdd4rc4qEkVkJIrQo1* zkqiqB?|y9Cdh$YQ7La3T+PbjQpjLaqg_n&c8jSw<)~?d$^iYehP zPDa3&mP65@JTL&Ikq(iFU`pf|dI}BxKSSL|g6%ODx`;x%=bnbkzq*?gb?nvAJ?Mb9 zUGWWa9V?foLt|m5|NaAHM*=dD7`3iJ)%~w}koHVN|4k24acEQ5zflS)I6Z7TigLhs zD)%&!rV#}0+;87jz;&ZF+6~p8bmhJTh{2M&n8FMd`(VeneHfh4LVxh6g2-Huh7t9iWTAUMf54~asVVgm9o{`Q!YfOKhdHcZNJ2Uim# z63obw*i16miCjY_Izi_WlAgg6jEM0Bydk5Npx{a{V-n{#+N+U$-g8zSwM9Oko?Z_c zI{xp|Y00;5!DX&}kF3LruGIud7v)Ddq5t`+cUQ*Gk`M)mLt;gmgOb~lShv()QSWWy zU9pY7LWtLC-tmAI(Eoa#pR9oRP1#b0?5lUl{vUAvmyQJ|-2YjgH}d>=ga0G^dH?rW z|C1FE?rykRq^88e#&)lHSmEZ4NJH!LYDm^cESsQm0A36|eB8N7vj5v{Vad*KU===L z5r)>H65gNppED}o!h8Q_%@}`&QEglQoign~)75jQV({D?+K8sn)>$wItV3E<2FNK- z=T#cwK^?i8l49@dAXH%>&(^JgpEcUf)}0klsKV9Ny)N;@o{BH{e682|dv-hi@!ojq2L0c@Ggt3KaYPy?CuNwZ)@XnTjRBfmpMEoT%3ztwF=;Z532!iA1^6wu+iH8e#C6q(8f@iGC zcuM9$N%J7z0M;%1_wUpM**)6HKfc0Qn2nPYMlpS3unqr3jdw2+`I9R_F?6JiBw zpoG>8v9gbik4M^W7E2lpuO&iOkwV8qJmH2yadEoddo&%&YvkE%+zYKc{kFWBYqEOPv2 zv@Y0AoH#)mc3}UXG4UXb$?|dt2be!UU*HCq)m`t87!BA zaNFL8NKXqBeS_v~iZ1Uf2?+^pXgE%jBXK#IE%yZl1QaoAu)!zn02-#1Bd=d?VP&m( zZQv>DBZyktU8$Tg40A>|X72D66wr#3BQV9hJe%$iI2;0R%80Y67i!4lYl0!UHlxKm`TK67@)1tJ~`EMA&v57f7-UmZzZ z=d0q0_pZ-!=ZB1^gi1|rzPM2TmX#uU6qOBtr(VL2fF4FV3f6jk2)VlzHM|sGV9?S0SH%vEAuDB zkYj;xlxoMu#)}3YWu@>p@!+ob#ND4 z0+*5*wWh)Bi|wbO<49)bl?-QJGb4gw9r%d>m_TS42hm<=$9#$)o2y&8{V!p}3LFJW zpZ=7bE*iA$Zt1T!&gNJgOr?P zNCU#g(~>i=&z-<2ML5OOu|#s5Gy01V@~b((vuIGVJIGfx_sZ06c~2_=!UK?54A7p& z%7!hu(sU<`Ny~w5*RQ!F z>`E@?f%0=54GvtK;D>*W{)+r1ASkLGvBYhno0$7ShbWr7HReT@Ff$WMKQLUaCj`m% z_3PIM685QXg}$*5I<*0^BPLMwfzOFn{*Rl15T``O0AYbgf;0=sT^;b9E}+0A{vfug zY#0vdD9kWZQq?Yg+RRCkCU|czyf@w4N%BT$dJKJ#RTS$4Thj*Fu1R-DpiS*iYln{f z>HGJi@K8p?$7D#^ca^S1QAF4aqk&aX(}Zjr3FWzGlgy(Wcy%J+nf?oQf$sr$e{pv0 z9UmXpnx4ehC8d!fCMuS{FE-FDgV_)E(!qrX*S)-iutk(643(t$38Cglb*02F43LB0c!I!@i7RxwZzLv462B_G z5{`8cMGU-O&CDZw&gcsg^y>Kp{a2Eg9aUSg`^r@h?NZDWRK|n|l5BquH5gWr2jnge zn7btym{zm+&n`B@bi|#Nyu4pw7~Ht=ci7ltotC|U*QZR8*gY~u@TLV@{^-Tuq?lf2%`i9!sJeezNf{Pam@oOJ`kdBIXZU3PuU~46RPza%D@TSNmUAItAIGaJXLKVl~1aT$%l}qAGw40`R$v0>|WJ$;q4h#++^pjxk z*5B^NGxsFM$i^Ya#;ZEoI$>m?UpuRCts_c$)e8KN3dI2>4Rou<4@;>nTet4R^neX! z5i#zi@NME%0)UPe&NCfFZWn+k_gwS&USePK7gW|aG`xmOZ7@;BjR8$qr5E8uF2)@# z_R^|GMr}j8uX9AnNyNQvGzH%-Ag)qyaL6v*A3DjbHqI4kux;7CUB%dV9VUmct>>Dd z*h2Tye!QB&Zp616`iqc~~f zo}t05xcCYNTvJmo5_`I5i5{lOvX7mU^Fj15BG-#3KA2lZ(~k5I=U()84LekaJwR}O zaC;1~QJ@u8T&y8CTPt>e@y3W;xLiC69Yy2&y)Hg<1rw7B+$^Z3NWKwFzt;+4UYImD z>6DnIDqvqjzg~V5lPBX6Bvna*Bo9o`#)B@Qm>{X{#N{s@v}T-U;z%xiYhy&jMMXQs z4L%ESwHcANP{_+_BsOeN({;l%PtoIT2UpZ%&&fS$fa_$n^z{YE96qY^X<{racJ6-nr!mSac_Mi%edbI~Qs@^*_rCcKY#Q-j+$ub^N*ZogWK2C&12 zI)HDI_ms3aNM68-96^y1Ml!yS4~(dhyxU^1{RPt;5LeZcj2(Sp5fv^ zQUZ2d{M#LoIIt)C9_Ba%#zDaLpIBZ&!ArRR?Q@7iR(nZ^9x$blAe#km9I!M91q-V` zS}Q5j$jCM~lhzU-h%uZ!nG-!5&8uP`!Gbq(tIn*%NK^AJ2E`N+E@I$2_u!L`yzKBf z8~fxp4xH>8VGW>`dqI6?Q9(e#0$pZYf1`|R$poQ|`vX*Ui=WPLP9bxt7;@TxHlVES z2^D2!ZJ?zO+$xneq(R?aZDvZ}ckB5vcfT!biQB>!SXt`jBb z>9dpOH8KmEh97)GhB&oz@*l?g=_oP@L+;Ze)@iJeC(LVPQUjB-u6+k4r*5NNgid}i+GSV)WivBT9ABG`BA#IIMiXVi-%r9TXp&hnFc%2?e=*x%-Us z-Yox9B^UMDvhBXNZ~v({tYCop90cbP)eMNQi~));%(a^O(%3u+Wh{s=*siSt6tW8C zK$5@WVao?wsg=E={*vyFDYGFPPe_m(28bw*;Y4a`$KAbn>@wUoh&zN9;gm08StEUu zV?oHEx#F4h%PWY17Pl7pgD7F_Z!yyUHQ)i_qXMsdjkw!F`)CSnf2g0q?I{EASOHYWP&sFCeh z_(Z$OzGtY;XiApHo!;>Xh?0)7?(mrYPfeUy$c3-lRJ#hnE%NHi1aA zf~C-XCUeR#4Y?F(`j8p%4ql0DIWu$U79b`Tx~!yD$nf+PyN%ZBYD%gl^{qwnE14eTRY84cQ{vdowL9#^k=>TT~$QRns16{ZUGm*g+IjteBktfC};TY zOSoSk#T8hUK(BWfTdvWR+_Tvuf!7cYAg2d-fDZ87-nm!Q-SOOtMf{iIV6!`b`Krn+ z*g*@?5&j2UsLe#a^ zBWexE*6Sc`6Kaby>^lW4tBo6xNV>_ELksPtx2t|z3 z$8TM78{^(&5I%Ey>!wn2`&TxI5t; zL%K+R&QwR^#`rW(F;Jn6L&Ej4};LSG}LIRt0bzZ&q6=^hzz&XL{|2Bc6 z>4Vb|_>2U>9`D0|cMEqBOkqAn_Cz@dOGtD%Jex1TCjzVdVy z<}=r!HF~JdGfdsHW+lYZzXX39$>;ml1YIfGKV7gu?uM(ZLTFZ27G_4Y$Wey1{&?NN z2OHMIzD4dNe0d__+3&iv!@3wO*un#+I!&ob9(Uo{2mj=Gq!R9rk~Q9&*7T{c@Bq{T z`pq5hFbnVo2Iwzv=7RsDQ2K-8;Pj9VRj_MqO^qrHQtbeTi6etz`_kpaUZ9(O3#(_C zw`_Pdsc5*}XOm#q*`0p3ZwCMfd1JKULT$3D(#)i$^UTycGHAf3;3h73bElU`B=fnp zKvZuApp-R(1A{@nFTKX2L!=6(mYEmATMqm|u1=-MJR+vw6_F)yA1>>h;6yHPg_xNO zUWdE#lwe2rKF-6byazS|ItsS7kFdESe}~)cd=~IRbQv5#uxh`?W1<61pEWzPi)wIM zFDECb1*eq#Nt_2d zt7goksNj3fg>|mrB)iWU^%%(aytGS~))$b^c=e*pECQMPZ{-O+Z#+=#44TUaacMwh zj2gP70jsv1C)wB-C8Ng~D-j63R|sBP-a(b?6vdp!;+4a_hU^JJNkp*hL>KN)6D9nJ zOi1YVIyFK$z(o8(q*?os)386|TQQ@w3B$gJJ%v>D26Vs4;c6~IN3o6sNVWF#Y``UR zTGrOLoNiTNJS+kZ9&-ByX!IRWS;4AMF(p+(%SYJ^i+EMHukXsMyU@18W`?w%PKoWK@HJ@LbJ3$K;luzlD ze}%9a2`pSrZUGtru~fzQdR|G%VI{kg6c~s}dL;Z38t`2>nGS+A0a|P_uN@uw(>-`t zQXx5cwh77=$(pfxS!dM(B{o53gX`G@C!7t};3Q2$6>A63`-=tVElN5TFf*^8U4(un z3Rnilwz|V8KX-xLBeJx>?04}TpK&j)@;YY^a)IwbaF$OPZ-#ee1LQ|S{>Z&JG$lPc z4p#DSPi1WOV)#nDv9YoH@@n(7=QPi|hr-+@0A|UitksTM)AN@ojbWlv@EZ(|bNA|V z0xmqnk$U-@&9*P^u%1Y`_vSA7HBXIUXV!*Yl_Xuvx}P#_fO6}DVMEAGUqI%Kfv(FT zsAGG{Z6px&?x9uv6>0Uzu#TuL9FizesZx5e zD$o-Ia6L0^oAN$b;RsJo9J%kGJA9;;*}zvN-1ZD`>>JFCy*!;3Psj;`5x!saYD&h~ z;hwIzQ}8|e&qJmGP)XkI8R!RO949kaP;EOnpwnW

fZrLfiZg%fz>wDybnDS5I*6pwjyt5-U~I6lw|@Y==4Htmp2rv<{Q^E^ zjEO4Q!3UXc)2E>BiXbiBLt2XI#g4F}%>ZKc{oNwOkOLuI=Hw>5VceB?06h)+Yk{JY z2s@}pkZsEyKm&5fxtiM=ZwTXRU?c0XBHGA&CGosl6)||b2`2K^)>c8Hvyj0%cxRqV z^w*^&pvs_CuNoKs>p>7dnLsBpHAbldAsVXT?SN~#8#B^z<;^~&u-q_USVI$&^|+0s zCl%5^QBrJdZSOqeHI__Q;(9hQ!P##mFQ>OW zmbHIX{4^9-DY1ic;0K7hkFp{ey)iE!b^iUN@qQ=f-%qw99R9JH+2pG)*CbxV8}{@n z3x*ALaC*BeEC%X308Kdpq6guTC*Rw>d_M&cE(J5ICx7>!iaLLCI}eF(NOUosBiJe_ zc#aXgSp2B?bX%0NLwq?Bf#VW z7#hegxeQxb$*%Yqp`N%(;vSj#pcMicnxLfswVtRD=Zc?-fk;<4t=YGGnTSO*?*O!RAMG2g=EBp>{&`7MGIq}K`RYYVaoQr z?w@n6=XtK@A9&7no$H)4*SYw8zu(X2zTfxze&6p~XWg6jE<@GzotnNhE|>n2(3h)@ zjg074-ti0dJte3f?{n?HOv+pR!=of4%PnHEqJ@sm#9fri^5;eO^c+i1#~en~`Qb*V zc`$)dVj)b^gbG{*gyok}2YdW9qNrI72(Pp*?ZS=ob8mIvPsXLQj0%LHmJUJD9t^!RIY4qocJ0M5l_6kGH8!y9@s0EDKOO z9r%{kco;l2I2V`sueBflabrNL7wy_dd(DiC4*by30)mintg#1&b95x12!vLt@E`$; zO)R17yl;(5Xi@a}%{rU(XQ+{BxlNmhXFkYhf5s;ym)SH=88%9O zvOF0ePetR7c*;hK#Kbp71R>6uJ7P z+ZY@Vq`Dax=a6K-^7mu2q<%9Ty8MX5VjyrC7b11_u;^7*=*^{l8p$KLwac(t6&GO| z<@{a!KjjlC3XaS1jNj>Qy86a?98E}qOPPj=uhA*n9eX-kyE!GUd}1b4tB+*B>Ip|g zI_-=a%CjQE7>uPk&WujA{N^U&3eKKKRn*jiIl6Z7O5 zFDOH5sfDGX0rRb^`}ezdu^dn*COlqHsa8*X4y<(ra)@DDEgiu4d*U*X0%YBC|NQ#5&FZB!j| zh=}ZZblRC@E_n5{^_)0d4fu;vi*1=Jv?J2Q+;EJxl@v|niRDStOy-5Eez9HqqjYNs zVy$%_*L5&3FhjUrgvgv%r`4`cFRoMdBCiJlN+nHNQvjxCLQZIdU94`-`Px-~J+hW- zN9>Xr+$pbl1=U}&)_$18yaly(9*^ztB5U3OnwipO2RQ|tI(htfBq-=`;X++A{S2!z zbzffHT=Y@LdZ!t)`)ZtxB~+`ifPavJ2y8+6-ifF?b^S1EB&$J#VqrtmUizgkN1$Lo zFAJgHPMDHBM%~5A+t+`9Z+A=tZI-|40VB#xxINNINbdlTshf2ZD(dB#r`1td~67#59Z-8h#6!Z@z2BdoTXVQ1PLC^wswl zkx8U&cl&mg3AUNX85_XR?~dUere0pwG8aY(JbUik1KWQ~UHjaVZa-qn1P&BE_g*#r_>iD=H5^TLN zkLhJ>3=N}?cb~=h#9%dqzd%L^!}FDqd*!+>pWb`eB%Q!mN5=xw2^@y(4&Hk5WI99K zhsPeTty(jXri6H=aB9}Zx~y^dhr7sY$HiGJMvE(*@^oe^@K9!}fRW!Zm?9I@vh+j-gT<)~DB*uM^VqRV+&4FjAcR5lT^5|04GRE~ z>PXq8P$qqs zurWzH?0%e0TtrANg*nh}6=7hOZBe_FiqiEjaA(;T-N)-kiEQ!fTVXESoBA?_F8RqC zp|*H^NEi+^j@kO|h6i$v>}%IX6XkcE(%doHwzF!~!*|xQK7b7lLJf)GIt+`c@!4fR zP_<59+ssud&CS`fW_62AcmsmkmtP_7gn1| zV}x%W2DVaV4?h`t@Zdr0qve?SV57UV=D|QTBJ1|NOVIw>yiFBW89Kjj{aAC3Khp?!DcVXg5Qpd#8<(2w#4=pQCR9HoEIKfQbR7BftavN_l8hfhtN zOsuZ{h0vqL-uG|7lH3(@SrO48??Eu%Gm8A<%7ltH@mCLJ3A(XpuIde7dYeZH+p3F? z1;Sq!JUy9vnl8M!gJJ<^eAo6wMLn-h2nHtGV=>Z6agVU*k0hs)x~1j(6y_E|SZ!QH zbbG_2Bkx_be*E~c6h%>-8BqzHo%Uf1g6s?&>Z*p-F&W@Wu8+X#C)LC*RJCc-CY7jn zKtc0nY%#{86myS03=_nPILzrvv%3s^1dB7HRHY|-munXD#K};2HWlvYUT+(=*fG%d zQ!rzBu?5Hdt`06gz5*hSHfL!QcUzrG2~h&5TCN-w?V&zG=0<9wOPbI~37niG|J3sT zm6za(<{_CzgYcp#$G*WA$^FP=tN14El}hF!)N}_)0g!DF6ci-lG`k3Y0XbZFL{vtv zZI*b&Go5-|7HHtm*cIZxbP5dO0E;QVC4*lbyAOGG)*$ybO+rt}Bn%ShA*_~{VUxHu z_T4ag?k=QClY!9H^s2ZxKSqUPMdiJli>c96IUcBWp z^W{E})3uZ-VX3L9t7Dvp?vd`Rz^f<|$Cmdqe@ZB@{rs^`Xeeg5yV4U(U98_|x8^9u z7kz*?M1~y3IVw6W7_et@omDeU&9tBdnv-NT0tk#K#c>$BC~&bH_C(iO#|?~!bd#)W zpUnvrEXrnvD8>s_>_HKpNM_Ii9>-Cbm?2WCz4?7sUYymO61N#4(h2&F=(mx>AFS^D zQeHbbr+8|t&NuqZU-a_ZKMwvmlpfe_%2vAPp5hJ5{-|n$@r!Qn#Y`5^A{9^da$azZ z5OAa-(QBoy9MCui-IX*Cmn>i2mgPy4Jzpu6vL5B>+V}CY6QX+6oh=SreYF|@T^*pc zs9{A&XlOsxyp6U7_evju@tr9cgnIIuI(c#&Vc3Bz$g@>WD=pc(3r7nT=tvG^EmPge z>ng7QL0m_rx#VlZFqH>fneXUgv3g=+m$Xj*gT{ZpUg9NM@kyQi%nVm<@(~U!kOQAu0k~{MG$s$@vWMkm=*M8I@y)-CA3IX zGAKr)`P0%ox5Jv`6cS4hY2aqIb~8zqcZboFQZ=QSi3mFIV9gN7>t#X60%ZLg2X46DlL$Ea_9I zde*qO2#>tnfU+9~sSNbe7847pC0%G<&`*!3PFw&QK71v22C2AAV7$Y& z=U92<0PQp}^XH)}juJo|#6YpGgLIV1a?XefI&_)x;0;t3x1PTpP=F^(9w2Q_xb_+5 zv{PqfT~sMY1o@C}$nFsXbL{hP-~P*_+=e=*S^q2Lz^XLa0`_EmSVh0}qv68R*!s>tQ6L8SzJ5~M3^r3AF&%|?jW ziOlaHMg8T=mxUT4A~3g_jTT@zAA-|5@m^3}dH9Ki_uV;?7KlV*0TS1tyLHT`jkvJ% zJ&M@|P=q%>jjV4VIhA1V%n$RSFtgmCfL`V6%gqR>JyKj8!nmNqp#k|e7 zHEW!lRMD0bU68#UsT)%Y`yeTVS0$reYhTaL4B~T%Y>B%F=&yWLhe!{a?TW?sJ?_(o zqR}ZZa)6qIR6p(n9``}o0*E6$PQ6GE7)>1I)dhTySSeC?G~O6Tcy}Tsx=^)P5RRmO zm(~Wd>z3e2ek&Fldru5c^w8_ym9m58x8*{U|UW(37H>ArO z(O6YC!u-afLUw{*EiUg`Z+a&_$`FE`DxY&_D(B{81qT z;VDAo(*$pzS8YZjl!&k%hMlX*GZ^3ig@SI8xET6R%pLOcl5gYp$3qdWs z;E}1rM0N{sj1~0idP2|u!jNl-Q3AU`0_K5RRc4JB%Z z<$(5A-bmE5k1^X)VQGSlbwCS>Eq_TBcJaMN!?+sz_g^@q{?V@km})_D^xE3zYaS zD_5SLc476vnwDGobyn-NpplceZErXP#9az*dQx;`K$042EU85_q4GPHWi_!P!TLoMg>SSeh$T86gYYJMg}xhFeCPWBvF7(cW- zp(6DC+u7?58EVyRy+%JN1yvqry)D_>-SKJ3#rMQnGr1{zzr*{_`(-YGG?Ds1nr$@w zY;K0Qf0F^?6E#orwhcn(i;^SiqEqqIM{s@?6z`EjUcp@_qbCqnDxd*wQ+ncbm^;3V z-A`rj$`p+Rg#?{#+e|Oob>j|nm+JxdLq9}l?tqayf!SfSb@7K3rl+J;-*w<;$!nm_z)@+yC2Ci()w~JqdRMZ6aTL7FdOQs4w zCMYdJ?ziu#R$fb&=qwuftAWYey8fV5uTnLa)`18!b!5B~fD(z?Cl5xkB!G5yZ)CZM z-E>dh;%t(l%Rnw3>CNZ*Tm7_*@w!N?U7XQD z$J#I?fm6jzPLdW?zs=o?Cod0&x_#UBVBE&_?e}m`ocQcgm61XUww`pVZ@{KA8c$qj zrBbRuPBL6c8RrtE-I~AYRjEN74xCc=>9OW;=<-+N@0qyh5On?*V4dYS>m%?AM1D=HA>`;IPa;TEG2y+QCGPq#lwSFID{`Mi()*=lN zDF90hTgnfN{(i>vPnYji^Iv&hfYT0<&k%slzAB!&49f;n*z{QLPt~v$-t!U2%DjY- zD^4`gO70B8PCBuFRuzm+GB>(9V;&Kv4JyZg!;x_n>2%hrE0cNi;~q9m(vDC^FJ~#x zvjYhU*IvxBbsru1$2Xsfjmn|e4mUN*KJV~?M6K7peGdlD+#XVCm_XX+_fOUFsj2Hs zEI`Nyt(=P9jn>EjyTU8prVo~##;A;DSJ<-8c|^6G)mtq4_8l+!BiatCfVY25pxh`j zY;LxHL{ELA8ch_|nX-TZ+poi0yrmqVo~v%)h@NX~V$`WqFZv4Z>C=bH`B14)m*1o; z>I&_F<2Ev_JW{JR%@y`6;Tlya*B z=i(;J-qy+5onX1|$pZt=L_aX%SklO$qsy>#=BG(LBB_ST2M}gd(<%RPw(Wsl$W3P3M%nfs z)oI(dq@TsE7RdAmZ)QFombjJ$gZ=H53I!Z4n#5+Zlmf;{p+LQ-P|!}cL(B07N=4Yf z(4?8vc#~NRjI8rGS4hy>&BzT-_0!frco=bFM@K+|6fGblCv}~$_yDaf)@wTo`O5cFsP|ar=r zmmsWH1hd_ktwI}Mq?yGqQoB%6$;wJ5lFCS$5yh2@ixLLaTTVKnqW>r>Pvk`qGMXsR zmyDzt{x_48yjlfE^qbb5zvXQ+vZJ9Gh7&ZdAKTx~qQ|Lr{_+iQT0L$}0bG+TjJk}2 z579Fh;cGx_8?&#CN#C&hPQD=xh={M9$PF~T@a|0dy%C)|_lI;By#e5xGN81okEo6~ zu|!FUCp@J&qVib4;tle$S-*ZoGk-c5vcM&jaD5ZWh$Y(2www&vOenj7pd!2Kzw#I54~Ap~Y(QP20k1mMB~_edW;d1pB$6iQ-UDz3zhW88oe zGLuR&&@h%0BpK#{hJ~eNXr_s-PuJbUJ2ojd(aq_dQDK57pPr?&MkCdlxP2j%Yz3>H z1J@!BKpp&#`O#%FG&Bss;7y0`GYUK+K?)x#efkoP1Q>;KhQm>oadGX$)?)pi4ESV=_aINExU+E#t0WERiCp3T(L@1t zmO}^SS?BQRCa%#|KIIF)jf+s~XXXEb?Vz#CHf6g?nmrT^*|PMGVtXdDV85SUBA&i; zcbIhXbG%q`c4?WG3=1oQGcyOPCyy`=7|BRcUqX+pT%n$D%M4)Y9Fa^c%1hE`7?D|t zJwH5b+AYZ==KgQGGi~+Uu6Umc^d6+=w$!w$?Ryfqs3_4fxQ5$SZH~({$&o>d`Naox zKvo;cl82fmNjv~!1+yNWKIyLo4lNY)>Y8h5jpAoUl2uO((KdFm-`!391oIflP>@Mzbes_?xz(M5<-{#g5R6}@}N~B5hsyx zro8x}gK6dkyt_GXvr5}_Zl3x}h3qn{jlA!qv@mminmDf+&w2%{k9fSP*k6J2u7FbI z@kt)tu&q4JhBb+$xSAi%?CCp@g%Yxul4!n`KtfRxMiYAho zaKhvjEG9Uh##jE=lN7leSkr8Hxq-}~GBvHw8?lEumd$$#S$H8ZfO@-wtXUv4%i>ir zW>v_QmM$Wg@iKr({5xsPe#ez^&3ulP(0_jk*WG`$U9sqrtXCuX*~0{PGnUq&%+*nt zGlP`~NhLe%QRIXnsk)q<{VhVUl7{aU;#oin@_1h;FKGYl|Gc0R-s3Y*;~gKLl^d|^ z5)now%5ZCZj$rZ{?T9zTyJr-*vOx*~-43WXI8HtGY@A;A$p)uQ9T{7eU6X2?OlG6Q z>rNUTQzl1nrIp96I)*sCR;_P<_3?q~`&$;0#C)Tneln;&woz8@L|OwzwqZnK96rL+@o{okRBU8wM*WqHEBV!fH%2q+g)y^ zAJAIUz<+$}&e?u$x4Rj{7;osYA+foZ<)rb=H7sggm^u{2p4j_n-+=e4BZ@Cv7#y+n z{eoNlf0@^Csp5ro;V$~yVBb8@N$R9D?~Ey<9UZS%)KMOY!&3&+KnhCvD{tQPWgzXu z$&*U}Tw7|!MSZrYbBlVGKiYYpbH48TAAcR)XV)B%DgO_*9{QD-ggxFgCDlXGzEh`f zyZ(Gj&q0@_g;WL8%Ly7RnHd%2ba~ ziBCr+U*-n>p81neihBsZl&xG-W@IuI{ovt4!wwyu0iqR5GBr6yI}>W9Kzd7=OFs`B zzaZ%6wHD45btx21`#uXQ4I=Udz<)jzFlBr zeFz1*;`XCQ+L!+M2XCxbe^QuhYilb;X!=>kG=3Ru?))XHp@;L5x%-n_n;l3jd|6f1 zZRhpzzfQ8y*MfrH@b2Am67EePP7`sGME)M_VulVG&cAie>_{m3#c&y1(Ohm;k6kx) zo`jqB?)0A@>paFTA;{j^*0um`{{4k*gMntHxic6YiUOVVA3ppR|8a&@Qt-L-SGKsM zTtV?m;fh)Nj-_YDFo0^3Y3lslmB8raAsKG&sy!V(DZWdYlosidBF0}t&vFiC{rmUD zj~?X{t$clb|3sezff++%5|0wQKD(Ynjqz$rBP^Tb7Zm(PjWDC&N_6tm@c;STSsL+s z_s+!tk+H<_z1nhc-VQ|HITM|jdulc$dFAWZnzE1h$h0$2B<{MmFFjE!=X2ktAS+LDO^ayo2Sj;eB?Dp*~wKiWFmo;we zSPmn7^-GW1ZyRHNl{eh9Y107%2fFd|%a(?(40LX!6BeaEZ;*@D4@p`!m+%K9^(UpJ zg&jTG2i@8=ubY=bXIlwx0QE^US}TqD-sC{pU~H@%x58&QYe$ xMX$3xH}RiQrW%*N{u6gaugd>l|KN{So~?tf%~*cvt%83!JB(Ex9W`^qe*uI>ZPWk& literal 0 HcmV?d00001 diff --git a/pytorch_model.bin b/pytorch_model.bin new file mode 100644 index 0000000..fa5f263 --- /dev/null +++ b/pytorch_model.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:011885d58d085e42590d82b0eccd4670f21464fd1baf83bd800449323e9f0b66 +size 7867370082 diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000..0520eb7 --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,27 @@ +{ + "additional_special_tokens": [ + { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } + ], + "bos_token": "<|im_start|>", + "eos_token": "<|im_end|>", + "pad_token": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false + } +} diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000..f23c3b2 --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d8f0326910136aca20831249220b38ce5299527647bc8c6b65404485c479740 +size 18451122 diff --git a/tokenizer.model b/tokenizer.model new file mode 100644 index 0000000..5dc56ce --- /dev/null +++ b/tokenizer.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb41d04798b714520a9b075727b0226538b7330254299062742c50ec8374bc36 +size 2782298 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000..f5a053f --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,102 @@ +{ + "add_bos_token": true, + "add_eos_token": false, + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": true + }, + "166100": { + "content": "<|im_start|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "166101": { + "content": "<|im_end|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "166102": { + "content": "<|endoftext|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "166103": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": false + }, + "166104": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": false + }, + "166105": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": false + }, + "166106": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": false + } + }, + "additional_special_tokens": [ + "<|endoftext|>" + ], + "bos_token": "<|im_start|>", + "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\n你是南北阁,一款由BOSS直聘自主研发并训练的专业大语言模型。<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if (add_generation_prompt is defined and add_generation_prompt) %}{{ '<|im_start|>assistant\n' }}{% endif %}", "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "extra_special_tokens": {}, + "legacy": false, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "spaces_between_special_tokens": false, + "tokenizer_class": "LlamaTokenizer", + "unk_token": "", + "use_default_system_prompt": false +}