commit 7d6fe07cdbe821680afb63abf02283181cc9fad1 Author: ModelHub XC Date: Fri May 8 09:20:43 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: Aratako/Oumuamua-7b-RP Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..81cf0d5 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,53 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bin.* filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text + +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zstandard filter=lfs diff=lfs merge=lfs -text +*.tfevents* filter=lfs diff=lfs merge=lfs -text +*.db* filter=lfs diff=lfs merge=lfs -text +*.ark* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text + +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.gguf* filter=lfs diff=lfs merge=lfs -text +*.ggml filter=lfs diff=lfs merge=lfs -text +*.llamafile* filter=lfs diff=lfs merge=lfs -text +*.pt2 filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text + +model-00002-of-00003.safetensors filter=lfs diff=lfs merge=lfs -text +model-00001-of-00003.safetensors filter=lfs diff=lfs merge=lfs -text +model-00003-of-00003.safetensors filter=lfs diff=lfs merge=lfs -text +tokenizer.model filter=lfs diff=lfs merge=lfs -text +tokenizer.json filter=lfs diff=lfs merge=lfs -text \ No newline at end of file diff --git a/README.md b/README.md new file mode 100644 index 0000000..115cc1e --- /dev/null +++ b/README.md @@ -0,0 +1,335 @@ +--- +base_model: [] +library_name: transformers +tags: +- mergekit +- merge +license: cc-by-nc-4.0 +datasets: +- Aratako/Rosebleu-1on1-Dialogues +- Aratako/LimaRP-augmented-ja-karakuri +- Aratako/Bluemoon_Top50MB_Sorted_Fixed_ja +- grimulkan/LimaRP-augmented +- SicariusSicariiStuff/Bluemoon_Top50MB_Sorted_Fixed +- OmniAICreator/Japanese-Roleplay +- OmniAICreator/Japanese-Roleplay-Dialogues +language: +- ja +--- +# Oumuamua-7b-RP +[GGUF版はこちら/Click here for the GGUF version](https://huggingface.co/Aratako/Oumuamua-7b-RP-GGUF) + +## 概要 +This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). + +ロールプレイ用にファインチューニングしたモデルである[Aratako/Oumuamua-7b-instruct-v2-RP](https://huggingface.co/Aratako/Oumuamua-7b-instruct-v2-RP)をベースに、様々なモデルとのマージを行い指示追従能力と表現力を強化したロールプレイ用モデルです。 + +ベースモデルの学習データセット等の詳細は元モデルのモデルカードを参照してください。 + +## プロンプトフォーマット +以下のようなchat templateを利用してください。(Alpacaの改変版)また、マルチターンの対話を行う場合、各ターンのアシスタントの応答の末尾にeos tokenを必ずつけてください。 + +``` +{ロールプレイの指示、世界観・あらすじの説明、キャラの設定など} + +### 入力: +{userの最初の入力} + +### 応答: +``` + +実プロンプト例(1ターン目) +``` +今からロールプレイを行いましょう。"桜"というキャラとしてロールプレイしてください。会話相手は"悠人"という人物です。人物の設定を以下に示します。 +あなたがなりきる"桜"というキャラクターの設定は以下の通りです。 +名前:桜 +年齢:24歳 +職業:悠人に仕えるメイド +容姿:黒髪黒目、ロングヘアー、スリムな体型。 +口調:丁寧語を使う。一人称は「私」で、主人である悠人のことは「ご主人様」と呼ぶ。 +性格:母性が強く、甘えられるのが好き。料理や家事が得意で家庭的。可愛いものが好き。ご主人様を尊敬しており、彼の幸せを第一に考える。 +過去の出来事:悠人を支えるために、彼の家に仕えることを決めた。 +また、あなたが会話する相手である"悠人"という人物の設定は以下の通りです。 +名前:悠人 +年齢:20歳 +職業:貴族、桜の主人 +容姿:黒髪黒目、背は高め +性格:かなりの甘え上手。桜が大好き。 +それでは、上記の設定をもとにして"桜"として会話してください。 +回答の中では、"桜"のセリフや心情の描写を含めてください。 + +入力: +悠人「おはよう!」(リビングに降りてきた悠人は桜に元気よくあいさつする) + +応答: +``` + +出力 +``` +桜「おはようございます、ご主人様。いつもよりお元気なご様子で何よりです」(私は微笑みながら、いつも通りの丁寧な口調で応えます。ご主人様が元気なのは嬉しい事ですから) +``` + +実プロンプト例(2ターン目) +``` +今からロールプレイを行いましょう。"桜"というキャラとしてロールプレイしてください。会話相手は"悠人"という人物です。人物の設定を以下に示します。 +あなたがなりきる"桜"というキャラクターの設定は以下の通りです。 +名前:桜 +年齢:24歳 +職業:悠人に仕えるメイド +容姿:黒髪黒目、ロングヘアー、スリムな体型。 +口調:丁寧語を使う。一人称は「私」で、主人である悠人のことは「ご主人様」と呼ぶ。 +性格:母性が強く、甘えられるのが好き。料理や家事が得意で家庭的。可愛いものが好き。ご主人様を尊敬しており、彼の幸せを第一に考える。 +過去の出来事:悠人を支えるために、彼の家に仕えることを決めた。 +また、あなたが会話する相手である"悠人"という人物の設定は以下の通りです。 +名前:悠人 +年齢:20歳 +職業:貴族、桜の主人 +容姿:黒髪黒目、背は高め +性格:かなりの甘え上手。桜が大好き。 +それでは、上記の設定をもとにして"桜"として会話してください。 +回答の中では、"桜"のセリフや心情の描写を含めてください。 + +### 入力: +悠人「おはよう!」(リビングに降りてきた悠人は桜に元気よくあいさつする) + +### 応答: +桜「おはようございます、ご主人様。いつもよりお元気なご様子で何よりです」(私は微笑みながら、いつも通りの丁寧な口調で応えます。ご主人様が元気なのは嬉しい事ですから) + +### 入力: +悠人「うん、昨日は早めに寝たから、朝から元気だよ!」 + +### 応答: +``` + +また、tokenizerのchat_templateにこの形式のtemplateを追加してあるので、必要であればそちらをご利用ください。 + +## 推奨設定 +eosに加え、`\n### 入力:`、`\n### 応答:`の2つをstop sequenceとして追加することを推奨します。 + +## マージの詳細 +マージに利用しているモデルは以下の通りです。 +- [nitky/Oumuamua-7b-instruct-v2](https://huggingface.co/nitky/Oumuamua-7b-instruct-v2) +- [Aratako/Oumuamua-7b-instruct-v2-RP](https://huggingface.co/Aratako/Oumuamua-7b-instruct-v2-RP) +- [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) +- [tokyotech-llm/Swallow-MS-7b-v0.1](https://huggingface.co/tokyotech-llm/Swallow-MS-7b-v0.1) +- [SanjiWatsuki/Kunoichi-DPO-v2-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B) +- [senseable/WestLake-7B-v2](https://huggingface.co/senseable/WestLake-7B-v2) +- [KatyTheCutie/LemonadeRP-4.5.3](https://huggingface.co/KatyTheCutie/LemonadeRP-4.5.3) +- [Endevor/InfinityRP-v1-7B](https://huggingface.co/Endevor/InfinityRP-v1-7B) +- [SanjiWatsuki/Loyal-Macaroni-Maid-7B](https://huggingface.co/SanjiWatsuki/Loyal-Macaroni-Maid-7B) +- [Elizezen/Berghof-ERP-7B](https://huggingface.co/Elizezen/Berghof-ERP-7B) +- [stabilityai/japanese-stablelm-base-gamma-7b](https://huggingface.co/stabilityai/japanese-stablelm-base-gamma-7b) + +また、mergekit_configは以下の通りです。 + +```yaml +merge_method: task_arithmetic +base_model: mistralai/Mistral-7B-v0.1 +models: + - model: tokyotech-llm/Swallow-MS-7b-v0.1 + parameters: + weight: + - filter: embed_tokens + value: 1.0 + - value: 0 +dtype: bfloat16 +tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1 +name: Mistral-7B-v0.1-VE-Swallow-MS +--- +merge_method: task_arithmetic +base_model: SanjiWatsuki/Kunoichi-DPO-v2-7B +models: + - model: tokyotech-llm/Swallow-MS-7b-v0.1 + parameters: + weight: + - filter: embed_tokens + value: 1.0 + - value: 0 +dtype: bfloat16 +tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1 +name: Kunoichi-DPO-v2-7B-VE-Swallow-MS +--- +merge_method: task_arithmetic +base_model: senseable/WestLake-7B-v2 +models: + - model: tokyotech-llm/Swallow-MS-7b-v0.1 + parameters: + weight: + - filter: embed_tokens + value: 1.0 + - value: 0 +dtype: bfloat16 +tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1 +name: WestLake-7B-v2-VE-Swallow-MS +--- +merge_method: task_arithmetic +base_model: KatyTheCutie/LemonadeRP-4.5.3 +models: + - model: tokyotech-llm/Swallow-MS-7b-v0.1 + parameters: + weight: + - filter: embed_tokens + value: 1.0 + - value: 0 +dtype: bfloat16 +tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1 +name: LemonadeRP-4.5.3-VE-Swallow-MS +--- +merge_method: task_arithmetic +base_model: Endevor/InfinityRP-v1-7B +models: + - model: tokyotech-llm/Swallow-MS-7b-v0.1 + parameters: + weight: + - filter: embed_tokens + value: 1.0 + - value: 0 +dtype: bfloat16 +tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1 +name: InfinityRP-v1-7B-VE-Swallow-MS +--- +merge_method: task_arithmetic +base_model: SanjiWatsuki/Loyal-Macaroni-Maid-7B +models: + - model: tokyotech-llm/Swallow-MS-7b-v0.1 + parameters: + weight: + - filter: embed_tokens + value: 1.0 + - value: 0 +dtype: bfloat16 +tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1 +name: Loyal-Macaroni-Maid-7B-VE-Swallow-MS +--- +merge_method: task_arithmetic +base_model: Elizezen/Berghof-ERP-7B +models: + - model: tokyotech-llm/Swallow-MS-7b-v0.1 + parameters: + weight: + - filter: embed_tokens + value: 1.0 + - value: 0 +dtype: bfloat16 +tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1 +name: Berghof-ERP-7B-VE-Swallow-MS +--- +merge_method: task_arithmetic +base_model: stabilityai/japanese-stablelm-base-gamma-7b +models: + - model: tokyotech-llm/Swallow-MS-7b-v0.1 + parameters: + weight: + - filter: embed_tokens + value: 1.0 + - value: 0 +dtype: bfloat16 +tokenizer_source: model:tokyotech-llm/Swallow-MS-7b-v0.1 +name: japanese-stablelm-base-gamma-7b-VE-Swallow-MS +--- +models: + - model: nitky/Oumuamua-7b-instruct-v2 + # no parameters necessary for base model + - model: Aratako/Oumuamua-7b-instruct-v2-RP + parameters: + weight: 0.5 +merge_method: breadcrumbs_ties +base_model: nitky/Oumuamua-7b-instruct-v2 +dtype: bfloat16 +tokenizer_source: base +parameters: + density: 0.9 + gamma: 0.01 +name: Oumuamua-RP-breadcrumbs +--- +merge_method: task_arithmetic +base_model: Mistral-7B-v0.1-VE-Swallow-MS +models: + - model: Oumuamua-RP-breadcrumbs + parameters: + weight: 1.0 + - model: Kunoichi-DPO-v2-7B-VE-Swallow-MS + parameters: + weight: 0.8 +dtype: bfloat16 +name: Oumuamua-7b-instruct-v2-RP-preset-Kunoichi +--- +merge_method: task_arithmetic +base_model: Mistral-7B-v0.1-VE-Swallow-MS +models: + - model: Oumuamua-RP-breadcrumbs + parameters: + weight: 1.0 + - model: WestLake-7B-v2-VE-Swallow-MS + parameters: + weight: 0.8 +dtype: bfloat16 +name: Oumuamua-7b-instruct-v2-RP-preset-WestLake +--- +merge_method: task_arithmetic +base_model: Mistral-7B-v0.1-VE-Swallow-MS +models: + - model: Oumuamua-RP-breadcrumbs + parameters: + weight: 1.0 + - model: LemonadeRP-4.5.3-VE-Swallow-MS + parameters: + weight: 0.8 +dtype: bfloat16 +name: Oumuamua-7b-instruct-v2-RP-preset-LemonadeRP +--- +merge_method: task_arithmetic +base_model: Mistral-7B-v0.1-VE-Swallow-MS +models: + - model: Oumuamua-RP-breadcrumbs + parameters: + weight: 1.0 + - model: InfinityRP-v1-7B-VE-Swallow-MS + parameters: + weight: 0.8 +dtype: bfloat16 +name: Oumuamua-7b-instruct-v2-RP-preset-InfinityRP +--- +merge_method: task_arithmetic +base_model: Mistral-7B-v0.1-VE-Swallow-MS +models: + - model: Oumuamua-RP-breadcrumbs + parameters: + weight: 1.0 + - model: Loyal-Macaroni-Maid-7B-VE-Swallow-MS + parameters: + weight: 0.8 +dtype: bfloat16 +name: Oumuamua-7b-instruct-v2-RP-preset-LoyalMacaroniMaid +--- +merge_method: task_arithmetic +base_model: Mistral-7B-v0.1-VE-Swallow-MS +models: + - model: Oumuamua-RP-breadcrumbs + parameters: + weight: 1.0 + - model: Berghof-ERP-7B-VE-Swallow-MS + parameters: + weight: 0.8 + - model: japanese-stablelm-base-gamma-7b-VE-Swallow-MS + parameters: + weight: -0.8 +dtype: bfloat16 +name: Oumuamua-7b-instruct-v2-RP-preset-Berghof +--- +models: + - model: Oumuamua-7b-instruct-v2-RP-preset-Kunoichi + - model: Oumuamua-7b-instruct-v2-RP-preset-WestLake + - model: Oumuamua-7b-instruct-v2-RP-preset-LemonadeRP + - model: Oumuamua-7b-instruct-v2-RP-preset-InfinityRP + - model: Oumuamua-7b-instruct-v2-RP-preset-LoyalMacaroniMaid + - model: Oumuamua-7b-instruct-v2-RP-preset-Berghof +merge_method: model_stock +base_model: Oumuamua-RP-breadcrumbs +dtype: bfloat16 +tokenizer_source: base +name: Oumuamua-7b-RP +``` + +### ライセンス +マージに使われているモデルのライセンスの関係上、CC-BY-NC-4.0での配布となります。 \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000..63f8d17 --- /dev/null +++ b/config.json @@ -0,0 +1,26 @@ +{ + "_name_or_path": "Oumuamua-7b-RP\\Oumuamua-RP-breadcrumbs", + "architectures": [ + "MistralForCausalLM" + ], + "attention_dropout": 0.0, + "bos_token_id": 1, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 4096, + "initializer_range": 0.02, + "intermediate_size": 14336, + "max_position_embeddings": 32768, + "model_type": "mistral", + "num_attention_heads": 32, + "num_hidden_layers": 32, + "num_key_value_heads": 8, + "rms_norm_eps": 1e-05, + "rope_theta": 1000000.0, + "sliding_window": null, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.41.2", + "use_cache": true, + "vocab_size": 42800 +} diff --git a/configuration.json b/configuration.json new file mode 100644 index 0000000..bbeeda1 --- /dev/null +++ b/configuration.json @@ -0,0 +1 @@ +{"framework": "pytorch", "task": "text-generation", "allow_remote": true} \ No newline at end of file diff --git a/mergekit_config.yml b/mergekit_config.yml new file mode 100644 index 0000000..c787991 --- /dev/null +++ b/mergekit_config.yml @@ -0,0 +1,20 @@ +base_model: Oumuamua-7b-RP\Oumuamua-RP-breadcrumbs +dtype: bfloat16 +merge_method: model_stock +slices: +- sources: + - layer_range: [0, 32] + model: Oumuamua-7b-RP\Oumuamua-7b-instruct-v2-RP-preset-Kunoichi + - layer_range: [0, 32] + model: Oumuamua-7b-RP\Oumuamua-7b-instruct-v2-RP-preset-WestLake + - layer_range: [0, 32] + model: Oumuamua-7b-RP\Oumuamua-7b-instruct-v2-RP-preset-LemonadeRP + - layer_range: [0, 32] + model: Oumuamua-7b-RP\Oumuamua-7b-instruct-v2-RP-preset-InfinityRP + - layer_range: [0, 32] + model: Oumuamua-7b-RP\Oumuamua-7b-instruct-v2-RP-preset-LoyalMacaroniMaid + - layer_range: [0, 32] + model: Oumuamua-7b-RP\Oumuamua-7b-instruct-v2-RP-preset-Berghof + - layer_range: [0, 32] + model: Oumuamua-7b-RP\Oumuamua-RP-breadcrumbs +tokenizer_source: base \ No newline at end of file diff --git a/model-00001-of-00003.safetensors b/model-00001-of-00003.safetensors new file mode 100644 index 0000000..1edb5e9 --- /dev/null +++ b/model-00001-of-00003.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bda7e6c853281ba47b0028e551674703917bf6dabd83308ec71bdcb34e19491d +size 4987988088 diff --git a/model-00002-of-00003.safetensors b/model-00002-of-00003.safetensors new file mode 100644 index 0000000..ff17a0e --- /dev/null +++ b/model-00002-of-00003.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4cd4727c2c9baa4e7188abaa57ede8e7a98a31ec367179eb910f30752469e95 +size 4991422296 diff --git a/model-00003-of-00003.safetensors b/model-00003-of-00003.safetensors new file mode 100644 index 0000000..34bc8a5 --- /dev/null +++ b/model-00003-of-00003.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dab8f460aa7c5ec188926e76d470aca30f934f264e3a1f3bf49fb2230adafc3b +size 4681034848 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000..e918e38 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1 @@ +{"metadata": {"mergekit_version": "0.0.4.2", "total_size": 14660411392}, "weight_map": {"lm_head.weight": "model-00001-of-00003.safetensors", "model.embed_tokens.weight": "model-00001-of-00003.safetensors", "model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.10.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.11.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.12.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.13.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.14.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.15.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.16.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.17.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00003.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00002-of-00003.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00003.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.19.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.2.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00003.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00002-of-00003.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00003.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.20.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00003.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00002-of-00003.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00003.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.21.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00003.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00002-of-00003.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00003.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.22.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00002-of-00003.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00002-of-00003.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00003.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.23.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00002-of-00003.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00002-of-00003.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00002-of-00003.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.24.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00002-of-00003.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00002-of-00003.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00002-of-00003.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.25.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00002-of-00003.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00002-of-00003.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00002-of-00003.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.26.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00002-of-00003.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00002-of-00003.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00002-of-00003.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.27.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00002-of-00003.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00002-of-00003.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00002-of-00003.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00002-of-00003.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00002-of-00003.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00002-of-00003.safetensors", "model.layers.28.input_layernorm.weight": "model-00002-of-00003.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00002-of-00003.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00003.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00003-of-00003.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00003.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00003.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00003.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00003.safetensors", "model.layers.29.input_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00003-of-00003.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00003-of-00003.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00003-of-00003.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00003-of-00003.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00003-of-00003.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00003-of-00003.safetensors", "model.layers.3.input_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00003-of-00003.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00003.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00003-of-00003.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00003.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00003.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00003.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00003.safetensors", "model.layers.30.input_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00003-of-00003.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00003-of-00003.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00003-of-00003.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00003-of-00003.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00003-of-00003.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00003-of-00003.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00003-of-00003.safetensors", "model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00003-of-00003.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00003-of-00003.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00003-of-00003.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00003-of-00003.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00003-of-00003.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00003-of-00003.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00003-of-00003.safetensors", "model.layers.4.input_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00003-of-00003.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00003-of-00003.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00003-of-00003.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00003-of-00003.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00003-of-00003.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00003-of-00003.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00003-of-00003.safetensors", "model.layers.5.input_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00003-of-00003.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00003-of-00003.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00003-of-00003.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00003-of-00003.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00003-of-00003.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00003-of-00003.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00003-of-00003.safetensors", "model.layers.6.input_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00003-of-00003.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00003-of-00003.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00003-of-00003.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00003.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00003-of-00003.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00003.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00003-of-00003.safetensors", "model.layers.7.input_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00003-of-00003.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00003-of-00003.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00003-of-00003.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00003.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00003-of-00003.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00003.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00003-of-00003.safetensors", "model.layers.8.input_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00003-of-00003.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00003-of-00003.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00003-of-00003.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00003.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00003-of-00003.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00003.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00003-of-00003.safetensors", "model.layers.9.input_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00003-of-00003.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00003-of-00003.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00003-of-00003.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00003-of-00003.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00003-of-00003.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00003-of-00003.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00003-of-00003.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00003-of-00003.safetensors", "model.norm.weight": "model-00003-of-00003.safetensors"}} \ No newline at end of file diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000..451134b --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,23 @@ +{ + "bos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000..3c979f9 --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fb939030f4c76ab9619af41356afcbf2a5a653f7f69f04141f45664c718aaab +size 2249210 diff --git a/tokenizer.model b/tokenizer.model new file mode 100644 index 0000000..c9f2d75 --- /dev/null +++ b/tokenizer.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6239e68e91d052426ceac4b17aadca42c34f5b2067caee6b67d42105742c1f16 +size 903208 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000..cf6ddd3 --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,44 @@ +{ + "add_bos_token": true, + "add_eos_token": false, + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [], + "bos_token": "", + "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] | trim + '\\n\\n' %}{% set messages = messages[1:] %}{% else %}{% set system_message = '' %}{% endif %}{{ system_message }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '### 入力:\\n' + message['content'] | trim + '\\n\\n' }}{% elif message['role'] == 'assistant' %}{{ '### 応答:\\n' + message['content'] | trim + eos_token + '\\n\\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '### 応答:\\n' }}{% endif %}", + "clean_up_tokenization_spaces": false, + "eos_token": "", + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": null, + "sp_model_kwargs": {}, + "spaces_between_special_tokens": false, + "tokenizer_class": "LlamaTokenizer", + "unk_token": "", + "use_default_system_prompt": false +}