From 89870f23d112c8437c1474306841f6d04674dbc9 Mon Sep 17 00:00:00 2001 From: Severian Date: Sun, 8 Oct 2023 23:55:31 +0000 Subject: [PATCH] Delete adapter_config.json --- adapter_config.json | 23 ----------------------- 1 file changed, 23 deletions(-) delete mode 100644 adapter_config.json diff --git a/adapter_config.json b/adapter_config.json deleted file mode 100644 index 216fb4a..0000000 --- a/adapter_config.json +++ /dev/null @@ -1,23 +0,0 @@ -{ - "alpha_pattern": {}, - "auto_mapping": null, - "base_model_name_or_path": "teknium/CollectiveCognition-v1.1-Mistral-7B", - "bias": "none", - "fan_in_fan_out": false, - "inference_mode": true, - "init_lora_weights": true, - "layers_pattern": null, - "layers_to_transform": null, - "lora_alpha": 32, - "lora_dropout": 0.1, - "modules_to_save": null, - "peft_type": "LORA", - "r": 16, - "rank_pattern": {}, - "revision": null, - "target_modules": [ - "q_proj", - "v_proj" - ], - "task_type": "CAUSAL_LM" -} \ No newline at end of file