Files
chatty-djinn-14B/model.safetensors.index.json

1 line
42 KiB
JSON
Raw Permalink Normal View History

{"metadata": {"mergekit_version": "0.0.4.1"}, "weight_map": {"model.layers.55.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.55.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.55.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.55.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.55.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.55.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.54.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.54.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.54.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.54.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.54.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.54.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.54.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.54.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.54.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.53.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.53.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.53.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.53.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.53.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.53.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.53.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.53.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.53.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.52.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.52.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.52.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.52.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.52.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.52.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.52.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.52.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.52.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.51.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.51.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.51.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.51.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.51.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.51.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.51.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.51.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.51.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.50.mlp.down_proj.weight": "model-00001-of-00003.safetensors", "model.layers.50.mlp.gate_proj.weight": "model-00001-of-00003.safetensors", "model.layers.50.mlp.up_proj.weight": "model-00001-of-00003.safetensors", "model.layers.50.post_attention_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.50.self_attn.o_proj.weight": "model-00001-of-00003.safetensors", "model.layers.50.self_attn.v_proj.weight": "model-00001-of-00003.safetensors", "model.layers.50.self_attn.k_proj.weight": "model-00001-of-00003.safetensors", "model.layers.50.self_attn.q_proj.weight": "model-00001-of-00003.safetensors", "model.layers.50.input_layernorm.weight": "model-00001-of-00003.safetensors", "model.layers.10.mlp.gate_proj.we