1 line
28 KiB
JSON
1 line
28 KiB
JSON
|
|
{"metadata": {"mergekit_version": "0.0.4.1"}, "weight_map": {"model.layers.35.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.35.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.35.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.35.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.35.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.35.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.35.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.35.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.35.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.34.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.34.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.34.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.34.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.34.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.34.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.34.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.34.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.34.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.33.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.33.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.33.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.33.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.33.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.33.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.33.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.33.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.33.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.32.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.32.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.32.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.32.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.32.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.32.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.32.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.32.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.32.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.31.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.30.self_attn.k
|