ModelHub XC 3bcd60ac5c 初始化项目,由ModelHub XC社区提供模型
Model: Inv/Konstanta-7B
Source: Original Platform
2026-04-24 07:39:36 +08:00

language, license, tags, base_model, model-index
language license tags base_model model-index
en
apache-2.0
merge
mergekit
lazymergekit
maywell/PiVoT-0.1-Evil-a
mlabonne/NeuralOmniBeagle-7B-v2
roleplay
rp
not-for-all-audiences
maywell/PiVoT-0.1-Evil-a
mlabonne/NeuralOmniBeagle-7B-v2
name results
Konstanta-7B
task dataset metrics source
type name
text-generation Text Generation
name type config split args
AI2 Reasoning Challenge (25-Shot) ai2_arc ARC-Challenge test
num_few_shot
25
type value name
acc_norm 70.05 normalized accuracy
url name
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-7B Open LLM Leaderboard
task dataset metrics source
type name
text-generation Text Generation
name type split args
HellaSwag (10-Shot) hellaswag validation
num_few_shot
10
type value name
acc_norm 87.5 normalized accuracy
url name
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-7B Open LLM Leaderboard
task dataset metrics source
type name
text-generation Text Generation
name type config split args
MMLU (5-Shot) cais/mmlu all test
num_few_shot
5
type value name
acc 65.06 accuracy
url name
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-7B Open LLM Leaderboard
task dataset metrics source
type name
text-generation Text Generation
name type config split args
TruthfulQA (0-shot) truthful_qa multiple_choice validation
num_few_shot
0
type value
mc2 65.43
url name
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-7B Open LLM Leaderboard
task dataset metrics source
type name
text-generation Text Generation
name type config split args
Winogrande (5-shot) winogrande winogrande_xl validation
num_few_shot
5
type value name
acc 82.16 accuracy
url name
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-7B Open LLM Leaderboard
task dataset metrics source
type name
text-generation Text Generation
name type config split args
GSM8k (5-shot) gsm8k main test
num_few_shot
5
type value name
acc 71.04 accuracy
url name
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-7B Open LLM Leaderboard

Konstanta-7B

Konstanta-7B is a merge of the following models using LazyMergekit:

This is a test merge that is supposed to improve Kunoichi by merging it with new Beagle model and PiVoT Evil, which both show good performance. Even though the model's name is in Russian, it is not really capable of properly using it, as it was not the main goal of the model.

🧩 Configuration

merge_method: dare_ties
dtype: bfloat16
parameters:
  int8_mask: true
base_model: SanjiWatsuki/Kunoichi-DPO-v2-7B
models:
  - model: SanjiWatsuki/Kunoichi-DPO-v2-7B
  - model: maywell/PiVoT-0.1-Evil-a
    parameters:
      density: 0.65
      weight: 0.15
  - model: mlabonne/NeuralOmniBeagle-7B-v2
    parameters:
      density: 0.85
      weight: 0.45

💻 Usage

!pip install -qU transformers accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "Inv/Konstanta-7B"
messages = [{"role": "user", "content": "What is a large language model?"}]

tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    torch_dtype=torch.float16,
    device_map="auto",
)

outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 73.54
AI2 Reasoning Challenge (25-Shot) 70.05
HellaSwag (10-Shot) 87.50
MMLU (5-Shot) 65.06
TruthfulQA (0-shot) 65.43
Winogrande (5-shot) 82.16
GSM8k (5-shot) 71.04
Description
Model synced from source: Inv/Konstanta-7B
Readme 565 KiB