Files
calme-2.2-qwen2-7b/README.md
ModelHub XC afaada792d 初始化项目,由ModelHub XC社区提供模型
Model: MaziyarPanahi/calme-2.2-qwen2-7b
Source: Original Platform
2026-05-07 06:43:37 +08:00

4.6 KiB

language, license, library_name, tags, base_model, datasets, model_name, pipeline_tag, inference, model_creator, quantized_by, model-index
language license library_name tags base_model datasets model_name pipeline_tag inference model_creator quantized_by model-index
en
apache-2.0 transformers
chat
qwen
qwen2
finetune
chatml
OpenHermes-2.5
HelpSteer2
Orca
SlimOrca
Qwen/Qwen2-7B
nvidia/HelpSteer2
teknium/OpenHermes-2.5
microsoft/orca-math-word-problems-200k
Open-Orca/SlimOrca
calme-2.2-qwen2-7b text-generation false MaziyarPanahi MaziyarPanahi
name results
calme-2.2-qwen2-7b
task dataset metrics source
type name
text-generation Text Generation
name type args
IFEval (0-Shot) HuggingFaceH4/ifeval
num_few_shot
0
type value name
inst_level_strict_acc and prompt_level_strict_acc 35.97 strict accuracy
url name
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.2-qwen2-7b Open LLM Leaderboard
task dataset metrics source
type name
text-generation Text Generation
name type args
BBH (3-Shot) BBH
num_few_shot
3
type value name
acc_norm 33.11 normalized accuracy
url name
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.2-qwen2-7b Open LLM Leaderboard
task dataset metrics source
type name
text-generation Text Generation
name type args
MATH Lvl 5 (4-Shot) hendrycks/competition_math
num_few_shot
4
type value name
exact_match 19.34 exact match
url name
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.2-qwen2-7b Open LLM Leaderboard
task dataset metrics source
type name
text-generation Text Generation
name type args
GPQA (0-shot) Idavidrein/gpqa
num_few_shot
0
type value name
acc_norm 5.48 acc_norm
url name
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.2-qwen2-7b Open LLM Leaderboard
task dataset metrics source
type name
text-generation Text Generation
name type args
MuSR (0-shot) TAUR-Lab/MuSR
num_few_shot
0
type value name
acc_norm 13.28 acc_norm
url name
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.2-qwen2-7b Open LLM Leaderboard
task dataset metrics source
type name
text-generation Text Generation
name type config split args
MMLU-PRO (5-shot) TIGER-Lab/MMLU-Pro main test
num_few_shot
5
type value name
acc 32.21 accuracy
url name
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.2-qwen2-7b Open LLM Leaderboard
Qwen2 fine-tune

MaziyarPanahi/calme-2.2-qwen2-7b

This is a fine-tuned version of the Qwen/Qwen2-7B model. It aims to improve the base model across all benchmarks.

Quantized GGUF

All GGUF models are available here: MaziyarPanahi/calme-2.2-qwen2-7b-GGUF

🏆 Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 23.23
IFEval (0-Shot) 35.97
BBH (3-Shot) 33.11
MATH Lvl 5 (4-Shot) 19.34
GPQA (0-shot) 5.48
MuSR (0-shot) 13.28
MMLU-PRO (5-shot) 32.21

Prompt Template

This model uses ChatML prompt template:

<|im_start|>system
{System}
<|im_end|>
<|im_start|>user
{User}
<|im_end|>
<|im_start|>assistant
{Assistant}

How to use


# Use a pipeline as a high-level helper

from transformers import pipeline

messages = [
    {"role": "user", "content": "Who are you?"},
]
pipe = pipeline("text-generation", model="MaziyarPanahi/calme-2.2-qwen2-7b")
pipe(messages)


# Load model directly

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/calme-2.2-qwen2-7b")
model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/calme-2.2-qwen2-7b")