Files
MINT-empathy-Qwen3-4B/README.md
ModelHub XC 076eb85c11 初始化项目,由ModelHub XC社区提供模型
Model: hongli-zhan/MINT-empathy-Qwen3-4B
Source: Original Platform
2026-05-01 05:04:31 +08:00

2.9 KiB

license, base_model, tags, language, pipeline_tag
license base_model tags language pipeline_tag
mit Qwen/Qwen3-4B
empathy
reinforcement-learning
grpo
dialogue
mint
emotional-support
en
text-generation

MINT-empathy-Qwen3-4B

This model is the Q + D_KL MINT checkpoint fine-tuned from Qwen/Qwen3-4B for multi-turn empathic dialogue.

MINT, short for Multi-turn Inter-tactic Novelty Training, is a reinforcement learning framework that optimizes empathic response quality together with cross-turn discourse-move novelty. In the paper, this checkpoint is the best overall released MINT model, delivering the strongest joint tradeoff between empathy quality and reduced cross-turn tactic repetition.

Key Results

On the Lend-an-Ear test set reported in the paper, which contains 315 supporter turns across 50 conversations:

  1. Aggregate empathy improves from 3.75 to 4.67 relative to the vanilla Qwen3-4B baseline.
  2. Tactic stickiness decreases from 0.57 to 0.42.
  3. This checkpoint provides the clearest joint gain in empathy and reduced cross-turn repetition.

Training Summary

Method GRPO via VERL
Reward Empathy quality + cross-turn tactic diversity
Base model Qwen/Qwen3-4B
KL coeff 0.01
Diversity weight 1.0
Response length 2048 tokens
Rollouts n=8 per prompt

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("hongli-zhan/MINT-empathy-Qwen3-4B")
tokenizer = AutoTokenizer.from_pretrained("hongli-zhan/MINT-empathy-Qwen3-4B")

With vLLM:

from vllm import LLM
llm = LLM(model="hongli-zhan/MINT-empathy-Qwen3-4B")

Intended Use and Limitations

This model is intended for research on empathic dialogue, discourse diversity, and supportive response generation. It is a research artifact, not a therapy system, and it was evaluated on fixed gold-history conversation contexts rather than fully interactive free-running conversations.

Citation

@article{zhan2026discourse,
  title={Discourse Diversity in Multi-Turn Empathic Dialogue},
  author={Zhan, Hongli and Gueorguieva, Emma S and Hernandez, Javier and Suh, Jina and Ong, Desmond C and Li, Junyi Jessy},
  journal={arXiv preprint arXiv:2604.11742},
  year={2026}
}