license, base_model, tags, language, pipeline_tag
license base_model tags language pipeline_tag
mit Qwen/Qwen3-1.7B
empathy
reinforcement-learning
grpo
dialogue
mint
emotional-support
en
text-generation

MINT-empathy-Qwen3-1.7B

This model is the Q + D_KL MINT checkpoint fine-tuned from Qwen/Qwen3-1.7B for multi-turn empathic dialogue.

MINT, short for Multi-turn Inter-tactic Novelty Training, is a reinforcement learning framework that optimizes empathic response quality together with cross-turn discourse-move novelty. At the 1.7B scale, this checkpoint provides the strongest aggregate empathy performance among the MINT variants reported in the paper.

Key Results

On the Lend-an-Ear test set reported in the paper, which contains 315 supporter turns across 50 conversations:

  1. Aggregate empathy improves from 3.60 to 4.54 relative to the vanilla Qwen3-1.7B baseline.
  2. In Table 2 of the paper, tactic stickiness is 0.51 for both the vanilla baseline and this checkpoint.
  3. This checkpoint should therefore be understood primarily as the strongest 1.7B empathy checkpoint, while the clearest reduction in tactic repetition appears at the 4B scale.

Training Summary

Method GRPO via VERL
Reward Empathy quality + cross-turn tactic diversity
Base model Qwen/Qwen3-1.7B
KL coeff 0.01
Diversity weight 1.0
Response length 2048 tokens
Rollouts n=8 per prompt

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("hongli-zhan/MINT-empathy-Qwen3-1.7B")
tokenizer = AutoTokenizer.from_pretrained("hongli-zhan/MINT-empathy-Qwen3-1.7B")

With vLLM:

from vllm import LLM
llm = LLM(model="hongli-zhan/MINT-empathy-Qwen3-1.7B")

Intended Use and Limitations

This model is intended for research on empathic dialogue, discourse diversity, and supportive response generation. It is a research artifact, not a therapy system, and at the 1.7B scale the main improvement is empathy quality rather than a substantial reduction in tactic repetition.

Citation

@article{zhan2026discourse,
  title={Discourse Diversity in Multi-Turn Empathic Dialogue},
  author={Zhan, Hongli and Gueorguieva, Emma S and Hernandez, Javier and Suh, Jina and Ong, Desmond C and Li, Junyi Jessy},
  journal={arXiv preprint arXiv:2604.11742},
  year={2026}
}
Description
Model synced from source: hongli-zhan/MINT-empathy-Qwen3-1.7B
Readme 2 MiB
Languages
Jinja 100%