Files

39 lines
2.3 KiB
Markdown
Raw Permalink Normal View History

2024-10-16 09:46:13 +00:00
---
2024-10-16 18:01:15 +08:00
license: apache-2.0
datasets:
- BAAI/IndustryInstruction_Health-Medicine
- BAAI/IndustryInstruction
base_model:
- MonteXiaofeng/CareBot_Medical_multi-llama3-8b-base
tags:
- 医疗对话模型
- 中英文多语种医疗对话模型
- chatmodel
---
2024-10-16 09:46:13 +00:00
2024-10-16 18:01:15 +08:00
This model is trained from the model: MonteXiaofeng/CareBot_Medical_multi-llama3-8b-base, training data is: BAAI/IndustryInstruction_Health-Medicine To enhance the model's ability to follow medical instructions and better adapt to specific medical scenarios, we conduct the supervised fine-tuning. This process involves using conversational-style data (comprising both queries and responses) to finetune the pretrained LLM. In the following sections, we will explore the details of data construction and training methods.
2024-10-16 09:46:13 +00:00
2024-10-16 18:01:15 +08:00
## Data Construction
2024-10-16 09:46:13 +00:00
2024-10-16 18:01:15 +08:00
Our SFT dataset comprises a diverse array of question types, including multiple-choice questions from medical exams, single-turn disease diagnoses, and multi-turn health consultations. It integrates data from seven publicly available sources: Chinese Medical Dialogue Data\footnote{https://github.com/Toyhom/Chinese-medical-dialogue-data}, Huatuo26M , MedDialog , ChatMed Consult Dataset , ChatDoctor , CMB\footnote{https://github.com/FreedomIntelligence/CMB}, and MedQA . We preserve portions of authentic doctor-patient conversations and augment the dataset by rewriting the remaining content. For these rewrites, we use real-world medical scenarios as prompts and generate responses via GPT-4. We believe this ensures the diversity of the SFT dataset, which can help the CareBot better adapt to different types of medical problems and patient situations, thereby improving its performance in a variety of scenarios.
2024-10-16 09:46:13 +00:00
2024-10-16 18:01:15 +08:00
## evaluation
2024-10-16 09:46:13 +00:00
2024-10-16 18:01:15 +08:00
evaluation on benchmark is bellow.
![image/png](https://cdn-uploads.huggingface.co/production/uploads/642f6c64f945a8a5c9ee5b5d/kqvLfcFtkw6lHcHtCySLr.png)
2024-10-16 09:46:13 +00:00
2024-10-16 18:01:15 +08:00
![image/png](https://cdn-uploads.huggingface.co/production/uploads/642f6c64f945a8a5c9ee5b5d/UiokfV8qcYEyCWEa__820.png)
gsb result with other medical LLMS
![image/png](https://cdn-uploads.huggingface.co/production/uploads/642f6c64f945a8a5c9ee5b5d/rOnnIoY9MaXPTFD_R10r1.png)
2026-01-11 12:58:16 +00:00
# Acknowledgements
This work is supported by the National Science and Technology Major Project (No. 2022ZD0116314).
本项目受新一代人工智能国家科技重大专项No. 2022ZD0116314支持。