初始化项目,由ModelHub XC社区提供模型
Model: Mushari440/Qwen3-8B-SFT-chatml Source: Original Platform
This commit is contained in:
51
README.md
Normal file
51
README.md
Normal file
@@ -0,0 +1,51 @@
|
||||
---
|
||||
library_name: transformers
|
||||
tags:
|
||||
- arabic
|
||||
- sft
|
||||
- qwen
|
||||
---
|
||||
|
||||
# Qwen3-8B-SFT
|
||||
|
||||
## Model Details
|
||||
|
||||
- **Developed by:** Mushari Alothman
|
||||
- **Model type:** Causal Language Model
|
||||
- **Language(s):** Arabic, English
|
||||
- **License:** Apache 2.0
|
||||
- **Finetuned from:** Qwen3-8B-Base
|
||||
|
||||
This is a supervised fine-tuned (SFT) Qwen3-8B model optimized for accurate, clean supervision across Arabic and English tasks.
|
||||
|
||||
## Intended Uses
|
||||
|
||||
### Direct Use
|
||||
- Arabic & English MCQ answering
|
||||
- Context-based QA / RAG
|
||||
- General instruction following
|
||||
|
||||
### Out-of-Scope Use
|
||||
- Safety-critical or real-time decision making
|
||||
- Generating factual guarantees without verification
|
||||
|
||||
## Training Summary
|
||||
|
||||
- **Training type:** Supervised Fine-Tuning (SFT)
|
||||
- **Precision:** bf16 mixed precision
|
||||
- **Data:** Curated Arabic & English datasets including:
|
||||
- MCQ
|
||||
- QA / RAG / context understanding
|
||||
- General instruction data
|
||||
|
||||
## How to Use
|
||||
|
||||
```python
|
||||
from transformers import AutoTokenizer, AutoModelForCausalLM
|
||||
|
||||
tokenizer = AutoTokenizer.from_pretrained("Mushari440/Qwen3-8B-SFT-chatml")
|
||||
model = AutoModelForCausalLM.from_pretrained("Mushari440/Qwen3-8B-SFT-chatml")
|
||||
|
||||
inputs = tokenizer("سؤال: ما عاصمة السعودية؟", return_tensors="pt")
|
||||
outputs = model.generate(**inputs, max_new_tokens=50)
|
||||
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
||||
Reference in New Issue
Block a user