58 lines
1.4 KiB
Markdown
58 lines
1.4 KiB
Markdown
|
|
---
|
||
|
|
pipeline_tag: text-generation
|
||
|
|
library_name: transformers
|
||
|
|
tags:
|
||
|
|
- qwen3
|
||
|
|
- gguf
|
||
|
|
- mandala
|
||
|
|
- education
|
||
|
|
- fine-tuned
|
||
|
|
language:
|
||
|
|
- ko
|
||
|
|
- en
|
||
|
|
license: apache-2.0
|
||
|
|
---
|
||
|
|
|
||
|
|
# Insighta Mandala v13
|
||
|
|
|
||
|
|
Fine-tuned Qwen3-4B model for generating mandala learning plans in JSON format.
|
||
|
|
|
||
|
|
## Model Details
|
||
|
|
|
||
|
|
- **Base model**: Qwen/Qwen3-4B
|
||
|
|
- **Fine-tuning**: LoRA on mandala learning plan generation task
|
||
|
|
- **Languages**: Korean, English
|
||
|
|
- **Output format**: Structured JSON (mandala chart format)
|
||
|
|
|
||
|
|
## Available Quantizations
|
||
|
|
|
||
|
|
| Format | Size | Description |
|
||
|
|
|--------|------|-------------|
|
||
|
|
| `model.safetensors` | ~8GB | Full F16 weights |
|
||
|
|
| `insighta-mandala-v13-Q8_0.gguf` | ~4GB | Q8_0 quantized GGUF |
|
||
|
|
| `insighta-mandala-v13-Q4_K_M.gguf` | ~2.4GB | Q4_K_M quantized GGUF (recommended for CPU) |
|
||
|
|
|
||
|
|
## Usage
|
||
|
|
|
||
|
|
### With llama-cpp-python
|
||
|
|
|
||
|
|
```python
|
||
|
|
from llama_cpp import Llama
|
||
|
|
|
||
|
|
llm = Llama(model_path="insighta-mandala-v13-Q4_K_M.gguf", n_ctx=4096)
|
||
|
|
output = llm(
|
||
|
|
"<|im_start|>user\nTOEFL 100점 만다라트 차트를 만들어줘<|im_end|>\n<|im_start|>assistant\n",
|
||
|
|
max_tokens=2048,
|
||
|
|
temperature=0.7,
|
||
|
|
)
|
||
|
|
print(output["choices"][0]["text"])
|
||
|
|
```
|
||
|
|
|
||
|
|
### With HF Space API
|
||
|
|
|
||
|
|
```bash
|
||
|
|
curl -X POST https://jamesjk4242-insighta-mandala-v13-api.hf.space/api/predict \
|
||
|
|
-H "Content-Type: application/json" \
|
||
|
|
-d '{"data": ["TOEFL 100점 만다라트 차트를 만들어줘", "You are a helpful assistant that generates mandala learning plans in JSON format.", 2048, 0.7, 0.9, true]}'
|
||
|
|
```
|