初始化项目,由ModelHub XC社区提供模型
Model: jamesjk4242/insighta-mandala-v13 Source: Original Platform
This commit is contained in:
57
README.md
Normal file
57
README.md
Normal file
@@ -0,0 +1,57 @@
|
||||
---
|
||||
pipeline_tag: text-generation
|
||||
library_name: transformers
|
||||
tags:
|
||||
- qwen3
|
||||
- gguf
|
||||
- mandala
|
||||
- education
|
||||
- fine-tuned
|
||||
language:
|
||||
- ko
|
||||
- en
|
||||
license: apache-2.0
|
||||
---
|
||||
|
||||
# Insighta Mandala v13
|
||||
|
||||
Fine-tuned Qwen3-4B model for generating mandala learning plans in JSON format.
|
||||
|
||||
## Model Details
|
||||
|
||||
- **Base model**: Qwen/Qwen3-4B
|
||||
- **Fine-tuning**: LoRA on mandala learning plan generation task
|
||||
- **Languages**: Korean, English
|
||||
- **Output format**: Structured JSON (mandala chart format)
|
||||
|
||||
## Available Quantizations
|
||||
|
||||
| Format | Size | Description |
|
||||
|--------|------|-------------|
|
||||
| `model.safetensors` | ~8GB | Full F16 weights |
|
||||
| `insighta-mandala-v13-Q8_0.gguf` | ~4GB | Q8_0 quantized GGUF |
|
||||
| `insighta-mandala-v13-Q4_K_M.gguf` | ~2.4GB | Q4_K_M quantized GGUF (recommended for CPU) |
|
||||
|
||||
## Usage
|
||||
|
||||
### With llama-cpp-python
|
||||
|
||||
```python
|
||||
from llama_cpp import Llama
|
||||
|
||||
llm = Llama(model_path="insighta-mandala-v13-Q4_K_M.gguf", n_ctx=4096)
|
||||
output = llm(
|
||||
"<|im_start|>user\nTOEFL 100점 만다라트 차트를 만들어줘<|im_end|>\n<|im_start|>assistant\n",
|
||||
max_tokens=2048,
|
||||
temperature=0.7,
|
||||
)
|
||||
print(output["choices"][0]["text"])
|
||||
```
|
||||
|
||||
### With HF Space API
|
||||
|
||||
```bash
|
||||
curl -X POST https://jamesjk4242-insighta-mandala-v13-api.hf.space/api/predict \
|
||||
-H "Content-Type: application/json" \
|
||||
-d '{"data": ["TOEFL 100점 만다라트 차트를 만들어줘", "You are a helpful assistant that generates mandala learning plans in JSON format.", 2048, 0.7, 0.9, true]}'
|
||||
```
|
||||
Reference in New Issue
Block a user