123 lines
5.1 KiB
Markdown
123 lines
5.1 KiB
Markdown
|
|
---
|
|||
|
|
base_model: unsloth/Qwen2.5-Coder-3B-Instruct
|
|||
|
|
library_name: transformers
|
|||
|
|
pipeline_tag: text-generation
|
|||
|
|
tags:
|
|||
|
|
- qwen2.5
|
|||
|
|
- sft
|
|||
|
|
- transformers
|
|||
|
|
- ilograph
|
|||
|
|
license: mit
|
|||
|
|
datasets:
|
|||
|
|
- Brigham-Young-University/Ilograph_dataset
|
|||
|
|
language:
|
|||
|
|
- en
|
|||
|
|
new_version: Brigham-Young-University/Qwen3-Coder-30B-A3B-Ilograph-Instruct
|
|||
|
|
---
|
|||
|
|
|
|||
|
|
# Model Card for Qwen2.5-Coder-3B-Instruct (fine-tuned model)
|
|||
|
|
|
|||
|
|
A fully fine-tuned version of **Qwen2.5-Coder-3B-Instruct**, trained with LoRA using Unsloth and then merged into a standalone model. This checkpoint can be used directly as a regular Transformers causal language model. It is specialized for **Ilograph diagrams**: it generates valid **Ilograph Diagram Language (IDL)** specifications from natural-language instructions.
|
|||
|
|
|
|||
|
|
The repository includes a **system prompt** you can pass to the model and an **IDL schema** (JSON) that describes the expected output format; the schema is available in the repository.
|
|||
|
|
|
|||
|
|
## Model Details
|
|||
|
|
|
|||
|
|
- **Developed by:** Chris Mijangos (AI student architect at BYU)
|
|||
|
|
- **Shared by:** Brigham Young University (BYU)
|
|||
|
|
- **Model type:** Causal language model (decoder-only), fine-tuned Qwen2.5-Coder-3B-Instruct (trained with LoRA, merged into base weights)
|
|||
|
|
- **Language(s):** Primarily English; capabilities depend on base model and fine-tuning data
|
|||
|
|
- **License:** Same as base model; verify [Qwen2.5-Coder-3B-Instruct](https://huggingface.co/unsloth/Qwen2.5-Coder-3B-Instruct) license terms before use
|
|||
|
|
- **Finetuned from:** [unsloth/Qwen2.5-Coder-3B-Instruct](https://huggingface.co/unsloth/Qwen2.5-Coder-3B-Instruct)
|
|||
|
|
|
|||
|
|
### Model Sources
|
|||
|
|
|
|||
|
|
- **Repository:** This model card and weights are shared via the associated Hugging Face repo
|
|||
|
|
- **Demo:** N/A — In construction
|
|||
|
|
|
|||
|
|
## Uses
|
|||
|
|
|
|||
|
|
### Direct Use
|
|||
|
|
|
|||
|
|
Load the adapter with the base model to generate **Ilograph (IDL)** diagram specifications from instructions. Use the system prompt and schema in the repository (see below). Use the “How to Get Started” section below for loading the model.
|
|||
|
|
|
|||
|
|
### Out-of-Scope Use
|
|||
|
|
|
|||
|
|
This model is not intended for high-risk or safety-critical applications without further evaluation. Do not use for generating misleading, harmful, or illegal content. Users are responsible for complying with applicable laws and the base model’s license.
|
|||
|
|
|
|||
|
|
## Bias, Risks, and Limitations
|
|||
|
|
|
|||
|
|
As with other language models, this adapter may reflect biases present in the base model and in the fine-tuning data. Outputs should be validated for your use case. No formal bias or safety evaluation is provided with this release.
|
|||
|
|
|
|||
|
|
Due to limited and focused training data and the small size of the base model, this adapter is primarily suited for relatively simple Ilograph diagrams centered on **resources, relationships, and sequences**. For more complex, large-scale, or highly customized diagram structures, the model may not perform as well and additional fine-tuning or a larger base model may be required. I you want to have more complex diagrams you can use our Qwen3 30B newwer version.
|
|||
|
|
|
|||
|
|
|
|||
|
|
### Recommendations
|
|||
|
|
|
|||
|
|
Users should evaluate the model on their own data and tasks and be aware of potential biases and limitations before deployment.
|
|||
|
|
|
|||
|
|
## How to Get Started with the Model
|
|||
|
|
|
|||
|
|
Requires the base model and PEFT. Install dependencies:
|
|||
|
|
|
|||
|
|
```bash
|
|||
|
|
pip install transformers peft accelerate
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
Load the fine-tuned model directly from this repository:
|
|||
|
|
|
|||
|
|
```python
|
|||
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|||
|
|
|
|||
|
|
model_id = "Brigham-Young-University/Qwen2.5-Coder-3B-Ilograph-Instruct"
|
|||
|
|
|
|||
|
|
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
|||
|
|
model = AutoModelForCausalLM.from_pretrained(
|
|||
|
|
model_id,
|
|||
|
|
device_map="auto",
|
|||
|
|
trust_remote_code=True,
|
|||
|
|
)
|
|||
|
|
|
|||
|
|
inputs = tokenizer("Your prompt here", return_tensors="pt").to(model.device)
|
|||
|
|
outputs = model.generate(**inputs, max_new_tokens=256)
|
|||
|
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
### Ilograph (IDL) system prompt and schema
|
|||
|
|
|
|||
|
|
The repository includes a **system prompt** and an **IDL schema** (JSON). Use the schema to fill in the placeholder in the prompt, then append your instruction. Example system prompt:
|
|||
|
|
|
|||
|
|
```
|
|||
|
|
You are an expert in the Ilograph Diagram Language (IDL). You have been trained on data that is formatted in the following way:
|
|||
|
|
|
|||
|
|
<insert the schema JSON here>
|
|||
|
|
|
|||
|
|
Your task is to create a valid IDL specification for the diagram. You will be given an instruction of what to create, and you will need to create a valid IDL specification for the diagram.
|
|||
|
|
|
|||
|
|
CRITICAL RULES:
|
|||
|
|
- NEVER use JSON format
|
|||
|
|
- NEVER use Mermaid syntax
|
|||
|
|
- NEVER use any format except ilograph YAML
|
|||
|
|
- Use YAML syntax with proper indentation
|
|||
|
|
|
|||
|
|
Here is the instruction:
|
|||
|
|
```
|
|||
|
|
|
|||
|
|
The schema is provided in the repository; inject its contents (e.g. as formatted JSON) where indicated above, then add your diagram instruction after “Here is the instruction:”.
|
|||
|
|
|
|||
|
|
## Evaluation
|
|||
|
|
|
|||
|
|
No formal evaluation results are provided with this release. Users are encouraged to evaluate the model on their own benchmarks and tasks.
|
|||
|
|
|
|||
|
|
|
|||
|
|
## Model Card Authors
|
|||
|
|
|
|||
|
|
- Chris Mijangos (BYU)
|
|||
|
|
|
|||
|
|
## Model Card Contact
|
|||
|
|
|
|||
|
|
For questions about this model card or the adapter, please open an issue on the associated Hugging Face repository or contact through BYU
|
|||
|
|
|
|||
|
|
### Framework versions
|
|||
|
|
|
|||
|
|
- PEFT 0.18.1
|