104 lines
2.6 KiB
Markdown
104 lines
2.6 KiB
Markdown
|
|
---
|
||
|
|
license: apache-2.0
|
||
|
|
base_model: qwen3-0.6B
|
||
|
|
tags:
|
||
|
|
- code-generation
|
||
|
|
- svg
|
||
|
|
- fine-tuned
|
||
|
|
- fp16
|
||
|
|
- vllm
|
||
|
|
- merged
|
||
|
|
language:
|
||
|
|
- en
|
||
|
|
pipeline_tag: text-generation
|
||
|
|
library_name: transformers
|
||
|
|
model_type: qwen
|
||
|
|
inference: true
|
||
|
|
torch_dtype: float16
|
||
|
|
widget:
|
||
|
|
- example_title: "Simple Circle"
|
||
|
|
text: "Create a red circle"
|
||
|
|
- example_title: "Rectangle with Border"
|
||
|
|
text: "Draw a blue rectangle with black border"
|
||
|
|
- example_title: "Complex Shape"
|
||
|
|
text: "Generate a star with 5 points in yellow"
|
||
|
|
---
|
||
|
|
|
||
|
|
# SVG Code Generator
|
||
|
|
|
||
|
|
This is a fine-tuned model for generating SVG code from natural language descriptions. The model has been merged with the base model weights and optimized in fp16 format.
|
||
|
|
|
||
|
|
## Model Details
|
||
|
|
|
||
|
|
- **Model Name**: model_v15
|
||
|
|
- **Base Model**: qwen3-0.6B
|
||
|
|
- **Training Method**: Fine-tuning with merged weights
|
||
|
|
- **Task**: Text-to-SVG code generation
|
||
|
|
- **Model Type**: Merged Qwen model
|
||
|
|
- **Precision**: fp16
|
||
|
|
- **Library**: Transformers, vLLM compatible
|
||
|
|
- **Format**: Merged model (not adapter-based)
|
||
|
|
|
||
|
|
## Usage
|
||
|
|
|
||
|
|
### With Transformers
|
||
|
|
|
||
|
|
Load the model directly using the transformers library:
|
||
|
|
|
||
|
|
```python
|
||
|
|
# Load base model and tokenizer
|
||
|
|
from transformers import AutoTokenizer, AutoModelForCausalLM
|
||
|
|
|
||
|
|
tokenizer = AutoTokenizer.from_pretrained("vinoku89/svg-code-generator")
|
||
|
|
model = AutoModelForCausalLM.from_pretrained("vinoku89/svg-code-generator")
|
||
|
|
|
||
|
|
|
||
|
|
# Generate SVG code
|
||
|
|
prompt = "Create a blue circle with radius 50"
|
||
|
|
inputs = tokenizer(prompt, return_tensors="pt")
|
||
|
|
|
||
|
|
# Generate with parameters
|
||
|
|
outputs = model.generate(
|
||
|
|
**inputs,
|
||
|
|
max_length=200,
|
||
|
|
temperature=0.7,
|
||
|
|
do_sample=True,
|
||
|
|
pad_token_id=tokenizer.eos_token_id
|
||
|
|
)
|
||
|
|
|
||
|
|
# Decode the generated SVG code
|
||
|
|
generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
||
|
|
svg_code = generated_text[len(prompt):].strip()
|
||
|
|
|
||
|
|
print("Generated SVG:")
|
||
|
|
print(svg_code)
|
||
|
|
```
|
||
|
|
|
||
|
|
### With vLLM
|
||
|
|
|
||
|
|
This model supports vLLM for high-performance inference in fp16 format.
|
||
|
|
|
||
|
|
## Training Data
|
||
|
|
|
||
|
|
The model was trained on SVG code generation tasks with natural language descriptions.
|
||
|
|
|
||
|
|
## Intended Use
|
||
|
|
|
||
|
|
This model is designed to generate SVG code from text descriptions for educational and creative purposes.
|
||
|
|
|
||
|
|
## Limitations
|
||
|
|
|
||
|
|
- Generated SVG may require validation
|
||
|
|
- Performance depends on prompt clarity
|
||
|
|
- Limited to SVG syntax and features seen during training
|
||
|
|
|
||
|
|
## Model Performance
|
||
|
|
|
||
|
|
The model has been fine-tuned specifically for SVG generation tasks with merged weights for optimal performance.
|
||
|
|
|
||
|
|
## Technical Details
|
||
|
|
|
||
|
|
- **Precision**: fp16 for memory efficiency
|
||
|
|
- **Compatibility**: vLLM supported for high-throughput inference
|
||
|
|
- **Architecture**: Merged fine-tuned weights (no adapters required)
|