113 lines
3.3 KiB
Markdown
113 lines
3.3 KiB
Markdown
---
|
|
license: apache-2.0
|
|
language:
|
|
- en
|
|
metrics:
|
|
- code_eval
|
|
- accuracy
|
|
base_model:
|
|
- Qwen/Qwen2.5-Coder-1.5B-Instruct
|
|
new_version: Qwen/Qwen2.5-Coder-1.5B-Instruct
|
|
pipeline_tag: text-generation
|
|
library_name: transformers
|
|
tags:
|
|
- cybersecurity
|
|
- mythos
|
|
- qween
|
|
- qween-security
|
|
- blue
|
|
- team
|
|
- blue-team
|
|
- cve
|
|
- ctf
|
|
- code
|
|
- code-security
|
|
---
|
|
|
|
---
|
|
language:
|
|
- en
|
|
- code
|
|
license: apache-2.0
|
|
tags:
|
|
- security
|
|
- exploit-development
|
|
- vulnerability-research
|
|
- php
|
|
- mybb
|
|
- cve
|
|
- python
|
|
- qwen
|
|
- fine-tuned
|
|
- cybersecurity
|
|
datasets:
|
|
- [your-dataset-name-if-uploaded]
|
|
metrics:
|
|
- accuracy
|
|
- code-eval
|
|
pipeline_tag: text-generation
|
|
library_name: transformers
|
|
base_model: Qwen/Qwen2.5-Coder-1.5B-Instruct
|
|
---
|
|
|
|
# Mythos Engine - Qwen 2.5 Coder 1.5B Security Fine-Tune
|
|
|
|
## 🔥 Model Description
|
|
|
|
Mythos Engine is a specialized fine-tune of **Qwen 2.5 Coder 1.5B Instruct** designed for **cybersecurity research, vulnerability analysis, and exploit development**. It has been trained on a curated dataset of 700+ high-reasoning security examples covering PHP internals, MyBB exploitation, deserialization chains, type juggling, and advanced Python exploit synthesis.
|
|
|
|
The model employs **Chain-of-Thought reasoning with self-correction loops** and mathematical logic notation to produce accurate, production-ready security code.
|
|
|
|
## 🎯 Intended Use
|
|
|
|
- **Security Research**: Analyzing CVEs and understanding exploit mechanics
|
|
- **Red Team Education**: Learning exploit development patterns
|
|
- **Blue Team Defense**: Understanding attack vectors to build better detections
|
|
- **CTF & Training**: Solving complex security challenges
|
|
|
|
**⚠️ Important**: This model is for **educational and authorized security testing only**. Do not use for unauthorized access or malicious purposes.
|
|
|
|
## 🧠 Training Details
|
|
|
|
| Aspect | Details |
|
|
| :--- | :--- |
|
|
| **Base Model** | Qwen/Qwen2.5-Coder-1.5B-Instruct |
|
|
| **Fine-Tuning Method** | QLoRA (4-bit quantization) with Unsloth |
|
|
| **Dataset Size** | 1000+ examples |
|
|
| **Epochs** | 4 |
|
|
| **Learning Rate** | 1e-5 |
|
|
| **Sequence Length** | 4096 |
|
|
| **Final Training Loss** | 2.02 |
|
|
|
|
## 📊 Dataset Composition
|
|
|
|
The training dataset includes:
|
|
|
|
- **40% PHP Vulnerabilities**: Type juggling, deserialization, filter chains, disable_functions bypasses
|
|
- **25% MyBB Exploits**: Admin CP RCE, SQL injection, XSS chains
|
|
- **20% Python Exploit Development**: C2 frameworks, scanners, injection techniques
|
|
- **10% Blue Team Detection**: Sigma/YARA rules, log analysis
|
|
- **5% Cryptographic Attacks**: Timing attacks, padding oracles, hash length extension
|
|
|
|
## 🚀 How to Use
|
|
|
|
```python
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
|
model = AutoModelForCausalLM.from_pretrained(
|
|
"expper/mythos-qwen-1.5b-final",
|
|
device_map="auto",
|
|
torch_dtype="auto"
|
|
)
|
|
tokenizer = AutoTokenizer.from_pretrained("expper/mythos-qwen-1.5b-final")
|
|
|
|
prompt = """<|im_start|>system
|
|
You are Mythos Engine, an elite security AI. Think step-by-step with self-correction.<|im_end|>
|
|
<|im_start|>user
|
|
Explain CVE-2022-43772 (MyBB Admin CP Avatar RCE) and write a PoC.<|im_end|>
|
|
<|im_start|>assistant
|
|
"""
|
|
|
|
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
|
outputs = model.generate(**inputs, max_new_tokens=1024, temperature=0.6)
|
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |