3.3 KiB
3.3 KiB
license, language, metrics, base_model, new_version, pipeline_tag, library_name, tags
| license | language | metrics | base_model | new_version | pipeline_tag | library_name | tags | |||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| apache-2.0 |
|
|
|
Qwen/Qwen2.5-Coder-1.5B-Instruct | text-generation | transformers |
|
language:
- en
- code license: apache-2.0 tags:
- security
- exploit-development
- vulnerability-research
- php
- mybb
- cve
- python
- qwen
- fine-tuned
- cybersecurity datasets:
- [your-dataset-name-if-uploaded] metrics:
- accuracy
- code-eval pipeline_tag: text-generation library_name: transformers base_model: Qwen/Qwen2.5-Coder-1.5B-Instruct
Mythos Engine - Qwen 2.5 Coder 1.5B Security Fine-Tune
🔥 Model Description
Mythos Engine is a specialized fine-tune of Qwen 2.5 Coder 1.5B Instruct designed for cybersecurity research, vulnerability analysis, and exploit development. It has been trained on a curated dataset of 700+ high-reasoning security examples covering PHP internals, MyBB exploitation, deserialization chains, type juggling, and advanced Python exploit synthesis.
The model employs Chain-of-Thought reasoning with self-correction loops and mathematical logic notation to produce accurate, production-ready security code.
🎯 Intended Use
- Security Research: Analyzing CVEs and understanding exploit mechanics
- Red Team Education: Learning exploit development patterns
- Blue Team Defense: Understanding attack vectors to build better detections
- CTF & Training: Solving complex security challenges
⚠️ Important: This model is for educational and authorized security testing only. Do not use for unauthorized access or malicious purposes.
🧠 Training Details
| Aspect | Details |
|---|---|
| Base Model | Qwen/Qwen2.5-Coder-1.5B-Instruct |
| Fine-Tuning Method | QLoRA (4-bit quantization) with Unsloth |
| Dataset Size | 1000+ examples |
| Epochs | 4 |
| Learning Rate | 1e-5 |
| Sequence Length | 4096 |
| Final Training Loss | 2.02 |
📊 Dataset Composition
The training dataset includes:
- 40% PHP Vulnerabilities: Type juggling, deserialization, filter chains, disable_functions bypasses
- 25% MyBB Exploits: Admin CP RCE, SQL injection, XSS chains
- 20% Python Exploit Development: C2 frameworks, scanners, injection techniques
- 10% Blue Team Detection: Sigma/YARA rules, log analysis
- 5% Cryptographic Attacks: Timing attacks, padding oracles, hash length extension
🚀 How to Use
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained(
"expper/mythos-qwen-1.5b-final",
device_map="auto",
torch_dtype="auto"
)
tokenizer = AutoTokenizer.from_pretrained("expper/mythos-qwen-1.5b-final")
prompt = """<|im_start|>system
You are Mythos Engine, an elite security AI. Think step-by-step with self-correction.<|im_end|>
<|im_start|>user
Explain CVE-2022-43772 (MyBB Admin CP Avatar RCE) and write a PoC.<|im_end|>
<|im_start|>assistant
"""
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=1024, temperature=0.6)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))