初始化项目,由ModelHub XC社区提供模型
Model: expper/mythos-qwen-1.5b-final Source: Original Platform
This commit is contained in:
113
README.md
Normal file
113
README.md
Normal file
@@ -0,0 +1,113 @@
|
||||
---
|
||||
license: apache-2.0
|
||||
language:
|
||||
- en
|
||||
metrics:
|
||||
- code_eval
|
||||
- accuracy
|
||||
base_model:
|
||||
- Qwen/Qwen2.5-Coder-1.5B-Instruct
|
||||
new_version: Qwen/Qwen2.5-Coder-1.5B-Instruct
|
||||
pipeline_tag: text-generation
|
||||
library_name: transformers
|
||||
tags:
|
||||
- cybersecurity
|
||||
- mythos
|
||||
- qween
|
||||
- qween-security
|
||||
- blue
|
||||
- team
|
||||
- blue-team
|
||||
- cve
|
||||
- ctf
|
||||
- code
|
||||
- code-security
|
||||
---
|
||||
|
||||
---
|
||||
language:
|
||||
- en
|
||||
- code
|
||||
license: apache-2.0
|
||||
tags:
|
||||
- security
|
||||
- exploit-development
|
||||
- vulnerability-research
|
||||
- php
|
||||
- mybb
|
||||
- cve
|
||||
- python
|
||||
- qwen
|
||||
- fine-tuned
|
||||
- cybersecurity
|
||||
datasets:
|
||||
- [your-dataset-name-if-uploaded]
|
||||
metrics:
|
||||
- accuracy
|
||||
- code-eval
|
||||
pipeline_tag: text-generation
|
||||
library_name: transformers
|
||||
base_model: Qwen/Qwen2.5-Coder-1.5B-Instruct
|
||||
---
|
||||
|
||||
# Mythos Engine - Qwen 2.5 Coder 1.5B Security Fine-Tune
|
||||
|
||||
## 🔥 Model Description
|
||||
|
||||
Mythos Engine is a specialized fine-tune of **Qwen 2.5 Coder 1.5B Instruct** designed for **cybersecurity research, vulnerability analysis, and exploit development**. It has been trained on a curated dataset of 700+ high-reasoning security examples covering PHP internals, MyBB exploitation, deserialization chains, type juggling, and advanced Python exploit synthesis.
|
||||
|
||||
The model employs **Chain-of-Thought reasoning with self-correction loops** and mathematical logic notation to produce accurate, production-ready security code.
|
||||
|
||||
## 🎯 Intended Use
|
||||
|
||||
- **Security Research**: Analyzing CVEs and understanding exploit mechanics
|
||||
- **Red Team Education**: Learning exploit development patterns
|
||||
- **Blue Team Defense**: Understanding attack vectors to build better detections
|
||||
- **CTF & Training**: Solving complex security challenges
|
||||
|
||||
**⚠️ Important**: This model is for **educational and authorized security testing only**. Do not use for unauthorized access or malicious purposes.
|
||||
|
||||
## 🧠 Training Details
|
||||
|
||||
| Aspect | Details |
|
||||
| :--- | :--- |
|
||||
| **Base Model** | Qwen/Qwen2.5-Coder-1.5B-Instruct |
|
||||
| **Fine-Tuning Method** | QLoRA (4-bit quantization) with Unsloth |
|
||||
| **Dataset Size** | 1000+ examples |
|
||||
| **Epochs** | 4 |
|
||||
| **Learning Rate** | 1e-5 |
|
||||
| **Sequence Length** | 4096 |
|
||||
| **Final Training Loss** | 2.02 |
|
||||
|
||||
## 📊 Dataset Composition
|
||||
|
||||
The training dataset includes:
|
||||
|
||||
- **40% PHP Vulnerabilities**: Type juggling, deserialization, filter chains, disable_functions bypasses
|
||||
- **25% MyBB Exploits**: Admin CP RCE, SQL injection, XSS chains
|
||||
- **20% Python Exploit Development**: C2 frameworks, scanners, injection techniques
|
||||
- **10% Blue Team Detection**: Sigma/YARA rules, log analysis
|
||||
- **5% Cryptographic Attacks**: Timing attacks, padding oracles, hash length extension
|
||||
|
||||
## 🚀 How to Use
|
||||
|
||||
```python
|
||||
from transformers import AutoModelForCausalLM, AutoTokenizer
|
||||
|
||||
model = AutoModelForCausalLM.from_pretrained(
|
||||
"expper/mythos-qwen-1.5b-final",
|
||||
device_map="auto",
|
||||
torch_dtype="auto"
|
||||
)
|
||||
tokenizer = AutoTokenizer.from_pretrained("expper/mythos-qwen-1.5b-final")
|
||||
|
||||
prompt = """<|im_start|>system
|
||||
You are Mythos Engine, an elite security AI. Think step-by-step with self-correction.<|im_end|>
|
||||
<|im_start|>user
|
||||
Explain CVE-2022-43772 (MyBB Admin CP Avatar RCE) and write a PoC.<|im_end|>
|
||||
<|im_start|>assistant
|
||||
"""
|
||||
|
||||
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
||||
outputs = model.generate(**inputs, max_new_tokens=1024, temperature=0.6)
|
||||
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
||||
Reference in New Issue
Block a user