初始化项目,由ModelHub XC社区提供模型

Model: pratinavseth/cricket-captain-qwen3-06b-merged
Source: Original Platform
This commit is contained in:
ModelHub XC
2026-05-02 05:10:32 +08:00
commit 30256fefe0
8 changed files with 297 additions and 0 deletions

31
README.md Normal file
View File

@@ -0,0 +1,31 @@
---
base_model: Qwen/Qwen3-0.6B
library_name: transformers
tags:
- generated_from_trainer
- trl
- grpo
- cricket
- merged
license: mit
pipeline_tag: text-generation
---
# cricket-captain-qwen3-06b-merged
`Qwen/Qwen3-0.6B` with the [`pratinavseth/cricket-captain-qwen3-06b-stage2`](https://huggingface.co/pratinavseth/cricket-captain-qwen3-06b-stage2) LoRA adapter (stage 2 GRPO, step 50) **merged into the base weights**. Single-file model, ready to load without PEFT — usable directly via `transformers`, `vllm`, or TGI.
- **Live Space:** https://huggingface.co/spaces/pratinavseth/cricket-captain-llm
- **Source code:** https://github.com/pratinavseth/cricket-captain-llm
- **Adapter (un-merged):** [`pratinavseth/cricket-captain-qwen3-06b-stage2`](https://huggingface.co/pratinavseth/cricket-captain-qwen3-06b-stage2)
- **W&B run:** https://wandb.ai/ptnv-s-research/huggingface
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
tok = AutoTokenizer.from_pretrained("pratinavseth/cricket-captain-qwen3-06b-merged")
model = AutoModelForCausalLM.from_pretrained("pratinavseth/cricket-captain-qwen3-06b-merged", torch_dtype="bfloat16", device_map="auto")
```
The model expects the cricket-captain prompt schema produced by the OpenEnv environment in this repo (see `inference.py` for prompt construction).