Model: pratinavseth/cricket-captain-qwen3-06b-merged Source: Original Platform
base_model, library_name, tags, license, pipeline_tag
| base_model | library_name | tags | license | pipeline_tag | |||||
|---|---|---|---|---|---|---|---|---|---|
| Qwen/Qwen3-0.6B | transformers |
|
mit | text-generation |
cricket-captain-qwen3-06b-merged
Qwen/Qwen3-0.6B with the pratinavseth/cricket-captain-qwen3-06b-stage2 LoRA adapter (stage 2 GRPO, step 50) merged into the base weights. Single-file model, ready to load without PEFT — usable directly via transformers, vllm, or TGI.
- Live Space: https://huggingface.co/spaces/pratinavseth/cricket-captain-llm
- Source code: https://github.com/pratinavseth/cricket-captain-llm
- Adapter (un-merged):
pratinavseth/cricket-captain-qwen3-06b-stage2 - W&B run: https://wandb.ai/ptnv-s-research/huggingface
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
tok = AutoTokenizer.from_pretrained("pratinavseth/cricket-captain-qwen3-06b-merged")
model = AutoModelForCausalLM.from_pretrained("pratinavseth/cricket-captain-qwen3-06b-merged", torch_dtype="bfloat16", device_map="auto")
The model expects the cricket-captain prompt schema produced by the OpenEnv environment in this repo (see inference.py for prompt construction).
Description
Languages
Jinja
100%