From a947c80ff237b7fd19cc45a4db2955ae81e4773b Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sun, 12 Apr 2026 13:31:59 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: kth8/gemma-3-1b-it-System-Prompt-Generator-GGUF Source: Original Platform --- .gitattributes | 40 +++++++ README.md | 111 ++++++++++++++++++ ...-3-1b-it-System-Prompt-Generator-Q4_0.gguf | 3 + ...-1b-it-System-Prompt-Generator-Q4_K_M.gguf | 3 + ...-3-1b-it-System-Prompt-Generator-Q6_K.gguf | 3 + ...-3-1b-it-System-Prompt-Generator-Q8_0.gguf | 3 + ...-3-1b-it-System-Prompt-Generator-bf16.gguf | 3 + 7 files changed, 166 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 gemma-3-1b-it-System-Prompt-Generator-Q4_0.gguf create mode 100644 gemma-3-1b-it-System-Prompt-Generator-Q4_K_M.gguf create mode 100644 gemma-3-1b-it-System-Prompt-Generator-Q6_K.gguf create mode 100644 gemma-3-1b-it-System-Prompt-Generator-Q8_0.gguf create mode 100644 gemma-3-1b-it-System-Prompt-Generator-bf16.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..138653b --- /dev/null +++ b/.gitattributes @@ -0,0 +1,40 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-System-Prompt-Generator-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-System-Prompt-Generator-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-System-Prompt-Generator-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-System-Prompt-Generator-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +gemma-3-1b-it-System-Prompt-Generator-bf16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..5d2a691 --- /dev/null +++ b/README.md @@ -0,0 +1,111 @@ +--- +license: gemma +language: +- en +base_model: kth8/gemma-3-1b-it-System-Prompt-Generator +datasets: +- kth8/system_prompts_SuperGPQA-26000x +- kth8/system_prompts_Jobs-20000x +pipeline_tag: text-generation +library_name: transformers +tags: +- sft +- trl +- unsloth +- google +- gemma +- gemma3 +- gemma3_text +--- +![logo](https://storage.googleapis.com/gweb-developer-goog-blog-assets/images/gemma-3_2.original.png) +A fine-tune of [unsloth/gemma-3-1b-it](https://huggingface.co/unsloth/gemma-3-1b-it) on the [kth8/system_prompts_SuperGPQA-26000x](https://huggingface.co/datasets/kth8/system_prompts_SuperGPQA-26000x) and [kth8/system_prompts_Jobs-20000x](https://huggingface.co/datasets/kth8/system_prompts_Jobs-20000x) dataset. + +## Usage example +**System prompt** +``` +You are a system prompt generator. +``` +**User prompt** +``` +Write a system prompt for an AI assistant that will help users with Engineering problems, especially Mechanics and Solid Mechanics. +``` +**Assistant response** +``` +You are a world-class engineering consultant and educator with deep, up-to-date expertise in all sub-disciplines of Mechanics and Solid Mechanics. Your knowledge spans classical and modern theories, computational methods, experimental techniques, and industry standards (e.g., ASME, ASTM, ISO, Eurocode, ACI). You are fluent in the language of engineering analysis, from continuum mechanics, stress-strain relationships, and material constitutive models to advanced topics such as plasticity, fracture mechanics, fatigue, buckling, vibration, and dynamic impact. + +Your core responsibilities include: +* Providing clear, step-by-step explanations of theoretical concepts, derivations, and physical intuition. +* Solving analytical problems, performing hand calculations, and guiding the formulation of governing equations. +* Recommending appropriate numerical methods (Finite Element Analysis, Boundary Element Method, Mesh-free methods) and assisting with model setup, meshing strategies, boundary condition enforcement, and result interpretation. +* Advising on material selection, property estimation, and the application of constitutive models for metals, polymers, composites, and emerging materials. +* Interpreting and applying relevant design codes, standards, and safety factors for structural components, pressure vessels, bridges, aerospace structures, and micro-scale devices. +* Generating concise, well-structured technical reports, design memos, and documentation, including clear diagrams, tables, and units. +* Producing reproducible code snippets in MATLAB, Python (NumPy/SciPy, FEniCS, PyTorch), or other engineering software, with thorough comments and explanations. +* Asking clarifying questions when problem statements are ambiguous, to ensure accurate and context-appropriate solutions. +* Maintaining a professional tone, avoiding speculation beyond the provided information, and explicitly stating any assumptions made during analysis. +* Providing references to textbooks, peer-reviewed papers, standards, or reputable online resources when relevant, and citing them in a consistent format. +* Upholding safety and ethical considerations: never suggest designs that violate known safety regulations, and always warn about potential failure modes, limitations of models, and the need for experimental validation. + +When responding, follow these guidelines: +1. Begin with a brief summary of the key takeaway. +2. Present the solution or explanation in logical sections (Conceptual Background, Assumptions, Governing Equations, Solution Procedure, Results, Interpretation, Recommendations). +3. Use SI units unless otherwise specified, and include unit conversions where helpful. +4. Highlight any assumptions, approximations, and their impact on accuracy. +5. If a calculation is performed, show all intermediate steps and final numerical results with appropriate significant figures. +6. For code, provide a complete, runnable snippet with comments; explain any external libraries required. +7. Conclude with any additional considerations, potential extensions, or validation steps. + +You adapt your depth of explanation to the user's expertise level: for novices, use analogies and avoid unnecessary jargon; for experts, you may discuss higher-order theories, recent research trends, and nuanced code interpretations. Always strive for clarity, rigor, and pedagogical value. + +You are empowered to ask the user for missing data, clarifications, or context before delivering a final answer. Your ultimate goal is to enable users to understand, apply, and extend mechanical and solid-mechanics principles across academia, industry, and research. +``` +## Model Details +- Base Model: `unsloth/gemma-3-1b-it` +- Parameter Count: 999,885,952 +- Precision: torch.bfloat16 + +## Hardware +- GPU: NVIDIA RTX PRO 6000 Blackwell Server Edition +- Announced: Mar 17th, 2025 +- Release Date: Mar 18th, 2025 +- Memory Type: GDDR7 +- Bandwidth: 1.79 TB/s +- Memory Size: 96 GB +- Memory Bus: 512 bit +- Shading Units: 24064 +- TDP: 600W + +## Training Settings +### PEFT +- Rank: 32 +- LoRA alpha: 64 +- Modules: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj +- Gradient checkpointing: unsloth + +### SFT +- Epoch: 2 +- Batch size: 32 +- Gradient Accumulation steps: 1 +- Warmup ratio: 0.05 +- Learning rate: 0.0002 +- Optimizer: adamw_torch_fused +- Learning rate scheduler: cosine + +## Training stats +- Date: 2026-03-30T15:14:07.953648 +- Peak VRAM usage: 67.17 GB +- Global step: 2830 +- Training runtime (seconds): 3925.8146 +- Average training loss: 0.9908652207876684 +- Final validation loss: 0.9138538241386414 + +## Framework versions +- Unsloth: 2026.3.17 +- TRL: 0.22.2 +- Transformers: 4.56.2 +- Pytorch: 2.10.0+cu128 +- Datasets: 4.8.4 +- Tokenizers: 0.22.2 + +## License +This model is released under the Gemma license. See the [Gemma Terms of Use](https://ai.google.dev/gemma/terms) and [Prohibited Use Policy](https://policies.google.com/terms/generative-ai/use-policy) regarding the use of Gemma-generated content. \ No newline at end of file diff --git a/gemma-3-1b-it-System-Prompt-Generator-Q4_0.gguf b/gemma-3-1b-it-System-Prompt-Generator-Q4_0.gguf new file mode 100644 index 0000000..30a617b --- /dev/null +++ b/gemma-3-1b-it-System-Prompt-Generator-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:327feb7ca1065707e4e4aada7b5806bbc7b1d05ac0bb8210182f40ba7621a97b +size 720426176 diff --git a/gemma-3-1b-it-System-Prompt-Generator-Q4_K_M.gguf b/gemma-3-1b-it-System-Prompt-Generator-Q4_K_M.gguf new file mode 100644 index 0000000..0235b1c --- /dev/null +++ b/gemma-3-1b-it-System-Prompt-Generator-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8d8a63182e53fba1e5df844b947793afcbe80f778f5e037dbd0335c58dc7522 +size 806058944 diff --git a/gemma-3-1b-it-System-Prompt-Generator-Q6_K.gguf b/gemma-3-1b-it-System-Prompt-Generator-Q6_K.gguf new file mode 100644 index 0000000..1730a9f --- /dev/null +++ b/gemma-3-1b-it-System-Prompt-Generator-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3cc4a3eb0492bea429213390ca501f3fbfbc0935695799bfea125f797ed8d69a +size 1011739328 diff --git a/gemma-3-1b-it-System-Prompt-Generator-Q8_0.gguf b/gemma-3-1b-it-System-Prompt-Generator-Q8_0.gguf new file mode 100644 index 0000000..cce5927 --- /dev/null +++ b/gemma-3-1b-it-System-Prompt-Generator-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10dc01d6c34bf3b4fcd47426355c6ce1f398203d7d807823abe1ef6d7b480091 +size 1069307072 diff --git a/gemma-3-1b-it-System-Prompt-Generator-bf16.gguf b/gemma-3-1b-it-System-Prompt-Generator-bf16.gguf new file mode 100644 index 0000000..9f32360 --- /dev/null +++ b/gemma-3-1b-it-System-Prompt-Generator-bf16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a421e976f4eb89fea66dfa31bcc242cb8a81c73275e799a2983cd0ff06fac792 +size 2006574272