From 2ce1af2397f8108e64b720dc25f532cdf35cf07c Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sat, 9 May 2026 16:53:43 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: cortexso/deepscaler Source: Original Platform --- .gitattributes | 10 ++++++ README.md | 29 ++++++++++++++++ configuration.json | 1 + deepscaler-1.5b-preview-q2_k.gguf | 3 ++ deepscaler-1.5b-preview-q3_k_l.gguf | 3 ++ deepscaler-1.5b-preview-q3_k_m.gguf | 3 ++ deepscaler-1.5b-preview-q3_k_s.gguf | 3 ++ deepscaler-1.5b-preview-q4_k_m.gguf | 3 ++ deepscaler-1.5b-preview-q4_k_s.gguf | 3 ++ deepscaler-1.5b-preview-q5_k_m.gguf | 3 ++ deepscaler-1.5b-preview-q5_k_s.gguf | 3 ++ deepscaler-1.5b-preview-q6_k.gguf | 3 ++ deepscaler-1.5b-preview-q8_0.gguf | 3 ++ metadata.yml | 4 +++ model.yml | 52 +++++++++++++++++++++++++++++ 15 files changed, 126 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 configuration.json create mode 100644 deepscaler-1.5b-preview-q2_k.gguf create mode 100644 deepscaler-1.5b-preview-q3_k_l.gguf create mode 100644 deepscaler-1.5b-preview-q3_k_m.gguf create mode 100644 deepscaler-1.5b-preview-q3_k_s.gguf create mode 100644 deepscaler-1.5b-preview-q4_k_m.gguf create mode 100644 deepscaler-1.5b-preview-q4_k_s.gguf create mode 100644 deepscaler-1.5b-preview-q5_k_m.gguf create mode 100644 deepscaler-1.5b-preview-q5_k_s.gguf create mode 100644 deepscaler-1.5b-preview-q6_k.gguf create mode 100644 deepscaler-1.5b-preview-q8_0.gguf create mode 100644 metadata.yml create mode 100644 model.yml diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..a691f42 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,10 @@ +deepscaler-1.5b-preview-q2_k.gguf filter=lfs diff=lfs merge=lfs -text +deepscaler-1.5b-preview-q3_k_l.gguf filter=lfs diff=lfs merge=lfs -text +deepscaler-1.5b-preview-q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text +deepscaler-1.5b-preview-q3_k_s.gguf filter=lfs diff=lfs merge=lfs -text +deepscaler-1.5b-preview-q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +deepscaler-1.5b-preview-q4_k_s.gguf filter=lfs diff=lfs merge=lfs -text +deepscaler-1.5b-preview-q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text +deepscaler-1.5b-preview-q5_k_s.gguf filter=lfs diff=lfs merge=lfs -text +deepscaler-1.5b-preview-q6_k.gguf filter=lfs diff=lfs merge=lfs -text +deepscaler-1.5b-preview-q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..dbe2a63 --- /dev/null +++ b/README.md @@ -0,0 +1,29 @@ +--- +license: mit +pipeline_tag: text-generation +tags: +- cortex.cpp +--- +## Overview +Deepscaler is an advanced AI model developed from the agentica-org's DeepScaleR-1.5B-Preview, designed to enhance the efficiency and scalability of various machine learning tasks. Its core purpose is to provide high-quality predictive analytics and data processing capabilities while optimizing resource usage. Deepscaler is particularly useful in scenarios such as natural language processing, computer vision, and more complex data interpretation tasks, making it suitable for applications in industries like finance, healthcare, and entertainment. Users can leverage its performance to achieve faster training times and improved accuracy in their models. Overall, Deepscaler's architecture allows it to deliver robust results with reduced computational overhead, making it an excellent choice for developers and organizations aiming to scale their AI solutions. +## Variants +| No | Variant | Cortex CLI command | +| --- | --- | --- | +| 1 | [Deepscaler-1.5b](https://huggingface.co/cortexso/deepscaler/tree/1.5b) | cortex run deepscaler:1.5b | +## Use it with Jan (UI) +1. Install **Jan** using [Quickstart](https://jan.ai/docs/quickstart) +2. Use in Jan model Hub: + ```bash + cortexso/deepscaler + ``` + +## Use it with Cortex (CLI) +1. Install **Cortex** using [Quickstart](https://cortex.jan.ai/docs/quickstart) +2. Run the model with command: + ```bash + cortex run deepscaler + ``` +## Credits +- **Author:** agentica-org +- **Converter:** [Homebrew](https://www.homebrew.ltd/) +- **Original License:** [LICENSE](https://huggingface.co/agentica-org/DeepScaleR-1.5B-Preview/blob/main/LICENSE) \ No newline at end of file diff --git a/configuration.json b/configuration.json new file mode 100644 index 0000000..bbeeda1 --- /dev/null +++ b/configuration.json @@ -0,0 +1 @@ +{"framework": "pytorch", "task": "text-generation", "allow_remote": true} \ No newline at end of file diff --git a/deepscaler-1.5b-preview-q2_k.gguf b/deepscaler-1.5b-preview-q2_k.gguf new file mode 100644 index 0000000..933f9a9 --- /dev/null +++ b/deepscaler-1.5b-preview-q2_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf16ea23a11890ba9c459c288fc32d17b3e54dbc93dbc2bf7163f16664822609 +size 752881312 diff --git a/deepscaler-1.5b-preview-q3_k_l.gguf b/deepscaler-1.5b-preview-q3_k_l.gguf new file mode 100644 index 0000000..1037c4a --- /dev/null +++ b/deepscaler-1.5b-preview-q3_k_l.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:348301bf126a7b1649983d10c9608d306b1965b851cea6af0db3f672d1c26726 +size 980441248 diff --git a/deepscaler-1.5b-preview-q3_k_m.gguf b/deepscaler-1.5b-preview-q3_k_m.gguf new file mode 100644 index 0000000..7950995 --- /dev/null +++ b/deepscaler-1.5b-preview-q3_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d178f3654212dd183319e89f86cab74042f589bae62f86fba941f65bb50ffbb8 +size 924457120 diff --git a/deepscaler-1.5b-preview-q3_k_s.gguf b/deepscaler-1.5b-preview-q3_k_s.gguf new file mode 100644 index 0000000..d55a121 --- /dev/null +++ b/deepscaler-1.5b-preview-q3_k_s.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43d680bfd259cd99bd7861351e55e8d38209ae2b96a220fa7ceff39423df67c1 +size 861223072 diff --git a/deepscaler-1.5b-preview-q4_k_m.gguf b/deepscaler-1.5b-preview-q4_k_m.gguf new file mode 100644 index 0000000..2fe009d --- /dev/null +++ b/deepscaler-1.5b-preview-q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ed31fa867cd2aadf61a2632ac49e439fe0c3526d218d9a2b0c9481da96b51a3 +size 1117321888 diff --git a/deepscaler-1.5b-preview-q4_k_s.gguf b/deepscaler-1.5b-preview-q4_k_s.gguf new file mode 100644 index 0000000..5d4ffb0 --- /dev/null +++ b/deepscaler-1.5b-preview-q4_k_s.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0fb7d20d782df27908333ffc1eaac52ec39c2f46991ea5abe90f11ad0cb9af6 +size 1071585952 diff --git a/deepscaler-1.5b-preview-q5_k_m.gguf b/deepscaler-1.5b-preview-q5_k_m.gguf new file mode 100644 index 0000000..3371f74 --- /dev/null +++ b/deepscaler-1.5b-preview-q5_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3bcf0e249050877384008a9080ce41ed1d438799ed92e61f9caf4d8e6320ac15 +size 1285495456 diff --git a/deepscaler-1.5b-preview-q5_k_s.gguf b/deepscaler-1.5b-preview-q5_k_s.gguf new file mode 100644 index 0000000..c5fb196 --- /dev/null +++ b/deepscaler-1.5b-preview-q5_k_s.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6c27caef49f057213508247a56adf82bea86ece10e5c50c3674859b670279dc +size 1259174560 diff --git a/deepscaler-1.5b-preview-q6_k.gguf b/deepscaler-1.5b-preview-q6_k.gguf new file mode 100644 index 0000000..1f74cbc --- /dev/null +++ b/deepscaler-1.5b-preview-q6_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a86c124377fd3b31502c988f877f29df9ca298dcb572d9e462eb0bc37b83ca0 +size 1464179872 diff --git a/deepscaler-1.5b-preview-q8_0.gguf b/deepscaler-1.5b-preview-q8_0.gguf new file mode 100644 index 0000000..29d7853 --- /dev/null +++ b/deepscaler-1.5b-preview-q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c911f55ee2ff7fc680eff8c15a7beff31f6b7d5eba89545cccc6c31d32e2b4fd +size 1894533280 diff --git a/metadata.yml b/metadata.yml new file mode 100644 index 0000000..ab77c58 --- /dev/null +++ b/metadata.yml @@ -0,0 +1,4 @@ +version: 1 +name: deepscaler +default: 1.5b +author: "agentica-org" diff --git a/model.yml b/model.yml new file mode 100644 index 0000000..7dc5803 --- /dev/null +++ b/model.yml @@ -0,0 +1,52 @@ +# BEGIN GENERAL GGUF METADATA +id: deepscaler +model: deepscaler +name: deepscaler +version: 1 +# END GENERAL GGUF METADATA + +# BEGIN INFERENCE PARAMETERS +# BEGIN REQUIRED +stop: +- <|im_end|> +# END REQUIRED + +# BEGIN OPTIONAL +stream: true +top_p: 0.9 +temperature: 0.7 +frequency_penalty: 0 +presence_penalty: 0 +max_tokens: 4096 +seed: -1 +dynatemp_range: 0 +dynatemp_exponent: 1 +top_k: 40 +min_p: 0.05 +tfs_z: 1 +typ_p: 1 +repeat_last_n: 64 +repeat_penalty: 1 +mirostat: false +mirostat_tau: 5 +mirostat_eta: 0.100000001 +penalize_nl: false +ignore_eos: false +n_probs: 0 +min_keep: 0 +# END OPTIONAL +# END INFERENCE PARAMETERS + +# BEGIN MODEL LOAD PARAMETERS +# BEGIN REQUIRED +engine: llama-cpp +prompt_template: | + <|im_start|>system + {system_message}<|im_end|> + <|im_start|>user + {prompt}<|im_end|> + <|im_start|>assistant +ctx_len: 4096 +ngl: 34 +# END REQUIRED +# END MODEL LOAD PARAMETERS