初始化项目,由ModelHub XC社区提供模型
Model: Open4bits/llama-nexora-vector-v0.1-GGUF Source: Original Platform
This commit is contained in:
43
.gitattributes
vendored
Normal file
43
.gitattributes
vendored
Normal file
@@ -0,0 +1,43 @@
|
|||||||
|
*.7z filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.arrow filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.bin filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ftz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.gz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.h5 filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.joblib filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.model filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.npy filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.npz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.onnx filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.ot filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.parquet filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pb filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pickle filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pkl filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pt filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.pth filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.rar filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
||||||
|
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tar filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tflite filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.tgz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.wasm filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.xz filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.zip filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*.zst filter=lfs diff=lfs merge=lfs -text
|
||||||
|
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||||
|
llama-nexora-vector-v0.1-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
||||||
|
llama-nexora-vector-v0.1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
||||||
|
llama-nexora-vector-v0.1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||||
|
llama-nexora-vector-v0.1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
||||||
|
llama-nexora-vector-v0.1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||||
|
llama-nexora-vector-v0.1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||||
|
llama-nexora-vector-v0.1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
||||||
|
llama-nexora-vector-v0.1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||||
301
README.md
Normal file
301
README.md
Normal file
@@ -0,0 +1,301 @@
|
|||||||
|
---
|
||||||
|
base_model:
|
||||||
|
- ArkAiLab-Adl/llama-nexora-vector-v0.1
|
||||||
|
license: llama3.2
|
||||||
|
language:
|
||||||
|
- en
|
||||||
|
pipeline_tag: text-generation
|
||||||
|
tags:
|
||||||
|
- nexora
|
||||||
|
- llama-nexora
|
||||||
|
- vector
|
||||||
|
- chat
|
||||||
|
- llama-3
|
||||||
|
- open4bits
|
||||||
|
---
|
||||||
|
<p align="center">
|
||||||
|
<img src="https://huggingface.co/ArkAiLab-Adl/llama-nexora-vector-v0.1/resolve/main/assets/llama-nexora-vector.jpg" alt="llama-nexora-vector-gguf"/>
|
||||||
|
</p>
|
||||||
|
|
||||||
|
# Llama-Nexora-Vector-v0.1 — GGUF
|
||||||
|
|
||||||
|
<p align="center">
|
||||||
|
<img src="https://img.shields.io/badge/status-beta-orange" alt="Status: Beta"/>
|
||||||
|
<img src="https://img.shields.io/badge/license-Llama%203.2%20Community-blue" alt="License: Llama 3.2 Community"/>
|
||||||
|
<img src="https://img.shields.io/badge/base_model-Llama--3.2--1B-blueviolet" alt="Base Model: Llama 3.2 1B"/>
|
||||||
|
<img src="https://img.shields.io/badge/output-SVG-green" alt="Output: SVG"/>
|
||||||
|
<img src="https://img.shields.io/badge/family-Llama--Nexora-red" alt="Family: Llama-Nexora"/>
|
||||||
|
<img src="https://img.shields.io/badge/format-GGUF-cyan" alt="Format: GGUF"/>
|
||||||
|
</p>
|
||||||
|
|
||||||
|
> This is the **official GGUF quantized release** of [llama-nexora-vector-v0.1](https://huggingface.co/ArkAiLab-Adl/llama-nexora-vector-v0.1), published by **[Open4bits](https://huggingface.co/Open4bits)** — the official quantization project under **ArkAiLabs**. Multiple quantization levels are provided to suit a wide range of hardware configurations. This is a beta release intended for research, prototyping, and early-stage development workflows only.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Table of Contents
|
||||||
|
|
||||||
|
- [Overview](#overview)
|
||||||
|
- [The Llama-Nexora Family](#the-llama-nexora-family)
|
||||||
|
- [Available Quantizations](#available-quantizations)
|
||||||
|
- [Which Quant Should I Use?](#which-quant-should-i-use)
|
||||||
|
- [Model Details](#model-details)
|
||||||
|
- [Requirements](#requirements)
|
||||||
|
- [Installation & Usage](#installation--usage)
|
||||||
|
- [Capabilities](#capabilities)
|
||||||
|
- [Limitations](#limitations)
|
||||||
|
- [Intended Use](#intended-use)
|
||||||
|
- [Usage Recommendations](#usage-recommendations)
|
||||||
|
- [Risks & Considerations](#risks--considerations)
|
||||||
|
- [Community & Support](#community--support)
|
||||||
|
- [License](#license)
|
||||||
|
- [Acknowledgements](#acknowledgements)
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Overview
|
||||||
|
|
||||||
|
**llama-nexora-vector-v0.1-GGUF** contains the official GGUF quantized versions of [llama-nexora-vector-v0.1](https://huggingface.co/ArkAiLab-Adl/llama-nexora-vector-v0.1) — an experimental text-to-vector model from the **Llama-Nexora family** that generates structured SVG graphics from natural language prompts.
|
||||||
|
|
||||||
|
These quantized releases are published by **[Open4bits](https://huggingface.co/Open4bits)**, the dedicated quantization project under ArkAiLabs, and are compatible with local inference tools such as **llama.cpp**, **Ollama**, and **LM Studio** on Windows, Linux, and macOS.
|
||||||
|
|
||||||
|
This release is in **beta** and is scoped to research, experimentation, and early-stage design tooling. All outputs should be validated before use in any downstream pipeline.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## The Llama-Nexora Family
|
||||||
|
|
||||||
|
This model is part of the **Llama-Nexora family** — a dedicated branch of Nexora models under **ArkAiLabs**, built on the Meta Llama architecture and focused on creative, efficient, and practical open AI systems.
|
||||||
|
|
||||||
|
| Model | Type | Link |
|
||||||
|
|---|---|---|
|
||||||
|
| **llama-nexora-vector-v0.1** | Original (Full Precision) | [ArkAiLab-Adl/llama-nexora-vector-v0.1](https://huggingface.co/ArkAiLab-Adl/llama-nexora-vector-v0.1) |
|
||||||
|
| **llama-nexora-vector-v0.1-GGUF** | GGUF (Windows / Linux / macOS) | *(this repo)* |
|
||||||
|
| **llama-nexora-vector-v0.1-mlx-4Bit** | MLX 4-Bit (Apple Silicon) | [Open4bits/llama-nexora-vector-v0.1-mlx-4Bit](https://huggingface.co/Open4bits/llama-nexora-vector-v0.1-mlx-4Bit) |
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Available Quantizations
|
||||||
|
|
||||||
|
All quantized files are available in this repository. Select the file that best matches your hardware and performance requirements.
|
||||||
|
|
||||||
|
### 2-bit
|
||||||
|
|
||||||
|
| Quantization | File Size | Description |
|
||||||
|
|---|---|---|
|
||||||
|
| **Q2_K** | 581 MB | Smallest size, lowest quality. Use only if very limited on RAM/VRAM. |
|
||||||
|
|
||||||
|
### 4-bit
|
||||||
|
|
||||||
|
| Quantization | File Size | Description |
|
||||||
|
|---|---|---|
|
||||||
|
| **Q4_K_S** | 776 MB | Small 4-bit quantization. Good balance of size and quality. |
|
||||||
|
| **Q4_0** | 771 MB | Legacy 4-bit format. Widely compatible. |
|
||||||
|
| **Q4_K_M** | 808 MB | Medium 4-bit quantization. Recommended for most users. |
|
||||||
|
|
||||||
|
### 5-bit
|
||||||
|
|
||||||
|
| Quantization | File Size | Description |
|
||||||
|
|---|---|---|
|
||||||
|
| **Q5_K_S** | 893 MB | Small 5-bit quantization. Better quality than Q4 with modest size increase. |
|
||||||
|
| **Q5_K_M** | 912 MB | Medium 5-bit quantization. Excellent quality-to-size ratio. |
|
||||||
|
|
||||||
|
### 6-bit
|
||||||
|
|
||||||
|
| Quantization | File Size | Description |
|
||||||
|
|---|---|---|
|
||||||
|
| **Q6_K** | 1.02 GB | High quality, close to full precision. Recommended if you have the RAM. |
|
||||||
|
|
||||||
|
### 8-bit
|
||||||
|
|
||||||
|
| Quantization | File Size | Description |
|
||||||
|
|---|---|---|
|
||||||
|
| **Q8_0** | 1.32 GB | Near full-precision quality. Best quality GGUF option available. |
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Which Quant Should I Use?
|
||||||
|
|
||||||
|
| Your Situation | Recommended Quant |
|
||||||
|
|---|---|
|
||||||
|
| Very limited RAM (< 2GB free) | Q2_K |
|
||||||
|
| General use / most users | Q4_K_M |
|
||||||
|
| Want better quality, have the space | Q5_K_M or Q6_K |
|
||||||
|
| Maximum quality, no size concern | Q8_0 |
|
||||||
|
| Legacy tooling / broad compatibility | Q4_0 |
|
||||||
|
|
||||||
|
> **Tip:** For most users, **Q4_K_M** offers the best balance between model size and output quality.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Model Details
|
||||||
|
|
||||||
|
| Property | Details |
|
||||||
|
|---|---|
|
||||||
|
| **Model Name** | llama-nexora-vector-v0.1-GGUF |
|
||||||
|
| **Model Family** | Llama-Nexora |
|
||||||
|
| **Model Type** | Text-to-SVG (Causal Language Model) |
|
||||||
|
| **Original Base Model** | [unsloth/Llama-3.2-1B-Instruct](https://huggingface.co/unsloth/Llama-3.2-1B-Instruct) |
|
||||||
|
| **Original Full Model** | [ArkAiLab-Adl/llama-nexora-vector-v0.1](https://huggingface.co/ArkAiLab-Adl/llama-nexora-vector-v0.1) |
|
||||||
|
| **Quantized By** | [Open4bits](https://huggingface.co/Open4bits) |
|
||||||
|
| **Output Format** | SVG |
|
||||||
|
| **Release Status** | Beta |
|
||||||
|
| **License** | Llama 3.2 Community License |
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Requirements
|
||||||
|
|
||||||
|
GGUF models can be run on **Windows, Linux, and macOS** (including Apple Silicon and Intel) using any of the following tools:
|
||||||
|
|
||||||
|
- **[llama.cpp](https://github.com/ggerganov/llama.cpp)** — CLI-based inference
|
||||||
|
- **[Ollama](https://ollama.com)** — Easy local model runner
|
||||||
|
- **[LM Studio](https://lmstudio.ai)** — GUI-based local inference app
|
||||||
|
- **[Jan](https://jan.ai)** — Open-source ChatGPT alternative for local use
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Installation & Usage
|
||||||
|
|
||||||
|
### llama.cpp
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Clone and build llama.cpp
|
||||||
|
git clone https://github.com/ggerganov/llama.cpp
|
||||||
|
cd llama.cpp && make
|
||||||
|
|
||||||
|
# Download the model (example: Q4_K_M)
|
||||||
|
huggingface-cli download Open4bits/llama-nexora-vector-v0.1-GGUF \
|
||||||
|
llama-nexora-vector-v0.1.Q4_K_M.gguf \
|
||||||
|
--local-dir ./models
|
||||||
|
|
||||||
|
# Run inference
|
||||||
|
./llama-cli -m ./models/llama-nexora-vector-v0.1.Q4_K_M.gguf \
|
||||||
|
-p "Generate an SVG of a simple red circle." \
|
||||||
|
-n 512
|
||||||
|
```
|
||||||
|
|
||||||
|
### Ollama
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Create a Modelfile
|
||||||
|
echo 'FROM ./llama-nexora-vector-v0.1.Q4_K_M.gguf' > Modelfile
|
||||||
|
|
||||||
|
# Create the model
|
||||||
|
ollama create llama-nexora-vector -f Modelfile
|
||||||
|
|
||||||
|
# Run it
|
||||||
|
ollama run llama-nexora-vector "Generate an SVG of a simple red circle."
|
||||||
|
```
|
||||||
|
|
||||||
|
### LM Studio
|
||||||
|
|
||||||
|
1. Open **LM Studio** and go to the Search tab.
|
||||||
|
2. Search for `Open4bits/llama-nexora-vector-v0.1-GGUF`.
|
||||||
|
3. Select your preferred quantization and download.
|
||||||
|
4. Load the model and start prompting.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Capabilities
|
||||||
|
|
||||||
|
llama-nexora-vector-v0.1-GGUF is designed to translate textual instructions into structured SVG code. The model is best suited for:
|
||||||
|
|
||||||
|
- Generating SVG markup for simple vector graphics
|
||||||
|
- Producing geometric shapes and basic illustrations
|
||||||
|
- Creating icons, shapes, logos, and simple illustrations
|
||||||
|
- Supporting rapid prototyping and concept design
|
||||||
|
- Producing lightweight scalable vector outputs
|
||||||
|
|
||||||
|
> **Tip:** The model performs best with concise, clearly scoped prompts focused on simple visual compositions.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Limitations
|
||||||
|
|
||||||
|
This is an early-stage beta release. Users should be aware of the following constraints before integrating the model:
|
||||||
|
|
||||||
|
- **High hallucination rate** — outputs may be invalid or non-renderable SVG
|
||||||
|
- **Limited generalization** — dataset size affects output consistency across diverse prompts
|
||||||
|
- **Weak complex scene handling** — highly detailed or multi-element prompts may produce poor results
|
||||||
|
- **Manual correction required** — outputs should be validated and post-processed before use
|
||||||
|
- **Not production-ready** — not suitable for safety-critical or automated pipelines
|
||||||
|
- **Quantization trade-off** — lower-bit quants (Q2, Q4) may show more quality degradation versus the full-precision model
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Intended Use
|
||||||
|
|
||||||
|
### ✅ Supported Use Cases
|
||||||
|
|
||||||
|
- Academic and applied research in text-to-vector generation
|
||||||
|
- Experimental AI-assisted design systems
|
||||||
|
- Educational exploration of structured output generation
|
||||||
|
- Lightweight SVG prototyping and ideation on local hardware
|
||||||
|
|
||||||
|
### ❌ Out-of-Scope Use Cases
|
||||||
|
|
||||||
|
- Production-grade or commercial vector asset pipelines
|
||||||
|
- High-precision design deliverables without human validation
|
||||||
|
- Automated systems where SVG correctness is required without manual review
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Usage Recommendations
|
||||||
|
|
||||||
|
To get the best results from this model:
|
||||||
|
|
||||||
|
1. **Keep prompts simple and specific** — avoid multi-scene or highly complex compositions
|
||||||
|
2. **Validate all SVG outputs** before rendering or integrating into any pipeline
|
||||||
|
3. **Post-process outputs** to correct syntax or structural issues
|
||||||
|
4. **Use iterative prompting** — refining prompts across multiple turns often yields better results
|
||||||
|
5. **Expect imperfections** — this is a beta model; treat outputs as drafts, not finals
|
||||||
|
6. **Human review is recommended** for all generated content
|
||||||
|
7. **Choose the right quant** — higher-bit quants yield better output quality if your hardware allows
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Risks & Considerations
|
||||||
|
|
||||||
|
Developers integrating this model should account for the following risks:
|
||||||
|
|
||||||
|
- Generation of malformed or non-functional SVG code
|
||||||
|
- Inconsistent instruction following across prompt variations
|
||||||
|
- Unpredictable outputs due to limited training data coverage
|
||||||
|
- Outputs may sometimes be invalid, incomplete, or require manual correction
|
||||||
|
- Quality degradation versus full-precision model, especially at lower bit widths
|
||||||
|
|
||||||
|
**Recommendation:** Implement downstream validation layers and SVG syntax checking before any rendering or integration. Human review is recommended for all generated content.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Community & Support
|
||||||
|
|
||||||
|
Join the community for updates, feedback, and discussion. Community feedback, testing, and contributions are welcome — this project will continue evolving through open research and real-world experimentation.
|
||||||
|
|
||||||
|
💬 **[Join our Discord Server](https://discord.gg/mwdrgYbzuG)**
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## License
|
||||||
|
|
||||||
|
This model is released under the **Llama 3.2 Community License**.
|
||||||
|
|
||||||
|
Use of this model is governed by the [Llama 3.2 Community License Agreement](https://www.llama.com/llama3_2/license/). Please review the license terms before use, modification, or distribution.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Acknowledgements
|
||||||
|
|
||||||
|
This quantized release is based on **[llama-nexora-vector-v0.1](https://huggingface.co/ArkAiLab-Adl/llama-nexora-vector-v0.1)** by ArkAiLabs, which itself is built upon **[Llama 3.2 1B Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct)** by Meta. Quantization was performed by **[Open4bits](https://huggingface.co/Open4bits)** using the [llama.cpp](https://github.com/ggerganov/llama.cpp) GGUF quantization toolchain. We thank the open-source AI community for their continued contributions that make projects like this possible.
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## About Open4bits
|
||||||
|
|
||||||
|
**[Open4bits](https://huggingface.co/Open4bits)** is the official quantization project under **ArkAiLabs**, dedicated to publishing efficient, accessible quantized versions of Nexora and Llama-Nexora models across multiple formats (GGUF, MLX) for local inference on a wide range of hardware.
|
||||||
|
|
||||||
|
## About Nexora & Llama-Nexora
|
||||||
|
|
||||||
|
**Nexora** is an experimental AI initiative under **ArkAiLabs**, focused on building lightweight, practical, and creative AI systems for real-world applications.
|
||||||
|
|
||||||
|
The **Llama-Nexora family** is a dedicated branch within Nexora, built on the Meta Llama architecture — focused on creative, efficient, and practical open AI systems that are accessible to the broader community.
|
||||||
3
llama-nexora-vector-v0.1-Q2_K.gguf
Normal file
3
llama-nexora-vector-v0.1-Q2_K.gguf
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:ee97114be253a7a867a55f89fae8c52c61d31b17106fb4a7f9b23a41baddfe02
|
||||||
|
size 580875168
|
||||||
3
llama-nexora-vector-v0.1-Q4_0.gguf
Normal file
3
llama-nexora-vector-v0.1-Q4_0.gguf
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:15817fb9df041aac47b16014069cb58281b6e9277047bc95ac6ce8bdb1f286cf
|
||||||
|
size 770929568
|
||||||
3
llama-nexora-vector-v0.1-Q4_K_M.gguf
Normal file
3
llama-nexora-vector-v0.1-Q4_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:5df0eaa66acf56e77ecde108f05db724d0f3ea510e41477490c5e0d6687402e5
|
||||||
|
size 807695232
|
||||||
3
llama-nexora-vector-v0.1-Q4_K_S.gguf
Normal file
3
llama-nexora-vector-v0.1-Q4_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:13000aa8e7dde2e12bb739358a7cfa9396c8fe193fcacd0be6ea49881289c0f0
|
||||||
|
size 775648160
|
||||||
3
llama-nexora-vector-v0.1-Q5_K_M.gguf
Normal file
3
llama-nexora-vector-v0.1-Q5_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:7b93ddb107d8ddd398f4c95a410f19cc9716905e192fdf6534a0d5e2bcb79275
|
||||||
|
size 911504288
|
||||||
3
llama-nexora-vector-v0.1-Q5_K_S.gguf
Normal file
3
llama-nexora-vector-v0.1-Q5_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:6fcb9244a5c8fb0ee1d483cf810ad936fdea33b46f5cb7338172f98f63104cf4
|
||||||
|
size 892564384
|
||||||
3
llama-nexora-vector-v0.1-Q6_K.gguf
Normal file
3
llama-nexora-vector-v0.1-Q6_K.gguf
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:4c8fdcb1d1e2b02dc28a128f62bbdf5af868f59fffe37d19839f9e6fe6f53e51
|
||||||
|
size 1021801376
|
||||||
3
llama-nexora-vector-v0.1-Q8_0.gguf
Normal file
3
llama-nexora-vector-v0.1-Q8_0.gguf
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
version https://git-lfs.github.com/spec/v1
|
||||||
|
oid sha256:7e05cda7de728516450a7e9a7a5d9ac67604d48c544e1beb89a00e0c22b1ebde
|
||||||
|
size 1321083392
|
||||||
Reference in New Issue
Block a user