Update README.md

This commit is contained in:
ai-modelscope
2024-11-26 20:13:21 +08:00
parent d2a1ebab4b
commit 8f12d2e626
15 changed files with 215 additions and 63 deletions

44
.gitattributes vendored
View File

@@ -1,47 +1,47 @@
*.7z filter=lfs diff=lfs merge=lfs -text *.7z filter=lfs diff=lfs merge=lfs -text
*.arrow filter=lfs diff=lfs merge=lfs -text *.arrow filter=lfs diff=lfs merge=lfs -text
*.bin filter=lfs diff=lfs merge=lfs -text *.bin filter=lfs diff=lfs merge=lfs -text
*.bin.* filter=lfs diff=lfs merge=lfs -text
*.bz2 filter=lfs diff=lfs merge=lfs -text *.bz2 filter=lfs diff=lfs merge=lfs -text
*.ckpt filter=lfs diff=lfs merge=lfs -text
*.ftz filter=lfs diff=lfs merge=lfs -text *.ftz filter=lfs diff=lfs merge=lfs -text
*.gz filter=lfs diff=lfs merge=lfs -text *.gz filter=lfs diff=lfs merge=lfs -text
*.h5 filter=lfs diff=lfs merge=lfs -text *.h5 filter=lfs diff=lfs merge=lfs -text
*.joblib filter=lfs diff=lfs merge=lfs -text *.joblib filter=lfs diff=lfs merge=lfs -text
*.lfs.* filter=lfs diff=lfs merge=lfs -text *.lfs.* filter=lfs diff=lfs merge=lfs -text
*.mlmodel filter=lfs diff=lfs merge=lfs -text
*.model filter=lfs diff=lfs merge=lfs -text *.model filter=lfs diff=lfs merge=lfs -text
*.msgpack filter=lfs diff=lfs merge=lfs -text *.msgpack filter=lfs diff=lfs merge=lfs -text
*.npy filter=lfs diff=lfs merge=lfs -text
*.npz filter=lfs diff=lfs merge=lfs -text
*.onnx filter=lfs diff=lfs merge=lfs -text *.onnx filter=lfs diff=lfs merge=lfs -text
*.ot filter=lfs diff=lfs merge=lfs -text *.ot filter=lfs diff=lfs merge=lfs -text
*.parquet filter=lfs diff=lfs merge=lfs -text *.parquet filter=lfs diff=lfs merge=lfs -text
*.pb filter=lfs diff=lfs merge=lfs -text *.pb filter=lfs diff=lfs merge=lfs -text
*.pickle filter=lfs diff=lfs merge=lfs -text
*.pkl filter=lfs diff=lfs merge=lfs -text
*.pt filter=lfs diff=lfs merge=lfs -text *.pt filter=lfs diff=lfs merge=lfs -text
*.pth filter=lfs diff=lfs merge=lfs -text *.pth filter=lfs diff=lfs merge=lfs -text
*.rar filter=lfs diff=lfs merge=lfs -text *.rar filter=lfs diff=lfs merge=lfs -text
*.safetensors filter=lfs diff=lfs merge=lfs -text
saved_model/**/* filter=lfs diff=lfs merge=lfs -text saved_model/**/* filter=lfs diff=lfs merge=lfs -text
*.tar.* filter=lfs diff=lfs merge=lfs -text *.tar.* filter=lfs diff=lfs merge=lfs -text
*.tar filter=lfs diff=lfs merge=lfs -text
*.tflite filter=lfs diff=lfs merge=lfs -text *.tflite filter=lfs diff=lfs merge=lfs -text
*.tgz filter=lfs diff=lfs merge=lfs -text *.tgz filter=lfs diff=lfs merge=lfs -text
*.wasm filter=lfs diff=lfs merge=lfs -text
*.xz filter=lfs diff=lfs merge=lfs -text *.xz filter=lfs diff=lfs merge=lfs -text
*.zip filter=lfs diff=lfs merge=lfs -text *.zip filter=lfs diff=lfs merge=lfs -text
*.zstandard filter=lfs diff=lfs merge=lfs -text
*.tfevents* filter=lfs diff=lfs merge=lfs -text
*.db* filter=lfs diff=lfs merge=lfs -text
*.ark* filter=lfs diff=lfs merge=lfs -text
**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text
**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text
**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text
*.safetensors filter=lfs diff=lfs merge=lfs -text
*.ckpt filter=lfs diff=lfs merge=lfs -text
*.gguf* filter=lfs diff=lfs merge=lfs -text
*.ggml filter=lfs diff=lfs merge=lfs -text
*.llamafile* filter=lfs diff=lfs merge=lfs -text
*.pt2 filter=lfs diff=lfs merge=lfs -text
*.mlmodel filter=lfs diff=lfs merge=lfs -text
*.npy filter=lfs diff=lfs merge=lfs -text
*.npz filter=lfs diff=lfs merge=lfs -text
*.pickle filter=lfs diff=lfs merge=lfs -text
*.pkl filter=lfs diff=lfs merge=lfs -text
*.tar filter=lfs diff=lfs merge=lfs -text
*.wasm filter=lfs diff=lfs merge=lfs -text
*.zst filter=lfs diff=lfs merge=lfs -text *.zst filter=lfs diff=lfs merge=lfs -text
*tfevents* filter=lfs diff=lfs merge=lfs -text *tfevents* filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q2_k.gguf filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q3_k_l.gguf filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q3_k_s.gguf filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q4_0.gguf filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q4_k_s.gguf filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q5_0.gguf filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q5_k_s.gguf filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q6_k.gguf filter=lfs diff=lfs merge=lfs -text
aya-expanse-8b-q8_0.gguf filter=lfs diff=lfs merge=lfs -text

197
README.md
View File

@@ -1,47 +1,162 @@
--- ---
license: Apache License 2.0 license: apache-2.0
language:
#model-type: - en
##如 gpt、phi、llama、chatglm、baichuan 等 - fr
#- gpt - de
- es
#domain: - it
##如 nlp、cv、audio、multi-modal - pt
#- nlp - ja
- ko
#language: - zh
##语言代码列表 https://help.aliyun.com/document_detail/215387.html?spm=a2c4g.11186623.0.0.9f8d7467kni6Aa - ar
#- cn - el
- fa
#metrics: - pl
##如 CIDEr、Blue、ROUGE 等 - id
#- CIDEr - cs
- he
#tags: - hi
##各种自定义,包括 pretrained、fine-tuned、instruction-tuned、RL-tuned 等训练方法和其他 - nl
#- pretrained - ro
- ru
#tools: - tr
##如 vllm、fastchat、llamacpp、AdaSeq 等 - uk
#- vllm - vi
base_model:
- CohereForAI/aya-expanse-8b
tags:
- gguf
- CohereForAI
- llama-cpp
- Aya-Expanse-8B-GGUF
pipeline_tag: text-generation
--- ---
### 当前模型的贡献者未提供更加详细的模型介绍。模型文件和权重,可浏览“模型文件”页面获取。 ## AYA Expanse 8B Model Files
#### 您可以通过如下git clone命令或者ModelScope SDK来下载模型
| File Name | Size | Description | Upload Status |
|------------------------------|---------|-----------------------|------------------|
| `.gitattributes` | 2.27 kB | Git attributes file | Uploaded |
| `README.md` | 4.53 kB | Documentation file | Updated 5 days ago |
| `aya-expanse-8b-q2_k.gguf` | 3.44 GB | Quantized Q2 model | Uploaded (LFS) |
| `aya-expanse-8b-q3_k_l.gguf` | 4.53 GB | Quantized Q3 (large) | Uploaded (LFS) |
| `aya-expanse-8b-q3_k_m.gguf` | 4.22 GB | Quantized Q3 (medium) | Uploaded (LFS) |
| `aya-expanse-8b-q3_k_s.gguf` | 3.87 GB | Quantized Q3 (small) | Uploaded (LFS) |
| `aya-expanse-8b-q4_0.gguf` | 4.8 GB | Quantized Q4 model | Uploaded (LFS) |
| `aya-expanse-8b-q4_k_m.gguf` | 5.06 GB | Quantized Q4 (medium) | Uploaded (LFS) |
| `aya-expanse-8b-q4_k_s.gguf` | 4.83 GB | Quantized Q4 (small) | Uploaded (LFS) |
| `aya-expanse-8b-q5_0.gguf` | 5.67 GB | Quantized Q5 model | Uploaded (LFS) |
| `aya-expanse-8b-q5_k_m.gguf` | 5.8 GB | Quantized Q5 (medium) | Uploaded (LFS) |
| `aya-expanse-8b-q5_k_s.gguf` | 5.67 GB | Quantized Q5 (small) | Uploaded (LFS) |
| `aya-expanse-8b-q6_k.gguf` | 6.6 GB | Quantized Q6 model | Uploaded (LFS) |
| `aya-expanse-8b-q8_0.gguf` | 8.54 GB | Quantized Q8 model | Uploaded (LFS) |
# Uploaded model
- **Developed by:** prithivMLmods
- **License:** apache-2.0
- **Quants from model :** CohereForAI/aya-expanse-8b
Here's a well-structured README.md documentation for running models with Ollama:
# Run with Ollama 🦙
> [!WARNING]
> **Note:** The model name mentioned in the README.md may or may not be same.
## Overview
Ollama is a powerful tool that allows you to run machine learning models effortlessly. This guide will help you download, install, and run your own GGUF models in just a few minutes.
## Table of Contents
| Section | Description |
|------------------------------|----------------------------------|
| [Download and Install Ollama](#download-and-install-ollama) | Instructions for downloading and installing Ollama. |
| [Steps to Run GGUF Models](#steps-to-run-gguf-models) | Guidelines for running GGUF models. |
| 1. [Create the Model File](#1-create-the-model-file) | Steps to create the model file. |
| 2. [Add the Template Command](#2-add-the-template-command) | Instructions for adding the template command. |
| 3. [Create and Patch the Model](#3-create-and-patch-the-model) | Process for creating and patching the model. |
| [Running the Model](#running-the-model) | How to run the model. |
| [Sample Usage](#sample-usage) | Examples of how to use the model. |
## Download and Install Ollama
To get started, download Ollama from [https://ollama.com/download](https://ollama.com/download) and install it on your Windows or Mac system.
## Steps to Run GGUF Models
### 1. Create the Model File
First, create a model file and name it appropriately. For example, you can name your model file `metallama`.
### 2. Add the Template Command
In your model file, include a `FROM` line that specifies the base model file you want to use. For instance:
SDK下载
```bash ```bash
#安装ModelScope FROM Llama-3.2-1B.F16.gguf
pip install modelscope
```
```python
#SDK模型下载
from modelscope import snapshot_download
model_dir = snapshot_download('prithivMLmods/Aya-Expanse-8B-GGUF')
```
Git下载
```
#Git模型下载
git clone https://www.modelscope.cn/prithivMLmods/Aya-Expanse-8B-GGUF.git
``` ```
<p style="color: lightgrey;">如果您是本模型的贡献者,我们邀请您根据<a href="https://modelscope.cn/docs/ModelScope%E6%A8%A1%E5%9E%8B%E6%8E%A5%E5%85%A5%E6%B5%81%E7%A8%8B%E6%A6%82%E8%A7%88" style="color: lightgrey; text-decoration: underline;">模型贡献文档</a>,及时完善模型卡片内容。</p> Ensure that the model file is in the same directory as your script.
### 3. Create and Patch the Model
Open your terminal and run the following command to create and patch your model:
```bash
ollama create metallama -f ./metallama
```
Once the process is successful, you will see a confirmation message.
To verify that the model was created successfully, you can list all models with:
```bash
ollama list
```
Make sure that `metallama` appears in the list of models.
---
## Running the Model
To run your newly created model, use the following command in your terminal:
```bash
ollama run metallama
```
### Sample Usage
In the command prompt, you can execute:
```bash
D:\>ollama run metallama
```
You can interact with the model like this:
```plaintext
>>> write a mini passage about space x
Space X, the private aerospace company founded by Elon Musk, is revolutionizing the field of space exploration.
With its ambitious goals to make humanity a multi-planetary species and establish a sustainable human presence in
the cosmos, Space X has become a leading player in the industry. The company's spacecraft, like the Falcon 9, have
demonstrated remarkable capabilities, allowing for the transport of crews and cargo into space with unprecedented
efficiency. As technology continues to advance, the possibility of establishing permanent colonies on Mars becomes
increasingly feasible, thanks in part to the success of reusable rockets that can launch multiple times without
sustaining significant damage. The journey towards becoming a multi-planetary species is underway, and Space X
plays a pivotal role in pushing the boundaries of human exploration and settlement.
```
---
## Conclusion
With these simple steps, you can easily download, install, and run your own models using Ollama. Whether you're exploring the capabilities of Llama or building your own custom models, Ollama makes it accessible and efficient.
- This README provides clear instructions and structured information to help users navigate the process of using Ollama effectively. Adjust any sections as needed based on your specific requirements or additional details you may want to include.
https://huggingface.co/prithivMLmods/Llama-3.2-1B-GGUF/resolve/main/Demo/gguf.gif

3
aya-expanse-8b-q2_k.gguf Normal file
View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:29d12456cd156762ac490558cf39f5004f49ce085cb3603588a88021212fd4b2
size 3438505408

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:690c85c052109d3ea3c57303342e2298c0de39dda3a4695e65685c21d8f4f605
size 4527975872

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:0e6bb47db51225994110be7d6a3368193c9f212bdddcd9018952768b97551a5d
size 4224937408

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:35d1ca5cdf6ddc207a70ab90a02571e873acb6866f1a27e1723a542550390961
size 3870518720

3
aya-expanse-8b-q4_0.gguf Normal file
View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:b3b852e75df95edd1d1ecc53ca2cd715e3dab43bd957c66ebbfc65479a64974c
size 4797459904

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:96e7e2feaeb229d97ae2bf31c9b93be0e1ab6f5836c179bc5287a5f3002ac1ce
size 5056982464

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:6d9288e864804a57772372fc83c43bf2bab20b901fc1ae1f964616d2bf204ec3
size 4828917184

3
aya-expanse-8b-q5_0.gguf Normal file
View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:67109311964532e316ddb484138a2d737c33564147b85c84442435d5bd558ea6
size 5669875136

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:a51ff84f315d57d976c57406f2b13a370896cb21cf31fe30996513d9e0a16dca
size 5803568576

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:bd008356780ad44756f5707fca770be38d01f906929c0ee3c037ea1b25e18ab5
size 5669875136

3
aya-expanse-8b-q6_k.gguf Normal file
View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:89d03c4428ec84708f3a33f346b71776b57e2f48889e8dadd754c46958706b10
size 6596816320

3
aya-expanse-8b-q8_0.gguf Normal file
View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:54ba366ce0a955a89964db43f155d8632282800381043983713870adc72532ab
size 8541072832

1
configuration.json Normal file
View File

@@ -0,0 +1 @@
{"framework": "pytorch", "task": "text-generation", "allow_remote": true}