base_model, datasets, inference, license, model_creator, model_name, pipeline_tag, quantized_by, tags, widget
base_model datasets inference license model_creator model_name pipeline_tag quantized_by tags widget
Felladrin/TinyMistral-248M-Alpaca
tatsu-lab/alpaca
false apache-2.0 Felladrin TinyMistral-248M-Alpaca text-generation afrideva
autotrain
text-generation
gguf
ggml
quantized
q2_k
q3_k_m
q4_k_m
q5_k_m
q6_k
q8_0
text
Below is an instruction that describes a task. Write a response that appropriately completes the request. ### Instruction: Tell me all you know about the Earth. ### Response:

Felladrin/TinyMistral-248M-Alpaca-GGUF

Quantized GGUF model files for TinyMistral-248M-Alpaca from Felladrin

Name Quant method Size
tinymistral-248m-alpaca.q2_k.gguf q2_k 115.26 MB
tinymistral-248m-alpaca.q3_k_m.gguf q3_k_m 130.08 MB
tinymistral-248m-alpaca.q4_k_m.gguf q4_k_m 155.67 MB
tinymistral-248m-alpaca.q5_k_m.gguf q5_k_m 179.23 MB
tinymistral-248m-alpaca.q6_k.gguf q6_k 204.26 MB
tinymistral-248m-alpaca.q8_0.gguf q8_0 264.32 MB

Original Model Card:

Locutusque's TinyMistral-248M Trained trained on the Alpaca dataset using AutoTrain

Prompt format

The training used these two following formats.

Prompt format #1

Below is an instruction that describes a task. Write a response that appropriately completes the request.

### Instruction:
<instruction>

### Response:

Prompt format #2

Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.

### Instruction:
<instruction>

### Input:
<input>

### Response:
Description
Model synced from source: afrideva/TinyMistral-248M-Alpaca-GGUF
Readme 26 KiB