2024-12-20 02:39:21 +00:00
|
|
|
---
|
2024-12-20 10:46:14 +08:00
|
|
|
language:
|
|
|
|
|
- en
|
|
|
|
|
tags:
|
|
|
|
|
- meta
|
|
|
|
|
- llama-3
|
|
|
|
|
- mlx
|
|
|
|
|
pipeline_tag: text-generation
|
2024-12-20 02:39:21 +00:00
|
|
|
---
|
|
|
|
|
|
2024-12-20 10:46:14 +08:00
|
|
|
# mlx-community/Llama-3-8B-Instruct-262k-unquantized
|
|
|
|
|
This model was converted to MLX format from [`gradientai/Llama-3-8B-Instruct-262k`]() using mlx-lm version **0.12.0**.
|
|
|
|
|
Refer to the [original model card](https://huggingface.co/gradientai/Llama-3-8B-Instruct-262k) for more details on the model.
|
|
|
|
|
## Use with mlx
|
|
|
|
|
|
2024-12-20 02:39:21 +00:00
|
|
|
```bash
|
2024-12-20 10:46:14 +08:00
|
|
|
pip install mlx-lm
|
2024-12-20 02:39:21 +00:00
|
|
|
```
|
2024-12-20 10:46:14 +08:00
|
|
|
|
2024-12-20 02:39:21 +00:00
|
|
|
```python
|
2024-12-20 10:46:14 +08:00
|
|
|
from mlx_lm import load, generate
|
2024-12-20 02:39:21 +00:00
|
|
|
|
2024-12-20 10:46:14 +08:00
|
|
|
model, tokenizer = load("mlx-community/Llama-3-8B-Instruct-262k-unquantized")
|
|
|
|
|
response = generate(model, tokenizer, prompt="hello", verbose=True)
|
|
|
|
|
```
|