初始化项目,由ModelHub XC社区提供模型

Model: Vlor999/mlx-UNfilteredAI-1B
Source: Original Platform
This commit is contained in:
ModelHub XC
2026-05-07 15:27:03 +08:00
commit 091d95710d
8 changed files with 277516 additions and 0 deletions

105
README.md Normal file
View File

@@ -0,0 +1,105 @@
---
license: mit
language:
- en
tags:
- UnfilteredAI
- mlx
library_name: mlx
base_model: UnfilteredAI/UNfilteredAI-1B
pipeline_tag: text-generation
---
# MLX UnfilteredAI-1B Model Card
This is the MLX version of the UnfilteredAI-1B model, optimized for efficient inference on Apple Silicon devices using Apple's MLX framework.
## Model Details
### Model Description
The MLX UnfilteredAI-1B is a converted version of the original UnfilteredAI-1B text generation model, adapted for Apple's MLX machine learning framework. This allows for fast and efficient text generation on macOS devices with M-series chips, leveraging the native capabilities of Apple Silicon.
- **Developed by:** UnfilteredAI
- **Model type:** Text generation language model
- **Language(s) (NLP):** English (primary), supports other languages
- **License:** MIT
- **Base model:** UnfilteredAI/UNfilteredAI-1B
### Model Sources
- **Repository:** https://huggingface.co/UnfilteredAI/UNfilteredAI-1B
- **MLX Conversion:** This repository
## Uses
### Direct Use
This model can be used for text generation tasks such as:
- Creative writing
- Conversational AI
- Educational and research applications
- Content generation without traditional filters
### Out-of-Scope Use
The model is uncensored and may generate sensitive, controversial, or harmful content. It should not be used for:
- Generating illegal or unethical content
- Misinformation or propaganda
- Any applications requiring content moderation
## Bias, Risks, and Limitations
As an uncensored model, it may exhibit biases present in the training data and generate inappropriate content. Users should be aware of potential risks including:
- Generation of biased or offensive text
- Potential for misuse in harmful applications
- Inconsistencies in output quality
### Recommendations
Use responsibly and with caution. Implement appropriate safeguards if deploying in production environments.
## How to Get Started with the Model
To use this model, you'll need to install the MLX library and mlx-lm:
```bash
# uv version
uv init .
uv add mlx mlx-lm
# pip version
pip install mlx mlx-lm
```
Then, load and use the model:
```python
from mlx_lm import load, generate
# Load the model from hugging-face
model, tokenizer = load("Vlor999/mlx-UNfilteredAI-1B")
# Generate text
prompt = "Hello, how are you?"
response = generate(model, tokenizer, prompt, max_tokens=100, verbose=True)
print(response)
```
Then run the model:
```bash
uv run python filename.py
# or by using the environment
source .venv/bin/activate
python filename.py
```
Or directly using mlx-lm:
```bash
uv run mlx_lm.chat --model Vlor999/mlx-UNfilteredAI-1B
# Example:
uv run mlx_lm.chat --model Vlor999/mlx-UNfilteredAI-1B --max-tokens=4096
# to have more informations about how to use mlx_lm you can run:
uv run mlx_lm.chat --help
```