101 lines
3.1 KiB
Markdown
101 lines
3.1 KiB
Markdown
---
|
|
license: other
|
|
license_name: helpingai
|
|
license_link: https://helpingai.co/license
|
|
pipeline_tag: text-generation
|
|
language:
|
|
- en
|
|
tags:
|
|
- HelpingAI
|
|
- Emotionally-Intelligent
|
|
- EQ-focused
|
|
- Conversational
|
|
- SLM
|
|
library_name: transformers
|
|
---
|
|
|
|
# HelpingAI3
|
|
|
|
## Model Description
|
|
|
|
**HelpingAI3** is an advanced language model developed to excel in emotionally intelligent conversations. Building upon the foundations of HelpingAI2.5, this model offers enhanced emotional understanding and contextual awareness.
|
|
|
|
## Model Details
|
|
|
|
- **Developed by**: HelpingAI
|
|
- **Model type**: Decoder-only large language model
|
|
- **Language**: English
|
|
- **License**: [HelpingAI License](https://helpingai.co/license)
|
|
|
|
## Training Data
|
|
|
|
HelpingAI3 was trained on a diverse dataset comprising:
|
|
|
|
- **Emotional Dialogues**: 15 million rows to enhance conversational intelligence.
|
|
- **Therapeutic Exchanges**: 3 million rows aimed at providing advanced emotional support.
|
|
- **Cultural Conversations**: 250,000 rows to improve global awareness.
|
|
- **Crisis Response Scenarios**: 1 million rows to better handle emergency situations.
|
|
|
|
## Training Procedure
|
|
|
|
The model underwent the following training processes:
|
|
|
|
- **Base Model**: Initiated from HelpingAI2.5.
|
|
- **Emotional Intelligence Training**: Employed Reinforcement Learning for Emotion Understanding (RLEU) and context-aware conversational fine-tuning.
|
|
- **Optimization**: Utilized mixed-precision training and advanced token efficiency techniques.
|
|
|
|
## Intended Use
|
|
|
|
HelpingAI3 is designed for:
|
|
|
|
- **AI Companionship & Emotional Support**: Offering empathetic interactions.
|
|
- **Therapy & Wellbeing Guidance**: Assisting in mental health support.
|
|
- **Personalized Learning**: Tailoring educational content to individual needs.
|
|
- **Professional AI Assistance**: Enhancing productivity in professional settings.
|
|
|
|
## Limitations
|
|
|
|
While HelpingAI3 strives for high emotional intelligence, users should be aware of potential limitations:
|
|
|
|
- **Biases**: The model may inadvertently reflect biases present in the training data.
|
|
- **Understanding Complex Emotions**: There might be challenges in accurately interpreting nuanced human emotions.
|
|
- **Not a Substitute for Professional Help**: For serious emotional or psychological issues, consulting a qualified professional is recommended.
|
|
|
|
## How to Use
|
|
|
|
### Using Transformers
|
|
|
|
```python
|
|
import torch
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
|
# Load the HelpingAI3 model
|
|
model = AutoModelForCausalLM.from_pretrained("HelpingAI/HelpingAI-3")
|
|
# Load the tokenizer
|
|
tokenizer = AutoTokenizer.from_pretrained("HelpingAI/HelpingAI-3")
|
|
|
|
# Define the chat input
|
|
chat = [
|
|
{"role": "system", "content": "You are HelpingAI, an emotional AI. Always answer my questions in the HelpingAI style."},
|
|
{"role": "user", "content": "Introduce yourself."}
|
|
]
|
|
|
|
inputs = tokenizer.apply_chat_template(
|
|
chat,
|
|
add_generation_prompt=True,
|
|
return_tensors="pt"
|
|
).to(model.device)
|
|
|
|
# Generate text
|
|
outputs = model.generate(
|
|
inputs,
|
|
max_new_tokens=256,
|
|
do_sample=True,
|
|
temperature=0.6,
|
|
top_p=0.9,
|
|
)
|
|
|
|
response = outputs[0][inputs.shape[-1]:]
|
|
print(tokenizer.decode(response, skip_special_tokens=True))
|
|
```
|