| --- |
| license: other |
| license_name: helpingai |
| license_link: https://helpingai.co/license |
| pipeline_tag: text-generation |
| language: |
| - en |
| tags: |
| - HelpingAI |
| - Emotionally-Intelligent |
| - EQ-focused |
| - Conversational |
| - SLM |
| library_name: transformers |
| --- |
| |
| # HelpingAI3 |
|
|
| ## Model Description |
|
|
| **HelpingAI3** is an advanced language model developed to excel in emotionally intelligent conversations. Building upon the foundations of HelpingAI2.5, this model offers enhanced emotional understanding and contextual awareness. |
|
|
| ## Model Details |
|
|
| - **Developed by**: HelpingAI |
| - **Model type**: Decoder-only large language model |
| - **Language**: English |
| - **License**: [HelpingAI License](https://helpingai.co/license) |
|
|
| ## Training Data |
|
|
| HelpingAI3 was trained on a diverse dataset comprising: |
|
|
| - **Emotional Dialogues**: 15 million rows to enhance conversational intelligence. |
| - **Therapeutic Exchanges**: 3 million rows aimed at providing advanced emotional support. |
| - **Cultural Conversations**: 250,000 rows to improve global awareness. |
| - **Crisis Response Scenarios**: 1 million rows to better handle emergency situations. |
|
|
| ## Training Procedure |
|
|
| The model underwent the following training processes: |
|
|
| - **Base Model**: Initiated from HelpingAI2.5. |
| - **Emotional Intelligence Training**: Employed Reinforcement Learning for Emotion Understanding (RLEU) and context-aware conversational fine-tuning. |
| - **Optimization**: Utilized mixed-precision training and advanced token efficiency techniques. |
|
|
| ## Intended Use |
|
|
| HelpingAI3 is designed for: |
|
|
| - **AI Companionship & Emotional Support**: Offering empathetic interactions. |
| - **Therapy & Wellbeing Guidance**: Assisting in mental health support. |
| - **Personalized Learning**: Tailoring educational content to individual needs. |
| - **Professional AI Assistance**: Enhancing productivity in professional settings. |
|
|
| ## Limitations |
|
|
| While HelpingAI3 strives for high emotional intelligence, users should be aware of potential limitations: |
|
|
| - **Biases**: The model may inadvertently reflect biases present in the training data. |
| - **Understanding Complex Emotions**: There might be challenges in accurately interpreting nuanced human emotions. |
| - **Not a Substitute for Professional Help**: For serious emotional or psychological issues, consulting a qualified professional is recommended. |
|
|
| ## How to Use |
|
|
| ### Using Transformers |
|
|
| ```python |
| import torch |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| |
| # Load the HelpingAI3 model |
| model = AutoModelForCausalLM.from_pretrained("HelpingAI/HelpingAI-3") |
| # Load the tokenizer |
| tokenizer = AutoTokenizer.from_pretrained("HelpingAI/HelpingAI-3") |
| |
| # Define the chat input |
| chat = [ |
| {"role": "system", "content": "You are HelpingAI, an emotional AI. Always answer my questions in the HelpingAI style."}, |
| {"role": "user", "content": "Introduce yourself."} |
| ] |
| |
| inputs = tokenizer.apply_chat_template( |
| chat, |
| add_generation_prompt=True, |
| return_tensors="pt" |
| ).to(model.device) |
| |
| # Generate text |
| outputs = model.generate( |
| inputs, |
| max_new_tokens=256, |
| do_sample=True, |
| temperature=0.6, |
| top_p=0.9, |
| ) |
| |
| response = outputs[0][inputs.shape[-1]:] |
| print(tokenizer.decode(response, skip_special_tokens=True)) |
| ``` |
|
|