Model Card for Model ID
Falcon3-1B-MentalHealth is a fine-tuned version of the tiiuae/Falcon3-1B-Instruct model, adapted for providing empathetic and contextually relevant responses to mental health-related queries. Since it is fine-tuned on an Instruct model, it's responses are contextually appropriate and reasonable. The model has been trained on a curated dataset to assist in mental health conversations, offering advice, guidance, and support for individuals dealing with issues like stress, anxiety, and depression. It provides a compassionate approach to mental health queries while focusing on promoting emotional well-being and mental health awareness.
Important Note
As Mental Health is a sensitive topic, it would be preferable to use the code snippet provided below in order to get optimal results. It is expected that this model will be used responsibly.
Falcon3-1B-Instruct Fine-Tuned for Mental Health (LoRA)
This is a LoRA adapter for the Falcon3-1B-Instruct LLM which has been merged with the respective base model. It was fine-tuned on the 'marmikpandya/mental-health' dataset.
Usage
Dependencies
pip install transformers accelerate torch peft bitsandbytes --quiet
Basic Usage
import torch
import re
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
# Load the model from Hugging Face
model_name = "ShivomH/Falcon3-1B-MentalHealth"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
# Move the model to GPU if available
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model.to(device)
def chat():
    print("Chat with your fine-tuned Falcon model (type 'exit' to quit):")
    system_instruction = (
        "### Instruction:\n"
        "You are an empathetic AI specialized in mental health support. "
        "Do not respond to topics that are unrelated to the medical domain. \n"
        "If a crisis situation is detected, suggest reaching out to a mental health professional immediately. "
        "Your responses should be clear, precise, supportive, comforting and free from speculation."
    )
    # Store short chat history for context
    chat_history = []
    while True:
        user_input = input("\nYou: ")
        if user_input.lower() == "exit":
            break
        # Maintain short chat history (last 3 exchanges)
        chat_history.append(f"User: {user_input}")
        chat_history = chat_history[-1:]
        prompt = f"{system_instruction}\n\n" + "\n".join(chat_history) + "\nAssistant:"
        inputs = tokenizer(prompt, return_tensors="pt").to("cuda" if torch.cuda.is_available() else "cpu")
        with torch.no_grad():
            output = model.generate(
                **inputs,
                max_new_tokens=100,
                pad_token_id=tokenizer.eos_token_id,
                temperature=0.5,
                top_p=0.85,
                repetition_penalty=1.2,
                do_sample=True,
                no_repeat_ngram_size=3,
                early_stopping=True
            )
        response = tokenizer.decode(output[0], skip_special_tokens=True).strip()
        if "Assistant:" in response:
            response = response.split("Assistant:", 1)[-1].strip()
        # Remove URLs from the response
        response = re.sub(r'http[s]?://\S+', '', response)
        print(f"Assistant: {response}")
chat()
Model Description
- Developed by: Shivom Hatalkar
 - Model type: Text-generation
 - Language(s) (NLP): English
 - License: apache-2.0
 - Finetuned from model: Falcon3-1B-Instruct
 
Bias, Risks, and Limitations
- Not a Substitute for Professional Care: This model is not a licensed mental health professional. Its responses may be incomplete, inaccurate, or unsuitable for serious conditions.
 - Inherent Biases - May reflect biases in training data (e.g., cultural assumptions, stigmatizing language).
 - Crisis Limitations - Not designed for crisis intervention (e.g., suicidal ideation, self-harm). Always direct users to human professionals or emergency services.
 - Over-Reliance Risk - Outputs could inadvertently worsen symptoms if users interpret them as definitive advice.
 - Intended Use - Assist with general emotional support, not diagnosis or treatment.
 
Training Hyperparameters
| Hyperparameter | Value | 
|---|---|
| Precision | float16 | 
| Optimizer | AdamW_32bit | 
| Learning rate | 2e-4 | 
| Weight decay | 1e-2 | 
| Batch size | 2 | 
| Training Epochs | 3 | 
| Quantization | 8-Bit | 
| LoRA Dropout | 0.1 | 
| LoRA Rank | 16 | 
| Warmup Ratio | 0.03 | 
- Downloads last month
 - 1