Falcon-7B finetuned for mental health conversations
Top 98.6% on sourcepulse
This repository provides a fine-tuned Falcon-7B Large Language Model (LLM) for mental health conversational applications. It addresses the need for accessible, empathetic AI support in mental well-being, targeting users seeking preliminary assistance or information. The project leverages QLoRA for efficient fine-tuning on a custom conversational dataset derived from reputable health sources.
How It Works
The project fine-tunes the sharded Falcon-7B LLM using the QLoRA technique, which significantly reduces memory requirements by quantizing model weights and applying low-rank adaptations. This approach allows for efficient training on consumer-grade hardware, such as an Nvidia T4 GPU, while achieving a reported training loss of 0.031 after 320 steps. The model is trained on a curated conversational dataset anonymized and pre-processed from mental health FAQs, blogs, and wiki articles.
Quick Start & Requirements
gradio_chatbot_app.ipynb
notebook.max_steps < 150
), Python, Google Colab Pro (optional but recommended for faster training).Highlighted Details
Maintenance & Community
The project is maintained by iamarunbrahma. Further queries can be directed via GitHub Issues or blog comments.
Licensing & Compatibility
The repository does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The project explicitly states that the chatbot is not a replacement for professional mental health care. The dataset is derived from public sources, and the quality of responses may vary. No specific performance benchmarks against other models are provided.
1 year ago
Inactive