| | --- |
| | base_model: teknium/OpenHermes-2.5-Mistral-7B |
| | library_name: peft |
| | pipeline_tag: text-generation |
| | tags: |
| | - base_model:adapter:teknium/OpenHermes-2.5-Mistral-7B |
| | - lora |
| | - medical |
| | - spinal-cord-injury |
| | - healthcare |
| | - assistant |
| | --- |
| | |
| | # SCI Assistant - Spinal Cord Injury Specialized AI Assistant |
| | A specialized AI assistant fine-tuned specifically for people with spinal cord injuries (SCI). This model is based on OpenHermes-2.5-Mistral-7B and has been trained using a two-phase approach with LoRA (Low-Rank Adaptation) to provide contextually appropriate and medically-informed responses for the SCI community. |
| |
|
| | ## Model Description |
| |
|
| | This model was fine-tuned using a two-phase training approach: |
| | 1. **Phase 1**: Domain pretraining on SCI-related medical texts and resources |
| | 2. **Phase 2**: Instruction tuning on conversational SCI-focused Q&A pairs |
| |
|
| | The model understands the unique challenges, medical realities, and daily life considerations of individuals living with spinal cord injuries. |
| |
|
| | ## Training Details |
| |
|
| | - **Base Model**: teknium/OpenHermes-2.5-Mistral-7B |
| | - **Training Method**: QLoRA (4-bit quantization with LoRA adapters) |
| | - **Training Data**: 119,117 total entries (35,779 domain text + 83,337 instruction pairs) |
| | - **Hardware**: RTX 4070 Super (8GB VRAM) |
| | - **Training Time**: ~20 hours total (Phase 1 + Phase 2) |
| |
|
| | ## Usage |
| |
|
| | ```python |
| | from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig |
| | from peft import PeftModel |
| | import torch |
| | |
| | # Load model |
| | bnb_config = BitsAndBytesConfig( |
| | load_in_4bit=True, |
| | bnb_4bit_compute_dtype=torch.float16, |
| | ) |
| | |
| | base_model = AutoModelForCausalLM.from_pretrained( |
| | "teknium/OpenHermes-2.5-Mistral-7B", |
| | quantization_config=bnb_config, |
| | device_map="auto" |
| | ) |
| | |
| | model = PeftModel.from_pretrained(base_model, "basiphobe/sci-assistant") |
| | tokenizer = AutoTokenizer.from_pretrained("basiphobe/sci-assistant") |
| | |
| | # Format prompt with SCI context |
| | system_context = "You are a specialized medical assistant for people with spinal cord injuries. Your responses should always consider the unique needs, challenges, and medical realities of individuals living with SCI." |
| | |
| | prompt = f"{system_context}\n\n### Instruction:\n{your_question}\n\n### Response:\n" |
| | |
| | # Generate response |
| | inputs = tokenizer(prompt, return_tensors="pt") |
| | outputs = model.generate(**inputs, max_new_tokens=256, temperature=0.7) |
| | response = tokenizer.decode(outputs[0][inputs["input_ids"].shape[1]:], skip_special_tokens=True) |
| | ``` |
| |
|
| | ## Intended Use |
| |
|
| | This model is designed to: |
| | - Provide SCI-specific information and guidance |
| | - Answer questions about daily life with spinal cord injuries |
| | - Offer practical advice for common SCI challenges |
| | - Support the SCI community with contextually appropriate responses |
| |
|
| | ## Limitations |
| |
|
| | - This model is for informational purposes only and should not replace professional medical advice |
| | - Always consult with healthcare providers for medical decisions |
| | - The model may not have information about the latest medical developments |
| | - Responses should be verified with medical professionals when making health-related decisions |
| |
|
| | ## Direct Use |
| |
|
| | This model can be used directly for: |
| | - Educational purposes about spinal cord injuries |
| | - Providing general information and support to the SCI community |
| | - Research into specialized medical AI assistants |
| | - Personal use by individuals seeking SCI-related information |
| |
|
| | The model is designed to provide contextually appropriate responses that consider the unique challenges and medical realities of spinal cord injuries. |
| |
|
| | ### Downstream Use |
| |
|
| | This model can be fine-tuned further for: |
| | - Integration into healthcare applications |
| | - Specialized medical chatbots for rehabilitation centers |
| | - Educational platforms for SCI awareness and training |
| | - Research applications in medical AI |
| | - Custom applications for SCI support organizations |
| |
|
| | When used in downstream applications, implementers should: |
| | - Maintain the medical disclaimer requirements |
| | - Ensure proper supervision by medical professionals |
| | - Implement appropriate safety measures and content filtering |
| | - Validate outputs for medical accuracy in their specific use case |
| |
|
| | ### Out-of-Scope Use |
| |
|
| | This model should NOT be used for: |
| | - **Medical diagnosis or treatment decisions** - Always consult healthcare professionals |
| | - **Emergency medical situations** - Seek immediate professional medical help |
| | - **Legal or financial advice** related to SCI cases |
| | - **Replacement for professional medical consultation** |
| | - **Clinical decision-making** without physician oversight |
| | - **Applications targeting vulnerable populations** without proper safeguards |
| | - **Commercial medical applications** without appropriate medical validation and oversight |
| |
|
| | ## Bias, Risks, and Limitations |
| |
|
| | ### Medical Limitations |
| | - **Not a substitute for medical professionals**: All medical advice should be verified with qualified healthcare providers |
| | - **Training data limitations**: May not include the most recent medical research or treatments |
| | - **Individual variation**: SCI affects individuals differently; responses may not apply to all cases |
| | - **Geographic bias**: Training data may be biased toward certain healthcare systems or regions |
| |
|
| | ### Technical Limitations |
| | - **Hallucination risk**: Like all language models, may generate plausible-sounding but incorrect information |
| | - **Context limitations**: Limited by input context window and may not retain information across long conversations |
| | - **Language limitations**: Primarily trained on English content |
| | - **Update lag**: Cannot access real-time medical research or current events |
| |
|
| | ### Bias Considerations |
| | - **Training data bias**: Reflects biases present in source medical literature and online content |
| | - **Demographic representation**: May not equally represent all demographics within the SCI community |
| | - **Healthcare access bias**: May reflect biases toward certain types of healthcare systems |
| | - **Severity bias**: May be more informed about certain types or severities of SCI |
| |
|
| | ### Risk Mitigation |
| | - Always include medical disclaimers when using this model |
| | - Implement content filtering for harmful or dangerous advice |
| | - Regular evaluation by medical professionals is recommended |
| | - Monitor outputs for accuracy and appropriateness |
| |
|
| | ### Recommendations |
| |
|
| | <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> |
| |
|
| | ### Recommendations |
| |
|
| | Users should be aware of the following recommendations: |
| |
|
| | **For Direct Users:** |
| | - Always verify medical information with qualified healthcare professionals |
| | - Use responses as educational/informational starting points, not definitive advice |
| | - Be aware that individual SCI experiences vary significantly |
| | - Seek immediate professional help for urgent medical concerns |
| |
|
| | **For Developers/Implementers:** |
| | - Implement clear medical disclaimers in any application using this model |
| | - Provide easy access to professional medical resources alongside model responses |
| | - Consider implementing content filtering for potentially harmful advice |
| | - Regular review by medical professionals is strongly recommended |
| | - Ensure compliance with relevant healthcare regulations (HIPAA, etc.) |
| |
|
| | **For Healthcare Organizations:** |
| | - Professional medical oversight is essential when implementing in clinical settings |
| | - Regular validation of model outputs against current medical standards |
| | - Integration should complement, not replace, professional medical consultation |
| | - Staff training on AI limitations and appropriate use cases |
| |
|
| | ## Training Details |
| |
|
| | ### Training Data |
| |
|
| | The training dataset consisted of 119,117 carefully curated entries focused on spinal cord injury information: |
| |
|
| | **Domain Pretraining Data (35,779 entries):** |
| | - Medical literature and research papers on SCI |
| | - Educational materials from reputable SCI organizations |
| | - Clinical guidelines and treatment protocols |
| | - Rehabilitation and therapy documentation |
| | - Patient education resources |
| |
|
| | **Instruction Tuning Data (83,337 entries):** |
| | - SCI-focused question-answer pairs |
| | - Conversational examples with appropriate medical context |
| | - Real-world scenarios and practical advice situations |
| | - Educational Q&A formatted for instruction following |
| |
|
| | All training data was filtered and curated to ensure: |
| | - Sources from reputable medical organizations and healthcare professionals |
| | - Content originally created or reviewed by medical professionals in the SCI field |
| | - Appropriate tone and sensitivity for SCI community |
| | - Removal of potentially harmful or dangerous advice |
| | - Proper medical disclaimers and context |
| |
|
| | **Note**: While the source materials were created by medical professionals, this model itself has not undergone independent medical validation. |
| |
|
| | ### Training Procedure |
| |
|
| | The model was trained using a two-phase approach with QLoRA (Quantized Low-Rank Adaptation): |
| |
|
| | **Phase 1 - Domain Pretraining:** |
| | - Focus: Medical terminology and SCI-specific knowledge |
| | - Duration: 2 epochs (~8 hours) |
| | - Data: 35,779 domain text entries |
| | - Objective: Adapt base model to SCI medical domain |
| |
|
| | **Phase 2 - Instruction Tuning:** |
| | - Focus: Conversational abilities and response formatting |
| | - Duration: 2 epochs (~12 hours) |
| | - Data: 83,337 instruction-response pairs |
| | - Objective: Teach appropriate response patterns and tone |
| |
|
| | #### Preprocessing |
| |
|
| | Training data underwent extensive preprocessing: |
| | - Content sourced from materials created by healthcare professionals |
| | - Sensitive content filtering and safety checks |
| | - Standardized formatting for instruction-following |
| | - Quality filtering to remove low-quality or inappropriate content |
| | - Tokenization optimization for efficient training |
| |
|
| | #### Training Hyperparameters |
| |
|
| | - **Training regime:** 4-bit quantization with LoRA adapters (QLoRA) |
| | - **Learning rate:** 2e-4 with cosine scheduling |
| | - **LoRA rank:** 16 |
| | - **LoRA alpha:** 32 |
| | - **LoRA dropout:** 0.05 |
| | - **Target modules:** q_proj, v_proj |
| | - **Batch size:** 4 with gradient accumulation |
| | - **Max sequence length:** 512 tokens |
| | - **Optimizer:** AdamW with weight decay |
| |
|
| | #### Speeds, Sizes, Times |
| |
|
| | - **Total training time:** ~20 hours (8h Phase 1 + 12h Phase 2) |
| | - **Hardware:** RTX 4070 Super (8GB VRAM) |
| | - **Final model size:** 30MB (LoRA adapter only) |
| | - **Base model size:** 7B parameters (not included in adapter) |
| | - **Training throughput:** ~3.5 samples/second average |
| | - **Memory usage:** 6-7GB VRAM during training |
| |
|
| | ## Evaluation |
| |
|
| | ### Testing Data, Factors & Metrics |
| |
|
| | #### Testing Data |
| |
|
| | The model was evaluated using: |
| | - Held-out test set of SCI-related questions (500 samples) |
| | - Manual review of response quality and appropriateness |
| | - Comparative analysis against general-purpose models on SCI topics |
| | - Assessment of domain-specific knowledge retention |
| |
|
| | **Note**: Evaluation was conducted by the model developer, not independent medical professionals. |
| |
|
| | #### Factors |
| |
|
| | Evaluation considered multiple factors: |
| | - **Medical accuracy**: Correctness of SCI-related information |
| | - **Appropriateness**: Sensitivity and tone for SCI community |
| | - **Contextual relevance**: Understanding of SCI-specific challenges |
| | - **Safety**: Avoidance of harmful or dangerous advice |
| | - **Completeness**: Comprehensive responses to complex questions |
| |
|
| | #### Metrics |
| |
|
| | - **Medical accuracy score**: Based on consistency with source medical literature (not independently validated) |
| | - **Appropriateness rating**: Developer assessment of tone and sensitivity (4.2/5.0 subjective rating) |
| | - **Response relevance**: SCI-specific context understanding (82% relevance score) |
| | - **Safety compliance**: No obviously harmful medical advice detected in test samples |
| | - **Response quality**: Perplexity improvements over base model for SCI domain |
| |
|
| | ### Results |
| |
|
| | **Quantitative Results:** |
| | - 40% improvement in SCI domain perplexity over base model |
| | - Responses demonstrate consistency with source medical literature |
| | - 95% safety compliance (no obviously harmful medical advice detected) |
| | - 82% average relevance score for SCI-specific contexts |
| |
|
| | **Qualitative Results:** |
| | - Responses demonstrate clear understanding of SCI terminology and concepts |
| | - Appropriate tone and sensitivity for disability community |
| | - Consistent inclusion of medical disclaimers |
| | - Good balance between being helpful and cautious about medical advice |
| |
|
| | **Limitations of Evaluation:** |
| | - Evaluation conducted by model developer, not independent medical experts |
| | - No formal clinical validation or testing with SCI patients |
| | - Results based on consistency with training sources, not independent medical verification |
| |
|
| | ## Environmental Impact |
| |
|
| | Training carbon emissions estimated using energy consumption data: |
| |
|
| | - **Hardware Type:** RTX 4070 Super (8GB VRAM) |
| | - **Hours used:** ~20 hours total training time |
| | - **Cloud Provider:** Local training (personal hardware) |
| | - **Compute Region:** North America |
| | - **Carbon Emitted:** Approximately 2.1 kg CO2eq (estimated based on local energy grid) |
| |
|
| | The use of QLoRA significantly reduced training time and energy consumption compared to full fine-tuning methods, making this a relatively efficient training approach. |
| |
|
| | ## Technical Specifications |
| |
|
| | ### Model Architecture and Objective |
| |
|
| | - **Base Architecture:** Mistral 7B transformer model |
| | - **Adaptation Method:** QLoRA (Quantized Low-Rank Adaptation) |
| | - **Objective:** Causal language modeling with SCI domain specialization |
| | - **Quantization:** 4-bit precision for memory efficiency |
| | - **LoRA Configuration:** Rank-16 adapters on attention projection layers |
| |
|
| | ### Compute Infrastructure |
| |
|
| | #### Hardware |
| |
|
| | - **GPU:** NVIDIA RTX 4070 Super (8GB VRAM) |
| | - **CPU:** Modern multi-core processor |
| | - **RAM:** 32GB system memory |
| | - **Storage:** NVMe SSD for fast data loading |
| |
|
| | #### Software |
| |
|
| | - **Framework:** Transformers 4.36+, PEFT 0.16.0 |
| | - **Training:** QLoRA with bitsandbytes quantization |
| | - **Environment:** Python 3.10+, PyTorch 2.0+, CUDA 12.1 |
| |
|
| | ## Citation |
| |
|
| | If you use this model in your research or applications, please cite: |
| |
|
| | **BibTeX:** |
| | ```bibtex |
| | @misc{sci_assistant_2025, |
| | title={SCI Assistant: A Specialized AI Assistant for Spinal Cord Injury Support}, |
| | author={basiphobe}, |
| | year={2025}, |
| | howpublished={Hugging Face Model Repository}, |
| | url={https://huggingface.co/basiphobe/sci-assistant} |
| | } |
| | ``` |
| |
|
| | **APA:** |
| | basiphobe. (2025). *SCI Assistant: A Specialized AI Assistant for Spinal Cord Injury Support*. Hugging Face. https://huggingface.co/basiphobe/sci-assistant |
| |
|
| | ## Glossary |
| |
|
| | **SCI**: Spinal Cord Injury - damage to the spinal cord that results in temporary or permanent changes in function |
| |
|
| | **QLoRA**: Quantized Low-Rank Adaptation - an efficient fine-tuning method that reduces memory requirements |
| |
|
| | **Domain Pretraining**: Training phase focused on learning domain-specific terminology and knowledge |
| |
|
| | **Instruction Tuning**: Training phase focused on learning conversational patterns and response formatting |
| |
|
| | **Perplexity**: A metric measuring how well a language model predicts text (lower is better) |
| |
|
| | **LoRA**: Low-Rank Adaptation - parameter-efficient fine-tuning technique |
| |
|
| | ## Model Card Authors |
| |
|
| | **Primary Author:** basiphobe |
| | **Model Development:** Individual research project for SCI community support |
| | **Data Sources:** Curated from medical literature and educational materials created by healthcare professionals |
| | **Validation Status:** Model has not undergone independent medical professional validation |
| |
|
| | ## Model Card Contact |
| |
|
| | For questions, issues, or feedback regarding this model: |
| | - **Hugging Face:** https://huggingface.co/basiphobe/sci-assistant |
| | - **Issues:** Please report issues through Hugging Face model repository |
| | - **Medical Concerns:** Always consult qualified healthcare professionals |
| |
|
| | **Important Note:** This model is provided for educational and informational purposes. Always seek professional medical advice for health-related questions and decisions. |
| | ### Framework versions |
| |
|
| | - PEFT 0.16.0 |