Update README.md
Browse files
README.md
CHANGED
|
@@ -1,201 +1,83 @@
|
|
| 1 |
-
|
| 2 |
-
library_name: transformers
|
| 3 |
-
tags: []
|
| 4 |
-
---
|
| 5 |
-
|
| 6 |
-
# Model Card for Model ID
|
| 7 |
-
|
| 8 |
-
<!-- Provide a quick summary of what the model is/does. -->
|
| 9 |
|
|
|
|
| 10 |
|
|
|
|
| 11 |
|
| 12 |
## Model Details
|
| 13 |
|
| 14 |
### Model Description
|
| 15 |
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
| 19 |
-
|
| 20 |
-
- **Developed by:** [More Information Needed]
|
| 21 |
-
- **Funded by [optional]:** [More Information Needed]
|
| 22 |
-
- **Shared by [optional]:** [More Information Needed]
|
| 23 |
-
- **Model type:** [More Information Needed]
|
| 24 |
-
- **Language(s) (NLP):** [More Information Needed]
|
| 25 |
-
- **License:** [More Information Needed]
|
| 26 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
| 27 |
|
| 28 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 29 |
|
| 30 |
-
|
| 31 |
|
| 32 |
-
- **Repository:**
|
| 33 |
-
- **Paper
|
| 34 |
-
- **Demo
|
| 35 |
|
| 36 |
## Uses
|
| 37 |
|
| 38 |
-
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 39 |
-
|
| 40 |
### Direct Use
|
| 41 |
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
[More Information Needed]
|
| 45 |
|
| 46 |
-
### Downstream Use
|
| 47 |
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
[More Information Needed]
|
| 51 |
|
| 52 |
### Out-of-Scope Use
|
| 53 |
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
[More Information Needed]
|
| 57 |
|
| 58 |
## Bias, Risks, and Limitations
|
| 59 |
|
| 60 |
-
|
| 61 |
-
|
| 62 |
-
[More Information Needed]
|
| 63 |
|
| 64 |
### Recommendations
|
| 65 |
|
| 66 |
-
|
| 67 |
-
|
| 68 |
-
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
| 69 |
|
| 70 |
## How to Get Started with the Model
|
| 71 |
|
| 72 |
-
|
| 73 |
-
|
| 74 |
-
[More Information Needed]
|
| 75 |
|
| 76 |
## Training Details
|
| 77 |
|
| 78 |
### Training Data
|
| 79 |
|
| 80 |
-
|
| 81 |
-
|
| 82 |
-
[More Information Needed]
|
| 83 |
|
| 84 |
-
### Training Procedure
|
| 85 |
|
| 86 |
-
|
| 87 |
-
|
| 88 |
-
#### Preprocessing [optional]
|
| 89 |
-
|
| 90 |
-
[More Information Needed]
|
| 91 |
|
|
|
|
| 92 |
|
| 93 |
#### Training Hyperparameters
|
| 94 |
|
| 95 |
-
-
|
| 96 |
-
|
| 97 |
-
#### Speeds, Sizes, Times [optional]
|
| 98 |
-
|
| 99 |
-
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
| 100 |
-
|
| 101 |
-
[More Information Needed]
|
| 102 |
|
| 103 |
## Evaluation
|
| 104 |
|
| 105 |
-
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 106 |
-
|
| 107 |
### Testing Data, Factors & Metrics
|
| 108 |
|
| 109 |
-
|
| 110 |
-
|
| 111 |
-
<!-- This should link to a Dataset Card if possible. -->
|
| 112 |
-
|
| 113 |
-
[More Information Needed]
|
| 114 |
-
|
| 115 |
-
#### Factors
|
| 116 |
-
|
| 117 |
-
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
| 118 |
-
|
| 119 |
-
[More Information Needed]
|
| 120 |
-
|
| 121 |
-
#### Metrics
|
| 122 |
-
|
| 123 |
-
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
| 124 |
-
|
| 125 |
-
[More Information Needed]
|
| 126 |
-
|
| 127 |
-
### Results
|
| 128 |
-
|
| 129 |
-
[More Information Needed]
|
| 130 |
-
|
| 131 |
-
#### Summary
|
| 132 |
-
|
| 133 |
-
|
| 134 |
-
|
| 135 |
-
## Model Examination [optional]
|
| 136 |
-
|
| 137 |
-
<!-- Relevant interpretability work for the model goes here -->
|
| 138 |
-
|
| 139 |
-
[More Information Needed]
|
| 140 |
|
| 141 |
## Environmental Impact
|
| 142 |
|
| 143 |
-
|
| 144 |
-
|
| 145 |
-
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
| 146 |
-
|
| 147 |
-
- **Hardware Type:** [More Information Needed]
|
| 148 |
-
- **Hours used:** [More Information Needed]
|
| 149 |
-
- **Cloud Provider:** [More Information Needed]
|
| 150 |
-
- **Compute Region:** [More Information Needed]
|
| 151 |
-
- **Carbon Emitted:** [More Information Needed]
|
| 152 |
-
|
| 153 |
-
## Technical Specifications [optional]
|
| 154 |
-
|
| 155 |
-
### Model Architecture and Objective
|
| 156 |
-
|
| 157 |
-
[More Information Needed]
|
| 158 |
-
|
| 159 |
-
### Compute Infrastructure
|
| 160 |
-
|
| 161 |
-
[More Information Needed]
|
| 162 |
-
|
| 163 |
-
#### Hardware
|
| 164 |
-
|
| 165 |
-
[More Information Needed]
|
| 166 |
-
|
| 167 |
-
#### Software
|
| 168 |
-
|
| 169 |
-
[More Information Needed]
|
| 170 |
-
|
| 171 |
-
## Citation [optional]
|
| 172 |
-
|
| 173 |
-
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 174 |
-
|
| 175 |
-
**BibTeX:**
|
| 176 |
-
|
| 177 |
-
[More Information Needed]
|
| 178 |
-
|
| 179 |
-
**APA:**
|
| 180 |
-
|
| 181 |
-
[More Information Needed]
|
| 182 |
-
|
| 183 |
-
## Glossary [optional]
|
| 184 |
-
|
| 185 |
-
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 186 |
-
|
| 187 |
-
[More Information Needed]
|
| 188 |
-
|
| 189 |
-
## More Information [optional]
|
| 190 |
-
|
| 191 |
-
[More Information Needed]
|
| 192 |
-
|
| 193 |
-
## Model Card Authors [optional]
|
| 194 |
-
|
| 195 |
-
[More Information Needed]
|
| 196 |
|
| 197 |
-
##
|
| 198 |
|
| 199 |
-
|
| 200 |
|
|
|
|
| 201 |
|
|
|
|
|
|
| 1 |
+
# Model Card for Frontida's T5-Small Model
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
|
| 3 |
+
## Introduction
|
| 4 |
|
| 5 |
+
This model card presents an overview of the `t5-small` model as adapted and fine-tuned for Frontida, a project dedicated to supporting new mothers through the challenges of postpartum depression. Frontida leverages the `t5-small` model to understand and respond to user queries with empathy and accuracy.
|
| 6 |
|
| 7 |
## Model Details
|
| 8 |
|
| 9 |
### Model Description
|
| 10 |
|
| 11 |
+
The `t5-small` model, developed by Google and fine-tuned by the Frontida team, serves as the backbone of our chatbot's natural language processing capabilities. This version of the T5 model is optimized for efficiency, enabling quick and reliable responses within our application. It has been adapted to specifically address the nuances and complexities of conversations surrounding postpartum depression.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
|
| 13 |
+
- **Developed by:** Google, with fine-tuning by the Frontida team
|
| 14 |
+
- **Model type:** Text-to-Text Transfer Transformer (T5) Small
|
| 15 |
+
- **Language(s) (NLP):** Primarily English, with plans to support additional languages
|
| 16 |
+
- **License:** Apache 2.0
|
| 17 |
+
- **Finetuned from:** Google’s original `t5-small` model
|
| 18 |
|
| 19 |
+
### Model Sources
|
| 20 |
|
| 21 |
+
- **Repository:** Available on Hugging Face (link to Frontida’s repository)
|
| 22 |
+
- **Paper:** "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer" (Raffel et al., 2019)
|
| 23 |
+
- **Demo:** Frontida Chatbot Interface (link to demo if available)
|
| 24 |
|
| 25 |
## Uses
|
| 26 |
|
|
|
|
|
|
|
| 27 |
### Direct Use
|
| 28 |
|
| 29 |
+
Frontida's `t5-small` model is directly used within our chatbot interface to provide immediate, contextually relevant support and video recommendations for mothers experiencing postpartum depression.
|
|
|
|
|
|
|
| 30 |
|
| 31 |
+
### Downstream Use
|
| 32 |
|
| 33 |
+
While primarily designed for direct interaction within Frontida, the model's applications can extend to other mental health support systems, offering a foundation for empathetic, AI-driven conversation.
|
|
|
|
|
|
|
| 34 |
|
| 35 |
### Out-of-Scope Use
|
| 36 |
|
| 37 |
+
The model is not intended for clinical diagnosis or as a substitute for professional healthcare advice.
|
|
|
|
|
|
|
| 38 |
|
| 39 |
## Bias, Risks, and Limitations
|
| 40 |
|
| 41 |
+
We acknowledge the potential for biases in AI models and have taken steps to mitigate such risks in `t5-small`. However, users should be aware of the model's limitations, particularly in understanding the full scope of an individual's emotional state.
|
|
|
|
|
|
|
| 42 |
|
| 43 |
### Recommendations
|
| 44 |
|
| 45 |
+
Users are encouraged to use Frontida as a supplementary support tool alongside traditional mental health resources. Ongoing model training and refinement are priorities to ensure the most empathetic and accurate responses.
|
|
|
|
|
|
|
| 46 |
|
| 47 |
## How to Get Started with the Model
|
| 48 |
|
| 49 |
+
To interact with Frontida's `t5-small` model, users can access our chatbot via the Frontida web application. Developers interested in exploring the model's architecture and training can visit our repository on Hugging Face.
|
|
|
|
|
|
|
| 50 |
|
| 51 |
## Training Details
|
| 52 |
|
| 53 |
### Training Data
|
| 54 |
|
| 55 |
+
The model was fine-tuned on a curated dataset comprising diverse conversations and texts related to mental health, specifically postpartum depression, ensuring a wide range of scenarios and emotions are covered.
|
|
|
|
|
|
|
| 56 |
|
| 57 |
+
### Training Procedure
|
| 58 |
|
| 59 |
+
#### Preprocessing
|
|
|
|
|
|
|
|
|
|
|
|
|
| 60 |
|
| 61 |
+
Text data was normalized and tokenized using standard NLP preprocessing techniques to ensure consistency and improve model understanding.
|
| 62 |
|
| 63 |
#### Training Hyperparameters
|
| 64 |
|
| 65 |
+
- Training regime details are provided in the repository, focusing on optimizing performance while maintaining the model's efficiency.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 66 |
|
| 67 |
## Evaluation
|
| 68 |
|
|
|
|
|
|
|
| 69 |
### Testing Data, Factors & Metrics
|
| 70 |
|
| 71 |
+
Evaluation was conducted using a separate test set, focusing on accuracy, empathy in responses, and relevance of video recommendations.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 72 |
|
| 73 |
## Environmental Impact
|
| 74 |
|
| 75 |
+
Efforts were made to minimize the carbon footprint during training, with details on compute usage and emissions available upon request.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 76 |
|
| 77 |
+
## Technical Specifications
|
| 78 |
|
| 79 |
+
Further details on the model's architecture, objective, and compute infrastructure are available in the Frontida repository.
|
| 80 |
|
| 81 |
+
## More Information
|
| 82 |
|
| 83 |
+
For additional details, including how to contribute to the model's development or integrate it into other applications, please visit the Frontida project page on Hugging Face.
|