| --- |
| {} |
| --- |
| # Model Card for Nexa Temp Mapping |
|
|
| ## Model Description |
| This model, named Nexa Temp Mapping, is fine-tuned from the Mistral-7B-Instruct-v0.2 model for specialized tasks in creating test cases for Temperature Mapping of areas. It incorporates enhancements using PEFT (Pretrained Encoder Fine-Tuning) techniques to optimize performance for specific applications. |
|
|
| ## Training Data |
| Describe the dataset used for training the model: |
| - **Source:** [Specify the source of the training data] |
| - **Size:** 50 Datapoints |
| - **Details:** Brief description of the dataset characteristics. |
|
|
| ## Intended Use |
| This model is intended for use in the creation of test cases to qualify equipment such as fridges, freezers, autoclaves and ovens. It is designed to improve the code model by including domain knowledge over Supplement 8 Temperature mapping of storage areas Technical supplement to WHO Technical Report Series, No. 961, 2011. Annex 9: Model guidance for the stoage and transport of time- and temperature-sensitive pharmaceutcial products. |
|
|
| ## How to Use |
| ```python |
| from transformers import AutoTokenizer, AutoModelForCausalLM |
| |
| tokenizer = AutoTokenizer.from_pretrained("PeterGordon/nexa-temp-mapping") |
| model = AutoModelForCausalLM.from_pretrained("PeterGordon/nexa-temp-mapping") |
| |
| text = "Your input text here" |
| encoded_input = tokenizer(text, return_tensors='pt') |
| output = model.generate(**encoded_input) |
| print(tokenizer.decode(output[0], skip_special_tokens=True)) |
| |
| --- |
| license: apache-2.0 |
| --- |
| |