license: cc-by-nc-4.0
language:
- en
metrics:
- bleu
- accuracy
- recall
- precision
- f1
tags:
- code
Model Card for Model ID
This modelcard aims to be a base template for new models. It has been generated using this raw template.
Model Details
Model Description
The Digital Navigator model is designed to assist users by generating natural language responses to input queries. It is a fine-tuned GPT-2 model, customized for providing assistanse to visitors of my university's website: https://cdm.uowm.gr/en/index/.
- Developed by: Papagiannakis Panagiotis
- Model type: GPT-2 (Generative Pre-trained Transformer 2)
- Language(s) (NLP): English
- License: CC BY-NC 4.0
- Finetuned from model [optional]: GPT-2
Direct Use
The Digital Navigator model can be directly used for generating conversational responses in English. It is intended for use in chatbots, virtual assistants, and other applications requiring natural language understanding and generation.
- Conversational AI
- Customer Support
- Virtual Assistance
Out-of-Scope Use
- Generating harmful, biased, or misleading content
- Use in high-stakes decision-making without human oversight
- Applications requiring high accuracy and context understanding
Bias, Risks, and Limitations
The model may generate biased or inappropriate content based on the training data. Users should be cautious of the following:
- Inherent biases in the training data
- Generation of factually incorrect information
- Limited understanding of context and nuance
Recommendations
Users (both direct and downstream) should be made aware of the risks, biases, and limitations of the model. It is recommended to monitor and evaluate the model's output in real-world applications to ensure it meets the desired criteria and ethical standards.
Training Details
Training Data
The model was fine-tuned on a dataset that includes conversational data. These data were collected from my university's website https://cdm.uowm.gr/en/index/.
Training Procedure
The model was trained using standard procedures for fine-tuning GPT-2 models.
Training Hyperparameters
- Training regime: [More Information Needed]
- Learning Rate: 5e-5
- Batch Size: 16
- Epochs: 30
- Block Size: 128
Evaluation
[More Information Needed]
Factors
[More Information Needed]
Metrics
The model was evaluated using metrics like BLEU score for language quality, accuracy for factual correctness,
Results
Summary
Due to the high complexity and small variety of the data:
- Training loss: 0.3999
- Validation loss: 1.5456
- Bleu: 0.39863
- Precision: 0.011
- Recall: 0.012
- Accuracy: 0.012
- F1: 0.011
Model Examination
- Hardware Type: [More Information Needed]
- Hours used: 18 minutes and 47 seconds
- Cloud Provider: Google Collab
Model Architecture and Objective
The model architecture is a transformer gpt. Its is an LLM model with a purpose of providing assistanse to visitors of my university's website.
Citation [optional]
Coming soon...
BibTeX:
Coming soon...
APA:
Coming soon...