anktechsol commited on
Commit
d951a54
·
verified ·
1 Parent(s): 9b3322a

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +62 -0
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ * * *
2
+ # For reference on model card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1
3
+ # Doc / guide: https://huggingface.co/docs/hub/model-cards
4
+ ## Model Card for AnkiGPT-small
5
+
6
+ # Model Card for AnkiGPT-small
7
+
8
+ This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
9
+
10
+ ## Model Details
11
+
12
+ ### Model Description
13
+
14
+ - **Developed by:** [anktechsol.com](www.anktechsol.com)
15
+ - **Shared by:** [More Information Needed - anktechsol]
16
+ - **Model type:** Causal Language Model
17
+ - **Language(s) (NLP):** English, potentially aspects of Indian languages/Hinglish due to fine-tuning data.
18
+ - **License:** (Specify the license of the fine-tuned model, often inherited from the base model or dataset. DialoGPT uses MIT license, check the dataset license.)
19
+ - **Finetuned from model [optional]:** `microsoft/DialoGPT-small`
20
+
21
+ ### Model Sources [optional]
22
+
23
+ - **Repository:** `https://huggingface.co/anktechsol/ankiGPT-small` (This will be the link after pushing to the hub)
24
+
25
+ ## Uses
26
+
27
+ ### Direct Use
28
+
29
+ This model can be used for text generation, particularly in conversational or narrative contexts, with a potential bias towards topics and linguistic styles present in the fine-tuning dataset (Indian conversational data).
30
+
31
+ ### Downstream Use
32
+
33
+ This model could potentially be used as a base for further fine-tuning on more specific Indian language tasks or domains.
34
+
35
+ ### Out-of-Scope Use
36
+
37
+ This model is not suitable for generating factual information, performing critical tasks requiring high accuracy, or deployment in sensitive applications without extensive further evaluation and mitigation of potential biases.
38
+
39
+ ## Bias, Risks, and Limitations
40
+
41
+ Based on initial testing, the model may exhibit repetitive text generation, especially for longer sequences. The model's knowledge and linguistic style are heavily influenced by the fine-tuning dataset, which may not cover all aspects of Indian languages or culture comprehensively. Biases present in the training data may be reflected in the model's output.
42
+
43
+ ### Recommendations
44
+
45
+ Users should be aware of the model's limitations in generating coherent long text and potential biases. It is recommended to experiment with different generation parameters (`max_length`, `no_repeat_ngram_size`, sampling strategies) to improve output quality. For any critical applications, thorough testing and human review of generated content are essential.
46
+
47
+ ## How to Get Started with the Model
48
+
49
+ Use the code below to get started with the model using the `transformers` library.
50
+
51
+ from transformers import pipeline
52
+
53
+ # Replace "anktechsol/ankiGPT-small" with your actual model ID on the Hugging Face Hub
54
+ generator = pipeline("text-generation", model="anktechsol/ankiGPT-small")
55
+
56
+ # A detailed prompt related to India
57
+ prompt = "Write a short story about a day in the life of a student in a bustling Indian city, describing their commute, interactions at school, and a cultural event they attend in the evening. Keep it in hinglish"
58
+
59
+ # Generate text with a reasonable max_length to allow for a detailed story
60
+ generated_text = generator(prompt, max_length=300, num_return_sequences=1)
61
+
62
+ print(generated_text[0]['generated_text'])