bhargavi909 commited on
Commit
7528475
·
verified ·
1 Parent(s): 1bb41f4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -7
README.md CHANGED
@@ -13,17 +13,18 @@ tags: []
13
 
14
  ### Model Description
15
 
16
- <!-- Provide a longer summary of what this model is. -->
 
17
 
18
  This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
19
 
20
  - **Developed by:** [Sree bhargavi balija]
21
  - **Funded by [optional]:** [self]
22
  - **Shared by [optional]:** [More Information Needed]
23
- - **Model type:** [More Information Needed]
24
- - **Language(s) (NLP):** [More Information Needed]
25
- - **License:** [More Information Needed]
26
- - **Finetuned from model [optional]:** [More Information Needed]
27
 
28
  ### Model Sources [optional]
29
 
@@ -69,7 +70,17 @@ Users (both direct and downstream) should be made aware of the risks, biases and
69
 
70
  ## How to Get Started with the Model
71
 
72
- Use the code below to get started with the model.
 
 
 
 
 
 
 
 
 
 
73
 
74
  [More Information Needed]
75
 
@@ -170,7 +181,12 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
170
 
171
  ## Citation [optional]
172
 
173
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
 
 
 
 
 
174
 
175
  **BibTeX:**
176
 
 
13
 
14
  ### Model Description
15
 
16
+ Initially, Pixprompt is the first open-source small LLM, Pixprompt combines a CLIP vision encoder and GPT-2 (125M) decoder, with optional LoRA adapters for efficient fine-tuning. It was originally trained to support image + prompt → text, and now fine-tuned on a curated set of financial data and news headlines fetched dynamically from the web.
17
+
18
 
19
  This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
20
 
21
  - **Developed by:** [Sree bhargavi balija]
22
  - **Funded by [optional]:** [self]
23
  - **Shared by [optional]:** [More Information Needed]
24
+ - **Model type:** [Multimodal Causal Language Model (CLIP + GPT2)]
25
+ - **Language(s) (NLP):** [English]
26
+ - **License:** [MIT]
27
+ - **Finetuned from model [optional]:** [bhargavi909/Pixprompt]
28
 
29
  ### Model Sources [optional]
30
 
 
70
 
71
  ## How to Get Started with the Model
72
 
73
+ from transformers import GPT2LMHeadModel, GPT2Tokenizer
74
+ from peft import PeftModel
75
+
76
+ base = GPT2LMHeadModel.from_pretrained("bhargavi909/Pixprompt")
77
+ model = PeftModel.from_pretrained(base, "./finetuned-financial-pixprompt")
78
+ tokenizer = GPT2Tokenizer.from_pretrained("bhargavi909/Pixprompt")
79
+
80
+ prompt = "The chart shows the impact of inflation"
81
+ inputs = tokenizer(prompt, return_tensors="pt")
82
+ outputs = model.generate(**inputs, max_length=100)
83
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
84
 
85
  [More Information Needed]
86
 
 
181
 
182
  ## Citation [optional]
183
 
184
+ @misc{pixprompt2024,
185
+ author = {Sree Bhargavi Balija},
186
+ title = {Pixprompt: A Multimodal GPT Model for Financial Text Generation},
187
+ year = {2024},
188
+ url = {https://huggingface.co/bhargavi909/Pixprompt},
189
+ }
190
 
191
  **BibTeX:**
192