AilexGPT commited on
Commit
cef8088
·
verified ·
1 Parent(s): 8db369e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +20 -14
app.py CHANGED
@@ -2,6 +2,10 @@ import streamlit as st
2
  import random
3
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
4
 
 
 
 
 
5
  tokenizer = AutoTokenizer.from_pretrained("Kaludi/chatgpt-gpt4-prompts-bart-large-cnn-samsum")
6
  model = AutoModelForSeq2SeqLM.from_pretrained("Kaludi/chatgpt-gpt4-prompts-bart-large-cnn-samsum", from_tf=True)
7
 
@@ -11,20 +15,22 @@ def generate(prompt, max_new_tokens):
11
  output = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
12
  return output[0]
13
 
 
 
14
  st.title("ChatGPT_Streamlit-Prompt Generator")
15
- st.write("This app generates ChatGPT/BingChat model trained by [Kaludi](https://huggingface.co/Kaludi/). Enter a role and a prompt will be generated based on it.")
16
- prompt = st.text_input("Enter a Role, Example: Virtual Assistant", placeholder="Text here", value="")
17
- max_new_tokens = st.slider("Select Max Tokens in Response", min_value=100, max_value=500, value=150, step=10)
18
- if st.button("Generate"):
 
19
  output = generate(prompt, max_new_tokens)
20
- st.write("Generated Prompt:", box=True)
21
- st.write("<div style='background-color: #2E2E2E; padding: 10px;'>{}</div>".format(output), unsafe_allow_html=True)
22
- st.write("")
23
- st.write("<div style='text-align: center; font-weight: bold;'>Examples:</div>",unsafe_allow_html=True, box=True)
24
- st.write("<style> .stBox span { background-color: #2E2E2E; } </style>", unsafe_allow_html=True)
25
- with open("examples.txt", "r") as f:
26
  examples = f.readlines()
27
- random_examples = random.sample(examples, 5)
28
- for example in random_examples:
29
- example = example.strip()
30
- st.write("<div style='background-color: #2E2E2E; padding: 10px; text-align: center;'>• {}</div>".format(example), unsafe_allow_html=True)
 
2
  import random
3
  from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
4
 
5
+ # URL des Logos
6
+ logo_url = "https://dein-logo-url-hier.com/logo.png" # Ersetze dies mit der tatsächlichen URL Deines Logos
7
+
8
+ # Modell und Tokenizer initialisieren
9
  tokenizer = AutoTokenizer.from_pretrained("Kaludi/chatgpt-gpt4-prompts-bart-large-cnn-samsum")
10
  model = AutoModelForSeq2SeqLM.from_pretrained("Kaludi/chatgpt-gpt4-prompts-bart-large-cnn-samsum", from_tf=True)
11
 
 
15
  output = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
16
  return output[0]
17
 
18
+ # Streamlit App Layout
19
+ st.markdown(f"<img src='{logo_url}' style='max-height: 100px;'>", unsafe_allow_html=True) # Logo einfügen
20
  st.title("ChatGPT_Streamlit-Prompt Generator")
21
+ st.write("Gib eine Rolle ein, und es wird ein Prompt basierend darauf generiert.")
22
+ prompt = st.text_input("Gib eine Rolle ein, Beispiel: Virtueller Assistent", placeholder="Text hier", value="")
23
+ max_new_tokens = st.slider("Wähle die maximale Anzahl an Tokens in der Antwort", min_value=100, max_value=500, value=150, step=10)
24
+
25
+ if st.button("Generieren"):
26
  output = generate(prompt, max_new_tokens)
27
+ st.write("Generierter Prompt:")
28
+ st.write(output)
29
+
30
+ # Beispiele anzeigen
31
+ st.write("Beispiele:")
32
+ with open("examples.txt", "r") as f: # Ersetze "examples.txt" mit dem tatsächlichen Pfad zu Deiner Beispieldatei
33
  examples = f.readlines()
34
+ random_examples = random.sample(examples, 5)
35
+ for example in random_examples:
36
+ st.write(f"• {example.strip()}")