Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -43,16 +43,19 @@ def main():
|
|
| 43 |
word_count = st.sidebar.slider("Number of Words", min_value=50, max_value=1000, value=200, step=50)
|
| 44 |
|
| 45 |
if st.sidebar.button("Generate Blog"):
|
| 46 |
-
|
| 47 |
-
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
|
|
|
|
|
|
|
|
|
|
| 53 |
|
| 54 |
-
st.write(aa)
|
| 55 |
-
st.write("Will Come here")
|
| 56 |
|
| 57 |
if __name__ == "__main__":
|
| 58 |
main()
|
|
|
|
| 43 |
word_count = st.sidebar.slider("Number of Words", min_value=50, max_value=1000, value=200, step=50)
|
| 44 |
|
| 45 |
if st.sidebar.button("Generate Blog"):
|
| 46 |
+
repo_id = "google/gemma-1.1-7b-it"
|
| 47 |
+
|
| 48 |
+
llm = HuggingFaceEndpoint(
|
| 49 |
+
repo_id=repo_id, max_length=128, temperature=0.5, huggingfacehub_api_token=HF_TOKEN
|
| 50 |
+
)
|
| 51 |
+
# llm_chain = LLMChain(prompt=prompt, llm=llm)
|
| 52 |
+
# print(llm_chain.run(question))
|
| 53 |
+
aa = llm_chain.run({"topic": topic,"words":word_count,"role":role})
|
| 54 |
+
st.write(aa)
|
| 55 |
+
|
| 56 |
|
| 57 |
+
# st.write(aa)
|
| 58 |
+
# st.write("Will Come here")
|
| 59 |
|
| 60 |
if __name__ == "__main__":
|
| 61 |
main()
|