vikas83 commited on
Commit
3034141
·
verified ·
1 Parent(s): a6dd84f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +12 -9
app.py CHANGED
@@ -1,16 +1,21 @@
1
  import gradio as gr
2
- from transformers import AutoModel, AutoTokenizer
3
 
4
- model_name = "vikas83/my-bert-model" # Your uploaded model
5
 
6
- model = AutoModel.from_pretrained(model_name)
 
7
  tokenizer = AutoTokenizer.from_pretrained(model_name)
8
 
 
 
 
 
9
  def generate_response(text):
10
- inputs = tokenizer(text, return_tensors="pt")
11
- outputs = model(**inputs)
12
- return "Model processed successfully!" # Modify as needed
13
 
 
14
  iface = gr.Interface(
15
  fn=generate_response,
16
  inputs="text",
@@ -19,6 +24,4 @@ iface = gr.Interface(
19
  description="Enter text and see the response from the model!"
20
  )
21
 
22
- # Enable public access
23
- public_url = iface.launch(share=True)
24
- print(f"Access the app at {public_url}")
 
1
  import gradio as gr
2
+ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
3
 
4
+ model_name = "vikas83/my-bert-model" # Your model
5
 
6
+ # Load the model & tokenizer
7
+ model = AutoModelForCausalLM.from_pretrained(model_name)
8
  tokenizer = AutoTokenizer.from_pretrained(model_name)
9
 
10
+ # Create a text generation pipeline
11
+ nlp_pipeline = pipeline("text-generation", model=model, tokenizer=tokenizer)
12
+
13
+ # Function to generate response
14
  def generate_response(text):
15
+ result = nlp_pipeline(text, max_length=50, do_sample=True)
16
+ return result[0]['generated_text']
 
17
 
18
+ # Gradio UI
19
  iface = gr.Interface(
20
  fn=generate_response,
21
  inputs="text",
 
24
  description="Enter text and see the response from the model!"
25
  )
26
 
27
+ iface.launch(server_name="0.0.0.0", server_port=7860)