dioarafl commited on
Commit
9df53e1
·
verified ·
1 Parent(s): ec4490e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +27 -1
app.py CHANGED
@@ -1,3 +1,29 @@
1
  import gradio as gr
 
2
 
3
- gr.load("models/meta-llama/Meta-Llama-3-8B").launch()
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  import gradio as gr
2
+ from transformers import AutoModelForCausalLM, AutoTokenizer
3
 
4
+ # Load model and tokenizer
5
+ model_name = "models/meta-llama/Meta-Llama-3-8B"
6
+ model = AutoModelForCausalLM.from_pretrained(model_name)
7
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
8
+
9
+ # Define function to translate code
10
+ def translate_code(input_code, prompt=""):
11
+ # Combine input code and prompt
12
+ input_text = f"{prompt}\n\n{input_code}"
13
+ # Tokenize input text
14
+ input_ids = tokenizer.encode(input_text, return_tensors="pt", max_length=1024, truncation=True)
15
+ # Generate output sequence
16
+ output = model.generate(input_ids, max_length=1024, num_return_sequences=1, temperature=0.7)
17
+ # Decode output sequence
18
+ translated_code = tokenizer.decode(output[0], skip_special_tokens=True)
19
+ return translated_code
20
+
21
+ # Launch Gradio interface
22
+ gr.Interface(
23
+ fn=translate_code,
24
+ inputs=["text", "text"],
25
+ outputs="text",
26
+ title="AI Code Translator",
27
+ description="Translate your code using Meta-Llama-3-8B model.",
28
+ theme="compact"
29
+ ).launch()