File size: 2,430 Bytes
0378eca
 
 
 
c38485a
0378eca
c38485a
0378eca
 
c38485a
0378eca
 
 
c38485a
0378eca
 
4f08328
 
3f78172
4f08328
 
0378eca
 
 
 
 
 
 
 
 
 
 
 
 
4f08328
 
 
 
0378eca
 
 
 
4f08328
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0378eca
 
c38485a
 
0378eca
c38485a
 
4f08328
 
 
 
0378eca
 
 
4f08328
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
import gradio as gr
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

print("πŸš€ Loading DeepSeek-Coder-1.3B model...")
model_name = "deepseek-ai/deepseek-coder-1.3b-instruct"

tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    trust_remote_code=True,
    torch_dtype=torch.float32
)
print("βœ… Model loaded successfully!")

def generate_code(prompt, max_length=512, temperature=0.7):
    # Fixed: Using proper string concatenation instead of f-string
    instruction_part = "### Instruction:"
    response_part = ""
    formatted_prompt = instruction_part + prompt + response_part
    
    inputs = tokenizer(formatted_prompt, return_tensors="pt")
    
    with torch.no_grad():
        outputs = model.generate(
            **inputs,
            max_new_tokens=max_length,
            temperature=temperature,
            do_sample=True,
            top_p=0.95,
            pad_token_id=tokenizer.eos_token_id
        )
    
    response = tokenizer.decode(outputs[0], skip_special_tokens=True)
    # Extract only the generated response
    if "### Response:" in response:
        return response.split("### Response:")[-1].strip()
    return response

demo = gr.Interface(
    fn=generate_code,
    inputs=[
        gr.Textbox(
            label="Coding Prompt", 
            placeholder="Enter your coding question or request...",
            lines=5
        ),
        gr.Slider(
            minimum=128, 
            maximum=1024, 
            value=512, 
            step=64, 
            label="Max Length"
        ),
        gr.Slider(
            minimum=0.1, 
            maximum=2.0, 
            value=0.7, 
            step=0.1, 
            label="Temperature"
        )
    ],
    outputs=gr.Textbox(label="Generated Code", lines=15),
    title="πŸš€ DeepSeek Coder 1.3B - Docker Edition",
    description="AI coding assistant powered by DeepSeek-Coder-1.3B running in Docker",
    examples=[
        ["Write a Python function for binary search"],
        ["Create a Flask REST API for user login"],
        ["Write a JavaScript function to debounce API calls"],
        ["Create a SQL query to find top 10 users by activity"]
    ],
    theme=gr.themes.Soft()
)

if __name__ == "__main__":
    demo.launch(
        server_name="0.0.0.0", 
        server_port=7860,
        show_error=True
    )