Spaces:
Runtime error
Runtime error
improve interface
Browse files
app.py
CHANGED
|
@@ -97,24 +97,24 @@ with gr.Blocks() as demo:
|
|
| 97 |
value="You are a helpful AI assistant.",
|
| 98 |
lines=2,
|
| 99 |
placeholder="Set the AI's behavior and context..."
|
| 100 |
-
)
|
| 101 |
-
context = gr.Textbox(
|
| 102 |
-
label="Context",
|
| 103 |
-
lines=2,
|
| 104 |
-
placeholder="Enter additional context information..."
|
| 105 |
-
)
|
| 106 |
-
max_tokens = gr.Slider(minimum=1, maximum=1024, value=256, step=1, label="Max Tokens")
|
| 107 |
-
temperature = gr.Slider(minimum=0.1, maximum=2.0, value=0.7, step=0.1, label="Temperature")
|
| 108 |
-
top_p = gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p")
|
| 109 |
-
use_pipeline = gr.Checkbox(label="Use Pipeline", value=False)
|
| 110 |
-
use_tool = gr.Checkbox(label="Use Function Calling", value=False)
|
| 111 |
-
with gr.Column(visible=False) as tool_options:
|
| 112 |
-
tool_definition = gr.Code(
|
| 113 |
-
label="Tool Definition (JSON)",
|
| 114 |
-
value=customtool,
|
| 115 |
-
lines=15,
|
| 116 |
-
language="json"
|
| 117 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 118 |
|
| 119 |
def user(user_message, history):
|
| 120 |
return "", history + [[user_message, None]]
|
|
|
|
| 97 |
value="You are a helpful AI assistant.",
|
| 98 |
lines=2,
|
| 99 |
placeholder="Set the AI's behavior and context..."
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 100 |
)
|
| 101 |
+
context = gr.Textbox(
|
| 102 |
+
label="Context",
|
| 103 |
+
lines=2,
|
| 104 |
+
placeholder="Enter additional context information..."
|
| 105 |
+
)
|
| 106 |
+
max_tokens = gr.Slider(minimum=1, maximum=1024, value=256, step=1, label="Max Tokens")
|
| 107 |
+
temperature = gr.Slider(minimum=0.1, maximum=2.0, value=0.7, step=0.1, label="Temperature")
|
| 108 |
+
top_p = gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p")
|
| 109 |
+
use_pipeline = gr.Checkbox(label="Use Pipeline", value=False)
|
| 110 |
+
use_tool = gr.Checkbox(label="Use Function Calling", value=False)
|
| 111 |
+
with gr.Column(visible=False) as tool_options:
|
| 112 |
+
tool_definition = gr.Code(
|
| 113 |
+
label="Tool Definition (JSON)",
|
| 114 |
+
value=customtool,
|
| 115 |
+
lines=15,
|
| 116 |
+
language="json"
|
| 117 |
+
)
|
| 118 |
|
| 119 |
def user(user_message, history):
|
| 120 |
return "", history + [[user_message, None]]
|
globe.py
CHANGED
|
@@ -14,14 +14,14 @@ presentation1 = """Try this model on [build.nvidia.com](https://build.nvidia.com
|
|
| 14 |
|
| 15 |
**Model Dates:** π€Nemotron-Mini-4B-Instruct was trained between February 2024 and Aug 2024.
|
| 16 |
|
| 17 |
-
|
| 18 |
|
| 19 |
[NVIDIA Community Model License](https://huggingface.co/nvidia/Nemotron-Mini-4B-Instruct/blob/main/nvidia-community-model-license-aug2024.pdf)"""
|
| 20 |
|
| 21 |
presentation2 = """
|
| 22 |
-
|
| 23 |
|
| 24 |
-
Nemotron-Mini-4B-Instruct uses a model embedding size of 3072, 32 attention heads, and an MLP intermediate dimension of 9216. It also uses Grouped-Query Attention (GQA) and Rotary Position Embeddings (RoPE).
|
| 25 |
|
| 26 |
**Architecture Type:** Transformer Decoder (auto-regressive language model)
|
| 27 |
|
|
|
|
| 14 |
|
| 15 |
**Model Dates:** π€Nemotron-Mini-4B-Instruct was trained between February 2024 and Aug 2024.
|
| 16 |
|
| 17 |
+
### License
|
| 18 |
|
| 19 |
[NVIDIA Community Model License](https://huggingface.co/nvidia/Nemotron-Mini-4B-Instruct/blob/main/nvidia-community-model-license-aug2024.pdf)"""
|
| 20 |
|
| 21 |
presentation2 = """
|
| 22 |
+
### Model Architecture
|
| 23 |
|
| 24 |
+
π€Nemotron-Mini-4B-Instruct uses a model embedding size of 3072, 32 attention heads, and an MLP intermediate dimension of 9216. It also uses Grouped-Query Attention (GQA) and Rotary Position Embeddings (RoPE).
|
| 25 |
|
| 26 |
**Architecture Type:** Transformer Decoder (auto-regressive language model)
|
| 27 |
|