Spaces:
Running
Running
Update src/backend/gradio_workflowbuilder/templates/component/index.js
Browse files
src/backend/gradio_workflowbuilder/templates/component/index.js
CHANGED
|
@@ -2612,6 +2612,51 @@ function Dl(t, e, l) {
|
|
| 2612 |
"AI & Language": {
|
| 2613 |
icon: "🤖",
|
| 2614 |
components: {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2615 |
OpenAIModel: {
|
| 2616 |
label: "OpenAI Model",
|
| 2617 |
icon: "🎯",
|
|
@@ -3480,7 +3525,51 @@ function Dl(t, e, l) {
|
|
| 3480 |
help: "Name shown in the workflow"
|
| 3481 |
}
|
| 3482 |
],
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3483 |
// AI & Language nodes
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3484 |
OpenAIModel: [
|
| 3485 |
{
|
| 3486 |
key: "display_name",
|
|
|
|
| 2612 |
"AI & Language": {
|
| 2613 |
icon: "🤖",
|
| 2614 |
components: {
|
| 2615 |
+
|
| 2616 |
+
|
| 2617 |
+
llmNode: {
|
| 2618 |
+
label: "Gemma-3-r1984-27B",
|
| 2619 |
+
icon: "🎯",
|
| 2620 |
+
color: "#4F46E5",
|
| 2621 |
+
defaultData: {
|
| 2622 |
+
display_name: "Gemma-3-r1984-27B",
|
| 2623 |
+
template: {
|
| 2624 |
+
prompt: {
|
| 2625 |
+
display_name: "Prompt",
|
| 2626 |
+
type: "string",
|
| 2627 |
+
value: "",
|
| 2628 |
+
is_handle: true
|
| 2629 |
+
},
|
| 2630 |
+
model: {
|
| 2631 |
+
display_name: "Model",
|
| 2632 |
+
type: "options",
|
| 2633 |
+
options: ["Gemma-3-r1984-27B"],
|
| 2634 |
+
value: "Gemma-3-r1984-27B"
|
| 2635 |
+
},
|
| 2636 |
+
temperature: {
|
| 2637 |
+
display_name: "Temperature",
|
| 2638 |
+
type: "number",
|
| 2639 |
+
value: 0.7,
|
| 2640 |
+
min: 0,
|
| 2641 |
+
max: 1,
|
| 2642 |
+
step: 0.1
|
| 2643 |
+
},
|
| 2644 |
+
response: {
|
| 2645 |
+
display_name: "Response",
|
| 2646 |
+
type: "string",
|
| 2647 |
+
is_handle: true
|
| 2648 |
+
}
|
| 2649 |
+
},
|
| 2650 |
+
resources: { cpu: 0.4, memory: "512Mi", gpu: "none" }
|
| 2651 |
+
}
|
| 2652 |
+
},
|
| 2653 |
+
|
| 2654 |
+
|
| 2655 |
+
|
| 2656 |
+
|
| 2657 |
+
|
| 2658 |
+
|
| 2659 |
+
|
| 2660 |
OpenAIModel: {
|
| 2661 |
label: "OpenAI Model",
|
| 2662 |
icon: "🎯",
|
|
|
|
| 3525 |
help: "Name shown in the workflow"
|
| 3526 |
}
|
| 3527 |
],
|
| 3528 |
+
|
| 3529 |
+
|
| 3530 |
+
|
| 3531 |
+
|
| 3532 |
+
|
| 3533 |
+
|
| 3534 |
// AI & Language nodes
|
| 3535 |
+
|
| 3536 |
+
llmNode: [
|
| 3537 |
+
{
|
| 3538 |
+
key: "display_name",
|
| 3539 |
+
label: "Display Name",
|
| 3540 |
+
type: "text",
|
| 3541 |
+
help: "Name shown in the workflow"
|
| 3542 |
+
},
|
| 3543 |
+
{
|
| 3544 |
+
key: "template.prompt.value",
|
| 3545 |
+
label: "Prompt",
|
| 3546 |
+
type: "textarea",
|
| 3547 |
+
help: "Text prompt sent to Gemma-3-27B"
|
| 3548 |
+
},
|
| 3549 |
+
{
|
| 3550 |
+
key: "template.temperature.value",
|
| 3551 |
+
label: "Temperature",
|
| 3552 |
+
type: "number",
|
| 3553 |
+
min: 0,
|
| 3554 |
+
max: 1,
|
| 3555 |
+
step: 0.1,
|
| 3556 |
+
help: "Sampling temperature"
|
| 3557 |
+
},
|
| 3558 |
+
{
|
| 3559 |
+
key: "template.model.value",
|
| 3560 |
+
label: "Model Variant",
|
| 3561 |
+
type: "select",
|
| 3562 |
+
options: ["Gemma-3-27B"],
|
| 3563 |
+
help: "Only one variant is pre-configured"
|
| 3564 |
+
}
|
| 3565 |
+
],
|
| 3566 |
+
|
| 3567 |
+
|
| 3568 |
+
|
| 3569 |
+
|
| 3570 |
+
|
| 3571 |
+
|
| 3572 |
+
|
| 3573 |
OpenAIModel: [
|
| 3574 |
{
|
| 3575 |
key: "display_name",
|