| import { ColorMode } from '$lib/enums/ui'; |
| import { Monitor, Moon, Sun } from '@lucide/svelte'; |
|
|
| export const SETTING_CONFIG_DEFAULT: Record<string, string | number | boolean> = { |
| |
| |
| apiKey: '', |
| systemMessage: '', |
| showSystemMessage: true, |
| theme: ColorMode.SYSTEM, |
| showThoughtInProgress: false, |
| disableReasoningParsing: false, |
| showRawOutputSwitch: false, |
| keepStatsVisible: false, |
| showMessageStats: true, |
| askForTitleConfirmation: false, |
| pasteLongTextToFileLen: 2500, |
| copyTextAttachmentsAsPlainText: false, |
| pdfAsImage: false, |
| disableAutoScroll: false, |
| renderUserContentAsMarkdown: false, |
| alwaysShowSidebarOnDesktop: false, |
| autoShowSidebarOnNewChat: true, |
| autoMicOnEmpty: false, |
| fullHeightCodeBlocks: false, |
| showRawModelNames: false, |
| mcpServers: '[]', |
| mcpServerUsageStats: '{}', |
| agenticMaxTurns: 10, |
| agenticMaxToolPreviewLines: 25, |
| showToolCallInProgress: false, |
| alwaysShowAgenticTurns: false, |
| |
| samplers: 'top_k;typ_p;top_p;min_p;temperature', |
| backend_sampling: false, |
| temperature: 0.8, |
| dynatemp_range: 0.0, |
| dynatemp_exponent: 1.0, |
| top_k: 40, |
| top_p: 0.95, |
| min_p: 0.05, |
| xtc_probability: 0.0, |
| xtc_threshold: 0.1, |
| typ_p: 1.0, |
| repeat_last_n: 64, |
| repeat_penalty: 1.0, |
| presence_penalty: 0.0, |
| frequency_penalty: 0.0, |
| dry_multiplier: 0.0, |
| dry_base: 1.75, |
| dry_allowed_length: 2, |
| dry_penalty_last_n: -1, |
| max_tokens: -1, |
| custom: '', |
| |
| pyInterpreterEnabled: false, |
| enableContinueGeneration: false |
| }; |
|
|
| export const SETTING_CONFIG_INFO: Record<string, string> = { |
| apiKey: 'Set the API Key if you are using <code>--api-key</code> option for the server.', |
| systemMessage: 'The starting message that defines how model should behave.', |
| showSystemMessage: 'Display the system message at the top of each conversation.', |
| theme: |
| 'Choose the color theme for the interface. You can choose between System (follows your device settings), Light, or Dark.', |
| pasteLongTextToFileLen: |
| 'On pasting long text, it will be converted to a file. You can control the file length by setting the value of this parameter. Value 0 means disable.', |
| copyTextAttachmentsAsPlainText: |
| 'When copying a message with text attachments, combine them into a single plain text string instead of a special format that can be pasted back as attachments.', |
| samplers: |
| 'The order at which samplers are applied, in simplified way. Default is "top_k;typ_p;top_p;min_p;temperature": top_k->typ_p->top_p->min_p->temperature', |
| backend_sampling: |
| 'Enable backend-based samplers. When enabled, supported samplers run on the accelerator backend for faster sampling.', |
| temperature: |
| 'Controls the randomness of the generated text by affecting the probability distribution of the output tokens. Higher = more random, lower = more focused.', |
| dynatemp_range: |
| 'Addon for the temperature sampler. The added value to the range of dynamic temperature, which adjusts probabilities by entropy of tokens.', |
| dynatemp_exponent: |
| 'Addon for the temperature sampler. Smoothes out the probability redistribution based on the most probable token.', |
| top_k: 'Keeps only k top tokens.', |
| top_p: 'Limits tokens to those that together have a cumulative probability of at least p', |
| min_p: |
| 'Limits tokens based on the minimum probability for a token to be considered, relative to the probability of the most likely token.', |
| xtc_probability: |
| 'XTC sampler cuts out top tokens; this parameter controls the chance of cutting tokens at all. 0 disables XTC.', |
| xtc_threshold: |
| 'XTC sampler cuts out top tokens; this parameter controls the token probability that is required to cut that token.', |
| typ_p: 'Sorts and limits tokens based on the difference between log-probability and entropy.', |
| repeat_last_n: 'Last n tokens to consider for penalizing repetition', |
| repeat_penalty: 'Controls the repetition of token sequences in the generated text', |
| presence_penalty: 'Limits tokens based on whether they appear in the output or not.', |
| frequency_penalty: 'Limits tokens based on how often they appear in the output.', |
| dry_multiplier: |
| 'DRY sampling reduces repetition in generated text even across long contexts. This parameter sets the DRY sampling multiplier.', |
| dry_base: |
| 'DRY sampling reduces repetition in generated text even across long contexts. This parameter sets the DRY sampling base value.', |
| dry_allowed_length: |
| 'DRY sampling reduces repetition in generated text even across long contexts. This parameter sets the allowed length for DRY sampling.', |
| dry_penalty_last_n: |
| 'DRY sampling reduces repetition in generated text even across long contexts. This parameter sets DRY penalty for the last n tokens.', |
| max_tokens: 'The maximum number of token per output. Use -1 for infinite (no limit).', |
| custom: 'Custom JSON parameters to send to the API. Must be valid JSON format.', |
| showThoughtInProgress: 'Expand thought process by default when generating messages.', |
| disableReasoningParsing: |
| 'Send reasoning_format=none to prevent server-side extraction of reasoning tokens into separate field', |
| showRawOutputSwitch: |
| 'Show toggle button to display messages as plain text instead of Markdown-formatted content', |
| keepStatsVisible: 'Keep processing statistics visible after generation finishes.', |
| showMessageStats: |
| 'Display generation statistics (tokens/second, token count, duration) below each assistant message.', |
| askForTitleConfirmation: |
| 'Ask for confirmation before automatically changing conversation title when editing the first message.', |
| pdfAsImage: |
| 'Parse PDF as image instead of text. Automatically falls back to text processing for non-vision models.', |
| disableAutoScroll: |
| 'Disable automatic scrolling while messages stream so you can control the viewport position manually.', |
| renderUserContentAsMarkdown: 'Render user messages using markdown formatting in the chat.', |
| alwaysShowSidebarOnDesktop: |
| 'Always keep the sidebar visible on desktop instead of auto-hiding it.', |
| autoShowSidebarOnNewChat: |
| 'Automatically show sidebar when starting a new chat. Disable to keep the sidebar hidden until you click on it.', |
| autoMicOnEmpty: |
| 'Automatically show microphone button instead of send button when textarea is empty for models with audio modality support.', |
| fullHeightCodeBlocks: |
| 'Always display code blocks at their full natural height, overriding any height limits.', |
| showRawModelNames: |
| 'Display full raw model identifiers (e.g. "unsloth/Qwen3.5-27B-GGUF:BF16") instead of parsed names with badges.', |
| mcpServers: |
| 'Configure MCP servers as a JSON list. Use the form in the MCP Client settings section to edit.', |
| mcpServerUsageStats: |
| 'Usage statistics for MCP servers. Tracks how many times tools from each server have been used.', |
| agenticMaxTurns: |
| 'Maximum number of tool execution cycles before stopping (prevents infinite loops).', |
| agenticMaxToolPreviewLines: |
| 'Number of lines shown in tool output previews (last N lines). Only these previews and the final LLM response persist after the agentic loop completes.', |
| showToolCallInProgress: |
| 'Automatically expand tool call details while executing and keep them expanded after completion.', |
| pyInterpreterEnabled: |
| 'Enable Python interpreter using Pyodide. Allows running Python code in markdown code blocks.', |
| enableContinueGeneration: |
| 'Enable "Continue" button for assistant messages. Currently works only with non-reasoning models.' |
| }; |
|
|
| export const SETTINGS_COLOR_MODES_CONFIG = [ |
| { value: ColorMode.SYSTEM, label: 'System', icon: Monitor }, |
| { value: ColorMode.LIGHT, label: 'Light', icon: Sun }, |
| { value: ColorMode.DARK, label: 'Dark', icon: Moon } |
| ]; |
|
|