id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
7,131
import re from typing import Callable, List, Optional, Union from sglang.backend.anthropic import Anthropic from sglang.backend.base_backend import BaseBackend from sglang.backend.openai import OpenAI from sglang.backend.runtime_endpoint import RuntimeEndpoint from sglang.backend.vertexai import VertexAI from sglang.global_config import global_config from sglang.lang.ir import ( SglExpr, SglExprList, SglFunction, SglGen, SglImage, SglRoleBegin, SglRoleEnd, SglSelect, ) class SglRoleEnd(SglExpr): def __init__(self, role): def __repr__(self): def user_end(): return SglRoleEnd("user")
null
7,132
import re from typing import Callable, List, Optional, Union from sglang.backend.anthropic import Anthropic from sglang.backend.base_backend import BaseBackend from sglang.backend.openai import OpenAI from sglang.backend.runtime_endpoint import RuntimeEndpoint from sglang.backend.vertexai import VertexAI from sglang.global_config import global_config from sglang.lang.ir import ( SglExpr, SglExprList, SglFunction, SglGen, SglImage, SglRoleBegin, SglRoleEnd, SglSelect, ) class SglRoleBegin(SglExpr): def __init__(self, role): super().__init__() self.role = role def __repr__(self): return f"RoleBegin({self.role})" def assistant_begin(): return SglRoleBegin("assistant")
null
7,133
import re from typing import Callable, List, Optional, Union from sglang.backend.anthropic import Anthropic from sglang.backend.base_backend import BaseBackend from sglang.backend.openai import OpenAI from sglang.backend.runtime_endpoint import RuntimeEndpoint from sglang.backend.vertexai import VertexAI from sglang.global_config import global_config from sglang.lang.ir import ( SglExpr, SglExprList, SglFunction, SglGen, SglImage, SglRoleBegin, SglRoleEnd, SglSelect, ) class SglRoleEnd(SglExpr): def __init__(self, role): super().__init__() self.role = role def __repr__(self): return f"RoleEnd({self.role})" def assistant_end(): return SglRoleEnd("assistant")
null
7,134
import base64 import json import threading import urllib.request from io import BytesIO from json import dumps import requests The provided code snippet includes necessary dependencies for implementing the `get_available_gpu_memory` function. Write a Python function `def get_available_gpu_memory(gpu_id, distributed=True)` to solve the following problem: Get available memory for cuda:gpu_id device. When distributed is True, the available memory is the minimum available memory of all GPUs. Here is the function: def get_available_gpu_memory(gpu_id, distributed=True): """ Get available memory for cuda:gpu_id device. When distributed is True, the available memory is the minimum available memory of all GPUs. """ import torch num_gpus = torch.cuda.device_count() assert gpu_id < num_gpus if torch.cuda.current_device() != gpu_id: print( f"WARNING: current device is not {gpu_id}, but {torch.cuda.current_device()}, ", "which may cause useless memory allocation for torch CUDA context.", ) free_gpu_memory, _ = torch.cuda.mem_get_info(gpu_id) if distributed: tensor = torch.tensor(free_gpu_memory, dtype=torch.float32).to( torch.device("cuda", gpu_id) ) torch.distributed.all_reduce(tensor, op=torch.distributed.ReduceOp.MIN) free_gpu_memory = tensor.item() return free_gpu_memory / (1 << 30)
Get available memory for cuda:gpu_id device. When distributed is True, the available memory is the minimum available memory of all GPUs.
7,135
import base64 import json import threading import urllib.request from io import BytesIO from json import dumps import requests The provided code snippet includes necessary dependencies for implementing the `is_same_type` function. Write a Python function `def is_same_type(values)` to solve the following problem: Return whether the elements in values are of the same type. Here is the function: def is_same_type(values): """Return whether the elements in values are of the same type.""" if len(values) <= 1: return True else: t = type(values[0]) return all(isinstance(v, t) for v in values[1:])
Return whether the elements in values are of the same type.
7,136
import base64 import json import threading import urllib.request from io import BytesIO from json import dumps import requests class HttpResponse: def __init__(self, resp): self.resp = resp def json(self): return json.loads(self.resp.read()) def status_code(self): return self.resp.status The provided code snippet includes necessary dependencies for implementing the `http_request` function. Write a Python function `def http_request( url, json=None, stream=False, auth_token=None, api_key=None, verify=None )` to solve the following problem: A faster version of requests.post with low-level urllib API. Here is the function: def http_request( url, json=None, stream=False, auth_token=None, api_key=None, verify=None ): """A faster version of requests.post with low-level urllib API.""" headers = {"Content-Type": "application/json; charset=utf-8"} # add the Authorization header if an auth token is provided if auth_token is not None: headers["Authorization"] = f"Bearer {auth_token}" # add the API Key header if an API key is provided if api_key is not None: headers["X-API-Key"] = api_key if stream: return requests.post(url, json=json, stream=True, headers=headers) else: req = urllib.request.Request(url, headers=headers) if json is None: data = None else: data = bytes(dumps(json), encoding="utf-8") resp = urllib.request.urlopen(req, data=data, cafile=verify) return HttpResponse(resp)
A faster version of requests.post with low-level urllib API.
7,137
import base64 import json import threading import urllib.request from io import BytesIO from json import dumps import requests The provided code snippet includes necessary dependencies for implementing the `encode_image_base64` function. Write a Python function `def encode_image_base64(image_path)` to solve the following problem: Encode an image in base64. Here is the function: def encode_image_base64(image_path): """Encode an image in base64.""" if isinstance(image_path, str): with open(image_path, "rb") as image_file: data = image_file.read() return base64.b64encode(data).decode("utf-8") elif isinstance(image_path, bytes): return base64.b64encode(image_path).decode("utf-8") else: # image_path is PIL.WebPImagePlugin.WebPImageFile image = image_path buffered = BytesIO() image.save(buffered, format="PNG") return base64.b64encode(buffered.getvalue()).decode("utf-8")
Encode an image in base64.
7,138
import base64 import json import threading import urllib.request from io import BytesIO from json import dumps import requests def _is_chinese_char(cp): """Checks whether CP is the codepoint of a CJK character.""" # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0x4E00 and cp <= 0x9FFF) or (cp >= 0x3400 and cp <= 0x4DBF) # or (cp >= 0x20000 and cp <= 0x2A6DF) # or (cp >= 0x2A700 and cp <= 0x2B73F) # or (cp >= 0x2B740 and cp <= 0x2B81F) # or (cp >= 0x2B820 and cp <= 0x2CEAF) # or (cp >= 0xF900 and cp <= 0xFAFF) or (cp >= 0x2F800 and cp <= 0x2FA1F) # ): # return True return False The provided code snippet includes necessary dependencies for implementing the `find_printable_text` function. Write a Python function `def find_printable_text(text)` to solve the following problem: Returns the longest printable substring of text that contains only entire words. Here is the function: def find_printable_text(text): """Returns the longest printable substring of text that contains only entire words.""" # Borrowed from https://github.com/huggingface/transformers/blob/061580c82c2db1de9139528243e105953793f7a2/src/transformers/generation/streamers.py#L99 # After the symbol for a new line, we flush the cache. if text.endswith("\n"): return text # If the last token is a CJK character, we print the characters. elif len(text) > 0 and _is_chinese_char(ord(text[-1])): return text # Otherwise if the penultimate token is a CJK character, we print the characters except for the last one. elif len(text) > 1 and _is_chinese_char(ord(text[-2])): return text[:-1] # Otherwise, prints until the last space char (simple heuristic to avoid printing incomplete words, # which may change with the subsequent token -- there are probably smarter ways to do this!) else: return text[: text.rfind(" ") + 1]
Returns the longest printable substring of text that contains only entire words.
7,139
import base64 import json import threading import urllib.request from io import BytesIO from json import dumps import requests The provided code snippet includes necessary dependencies for implementing the `run_with_timeout` function. Write a Python function `def run_with_timeout(func, args=(), kwargs=None, timeout=None)` to solve the following problem: Run a function with timeout. Here is the function: def run_with_timeout(func, args=(), kwargs=None, timeout=None): """Run a function with timeout.""" ret_value = [] def _target_func(): ret_value.append(func(*args, **(kwargs or {}))) t = threading.Thread(target=_target_func) t.start() t.join(timeout=timeout) if t.is_alive(): raise TimeoutError() if not ret_value: raise RuntimeError() return ret_value[0]
Run a function with timeout.
7,140
import uuid from typing import Any, Callable, Dict, List, Optional, Union from sglang.backend.base_backend import BaseBackend from sglang.global_config import global_config from sglang.lang.interpreter import ProgramState, ProgramStateGroup from sglang.lang.ir import ( SglArgument, SglCommitLazy, SglConcateAndAppend, SglConstantText, SglExpr, SglExprList, SglFork, SglFunction, SglGen, SglGetForkItem, SglRoleBegin, SglRoleEnd, SglSelect, SglVariable, SglVarScopeBegin, SglVarScopeEnd, ) class TracerProgramState(ProgramState): def __init__(self, backend, arguments, only_trace_prefix): self.pid = uuid.uuid4().hex self.backend = backend self.arguments: Dict[str, Any] = arguments self.only_trace_prefix = only_trace_prefix if hasattr(backend, "endpoint"): self.backend = backend.endpoint self.nodes = [] self.last_node = None self.variables = {} self.ret_value = None # For completion # For chat self.messages_ = [] self.cur_role = None self.chat_template = self.backend.get_chat_template() # For multi states self.child_states = [] cur_scope = TracingScope.get_current_scope() if cur_scope is not None: cur_scope.add_child_state(self) ################################## ########### Public API ########### ################################## def fork(self, number: int, position_ids_offset: Optional[List[int]] = None): if self.only_trace_prefix: raise StopTracing() fork_node = SglFork(number) fork_node.prev_node = self.last_node states = [ TracerProgramState(self.backend, self.arguments, self.only_trace_prefix) for _ in range(number) ] for i in range(number): node = SglGetForkItem(i) node.prev_node = fork_node states[i].last_node = node states[i].variables = dict(self.variables) states[i].messages_ = list(self.messages_) states[i].cur_role = self.cur_role states[i].chat_template = self.chat_template state_group = ProgramStateGroup(states, self) return state_group ################################## ########## Internal API ########## ################################## def _append_node(self, other: SglExpr): self.nodes.append(other) other.prev_node = self.last_node self.last_node = other def _execute(self, other: SglExpr): if isinstance(other, str): other = SglConstantText(other) other.pid = self.pid if isinstance(other, SglConstantText): self._execute_fill(other) elif isinstance(other, SglGen): self._execute_gen(other) elif isinstance(other, SglSelect): self._execute_select(other) elif isinstance(other, SglExprList): for x in other.expr_list: self._execute(x) elif isinstance(other, SglRoleBegin): self._execute_role_begin(other) elif isinstance(other, SglRoleEnd): self._execute_role_end(other) elif isinstance(other, SglVarScopeBegin): self._execute_var_scope_begin(other) elif isinstance(other, SglVarScopeEnd): self._execute_var_scope_end(other) else: if self.only_trace_prefix: raise StopTracing() else: self._append_node(other) return self def __iadd__(self, other): self._execute(other) return self def _execute_fill(self, expr: SglConstantText): if isinstance(expr, str): expr = SglConstantText(expr) self._append_node(expr) def _execute_gen(self, expr: SglGen): name = expr.name if expr.name is not None else "gen_" + str(len(self.variables)) new_node = SglVariable(name, source=expr) self.variables[name] = new_node self._append_node(expr) def _execute_select(self, expr: SglSelect): name = ( expr.name if expr.name is not None else "select_" + str(len(self.variables)) ) new_node = SglVariable(name, source=expr) self.variables[name] = new_node self._append_node(expr) def _execute_role_begin(self, expr: SglRoleBegin): assert self.cur_role is None, "Nested roles are not allowed." if len(self.messages_) == 0 and expr.role != "system": # Insert default system message default_system = self.chat_template.default_system_prompt if default_system: self._execute_role_begin(SglRoleBegin("system")) self._execute_fill(default_system) self._execute_role_end(SglRoleEnd("system")) self.cur_role = expr.role prefix, suffix = self.chat_template.get_prefix_and_suffix( expr.role, self.messages_ ) self._execute_fill(prefix) def _execute_role_end(self, expr: SglRoleEnd): prefix, suffix = self.chat_template.get_prefix_and_suffix( expr.role, self.messages_ ) self._execute_fill(suffix) self.messages_.append({"role": expr.role, "content": ""}) self.cur_role = None def _execute_var_scope_end(self, expr: SglVarScopeEnd): new_node = SglVariable(name, source=self.last_node) self.variables[name] = new_node def get_var(self, name): ret = self.arguments.get(name, None) if ret is not None: return ret v = self.variables[name] return SglVariable(v.name, v.source) def flatten_nodes(self): def traverse(cur): if isinstance(cur, SglExprList): for child in cur.expr_list: traverse(child) else: ret.append(cur) ret = [] for x in self.nodes: traverse(x) return ret def __del__(self): pass class TracingScope: cur_scope = None def __init__(self, tracer_state: TracerProgramState): self.tracer_state = tracer_state self.last_scope = TracingScope.cur_scope def __enter__(self): TracingScope.cur_scope = self return self def __exit__(self, exc_type, exc_value, traceback): TracingScope.cur_scope = self.last_scope def get_current_scope(): return TracingScope.cur_scope def add_child_state(self, state: TracerProgramState): cur_scope = self while cur_scope != None: cur_scope.tracer_state.child_states.append(state) cur_scope = cur_scope.last_scope class BaseBackend: def __init__(self) -> None: self.support_concate_and_append = False self.chat_template = get_chat_template("default") def get_model_name(self): raise NotImplementedError() def get_chat_template(self): return self.chat_template def cache_prefix(self, prefix_str: str): pass def uncache_prefix(self, rid: str): pass def end_request(self, rid: Union[str, List[str]]): pass def begin_program(self, s: StreamExecutor): pass def end_program(self, s: Union[StreamExecutor, List[StreamExecutor]]): pass def commit_lazy_operations(self, s: StreamExecutor): pass def fork_program( self, src: StreamExecutor, dst: List[StreamExecutor], position_ids_offset: Optional[List[int]] = None, ): pass def fill_image(self, s: StreamExecutor): pass def generate( self, s: StreamExecutor, sampling_params: SglSamplingParams, ): raise NotImplementedError() def generate_stream( self, s: StreamExecutor, sampling_params: SglSamplingParams, ): raise NotImplementedError() def select( self, s: StreamExecutor, choices: List[str], temperature: float, ): raise NotImplementedError() def concatenate_and_append(self, src_rids: List[str], dst_rid: str): raise NotImplementedError() def shutdown(self): pass def flush_cache(self): pass def get_server_args(self): pass class SglArgument(SglExpr): def __init__(self, name: str, value: str): super().__init__() self.name = name self.value = value def __repr__(self): return f"Argument(name={self.name}, value={repr(self.value)})" def __len__(self): return len(self.value) def __getitem__(self, i): return self.value[i] def __int__(self): return self.value def __bool__(self): return self.value def __format__(self, *args): raise TypeError( "Cannot put argument inside a f-string. " "This is not compatible with the tracer. " ) def trace_program(program, arguments, backend): # Create dummy backend if backend is None: backend = BaseBackend() # Create dummy arguments dummy_arguments = { name: SglArgument(name, None) for name in program.arg_names if name not in arguments } arguments.update(dummy_arguments) arguments.update(program.bind_arguments) # Trace tracer = TracerProgramState(backend, arguments, only_trace_prefix=False) with TracingScope(tracer): tracer.ret_value = program.func(tracer, **arguments) return tracer
null
7,141
import asyncio import multiprocessing import queue import threading import uuid from concurrent.futures import ThreadPoolExecutor from contextlib import contextmanager from typing import Any, Callable, Dict, List, Optional, Union import tqdm from sglang.global_config import global_config from sglang.lang.ir import ( SglCommitLazy, SglConcateAndAppend, SglConstantText, SglExpr, SglExprList, SglFunction, SglGen, SglImage, SglRoleBegin, SglRoleEnd, SglSelect, SglVariable, SglVarScopeBegin, SglVarScopeEnd, ) from sglang.utils import encode_image_base64 def run_program( program, backend, func_args, func_kwargs, default_sampling_para, stream, sync=False ): if hasattr(backend, "endpoint"): backend = backend.endpoint assert backend is not None, "Please specify a backend" func_kwargs.update(program.bind_arguments) stream_executor = StreamExecutor( backend, func_kwargs, default_sampling_para, chat_template=None, stream=stream, api_num_spec_tokens=program.api_num_spec_tokens, ) state = ProgramState(stream_executor) if stream: t = threading.Thread( target=run_internal, args=(state, program, func_args, func_kwargs, sync) ) t.start() return state else: run_internal(state, program, func_args, func_kwargs, sync) return state def pin_program(program, backend): if global_config.enable_prefix_sharing and program.pin_prefix_rid is None: # TODO: handle multiple backends from sglang.lang.tracer import extract_prefix_by_tracing prefix = extract_prefix_by_tracing(program, backend) if prefix and len(prefix) > 64: prefix_rid = backend.cache_prefix(prefix) program.pin_prefix_rid = prefix_rid return prefix_rid return None def run_program_batch( program, backend, batch_arguments, default_sampling_para, num_threads, progress_bar, ): if hasattr(backend, "endpoint"): backend = backend.endpoint # Extract prefix by tracing and cache it if len(batch_arguments) > 1: pin_program(program, backend) # Run all programs if num_threads == "auto": num_threads = max(96, multiprocessing.cpu_count() * 16) num_threads = min(num_threads, len(batch_arguments)) if num_threads == 1: rets = [] if progress_bar: for arguments in tqdm.tqdm(batch_arguments): rets.append( run_program( program, backend, (), arguments, default_sampling_para, False, True, ) ) else: for arguments in batch_arguments: rets.append( run_program( program, backend, (), arguments, default_sampling_para, False, True, ) ) else: if progress_bar: pbar = tqdm.tqdm(total=len(batch_arguments)) with ThreadPoolExecutor(num_threads) as executor: futures = [] for arguments in batch_arguments: futures.append( executor.submit( run_program, program, backend, (), arguments, default_sampling_para, False, True, ) ) if progress_bar: futures[-1].add_done_callback(lambda _: pbar.update()) rets = [f.result() for f in futures] rets[-1].sync() if progress_bar: pbar.close() return rets
null
7,142
import asyncio import multiprocessing import queue import threading import uuid from concurrent.futures import ThreadPoolExecutor from contextlib import contextmanager from typing import Any, Callable, Dict, List, Optional, Union import tqdm from sglang.global_config import global_config from sglang.lang.ir import ( SglCommitLazy, SglConcateAndAppend, SglConstantText, SglExpr, SglExprList, SglFunction, SglGen, SglImage, SglRoleBegin, SglRoleEnd, SglSelect, SglVariable, SglVarScopeBegin, SglVarScopeEnd, ) from sglang.utils import encode_image_base64 def unpin_program(program, backend): pass
null
7,143
import multiprocessing from concurrent.futures import ThreadPoolExecutor from queue import Queue from typing import List, Union from sglang.global_config import global_config from sglang.lang.interpreter import ProgramState, StreamExecutor, pin_program from sglang.lang.ir import ( SglArgument, SglConstantText, SglExpr, SglSamplingParams, SglVariable, ) class CompiledFunction: def __init__(self, tracer, function): self.function = function self.last_node = CompGraphNode(tracer.last_node) self.expr_to_node = {} self.build_graph(tracer) self.topological_sort() def build_graph(self, tracer): self.nodes = [self.last_node] self.expr_to_node[tracer.last_node] = self.nodes[-1] rename_pid = {} visited = set([tracer.last_node]) head = 0 while head < len(self.nodes): cur_node = self.nodes[head] # add prev node prev_node = cur_node.expr.prev_node if prev_node is not None: if prev_node not in visited: visited.add(prev_node) self.nodes.append(CompGraphNode(prev_node)) self.expr_to_node[prev_node] = self.nodes[-1] cur_node.prev_node = self.expr_to_node[prev_node] self.expr_to_node[prev_node].add_next_node(cur_node) # add source node if isinstance(cur_node.expr, SglVariable): if cur_node.expr.name in tracer.variables: source = tracer.variables[cur_node.expr.name].source else: source = cur_node.expr.source if source not in visited: visited.add(source) self.nodes.append(CompGraphNode(source)) self.expr_to_node[source] = self.nodes[-1] cur_node.source_node = self.expr_to_node[source] self.expr_to_node[source].add_next_node(cur_node) head += 1 # rename pid if cur_node.expr.pid not in rename_pid: rename_pid[cur_node.expr.pid] = len(rename_pid) cur_node.expr.pid = rename_pid[cur_node.expr.pid] def topological_sort(self): prevd = {} cand = Queue() for x in self.nodes: prevd[x] = (x.prev_node is not None) + (x.source_node is not None) if prevd[x] == 0: cand.put(x) new_list = [] while cand.qsize() > 0: head = cand.get() new_list.append(head) for x in head.next_nodes: prevd[x] -= 1 if prevd[x] == 0: cand.put(x) self.nodes = new_list def print_graph( self, ): for node in self.nodes: print(node) def run_internal( self, backend, kwargs, default_sampling_para, ): stream_executor_ids = set([x.expr.pid for x in self.nodes]) stream_executors = {} for x in stream_executor_ids: arguments = kwargs if x == self.last_node.expr.pid else {} stream_executors[x] = StreamExecutor( backend, arguments, default_sampling_para, None, False ) for node in self.nodes: se_id = node.expr.pid expr = node.expr if isinstance(expr, SglVariable): # Make a copy for SglVariable expr = SglVariable(expr.name, expr.source) expr.source_stream_executor = stream_executors[ node.source_node.expr.pid ] elif isinstance(expr, SglArgument): # Substitute SglArgument expr = kwargs[expr.name] stream_executors[se_id].submit(expr) for stream_executor in stream_executors.values(): stream_executor.end() return ProgramState(stream_executors[self.last_node.expr.pid]) def run( self, *, max_new_tokens: int = 16, stop: Union[str, List[str]] = (), temperature: float = 1.0, top_p: float = 1.0, top_k: int = -1, frequency_penalty: float = 0.0, presence_penalty: float = 0.0, backend=None, **kwargs, ): backend = backend or global_config.default_backend kwargs.update(self.function.bind_arguments) default_sampling_para = SglSamplingParams( max_new_tokens=max_new_tokens, stop=stop, temperature=temperature, top_p=top_p, top_k=top_k, frequency_penalty=frequency_penalty, presence_penalty=presence_penalty, ) return self.run_internal(backend, kwargs, default_sampling_para) def run_batch( self, batch_kwargs, *, max_new_tokens: int = 16, stop: Union[str, List[str]] = (), temperature: float = 1.0, top_p: float = 1.0, top_k: int = -1, frequency_penalty: float = 0.0, presence_penalty: float = 0.0, backend=None, num_threads: Union[str, int] = "auto", ): assert isinstance(batch_kwargs, (list, tuple)) if len(batch_kwargs) == 0: return [] assert isinstance(batch_kwargs[0], dict) backend = backend or global_config.default_backend default_sampling_para = SglSamplingParams( max_new_tokens=max_new_tokens, stop=stop, temperature=temperature, top_p=top_p, top_k=top_k, frequency_penalty=frequency_penalty, presence_penalty=presence_penalty, ) # Extract prefix by tracing and cache it if len(batch_kwargs) > 1: pin_program(self.function, backend) # Run all programs if num_threads == "auto": num_threads = multiprocessing.cpu_count() num_threads = min(num_threads, len(batch_kwargs)) if num_threads == 1: rets = [] for arguments in batch_kwargs: rets.append( self.run_internal(backend, arguments, default_sampling_para) ) else: with ThreadPoolExecutor(num_threads) as executor: futures = [] for arguments in batch_kwargs: futures.append( executor.submit( self.run_internal, backend, arguments, default_sampling_para ) ) rets = [f.result() for f in futures] rets[-1].sync() return rets def compile_func(function, backend): tracer = function.trace(backend=backend) compiler = CompiledFunction(tracer, function) return compiler
null
7,144
from dataclasses import dataclass, field from enum import Enum, auto from typing import Callable, Dict, List, Optional, Tuple chat_template_registry: Dict[str, ChatTemplate] = {} def register_chat_template(template): chat_template_registry[template.name] = template
null
7,145
from dataclasses import dataclass, field from enum import Enum, auto from typing import Callable, Dict, List, Optional, Tuple matching_function_registry: List[Callable] = [] def register_chat_template_matching_function(func): matching_function_registry.append(func)
null
7,146
from dataclasses import dataclass, field from enum import Enum, auto from typing import Callable, Dict, List, Optional, Tuple matching_function_registry: List[Callable] = [] def get_chat_template(name): return chat_template_registry[name] def get_chat_template_by_model_path(model_path): for matching_func in matching_function_registry: template = matching_func(model_path) if template is not None: return template return get_chat_template("default")
null
7,147
from dataclasses import dataclass, field from enum import Enum, auto from typing import Callable, Dict, List, Optional, Tuple def get_chat_template(name): return chat_template_registry[name] def match_vicuna(model_path: str): if "vicuna" in model_path.lower(): return get_chat_template("vicuna_v1.1") if "llava-v1.5" in model_path.lower(): return get_chat_template("vicuna_v1.1")
null
7,148
from dataclasses import dataclass, field from enum import Enum, auto from typing import Callable, Dict, List, Optional, Tuple def get_chat_template(name): return chat_template_registry[name] def match_llama2_chat(model_path: str): model_path = model_path.lower() if "llama-2" in model_path and "chat" in model_path: return get_chat_template("llama-2-chat") if ( "mistral" in model_path or "mixtral" in model_path ) and "instruct" in model_path: return get_chat_template("llama-2-chat") if "codellama" in model_path and "instruct" in model_path: return get_chat_template("llama-2-chat")
null
7,149
from dataclasses import dataclass, field from enum import Enum, auto from typing import Callable, Dict, List, Optional, Tuple def get_chat_template(name): def match_chat_ml(model_path: str): model_path = model_path.lower() if "tinyllama" in model_path: return get_chat_template("chatml") if "qwen" in model_path and "chat" in model_path: return get_chat_template("chatml") if "llava-v1.6-34b" in model_path: return get_chat_template("chatml-llava")
null
7,150
from dataclasses import dataclass, field from enum import Enum, auto from typing import Callable, Dict, List, Optional, Tuple def get_chat_template(name): return chat_template_registry[name] def match_chat_yi(model_path: str): model_path = model_path.lower() if "yi" in model_path: return get_chat_template("yi")
null
7,151
from dataclasses import dataclass, field from enum import Enum, auto from typing import Callable, Dict, List, Optional, Tuple def get_chat_template(name): def match_gemma_it(model_path: str): model_path = model_path.lower() if "gemma" in model_path and "it" in model_path: return get_chat_template("gemma-it")
null
7,152
import logging import time from typing import Callable, List, Optional, Union import numpy as np from sglang.backend.base_backend import BaseBackend from sglang.lang.chat_template import ChatTemplate, get_chat_template_by_model_path from sglang.lang.interpreter import StreamExecutor from sglang.lang.ir import SglSamplingParams The provided code snippet includes necessary dependencies for implementing the `create_logit_bias_int` function. Write a Python function `def create_logit_bias_int(tokenizer)` to solve the following problem: Get logit bias for integer numbers. Here is the function: def create_logit_bias_int(tokenizer): """Get logit bias for integer numbers.""" int_token_ids = [] tokens = tokenizer._mergeable_ranks for token, token_id in tokens.items(): s = tokenizer.decode([token_id]) if all([c.isdigit() for c in s]) or s in [" "]: int_token_ids.append(token_id) if len(int_token_ids) >= 300: # OpenAI API limit break special_tokens = tokenizer._special_tokens mask = {t: 100 for t in int_token_ids[:299]} mask[special_tokens["<|endoftext|>"]] = 100 return mask
Get logit bias for integer numbers.
7,153
import logging import time from typing import Callable, List, Optional, Union import numpy as np from sglang.backend.base_backend import BaseBackend from sglang.lang.chat_template import ChatTemplate, get_chat_template_by_model_path from sglang.lang.interpreter import StreamExecutor from sglang.lang.ir import SglSamplingParams try: import openai import tiktoken except ImportError as e: openai = tiktoken = e logger = logging.getLogger("openai") try: import openai import tiktoken except ImportError as e: openai = tiktoken = e def openai_completion(client, retries=3, is_chat=None, prompt=None, **kwargs): for attempt in range(retries): try: if is_chat: if "stop" in kwargs and kwargs["stop"] is None: kwargs.pop("stop") ret = client.chat.completions.create(messages=prompt, **kwargs) comp = ret.choices[0].message.content else: ret = client.completions.create(prompt=prompt, **kwargs) if isinstance(prompt, (list, tuple)): comp = [c.text for c in ret.choices] else: comp = ret.choices[0].text break except (openai.APIError, openai.APIConnectionError, openai.RateLimitError) as e: logger.error(f"OpenAI Error: {e}. Waiting 5 seconds...") time.sleep(5) if attempt == retries - 1: raise e except Exception as e: logger.error(f"RuntimeError {e}.") raise e return comp
null
7,154
import logging import time from typing import Callable, List, Optional, Union import numpy as np from sglang.backend.base_backend import BaseBackend from sglang.lang.chat_template import ChatTemplate, get_chat_template_by_model_path from sglang.lang.interpreter import StreamExecutor from sglang.lang.ir import SglSamplingParams try: import openai import tiktoken except ImportError as e: openai = tiktoken = e logger = logging.getLogger("openai") try: import openai import tiktoken except ImportError as e: openai = tiktoken = e def openai_completion_stream(client, retries=3, is_chat=None, prompt=None, **kwargs): for attempt in range(retries): try: if is_chat: if "stop" in kwargs and kwargs["stop"] is None: kwargs.pop("stop") generator = client.chat.completions.create( messages=prompt, stream=True, **kwargs ) for ret in generator: try: content = ret.choices[0].delta.content except IndexError: content = None yield content or "", {} else: generator = client.completions.create( prompt=prompt, stream=True, **kwargs ) for ret in generator: content = ret.choices[0].text yield content or "", {} break except (openai.APIError, openai.APIConnectionError, openai.RateLimitError) as e: logger.error(f"OpenAI Error: {e}. Waiting 5 seconds...") time.sleep(5) if attempt == retries - 1: raise e except Exception as e: logger.error(f"RuntimeError {e}.") raise e
null
7,155
import base64 import os import random import socket import sys import time import traceback from io import BytesIO from typing import List, Optional import numpy as np import requests import torch import torch.distributed as dist is_show_cost_time = False def mark_cost_time(func_name): def inner_func(func): def time_func(*args, **kwargs): if dist.get_rank() in [0, 1] and is_show_cost_time: torch.cuda.synchronize() start_time = time.time() ans = func(*args, **kwargs) torch.cuda.synchronize() print(func_name, "cost time:", (time.time() - start_time) * 1000) return ans else: torch.cuda.synchronize() ans = func(*args, **kwargs) torch.cuda.synchronize() return ans return time_func return inner_func
null
7,156
import base64 import os import random import socket import sys import time import traceback from io import BytesIO from typing import List, Optional import numpy as np import requests import torch import torch.distributed as dist time_mark = {} def mark_start(key): torch.cuda.synchronize() global time_mark time_mark[key] = time.time() return
null
7,157
import base64 import os import random import socket import sys import time import traceback from io import BytesIO from typing import List, Optional import numpy as np import requests import torch import torch.distributed as dist time_mark = {} def mark_end(key, print_min_cost=0.0): torch.cuda.synchronize() global time_mark cost_time = (time.time() - time_mark[key]) * 1000 if cost_time > print_min_cost: print(f"cost {key}:", cost_time)
null
7,158
import base64 import os import random import socket import sys import time import traceback from io import BytesIO from typing import List, Optional import numpy as np import requests import torch import torch.distributed as dist def calculate_time(show=False, min_cost_ms=0.0): def wrapper(func): def inner_func(*args, **kwargs): torch.cuda.synchronize() if show: start_time = time.time() result = func(*args, **kwargs) torch.cuda.synchronize() if show: cost_time = (time.time() - start_time) * 1000 if cost_time > min_cost_ms: print(f"Function {func.__name__} took {cost_time} ms to run.") return result return inner_func return wrapper
null
7,159
import base64 import os import random import socket import sys import time import traceback from io import BytesIO from typing import List, Optional import numpy as np import requests import torch import torch.distributed as dist def set_random_seed(seed: int) -> None: random.seed(seed) torch.manual_seed(seed) if torch.cuda.is_available(): torch.cuda.manual_seed_all(seed)
null
7,160
import base64 import os import random import socket import sys import time import traceback from io import BytesIO from typing import List, Optional import numpy as np import requests import torch import torch.distributed as dist def get_int_token_logit_bias(tokenizer, vocab_size): from transformers import LlamaTokenizer, LlamaTokenizerFast # a bug when model's vocab size > tokenizer.vocab_size vocab_size = tokenizer.vocab_size logit_bias = np.zeros(vocab_size, dtype=np.float32) for t_id in range(vocab_size): ss = tokenizer.decode([t_id]).strip() if not (ss.isdigit() or len(ss) == 0 or t_id == tokenizer.eos_token_id): logit_bias[t_id] = -1e5 return logit_bias
null
7,161
import interegular from sglang.srt.constrained import FSMInfo, disk_cache, make_deterministic_fsm from sglang.srt.constrained.base_cache import BaseCache IP_REGEX = r"((25[0-5]|2[0-4]\d|[01]?\d\d?)\.){3}(25[0-5]|2[0-4]\d|[01]?\d\d?)" class JumpForwardMap: def __init__(self, regex_string): def _init_state_to_jump_forward(regex_string): regex_pattern = interegular.parse_pattern(regex_string) regex_fsm, _ = make_deterministic_fsm(regex_pattern.to_fsm().reduce()) fsm_info: FSMInfo = regex_fsm.fsm_info symbol_to_id = fsm_info.alphabet_symbol_mapping id_to_symbol = {} for symbol, id_ in symbol_to_id.items(): id_to_symbol.setdefault(id_, []).append(symbol) transitions = fsm_info.transitions dirty_states = set() state_to_jump_forward = {} for (state, id_), next_state in transitions.items(): if state in dirty_states: continue if state in state_to_jump_forward: dirty_states.add(state) del state_to_jump_forward[state] continue if len(id_to_symbol[id_]) > 1: dirty_states.add(state) continue state_to_jump_forward[state] = (id_to_symbol[id_][0], next_state) return state_to_jump_forward self.state_to_jump_forward = _init_state_to_jump_forward(regex_string) def valid_states(self): return self.state_to_jump_forward.keys() def jump_forward(self, state): if state not in self.state_to_jump_forward: return None jump_forward_str = "" next_state = None while state in self.state_to_jump_forward: symbol, next_state = self.state_to_jump_forward[state] jump_forward_str += symbol state = next_state return jump_forward_str, next_state def test_main(): regex_string = r"The google's DNS sever address is " + IP_REGEX jump_forward_map = JumpForwardMap(regex_string) for state in jump_forward_map.valid_states(): print(state, f'"{jump_forward_map.jump_forward(state)}"')
null
7,162
import torch import triton import triton.language as tl from sglang.srt.layers.context_flashattention_nopad import context_attention_fwd from sglang.srt.utils import wrap_kernel_launcher CUDA_CAPABILITY = torch.cuda.get_device_capability() def _fwd_kernel( Q_Extend, K_Extend, V_Extend, O_Extend, K_Buffer, V_Buffer, Req_to_tokens, B_req_idx, B_Seq_Len, B_Start_Loc_Extend, B_Seq_Len_Extend, sm_scale, kv_group_num, stride_qbs, stride_qh, stride_kbs, stride_kh, stride_vbs, stride_vh, stride_obs, stride_oh, stride_buf_kbs, stride_buf_kh, stride_buf_vbs, stride_buf_vh, stride_req_to_tokens_b, BLOCK_DMODEL: tl.constexpr, BLOCK_M: tl.constexpr, BLOCK_N: tl.constexpr, ): cur_seq = tl.program_id(0) cur_head = tl.program_id(1) cur_block_m = tl.program_id(2) cur_kv_head = cur_head // kv_group_num cur_seq_len = tl.load(B_Seq_Len + cur_seq) cur_seq_len_extend = tl.load(B_Seq_Len_Extend + cur_seq) cur_seq_len_prefix = cur_seq_len - cur_seq_len_extend cur_seq_prefix_start_in_loc = 0 cur_seq_extend_start_contiguous = tl.load(B_Start_Loc_Extend + cur_seq) cur_batch_req_idx = tl.load(B_req_idx + cur_seq) offs_d = tl.arange(0, BLOCK_DMODEL) offs_m = tl.arange(0, BLOCK_M) mask_m = (cur_block_m * BLOCK_M + offs_m) < cur_seq_len_extend offs_q = ( (cur_seq_extend_start_contiguous + cur_block_m * BLOCK_M + offs_m[:, None]) * stride_qbs + cur_head * stride_qh + offs_d[None, :] ) q = tl.load(Q_Extend + offs_q, mask=mask_m[:, None], other=0.0) # stage1: compute scores with prefix offs_n = tl.arange(0, BLOCK_N) acc = tl.zeros([BLOCK_M, BLOCK_DMODEL], dtype=tl.float32) deno = tl.zeros([BLOCK_M], dtype=tl.float32) e_max = tl.zeros([BLOCK_M], dtype=tl.float32) - float("inf") for start_n in range(0, cur_seq_len_prefix, BLOCK_N): start_n = tl.multiple_of(start_n, BLOCK_N) mask_n = (start_n + offs_n) < cur_seq_len_prefix offs_b_loc_prefix = cur_batch_req_idx * stride_req_to_tokens_b + ( cur_seq_prefix_start_in_loc + start_n + offs_n ) offs_kv_loc = tl.load(Req_to_tokens + offs_b_loc_prefix, mask=mask_n, other=0) # load k in transposed way offs_buf_k = ( offs_kv_loc[None, :] * stride_buf_kbs + cur_kv_head * stride_buf_kh + offs_d[:, None] ) k = tl.load(K_Buffer + offs_buf_k, mask=mask_n[None, :], other=0.0) qk = tl.zeros([BLOCK_M, BLOCK_N], dtype=tl.float32) qk += tl.dot(q, k) qk *= sm_scale qk = tl.where(mask_m[:, None] & mask_n[None, :], qk, float("-inf")) n_e_max = tl.maximum(tl.max(qk, 1), e_max) re_scale = tl.exp(e_max - n_e_max) p = tl.exp(qk - n_e_max[:, None]) deno = deno * re_scale + tl.sum(p, 1) offs_buf_v = ( offs_kv_loc[:, None] * stride_buf_vbs + cur_kv_head * stride_buf_vh + offs_d[None, :] ) v = tl.load(V_Buffer + offs_buf_v, mask=mask_n[:, None], other=0.0) p = p.to(v.dtype) acc = acc * re_scale[:, None] + tl.dot(p, v) e_max = n_e_max # stage2: compute the trianlge part cur_block_m_end = tl.minimum(cur_seq_len_extend, (cur_block_m + 1) * BLOCK_M) for start_n in range(0, cur_block_m_end, BLOCK_N): start_n = tl.multiple_of(start_n, BLOCK_N) mask_n = (start_n + offs_n) < cur_block_m_end # load k in transposed way offs_k = ( (cur_seq_extend_start_contiguous + start_n + offs_n[None, :]) * stride_kbs + cur_kv_head * stride_kh + offs_d[:, None] ) k = tl.load(K_Extend + offs_k, mask=mask_n[None, :], other=0.0) qk = tl.zeros([BLOCK_M, BLOCK_N], dtype=tl.float32) qk += tl.dot(q, k) qk *= sm_scale mask_causual = (cur_block_m * BLOCK_M + offs_m[:, None]) >= ( start_n + offs_n[None, :] ) mask_causual &= mask_m[:, None] & mask_n[None, :] qk = tl.where(mask_causual, qk, float("-inf")) n_e_max = tl.maximum(tl.max(qk, 1), e_max) re_scale = tl.exp(e_max - n_e_max) p = tl.exp(qk - n_e_max[:, None]) deno = deno * re_scale + tl.sum(p, 1) offs_v = ( (cur_seq_extend_start_contiguous + start_n + offs_n[:, None]) * stride_vbs + cur_kv_head * stride_vh + offs_d[None, :] ) v = tl.load(V_Extend + offs_v, mask=mask_n[:, None], other=0.0) p = p.to(v.dtype) acc = acc * re_scale[:, None] + tl.dot(p, v) e_max = n_e_max offs_o = ( (cur_seq_extend_start_contiguous + cur_block_m * BLOCK_M + offs_m[:, None]) * stride_obs + cur_head * stride_oh + offs_d[None, :] ) tl.store(O_Extend + offs_o, acc / deno[:, None], mask=mask_m[:, None]) cached_kernel = None def wrap_kernel_launcher(kernel): """A faster launcher for triton kernels.""" import torch.distributed as dist if dist.is_initialized(): rank = dist.get_rank() else: rank = 0 kernels = kernel.cache[rank].values() kernel = next(iter(kernels)) # Different trition versions use different low-level names if hasattr(kernel, "cu_function"): kfunction = kernel.cu_function else: kfunction = kernel.function if hasattr(kernel, "c_wrapper"): run = kernel.c_wrapper else: run = kernel.run add_cluster_dim = True def ret_func(grid, num_warps, *args): nonlocal add_cluster_dim try: if add_cluster_dim: run( grid[0], grid[1], grid[2], num_warps, 1, 1, 1, 1, kernel.shared, 0, kfunction, None, None, kernel, *args, ) else: run( grid[0], grid[1], grid[2], num_warps, kernel.shared, 0, kfunction, None, None, kernel, *args, ) except TypeError: add_cluster_dim = not add_cluster_dim ret_func(grid, num_warps, *args) return ret_func The provided code snippet includes necessary dependencies for implementing the `extend_attention_fwd` function. Write a Python function `def extend_attention_fwd( q_extend, k_extend, v_extend, o_extend, k_buffer, v_buffer, req_to_tokens, b_req_idx, b_start_loc, b_seq_len, b_seq_len_prefix, b_start_loc_extend, b_seq_len_extend, max_len_in_batch, max_len_extend, )` to solve the following problem: q_extend, k_extend, v_extend, o_extend: contiguous tensors k_buffer, v_buffer: (prefix + extend) tensors in mem_manager Here is the function: def extend_attention_fwd( q_extend, k_extend, v_extend, o_extend, k_buffer, v_buffer, req_to_tokens, b_req_idx, b_start_loc, b_seq_len, b_seq_len_prefix, b_start_loc_extend, b_seq_len_extend, max_len_in_batch, max_len_extend, ): """ q_extend, k_extend, v_extend, o_extend: contiguous tensors k_buffer, v_buffer: (prefix + extend) tensors in mem_manager """ Lq, Lk, Lv, Lo = ( q_extend.shape[-1], k_extend.shape[-1], v_extend.shape[-1], o_extend.shape[-1], ) assert Lq == Lk and Lk == Lv and Lv == Lo assert Lq in {16, 32, 64, 128, 256} if CUDA_CAPABILITY[0] >= 8: BLOCK_M, BLOCK_N = (128, 128) if Lq <= 128 else (64, 64) else: BLOCK_M, BLOCK_N = (64, 64) if Lq <= 128 else (32, 32) sm_scale = 1.0 / (Lq**0.5) batch_size, head_num = b_seq_len.shape[0], q_extend.shape[1] kv_group_num = q_extend.shape[1] // k_extend.shape[1] grid = (batch_size, head_num, triton.cdiv(max_len_extend, BLOCK_M)) num_warps = 4 if Lk <= 64 else 8 num_stages = 1 global cached_kernel if cached_kernel: cached_kernel( grid, num_warps, q_extend, k_extend, v_extend, o_extend, k_buffer, v_buffer, req_to_tokens, b_req_idx, b_seq_len, b_start_loc_extend, b_seq_len_extend, sm_scale, kv_group_num, q_extend.stride(0), q_extend.stride(1), k_extend.stride(0), k_extend.stride(1), v_extend.stride(0), v_extend.stride(1), o_extend.stride(0), o_extend.stride(1), k_buffer.stride(0), k_buffer.stride(1), v_buffer.stride(0), v_buffer.stride(1), req_to_tokens.stride(0), ) return _fwd_kernel[grid]( q_extend, k_extend, v_extend, o_extend, k_buffer, v_buffer, req_to_tokens, b_req_idx, b_seq_len, b_start_loc_extend, b_seq_len_extend, sm_scale, kv_group_num, q_extend.stride(0), q_extend.stride(1), k_extend.stride(0), k_extend.stride(1), v_extend.stride(0), v_extend.stride(1), o_extend.stride(0), o_extend.stride(1), k_buffer.stride(0), k_buffer.stride(1), v_buffer.stride(0), v_buffer.stride(1), req_to_tokens.stride(0), BLOCK_DMODEL=Lq, BLOCK_M=BLOCK_M, BLOCK_N=BLOCK_N, num_warps=num_warps, num_stages=num_stages, ) cached_kernel = wrap_kernel_launcher(_fwd_kernel)
q_extend, k_extend, v_extend, o_extend: contiguous tensors k_buffer, v_buffer: (prefix + extend) tensors in mem_manager
7,163
import torch import triton import triton.language as tl from sglang.srt.layers.context_flashattention_nopad import context_attention_fwd from sglang.srt.utils import wrap_kernel_launcher def context_attention_fwd(q, k, v, o, b_start_loc, b_seq_len, max_input_len): def redundant_attention( q_extend, k_extend, v_extend, o_extend, k_buffer, v_buffer, req_to_tokens, b_req_idx, b_start_loc, b_seq_len, b_seq_len_prefix, max_len_in_batch, ): total_token_num = k_buffer.shape[0] B, H_Q, D = b_req_idx.shape[0], q_extend.shape[-2], q_extend.shape[-1] q_buffer = torch.empty( (total_token_num, H_Q, D), dtype=q_extend.dtype, device=q_extend.device ) pt = 0 for i in range(B): cur_seq_len_extend = b_seq_len[i] - b_seq_len_prefix[i] pl, pr = b_start_loc[i] + b_seq_len_prefix[i], b_start_loc[i] + b_seq_len[i] q_buffer[pl:pr] = q_extend[pt : pt + cur_seq_len_extend] pt += cur_seq_len_extend o_buffer = torch.empty_like(q_buffer) context_attention_fwd( q_buffer, k_buffer, v_buffer, o_buffer, b_start_loc, b_seq_len, max_len_in_batch ) pt = 0 for i in range(B): cur_seq_len_extend = b_seq_len[i] - b_seq_len_prefix[i] pl, pr = b_start_loc[i] + b_seq_len_prefix[i], b_start_loc[i] + b_seq_len[i] o_extend[pt : pt + cur_seq_len_extend] = o_buffer[pl:pr] pt += cur_seq_len_extend
null
7,164
import torch import triton import triton.language as tl from sglang.srt.managers.router.model_runner import global_server_args_dict from sglang.srt.utils import wrap_kernel_launcher def _token_att_m_fwd( q, k_buffer, att_out, Req_to_tokens, B_req_idx, B_Start_Loc, B_Seqlen, max_len_in_batch, ): BLOCK = 32 # shape constraints Lq, Lk = q.shape[-1], k_buffer.shape[-1] assert Lq == Lk assert Lk in {16, 32, 64, 128, 256} sm_scale = 1.0 / (Lk**0.5) batch, head_num = B_req_idx.shape[0], q.shape[1] grid = (batch, head_num, triton.cdiv(max_len_in_batch, BLOCK)) kv_group_num = q.shape[1] // k_buffer.shape[1] if kv_group_num == 1: num_warps = 4 else: num_warps = 2 global cached_kernel_stage1 if cached_kernel_stage1: cached_kernel_stage1( grid, num_warps, q, k_buffer, sm_scale, Req_to_tokens, B_req_idx, B_Start_Loc, B_Seqlen, att_out, Req_to_tokens.stride(0), q.stride(0), q.stride(1), k_buffer.stride(0), k_buffer.stride(1), att_out.stride(0), ) return _fwd_kernel_stage1[grid]( q, k_buffer, sm_scale, Req_to_tokens, B_req_idx, B_Start_Loc, B_Seqlen, att_out, Req_to_tokens.stride(0), q.stride(0), q.stride(1), k_buffer.stride(0), k_buffer.stride(1), att_out.stride(0), kv_group_num=kv_group_num, BLOCK_DMODEL=Lk, BLOCK_N=BLOCK, num_warps=num_warps, num_stages=1, ) cached_kernel_stage1 = wrap_kernel_launcher(_fwd_kernel_stage1) def _token_softmax_reducev_fwd( logics, v_buffer, o, req_to_tokens, b_req_idx, b_start_loc, b_seq_len, other_kv_index, ): BLOCK = 64 batch, head = b_seq_len.shape[0], logics.shape[0] grid = (batch, head, 1) kv_group_num = logics.shape[0] // v_buffer.shape[1] num_warps = 1 global cached_kernel_stage2 if cached_kernel_stage2: cached_kernel_stage2( grid, num_warps, logics, v_buffer, o, req_to_tokens, b_req_idx, b_start_loc, b_seq_len, logics.stride(0), v_buffer.stride(0), v_buffer.stride(1), o.stride(0), o.stride(1), req_to_tokens.stride(0), other_kv_index, ) return _fwd_kernel_stage2[grid]( logics, v_buffer, o, req_to_tokens, b_req_idx, b_start_loc, b_seq_len, logics.stride(0), v_buffer.stride(0), v_buffer.stride(1), o.stride(0), o.stride(1), req_to_tokens.stride(0), other_kv_index, kv_group_num=kv_group_num, BLOCK_DMODEL=v_buffer.shape[-1], BLOCK_N=BLOCK, num_warps=num_warps, num_stages=3, ) cached_kernel_stage2 = wrap_kernel_launcher(_fwd_kernel_stage2) def token_attention_fwd( q, k_buffer, v_buffer, o, req_to_token, b_req_idx, b_start_loc, b_seq_len, max_len_in_batch, other_kv_index, total_num_tokens, att_m=None, ): if att_m is None: att_m = torch.empty( (q.shape[-2], total_num_tokens), dtype=REDUCE_TORCH_TYPE, device="cuda" ) _token_att_m_fwd( q, k_buffer, att_m, req_to_token, b_req_idx, b_start_loc, b_seq_len, max_len_in_batch, ) _token_softmax_reducev_fwd( att_m, v_buffer, o, req_to_token, b_req_idx, b_start_loc, b_seq_len, other_kv_index, )
null
7,165
import json import os import warnings from typing import List, Optional, Tuple, Union from huggingface_hub import snapshot_download from sglang.srt.utils import is_multimodal_model from transformers import ( AutoConfig, AutoProcessor, AutoTokenizer, PreTrainedTokenizer, PreTrainedTokenizerFast, ) def download_from_hf(model_path: str): if os.path.exists(model_path): return model_path return snapshot_download(model_path, allow_patterns=["*.json", "*.bin", "*.model"])
null
7,166
import json import os import warnings from typing import List, Optional, Tuple, Union from huggingface_hub import snapshot_download from sglang.srt.utils import is_multimodal_model from transformers import ( AutoConfig, AutoProcessor, AutoTokenizer, PreTrainedTokenizer, PreTrainedTokenizerFast, ) def get_config_json(model_path: str): with open(os.path.join(model_path, "config.json")) as f: config = json.load(f) return config
null
7,167
import json import os import warnings from typing import List, Optional, Tuple, Union from huggingface_hub import snapshot_download from sglang.srt.utils import is_multimodal_model from transformers import ( AutoConfig, AutoProcessor, AutoTokenizer, PreTrainedTokenizer, PreTrainedTokenizerFast, ) def get_config(model: str, trust_remote_code: bool, revision: Optional[str] = None): config = AutoConfig.from_pretrained( model, trust_remote_code=trust_remote_code, revision=revision ) return config
null
7,168
import json import os import warnings from typing import List, Optional, Tuple, Union from huggingface_hub import snapshot_download from sglang.srt.utils import is_multimodal_model from transformers import ( AutoConfig, AutoProcessor, AutoTokenizer, PreTrainedTokenizer, PreTrainedTokenizerFast, ) CONTEXT_LENGTH_KEYS = [ "max_sequence_length", "seq_length", "max_position_embeddings", "max_seq_len", "model_max_length", ] The provided code snippet includes necessary dependencies for implementing the `get_context_length` function. Write a Python function `def get_context_length(config)` to solve the following problem: Get the context length of a model from a huggingface model config. Here is the function: def get_context_length(config): """Get the context length of a model from a huggingface model config.""" rope_scaling = getattr(config, "rope_scaling", None) if rope_scaling: rope_scaling_factor = config.rope_scaling["factor"] else: rope_scaling_factor = 1 for key in CONTEXT_LENGTH_KEYS: val = getattr(config, key, None) if val is not None: return int(rope_scaling_factor * val) return 2048
Get the context length of a model from a huggingface model config.
7,169
import json import os import warnings from typing import List, Optional, Tuple, Union from huggingface_hub import snapshot_download from sglang.srt.utils import is_multimodal_model from transformers import ( AutoConfig, AutoProcessor, AutoTokenizer, PreTrainedTokenizer, PreTrainedTokenizerFast, ) _FAST_LLAMA_TOKENIZER = "hf-internal-testing/llama-tokenizer" def get_processor( tokenizer_name: str, *args, tokenizer_mode: str = "auto", trust_remote_code: bool = False, tokenizer_revision: Optional[str] = None, **kwargs, ): processor = AutoProcessor.from_pretrained( tokenizer_name, *args, trust_remote_code=trust_remote_code, tokenizer_revision=tokenizer_revision, **kwargs, ) return processor def is_multimodal_model(model): if isinstance(model, str): return "llava" in model or "yi-vl" in model from sglang.srt.model_config import ModelConfig if isinstance(model, ModelConfig): model_path = model.path.lower() return "llava" in model_path or "yi-vl" in model_path raise Exception("unrecognized type") The provided code snippet includes necessary dependencies for implementing the `get_tokenizer` function. Write a Python function `def get_tokenizer( tokenizer_name: str, *args, tokenizer_mode: str = "auto", trust_remote_code: bool = False, tokenizer_revision: Optional[str] = None, **kwargs, ) -> Union[PreTrainedTokenizer, PreTrainedTokenizerFast]` to solve the following problem: Gets a tokenizer for the given model name via Huggingface. Here is the function: def get_tokenizer( tokenizer_name: str, *args, tokenizer_mode: str = "auto", trust_remote_code: bool = False, tokenizer_revision: Optional[str] = None, **kwargs, ) -> Union[PreTrainedTokenizer, PreTrainedTokenizerFast]: """Gets a tokenizer for the given model name via Huggingface.""" if is_multimodal_model(tokenizer_name): processor = get_processor( tokenizer_name, *args, trust_remote_code=trust_remote_code, tokenizer_revision=tokenizer_revision, **kwargs, ) tokenizer = processor.tokenizer return tokenizer if tokenizer_mode == "slow": if kwargs.get("use_fast", False): raise ValueError("Cannot use the fast tokenizer in slow tokenizer mode.") kwargs["use_fast"] = False if ( "llama" in tokenizer_name.lower() and kwargs.get("use_fast", True) and tokenizer_name != _FAST_LLAMA_TOKENIZER ): pass # warnings.warn( # "For some LLaMA V1 models, initializing the fast tokenizer may " # "take a long time. To reduce the initialization time, consider " # f"using '{_FAST_LLAMA_TOKENIZER}' instead of the original " # "tokenizer." # ) try: tokenizer = AutoTokenizer.from_pretrained( tokenizer_name, *args, trust_remote_code=trust_remote_code, tokenizer_revision=tokenizer_revision, **kwargs, ) except TypeError as e: # The LLaMA tokenizer causes a protobuf error in some environments. err_msg = ( "Failed to load the tokenizer. If you are using a LLaMA V1 model " f"consider using '{_FAST_LLAMA_TOKENIZER}' instead of the " "original tokenizer." ) raise RuntimeError(err_msg) from e except ValueError as e: # If the error pertains to the tokenizer class not existing or not # currently being imported, suggest using the --trust-remote-code flag. if not trust_remote_code and ( "does not exist or is not currently imported." in str(e) or "requires you to execute the tokenizer file" in str(e) ): err_msg = ( "Failed to load the tokenizer. If the tokenizer is a custom " "tokenizer not yet available in the HuggingFace transformers " "library, consider setting `trust_remote_code=True` in LLM " "or using the `--trust-remote-code` flag in the CLI." ) raise RuntimeError(err_msg) from e else: raise e if not isinstance(tokenizer, PreTrainedTokenizerFast): warnings.warn( "Using a slow tokenizer. This might cause a significant " "slowdown. Consider using a fast tokenizer instead." ) return tokenizer
Gets a tokenizer for the given model name via Huggingface.
7,170
import asyncio import dataclasses import json import multiprocessing as mp import os import sys import threading import time from typing import List, Optional, Union import aiohttp import psutil import pydantic import requests import uvicorn import uvloop from fastapi import FastAPI, HTTPException, Request from fastapi.responses import Response, StreamingResponse from pydantic import BaseModel from sglang.backend.runtime_endpoint import RuntimeEndpoint from sglang.srt.constrained import disable_cache from sglang.srt.conversation import ( Conversation, SeparatorStyle, chat_template_exists, generate_chat_conv, register_conv_template, ) from sglang.srt.hf_transformers_utils import get_tokenizer from sglang.srt.managers.detokenizer_manager import start_detokenizer_process from sglang.srt.managers.io_struct import DetokenizeReqInput, GenerateReqInput from sglang.srt.managers.openai_protocol import ( ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, DeltaMessage, LogProbs, UsageInfo, ) from sglang.srt.managers.router.manager import start_router_process from sglang.srt.managers.tokenizer_manager import TokenizerManager from sglang.srt.server_args import PortArgs, ServerArgs from sglang.srt.utils import handle_port_init from starlette.middleware.base import BaseHTTPMiddleware from starlette.responses import JSONResponse The provided code snippet includes necessary dependencies for implementing the `health` function. Write a Python function `async def health() -> Response` to solve the following problem: Health check. Here is the function: async def health() -> Response: """Health check.""" return Response(status_code=200)
Health check.
7,171
import asyncio import dataclasses import json import multiprocessing as mp import os import sys import threading import time from typing import List, Optional, Union import aiohttp import psutil import pydantic import requests import uvicorn import uvloop from fastapi import FastAPI, HTTPException, Request from fastapi.responses import Response, StreamingResponse from pydantic import BaseModel from sglang.backend.runtime_endpoint import RuntimeEndpoint from sglang.srt.constrained import disable_cache from sglang.srt.conversation import ( Conversation, SeparatorStyle, chat_template_exists, generate_chat_conv, register_conv_template, ) from sglang.srt.hf_transformers_utils import get_tokenizer from sglang.srt.managers.detokenizer_manager import start_detokenizer_process from sglang.srt.managers.io_struct import DetokenizeReqInput, GenerateReqInput from sglang.srt.managers.openai_protocol import ( ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, DeltaMessage, LogProbs, UsageInfo, ) from sglang.srt.managers.router.manager import start_router_process from sglang.srt.managers.tokenizer_manager import TokenizerManager from sglang.srt.server_args import PortArgs, ServerArgs from sglang.srt.utils import handle_port_init from starlette.middleware.base import BaseHTTPMiddleware from starlette.responses import JSONResponse tokenizer_manager = None async def get_model_info(): result = { "model_path": tokenizer_manager.model_path, } return result
null
7,172
import asyncio import dataclasses import json import multiprocessing as mp import os import sys import threading import time from typing import List, Optional, Union import aiohttp import psutil import pydantic import requests import uvicorn import uvloop from fastapi import FastAPI, HTTPException, Request from fastapi.responses import Response, StreamingResponse from pydantic import BaseModel from sglang.backend.runtime_endpoint import RuntimeEndpoint from sglang.srt.constrained import disable_cache from sglang.srt.conversation import ( Conversation, SeparatorStyle, chat_template_exists, generate_chat_conv, register_conv_template, ) from sglang.srt.hf_transformers_utils import get_tokenizer from sglang.srt.managers.detokenizer_manager import start_detokenizer_process from sglang.srt.managers.io_struct import DetokenizeReqInput, GenerateReqInput from sglang.srt.managers.openai_protocol import ( ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, DeltaMessage, LogProbs, UsageInfo, ) from sglang.srt.managers.router.manager import start_router_process from sglang.srt.managers.tokenizer_manager import TokenizerManager from sglang.srt.server_args import PortArgs, ServerArgs from sglang.srt.utils import handle_port_init from starlette.middleware.base import BaseHTTPMiddleware from starlette.responses import JSONResponse tokenizer_manager = None async def get_server_args(): return dataclasses.asdict(tokenizer_manager.server_args)
null
7,173
import asyncio import dataclasses import json import multiprocessing as mp import os import sys import threading import time from typing import List, Optional, Union import aiohttp import psutil import pydantic import requests import uvicorn import uvloop from fastapi import FastAPI, HTTPException, Request from fastapi.responses import Response, StreamingResponse from pydantic import BaseModel from sglang.backend.runtime_endpoint import RuntimeEndpoint from sglang.srt.constrained import disable_cache from sglang.srt.conversation import ( Conversation, SeparatorStyle, chat_template_exists, generate_chat_conv, register_conv_template, ) from sglang.srt.hf_transformers_utils import get_tokenizer from sglang.srt.managers.detokenizer_manager import start_detokenizer_process from sglang.srt.managers.io_struct import DetokenizeReqInput, GenerateReqInput from sglang.srt.managers.openai_protocol import ( ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, DeltaMessage, LogProbs, UsageInfo, ) from sglang.srt.managers.router.manager import start_router_process from sglang.srt.managers.tokenizer_manager import TokenizerManager from sglang.srt.server_args import PortArgs, ServerArgs from sglang.srt.utils import handle_port_init from starlette.middleware.base import BaseHTTPMiddleware from starlette.responses import JSONResponse tokenizer_manager = None async def flush_cache(): await tokenizer_manager.flush_cache() return Response( content="Cache flushed.\nPlease check backend logs for more details. " "(When there are running or waiting requests, the operation will not be performed.)\n", status_code=200, )
null
7,174
import asyncio import dataclasses import json import multiprocessing as mp import os import sys import threading import time from typing import List, Optional, Union import aiohttp import psutil import pydantic import requests import uvicorn import uvloop from fastapi import FastAPI, HTTPException, Request from fastapi.responses import Response, StreamingResponse from pydantic import BaseModel from sglang.backend.runtime_endpoint import RuntimeEndpoint from sglang.srt.constrained import disable_cache from sglang.srt.conversation import ( Conversation, SeparatorStyle, chat_template_exists, generate_chat_conv, register_conv_template, ) from sglang.srt.hf_transformers_utils import get_tokenizer from sglang.srt.managers.detokenizer_manager import start_detokenizer_process from sglang.srt.managers.io_struct import DetokenizeReqInput, GenerateReqInput from sglang.srt.managers.openai_protocol import ( ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, DeltaMessage, LogProbs, UsageInfo, ) from sglang.srt.managers.router.manager import start_router_process from sglang.srt.managers.tokenizer_manager import TokenizerManager from sglang.srt.server_args import PortArgs, ServerArgs from sglang.srt.utils import handle_port_init from starlette.middleware.base import BaseHTTPMiddleware from starlette.responses import JSONResponse def jsonify_pydantic_model(obj: BaseModel): if IS_PYDANTIC_1: return obj.json(ensure_ascii=False) return obj.model_dump_json() async def stream_generator(obj: GenerateReqInput): async for out in tokenizer_manager.generate_request(obj): if obj.return_logprob and obj.return_text_in_logprobs: out["meta_info"]["token_logprob"] = await detokenize_logprob_tokens( out["meta_info"]["token_logprob"] ) yield out async def make_openai_style_logprobs(token_logprobs): ret_logprobs = LogProbs() for token_text, token_logprob in token_logprobs: ret_logprobs.tokens.append(token_text) ret_logprobs.token_logprobs.append(token_logprob) # Not supported yet. ret_logprobs.top_logprobs.append({}) ret_logprobs.text_offset.append(-1) return ret_logprobs async def generate_request(obj: GenerateReqInput): obj.post_init() if obj.stream: async def stream_results(): async for out in stream_generator(obj): yield f"data: {json.dumps(out, ensure_ascii=False)}\n\n" yield "data: [DONE]\n\n" return StreamingResponse(stream_results(), media_type="text/event-stream") ret = await tokenizer_manager.generate_request(obj).__anext__() if obj.return_logprob and obj.return_text_in_logprobs: ret["meta_info"]["token_logprob"] = await detokenize_logprob_tokens( ret["meta_info"]["token_logprob"] ) return ret class GenerateReqInput: # The input prompt text: Union[List[str], str] # The image input image_data: Optional[Union[List[str], str]] = None # The sampling_params sampling_params: Union[List[Dict], Dict] = None # The request id rid: Optional[Union[List[str], str]] = None # Whether to return logprobs return_logprob: Optional[Union[List[bool], bool]] = None # The start location of the prompt for return_logprob logprob_start_len: Optional[Union[List[int], int]] = None # Whether to detokenize tokens in logprobs return_text_in_logprobs: bool = False # Whether to stream output stream: bool = False def post_init(self): is_single = isinstance(self.text, str) if is_single: if self.sampling_params is None: self.sampling_params = {} if self.rid is None: self.rid = uuid.uuid4().hex if self.return_logprob is None: self.return_logprob = False if self.logprob_start_len is None: self.logprob_start_len = 0 else: num = len(self.text) if self.image_data is None: self.image_data = [None] * num elif not isinstance(self.image_data, list): self.image_data = [self.image_data] * num if self.sampling_params is None: self.sampling_params = [{}] * num elif not isinstance(self.sampling_params, list): self.sampling_params = [self.sampling_params] * num if self.rid is None: self.rid = [uuid.uuid4().hex for _ in range(num)] else: assert isinstance(self.rid, list) if self.return_logprob is None: self.return_logprob = [False] * num elif not isinstance(self.return_logprob, list): self.return_logprob = [self.return_logprob] * num if self.logprob_start_len is None: self.logprob_start_len = [0] * num elif not isinstance(self.logprob_start_len, list): self.logprob_start_len = [self.logprob_start_len] * num class UsageInfo(BaseModel): prompt_tokens: int = 0 total_tokens: int = 0 completion_tokens: Optional[int] = 0 class CompletionRequest(BaseModel): model: str prompt: Union[str, List[str]] suffix: Optional[str] = None max_tokens: Optional[int] = 16 temperature: Optional[float] = 0.7 top_p: Optional[float] = 1.0 n: Optional[int] = 1 stream: Optional[bool] = False logprobs: Optional[int] = None echo: Optional[bool] = False stop: Optional[Union[str, List[str]]] = Field(default_factory=list) presence_penalty: Optional[float] = 0.0 frequency_penalty: Optional[float] = 0.0 best_of: Optional[int] = None logit_bias: Optional[Dict[str, float]] = None user: Optional[str] = None # Extra parameters for SRT backend only and will be ignored by OpenAI models. regex: Optional[str] = None class CompletionResponseChoice(BaseModel): index: int text: str logprobs: Optional[LogProbs] = None finish_reason: Optional[str] = None class CompletionResponse(BaseModel): id: str object: str = "text_completion" created: int = Field(default_factory=lambda: int(time.time())) model: str choices: List[CompletionResponseChoice] usage: UsageInfo class CompletionResponseStreamChoice(BaseModel): index: int text: str logprobs: Optional[LogProbs] = None finish_reason: Optional[str] = None class CompletionStreamResponse(BaseModel): id: str object: str = "text_completion" created: int = Field(default_factory=lambda: int(time.time())) model: str choices: List[CompletionResponseStreamChoice] usage: UsageInfo async def v1_completions(raw_request: Request): request_json = await raw_request.json() request = CompletionRequest(**request_json) # TODO: Validate the request and return HTTPStatus.BAD_REQUEST if invalid. assert request.n == 1 adapted_request = GenerateReqInput( text=request.prompt, sampling_params={ "temperature": request.temperature, "max_new_tokens": request.max_tokens, "stop": request.stop, "top_p": request.top_p, "presence_penalty": request.presence_penalty, "frequency_penalty": request.frequency_penalty, "regex": request.regex, }, return_logprob=request.logprobs is not None, return_text_in_logprobs=True, stream=request.stream, ) adapted_request.post_init() if adapted_request.stream: async def gnerate_stream_resp(): stream_buffer = "" n_prev_token = 0 async for content in stream_generator(adapted_request): text = content["text"] prompt_tokens = content["meta_info"]["prompt_tokens"] completion_tokens = content["meta_info"]["completion_tokens"] if not stream_buffer: # The first chunk if request.echo: # Prepend prompt in response text. text = request.prompt + text else: # Skip prompt tokens if echo is disabled. n_prev_token = prompt_tokens if request.logprobs is not None: logprobs = await make_openai_style_logprobs( content["meta_info"]["token_logprob"][n_prev_token:] ) n_prev_token = len(content["meta_info"]["token_logprob"]) else: logprobs = None delta = text[len(stream_buffer) :] stream_buffer = content["text"] choice_data = CompletionResponseStreamChoice( index=0, text=delta, logprobs=logprobs, finish_reason=None, ) chunk = CompletionStreamResponse( id=content["meta_info"]["id"], object="text_completion", choices=[choice_data], model=request.model, usage=UsageInfo( prompt_tokens=prompt_tokens, completion_tokens=completion_tokens, total_tokens=prompt_tokens + completion_tokens, ), ) yield f"data: {jsonify_pydantic_model(chunk)}\n\n" yield "data: [DONE]\n\n" return StreamingResponse(gnerate_stream_resp(), media_type="text/event-stream") # Non-streaming response. ret = await generate_request(adapted_request) ret = ret[0] if isinstance(ret, list) else ret prompt_tokens = ret["meta_info"]["prompt_tokens"] completion_tokens = ret["meta_info"]["completion_tokens"] text = ret["text"] token_logprob_pos = prompt_tokens if request.echo: token_logprob_pos = 0 text = request.prompt + text else: token_logprob_pos = prompt_tokens logprobs = ( await make_openai_style_logprobs( ret["meta_info"]["token_logprob"][token_logprob_pos:] ) if request.logprobs is not None else None ) choice_data = CompletionResponseChoice( index=0, text=text, logprobs=logprobs, finish_reason=None, # TODO(comaniac): Add finish reason. ) response = CompletionResponse( id=ret["meta_info"]["id"], model=request.model, choices=[choice_data], usage=UsageInfo( prompt_tokens=prompt_tokens, completion_tokens=completion_tokens, total_tokens=prompt_tokens + completion_tokens, ), ) return response
null
7,175
import asyncio import dataclasses import json import multiprocessing as mp import os import sys import threading import time from typing import List, Optional, Union import aiohttp import psutil import pydantic import requests import uvicorn import uvloop from fastapi import FastAPI, HTTPException, Request from fastapi.responses import Response, StreamingResponse from pydantic import BaseModel from sglang.backend.runtime_endpoint import RuntimeEndpoint from sglang.srt.constrained import disable_cache from sglang.srt.conversation import ( Conversation, SeparatorStyle, chat_template_exists, generate_chat_conv, register_conv_template, ) from sglang.srt.hf_transformers_utils import get_tokenizer from sglang.srt.managers.detokenizer_manager import start_detokenizer_process from sglang.srt.managers.io_struct import DetokenizeReqInput, GenerateReqInput from sglang.srt.managers.openai_protocol import ( ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, DeltaMessage, LogProbs, UsageInfo, ) from sglang.srt.managers.router.manager import start_router_process from sglang.srt.managers.tokenizer_manager import TokenizerManager from sglang.srt.server_args import PortArgs, ServerArgs from sglang.srt.utils import handle_port_init from starlette.middleware.base import BaseHTTPMiddleware from starlette.responses import JSONResponse tokenizer_manager = None chat_template_name = None def jsonify_pydantic_model(obj: BaseModel): if IS_PYDANTIC_1: return obj.json(ensure_ascii=False) return obj.model_dump_json() async def stream_generator(obj: GenerateReqInput): async for out in tokenizer_manager.generate_request(obj): if obj.return_logprob and obj.return_text_in_logprobs: out["meta_info"]["token_logprob"] = await detokenize_logprob_tokens( out["meta_info"]["token_logprob"] ) yield out async def generate_request(obj: GenerateReqInput): obj.post_init() if obj.stream: async def stream_results(): async for out in stream_generator(obj): yield f"data: {json.dumps(out, ensure_ascii=False)}\n\n" yield "data: [DONE]\n\n" return StreamingResponse(stream_results(), media_type="text/event-stream") ret = await tokenizer_manager.generate_request(obj).__anext__() if obj.return_logprob and obj.return_text_in_logprobs: ret["meta_info"]["token_logprob"] = await detokenize_logprob_tokens( ret["meta_info"]["token_logprob"] ) return ret def generate_chat_conv( request: ChatCompletionRequest, template_name: str ) -> Conversation: conv = chat_templates[template_name].copy() conv = Conversation( name=conv.name, system_template=conv.system_template, system_message=conv.system_message, roles=conv.roles, messages=list(conv.messages), # prevent in-place modification offset=conv.offset, sep_style=SeparatorStyle(conv.sep_style), sep=conv.sep, sep2=conv.sep2, stop_str=conv.stop_str, image_data=[], ) if isinstance(request.messages, str): raise ValueError("The messages should be a list of dict.") for message in request.messages: msg_role = message.role if msg_role == "system": conv.system_message = message.content elif msg_role == "user": # Handle the various types of Chat Request content types here. role = conv.roles[0] if isinstance(message.content, str): conv.append_message(conv.roles[0], message.content) else: real_content = "" for content in message.content: if content.type == "text": real_content += content.text elif content.type == "image_url": # NOTE: Only works for llava real_content += "<image>\n" conv.append_image(content.image_url.url) conv.append_message(conv.roles[0], real_content) elif msg_role == "assistant": conv.append_message(conv.roles[1], message.content) else: raise ValueError(f"Unknown role: {msg_role}") # Add a blank message for the assistant. conv.append_message(conv.roles[1], None) return conv class GenerateReqInput: # The input prompt text: Union[List[str], str] # The image input image_data: Optional[Union[List[str], str]] = None # The sampling_params sampling_params: Union[List[Dict], Dict] = None # The request id rid: Optional[Union[List[str], str]] = None # Whether to return logprobs return_logprob: Optional[Union[List[bool], bool]] = None # The start location of the prompt for return_logprob logprob_start_len: Optional[Union[List[int], int]] = None # Whether to detokenize tokens in logprobs return_text_in_logprobs: bool = False # Whether to stream output stream: bool = False def post_init(self): is_single = isinstance(self.text, str) if is_single: if self.sampling_params is None: self.sampling_params = {} if self.rid is None: self.rid = uuid.uuid4().hex if self.return_logprob is None: self.return_logprob = False if self.logprob_start_len is None: self.logprob_start_len = 0 else: num = len(self.text) if self.image_data is None: self.image_data = [None] * num elif not isinstance(self.image_data, list): self.image_data = [self.image_data] * num if self.sampling_params is None: self.sampling_params = [{}] * num elif not isinstance(self.sampling_params, list): self.sampling_params = [self.sampling_params] * num if self.rid is None: self.rid = [uuid.uuid4().hex for _ in range(num)] else: assert isinstance(self.rid, list) if self.return_logprob is None: self.return_logprob = [False] * num elif not isinstance(self.return_logprob, list): self.return_logprob = [self.return_logprob] * num if self.logprob_start_len is None: self.logprob_start_len = [0] * num elif not isinstance(self.logprob_start_len, list): self.logprob_start_len = [self.logprob_start_len] * num class UsageInfo(BaseModel): prompt_tokens: int = 0 total_tokens: int = 0 completion_tokens: Optional[int] = 0 class ChatCompletionRequest(BaseModel): model: str messages: Union[str, List[ChatCompletionMessageParam]] temperature: Optional[float] = 0.7 top_p: Optional[float] = 1.0 n: Optional[int] = 1 max_tokens: Optional[int] = 16 stop: Optional[Union[str, List[str]]] = Field(default_factory=list) stream: Optional[bool] = False presence_penalty: Optional[float] = 0.0 frequency_penalty: Optional[float] = 0.0 logit_bias: Optional[Dict[str, float]] = None user: Optional[str] = None best_of: Optional[int] = None # Extra parameters for SRT backend only and will be ignored by OpenAI models. regex: Optional[str] = None class ChatMessage(BaseModel): role: Optional[str] = None content: Optional[str] = None class ChatCompletionResponseChoice(BaseModel): index: int message: ChatMessage finish_reason: Optional[str] = None class ChatCompletionResponse(BaseModel): id: str object: str = "chat.completion" created: int = Field(default_factory=lambda: int(time.time())) model: str choices: List[ChatCompletionResponseChoice] usage: UsageInfo class DeltaMessage(BaseModel): role: Optional[str] = None content: Optional[str] = None class ChatCompletionResponseStreamChoice(BaseModel): index: int delta: DeltaMessage finish_reason: Optional[str] = None class ChatCompletionStreamResponse(BaseModel): id: str object: str = "chat.completion.chunk" created: int = Field(default_factory=lambda: int(time.time())) model: str choices: List[ChatCompletionResponseStreamChoice] async def v1_chat_completions(raw_request: Request): request_json = await raw_request.json() request = ChatCompletionRequest(**request_json) # TODO: Validate the request and return HTTPStatus.BAD_REQUEST if invalid. assert request.n == 1 # Prep the data needed for the underlying GenerateReqInput: # - prompt: The full prompt string. # - stop: Custom stop tokens. # - image_data: None or a list of image strings (URLs or base64 strings). # None skips any image processing in GenerateReqInput. if not isinstance(request.messages, str): # Apply chat template and its stop strings. if chat_template_name is None: # This flow doesn't support the full OpenAI spec. Verify messages # has the right type before proceeding: for m in request.messages: if not isinstance(m.content, str): raise HTTPException( status_code=503, detail="Structured content requests not supported with " "HuggingFace Chat Templates. " "Make sure the server specifies a sglang chat template.", ) prompt = tokenizer_manager.tokenizer.apply_chat_template( request.messages, tokenize=False, add_generation_prompt=True ) stop = request.stop image_data = None else: conv = generate_chat_conv(request, chat_template_name) prompt = conv.get_prompt() image_data = conv.image_data stop = conv.stop_str or [] if request.stop: if isinstance(request.stop, str): stop.append(request.stop) else: stop.extend(request.stop) else: # Use the raw prompt and stop strings if the messages is already a string. prompt = request.messages stop = request.stop image_data = None adapted_request = GenerateReqInput( text=prompt, image_data=image_data, sampling_params={ "temperature": request.temperature, "max_new_tokens": request.max_tokens, "stop": stop, "top_p": request.top_p, "presence_penalty": request.presence_penalty, "frequency_penalty": request.frequency_penalty, "regex": request.regex, }, stream=request.stream, ) adapted_request.post_init() if adapted_request.stream: async def gnerate_stream_resp(): is_first = True stream_buffer = "" async for content in stream_generator(adapted_request): if is_first: # First chunk with role is_first = False choice_data = ChatCompletionResponseStreamChoice( index=0, delta=DeltaMessage(role="assistant"), finish_reason=None, ) chunk = ChatCompletionStreamResponse( id=content["meta_info"]["id"], choices=[choice_data], model=request.model, ) yield f"data: {jsonify_pydantic_model(chunk)}\n\n" text = content["text"] delta = text[len(stream_buffer) :] stream_buffer = text choice_data = ChatCompletionResponseStreamChoice( index=0, delta=DeltaMessage(content=delta), finish_reason=None ) chunk = ChatCompletionStreamResponse( id=content["meta_info"]["id"], choices=[choice_data], model=request.model, ) yield f"data: {jsonify_pydantic_model(chunk)}\n\n" yield "data: [DONE]\n\n" return StreamingResponse(gnerate_stream_resp(), media_type="text/event-stream") # Non-streaming response. ret = await generate_request(adapted_request) prompt_tokens = ret["meta_info"]["prompt_tokens"] completion_tokens = ret["meta_info"]["completion_tokens"] choice_data = ChatCompletionResponseChoice( index=0, message=ChatMessage(role="assistant", content=ret["text"]), finish_reason=None, # TODO(comaniac): Add finish reason. ) response = ChatCompletionResponse( id=ret["meta_info"]["id"], model=request.model, choices=[choice_data], usage=UsageInfo( prompt_tokens=prompt_tokens, completion_tokens=completion_tokens, total_tokens=prompt_tokens + completion_tokens, ), ) return response
null
7,176
import asyncio import dataclasses import json import multiprocessing as mp import os import sys import threading import time from typing import List, Optional, Union import aiohttp import psutil import pydantic import requests import uvicorn import uvloop from fastapi import FastAPI, HTTPException, Request from fastapi.responses import Response, StreamingResponse from pydantic import BaseModel from sglang.backend.runtime_endpoint import RuntimeEndpoint from sglang.srt.constrained import disable_cache from sglang.srt.conversation import ( Conversation, SeparatorStyle, chat_template_exists, generate_chat_conv, register_conv_template, ) from sglang.srt.hf_transformers_utils import get_tokenizer from sglang.srt.managers.detokenizer_manager import start_detokenizer_process from sglang.srt.managers.io_struct import DetokenizeReqInput, GenerateReqInput from sglang.srt.managers.openai_protocol import ( ChatCompletionRequest, ChatCompletionResponse, ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice, ChatCompletionStreamResponse, ChatMessage, CompletionRequest, CompletionResponse, CompletionResponseChoice, CompletionResponseStreamChoice, CompletionStreamResponse, DeltaMessage, LogProbs, UsageInfo, ) from sglang.srt.managers.router.manager import start_router_process from sglang.srt.managers.tokenizer_manager import TokenizerManager from sglang.srt.server_args import PortArgs, ServerArgs from sglang.srt.utils import handle_port_init from starlette.middleware.base import BaseHTTPMiddleware from starlette.responses import JSONResponse API_KEY_HEADER_NAME = "X-API-Key" class APIKeyValidatorMiddleware(BaseHTTPMiddleware): def __init__(self, app, api_key: str): async def dispatch(self, request: Request, call_next): app = FastAPI() tokenizer_manager = None chat_template_name = None class SeparatorStyle(IntEnum): class Conversation: def get_prompt(self) -> str: def set_system_message(self, system_message: str): def append_message(self, role: str, message: str): def append_image(self, image: str): def update_last_message(self, message: str): def to_gradio_chatbot(self): def to_openai_api_messages(self): def copy(self): def dict(self): def register_conv_template(template: Conversation, override: bool = False): def chat_template_exists(template_name: str) -> bool: register_conv_template( Conversation( name="llama-2", system_template="[INST] <<SYS>>\n{system_message}\n<</SYS>>\n\n", roles=("[INST]", "[/INST]"), sep_style=SeparatorStyle.LLAMA2, sep=" ", sep2=" </s><s>", stop_str=["[INST]", "[/INST]", "<<SYS>>", "<</SYS>>"], ) ) register_conv_template( Conversation( name="chatml", system_template="<|im_start|>system\n{system_message}", system_message="You are an AI assistant.", roles=("<|im_start|>user", "<|im_start|>assistant"), sep_style=SeparatorStyle.CHATML, sep="<|im_end|>", stop_str=["<|endoftext|>", "<|im_end|>"], ) ) register_conv_template( Conversation( name="vicuna_v1.1", system_message="A chat between a curious user and an artificial intelligence assistant. " "The assistant gives helpful, detailed, and polite answers to the user's questions.", roles=("USER", "ASSISTANT"), sep_style=SeparatorStyle.ADD_COLON_TWO, sep=" ", sep2="</s>", ) ) def start_detokenizer_process( server_args: ServerArgs, port_args: PortArgs, pipe_writer, ): def start_router_process( server_args: ServerArgs, port_args: PortArgs, pipe_writer, ): class TokenizerManager: def __init__( self, server_args: ServerArgs, port_args: PortArgs, ): async def get_pixel_values(self, image_data): async def generate_request(self, obj: GenerateReqInput): async def detokenize(self, obj: DetokenizeReqInput): async def flush_cache(self): async def create_handle_loop(self): async def handle_loop(self): class PortArgs: def handle_port_init( port: Optional[int] = None, additional_ports: Optional[List[int]] = None, tp_size: int = 1, ): def launch_server(server_args, pipe_finish_writer): global tokenizer_manager global chat_template_name # disable disk cache if needed if server_args.disable_disk_cache: disable_cache() # Handle ports server_args.port, server_args.additional_ports = handle_port_init( server_args.port, server_args.additional_ports, server_args.tp_size ) port_args = PortArgs( tokenizer_port=server_args.additional_ports[0], router_port=server_args.additional_ports[1], detokenizer_port=server_args.additional_ports[2], nccl_port=server_args.additional_ports[3], model_rpc_ports=server_args.additional_ports[4:], ) # Load chat template if needed if server_args.chat_template is not None: print(f"Use chat template: {server_args.chat_template}") if not chat_template_exists(server_args.chat_template): if not os.path.exists(server_args.chat_template): raise RuntimeError( f"Chat template {server_args.chat_template} is not a built-in template name " "or a valid chat template file path." ) with open(server_args.chat_template, "r") as filep: template = json.load(filep) try: sep_style = SeparatorStyle[template["sep_style"]] except KeyError: raise ValueError( f"Unknown separator style: {template['sep_style']}" ) from None register_conv_template( Conversation( name=template["name"], system_template=template["system"] + "\n{system_message}", system_message=template.get("system_message", ""), roles=(template["user"], template["assistant"]), sep_style=sep_style, sep=template.get("sep", "\n"), stop_str=template["stop_str"], ), override=True, ) chat_template_name = template["name"] else: chat_template_name = server_args.chat_template # Launch processes tokenizer_manager = TokenizerManager(server_args, port_args) pipe_router_reader, pipe_router_writer = mp.Pipe(duplex=False) pipe_detoken_reader, pipe_detoken_writer = mp.Pipe(duplex=False) proc_router = mp.Process( target=start_router_process, args=( server_args, port_args, pipe_router_writer, ), ) proc_router.start() proc_detoken = mp.Process( target=start_detokenizer_process, args=( server_args, port_args, pipe_detoken_writer, ), ) proc_detoken.start() # Wait for the model to finish loading router_init_state = pipe_router_reader.recv() detoken_init_state = pipe_detoken_reader.recv() if router_init_state != "init ok" or detoken_init_state != "init ok": proc_router.kill() proc_detoken.kill() print("router init state:", router_init_state) print("detoken init state:", detoken_init_state) sys.exit(1) assert proc_router.is_alive() and proc_detoken.is_alive() if server_args.api_key and server_args.api_key != "": app.add_middleware(APIKeyValidatorMiddleware, api_key=server_args.api_key) def _launch_server(): uvicorn.run( app, host=server_args.host, port=server_args.port, log_level=server_args.log_level, timeout_keep_alive=5, loop="uvloop", ) def _wait_and_warmup(): headers = {} url = server_args.url() if server_args.api_key and server_args.api_key != "": headers[API_KEY_HEADER_NAME] = server_args.api_key for _ in range(120): time.sleep(0.5) try: requests.get(url + "/get_model_info", timeout=5, headers=headers) break except requests.exceptions.RequestException as e: pass else: if pipe_finish_writer is not None: pipe_finish_writer.send(str(e)) else: print(e, flush=True) return # Warmup try: # print("Warmup...", flush=True) res = requests.post( url + "/generate", json={ "text": "Say this is a warmup request.", "sampling_params": { "temperature": 0, "max_new_tokens": 16, }, }, headers=headers, timeout=60, ) # print(f"Warmup done. model response: {res.json()['text']}") # print("=" * 20, "Server is ready", "=" * 20, flush=True) except requests.exceptions.RequestException as e: if pipe_finish_writer is not None: pipe_finish_writer.send(str(e)) else: print(e, flush=True) return if pipe_finish_writer is not None: pipe_finish_writer.send("init ok") t = threading.Thread(target=_wait_and_warmup) t.start() try: _launch_server() finally: t.join()
null
7,177
import ast import base64 import math from io import BytesIO import numpy as np from PIL import Image def select_best_resolution(original_size, possible_resolutions): """ Selects the best resolution from a list of possible resolutions based on the original size. Args: original_size (tuple): The original size of the image in the format (width, height). possible_resolutions (list): A list of possible resolutions in the format [(width1, height1), (width2, height2), ...]. Returns: tuple: The best fit resolution in the format (width, height). """ original_width, original_height = original_size best_fit = None max_effective_resolution = 0 min_wasted_resolution = float("inf") for width, height in possible_resolutions: scale = min(width / original_width, height / original_height) downscaled_width, downscaled_height = int(original_width * scale), int( original_height * scale ) effective_resolution = min( downscaled_width * downscaled_height, original_width * original_height ) wasted_resolution = (width * height) - effective_resolution if effective_resolution > max_effective_resolution or ( effective_resolution == max_effective_resolution and wasted_resolution < min_wasted_resolution ): max_effective_resolution = effective_resolution min_wasted_resolution = wasted_resolution best_fit = (width, height) return best_fit The provided code snippet includes necessary dependencies for implementing the `get_anyres_image_grid_shape` function. Write a Python function `def get_anyres_image_grid_shape(image_size, grid_pinpoints, patch_size)` to solve the following problem: Calculate the shape of the image patch grid after the preprocessing for images of any resolution. Args: image_size (tuple): The size of the input image in the format (width, height). grid_pinpoints (str): A string representation of a list of possible resolutions. patch_size (int): The size of each image patch. Returns: tuple: The shape of the image patch grid in the format (width, height). Here is the function: def get_anyres_image_grid_shape(image_size, grid_pinpoints, patch_size): """ Calculate the shape of the image patch grid after the preprocessing for images of any resolution. Args: image_size (tuple): The size of the input image in the format (width, height). grid_pinpoints (str): A string representation of a list of possible resolutions. patch_size (int): The size of each image patch. Returns: tuple: The shape of the image patch grid in the format (width, height). """ if type(grid_pinpoints) is list: possible_resolutions = grid_pinpoints else: possible_resolutions = ast.literal_eval(grid_pinpoints) width, height = select_best_resolution(image_size, possible_resolutions) return width // patch_size, height // patch_size
Calculate the shape of the image patch grid after the preprocessing for images of any resolution. Args: image_size (tuple): The size of the input image in the format (width, height). grid_pinpoints (str): A string representation of a list of possible resolutions. patch_size (int): The size of each image patch. Returns: tuple: The shape of the image patch grid in the format (width, height).
7,178
import ast import base64 import math from io import BytesIO import numpy as np from PIL import Image def load_image_from_base64(image): return Image.open(BytesIO(base64.b64decode(image)))
null
7,179
import ast import base64 import math from io import BytesIO import numpy as np from PIL import Image The provided code snippet includes necessary dependencies for implementing the `unpad_image` function. Write a Python function `def unpad_image(tensor, original_size)` to solve the following problem: Unpads a PyTorch tensor of a padded and resized image. Args: tensor (torch.Tensor): The image tensor, assumed to be in CxHxW format. original_size (tuple): The original size of the image (height, width). Returns: torch.Tensor: The unpadded image tensor. Here is the function: def unpad_image(tensor, original_size): """ Unpads a PyTorch tensor of a padded and resized image. Args: tensor (torch.Tensor): The image tensor, assumed to be in CxHxW format. original_size (tuple): The original size of the image (height, width). Returns: torch.Tensor: The unpadded image tensor. """ original_width, original_height = original_size current_height, current_width = tensor.shape[1:] original_aspect_ratio = original_width / original_height current_aspect_ratio = current_width / current_height if original_aspect_ratio > current_aspect_ratio: scale_factor = current_width / original_width new_height = int(original_height * scale_factor) padding = (current_height - new_height) // 2 unpadded_tensor = tensor[:, padding : current_height - padding, :] else: scale_factor = current_height / original_height new_width = int(original_width * scale_factor) padding = (current_width - new_width) // 2 unpadded_tensor = tensor[:, :, padding : current_width - padding] return unpadded_tensor
Unpads a PyTorch tensor of a padded and resized image. Args: tensor (torch.Tensor): The image tensor, assumed to be in CxHxW format. original_size (tuple): The original size of the image (height, width). Returns: torch.Tensor: The unpadded image tensor.
7,180
import ast import base64 import math from io import BytesIO import numpy as np from PIL import Image The provided code snippet includes necessary dependencies for implementing the `unpad_image_shape` function. Write a Python function `def unpad_image_shape(current_height, current_width, original_size)` to solve the following problem: Unpads a PyTorch tensor of a padded and resized image and returns the new shape. Here is the function: def unpad_image_shape(current_height, current_width, original_size): """ Unpads a PyTorch tensor of a padded and resized image and returns the new shape. """ original_width, original_height = original_size original_aspect_ratio = original_width / original_height current_aspect_ratio = current_width / current_height if original_aspect_ratio > current_aspect_ratio: scale_factor = current_width / original_width new_height = int(original_height * scale_factor) padding = (current_height - new_height) // 2 new_shape = (current_height - 2 * padding, current_width) else: scale_factor = current_height / original_height new_width = int(original_width * scale_factor) padding = (current_width - new_width) // 2 new_shape = (current_height, current_width - 2 * padding) return new_shape
Unpads a PyTorch tensor of a padded and resized image and returns the new shape.
7,181
import ast import base64 import math from io import BytesIO import numpy as np from PIL import Image def process_anyres_image(image, processor, grid_pinpoints): def expand2square(pil_img, background_color): def process_images(images, image_processor, model_cfg): image_aspect_ratio = getattr(model_cfg, "image_aspect_ratio", None) new_images = [] if image_aspect_ratio == "pad": for image in images: image = expand2square( image, tuple(int(x * 255) for x in image_processor.image_mean) ) image = image_processor.preprocess(image)["pixel_values"][0] new_images.append(image) elif image_aspect_ratio == "anyres": for image in images: image = process_anyres_image( image, image_processor, model_cfg.image_grid_pinpoints ) new_images.append(image) else: return image_processor(images)["pixel_values"] if all(x.shape == new_images[0].shape for x in new_images): new_images = np.stack(new_images, axis=0) return new_images
null
7,182
from typing import List, Optional import numpy as np import torch from sglang.srt.managers.router.infer_batch import ForwardMode from sglang.srt.managers.router.model_runner import InputMetadata from sglang.srt.mm_utils import ( get_anyres_image_grid_shape, unpad_image, unpad_image_shape, ) from sglang.srt.models.llama2 import LlamaForCausalLM from torch import nn from transformers import CLIPVisionModel, LlamaConfig, LlavaConfig from transformers.models.llava.modeling_llava import LlavaMultiModalProjector from vllm.model_executor.layers.linear import LinearMethodBase from vllm.model_executor.weight_utils import ( default_weight_loader, hf_model_weights_iterator, ) def clip_vision_embed_forward(self, pixel_values: torch.FloatTensor) -> torch.Tensor: batch_size = pixel_values.shape[0] # Move this conv layer to CPU to avoid a bug in torch >= 2.1 on A10G. global first_call if first_call: self.patch_embedding.cpu().float() first_call = False pixel_values = pixel_values.to(dtype=torch.float32, device="cpu") patch_embeds = self.patch_embedding(pixel_values).cuda().half() patch_embeds = patch_embeds.flatten(2).transpose(1, 2) class_embeds = self.class_embedding.expand(batch_size, 1, -1) embeddings = torch.cat([class_embeds, patch_embeds], dim=1) embeddings = embeddings + self.position_embedding(self.position_ids) return embeddings def monkey_path_clip_vision_embed_forward(): import transformers setattr( transformers.models.clip.modeling_clip.CLIPVisionEmbeddings, "forward", clip_vision_embed_forward, )
null
7,183
import asyncio import concurrent.futures import dataclasses import multiprocessing as mp import os from typing import List import numpy as np import transformers import uvloop import zmq import zmq.asyncio from sglang.srt.hf_transformers_utils import ( get_config, get_context_length, get_processor, get_tokenizer, ) from sglang.srt.managers.io_struct import ( BatchStrOut, DetokenizeReqInput, FlushCacheReq, GenerateReqInput, TokenizedGenerateReqInput, ) from sglang.srt.mm_utils import expand2square, process_anyres_image from sglang.srt.sampling_params import SamplingParams from sglang.srt.server_args import PortArgs, ServerArgs from sglang.srt.utils import get_exception_traceback, is_multimodal_model, load_image global global_processor def get_processor( tokenizer_name: str, *args, tokenizer_mode: str = "auto", trust_remote_code: bool = False, tokenizer_revision: Optional[str] = None, **kwargs, ): processor = AutoProcessor.from_pretrained( tokenizer_name, *args, trust_remote_code=trust_remote_code, tokenizer_revision=tokenizer_revision, **kwargs, ) return processor class ServerArgs: model_path: str tokenizer_path: Optional[str] = None host: str = "127.0.0.1" port: int = 30000 additional_ports: Optional[Union[List[int], int]] = None load_format: str = "auto" tokenizer_mode: str = "auto" chat_template: Optional[str] = None trust_remote_code: bool = True mem_fraction_static: Optional[float] = None max_prefill_num_token: Optional[int] = None context_length: Optional[int] = None tp_size: int = 1 schedule_heuristic: str = "lpm" schedule_conservativeness: float = 1.0 attention_reduce_in_fp32: bool = False random_seed: int = 42 stream_interval: int = 8 disable_log_stats: bool = False log_stats_interval: int = 10 log_level: str = "info" # optional modes disable_radix_cache: bool = False enable_flashinfer: bool = False disable_regex_jump_forward: bool = False disable_disk_cache: bool = False api_key: str = "" def __post_init__(self): if self.tokenizer_path is None: self.tokenizer_path = self.model_path if self.mem_fraction_static is None: if self.tp_size >= 8: self.mem_fraction_static = 0.80 elif self.tp_size >= 4: self.mem_fraction_static = 0.82 elif self.tp_size >= 2: self.mem_fraction_static = 0.85 else: self.mem_fraction_static = 0.90 if isinstance(self.additional_ports, int): self.additional_ports = [self.additional_ports] elif self.additional_ports is None: self.additional_ports = [] def add_cli_args(parser: argparse.ArgumentParser): parser.add_argument( "--model-path", type=str, help="The path of the model weights. This can be a local folder or a Hugging Face repo ID.", required=True, ) parser.add_argument( "--tokenizer-path", type=str, default=ServerArgs.tokenizer_path, help="The path of the tokenizer.", ) parser.add_argument("--host", type=str, default=ServerArgs.host) parser.add_argument("--port", type=int, default=ServerArgs.port) # we want to be able to pass a list of ports parser.add_argument( "--additional-ports", type=int, nargs="*", default=[], help="Additional ports specified for launching server.", ) parser.add_argument( "--load-format", type=str, default=ServerArgs.load_format, choices=["auto", "pt", "safetensors", "npcache", "dummy"], help="The format of the model weights to load. " '"auto" will try to load the weights in the safetensors format ' "and fall back to the pytorch bin format if safetensors format " "is not available. " '"pt" will load the weights in the pytorch bin format. ' '"safetensors" will load the weights in the safetensors format. ' '"npcache" will load the weights in pytorch format and store ' "a numpy cache to speed up the loading. " '"dummy" will initialize the weights with random values, ' "which is mainly for profiling.", ) parser.add_argument( "--tokenizer-mode", type=str, default=ServerArgs.tokenizer_mode, choices=["auto", "slow"], help="Tokenizer mode. 'auto' will use the fast " "tokenizer if available, and 'slow' will " "always use the slow tokenizer.", ) parser.add_argument( "--chat-template", type=str, default=ServerArgs.chat_template, help="The buliltin chat template name or the path of the chat template file. This is only used for OpenAI-compatible API server", ) parser.add_argument( "--trust-remote-code", action="store_true", help="Whether or not to allow for custom models defined on the Hub in their own modeling files.", ) parser.add_argument( "--mem-fraction-static", type=float, default=ServerArgs.mem_fraction_static, help="The fraction of the memory used for static allocation (model weights and KV cache memory pool). Use a smaller value if you see out-of-memory errors.", ) parser.add_argument( "--max-prefill-num-token", type=int, default=ServerArgs.max_prefill_num_token, help="The maximum number of tokens in a prefill batch. The real bound will be the maximum of this value and the model's maximum context length.", ) parser.add_argument( "--context-length", type=int, default=ServerArgs.context_length, help="The model's maximum context length. Use this to reduce the context length to save memory. Defaults to None (will use the value from the model's config.json instead).", ) parser.add_argument( "--tp-size", type=int, default=ServerArgs.tp_size, help="Tensor parallelism degree.", ) parser.add_argument( "--schedule-heuristic", type=str, default=ServerArgs.schedule_heuristic, help="Schudule mode: [lpm, weight, random, fcfs]", ) parser.add_argument( "--schedule-conservativeness", type=float, default=ServerArgs.schedule_conservativeness, help="How conservative the schedule policy is. A larger value means more conservative scheduling. Use a larger value if you see requests being retracted frequently.", ) parser.add_argument( "--random-seed", type=int, default=ServerArgs.random_seed, help="Random seed.", ) parser.add_argument( "--attention-reduce-in-fp32", action="store_true", help="Cast the intermidiate attention results to fp32 to avoid possible crashes related to fp16.", ) parser.add_argument( "--stream-interval", type=int, default=ServerArgs.stream_interval, help="The interval (or buffer size) for streaming in terms of the token length. A smaller value makes streaming smoother, while a larger value makes the throughput higher", ) parser.add_argument( "--log-level", type=str, default=ServerArgs.log_level, help="Log level", ) parser.add_argument( "--disable-log-stats", action="store_true", help="Disable logging throughput stats.", ) parser.add_argument( "--log-stats-interval", type=int, default=ServerArgs.log_stats_interval, help="Log stats interval in second.", ) # optional modes parser.add_argument( "--disable-radix-cache", action="store_true", help="Disable RadixAttention", ) parser.add_argument( "--enable-flashinfer", action="store_true", help="Enable flashinfer inference kernels", ) parser.add_argument( "--disable-regex-jump-forward", action="store_true", help="Disable regex jump-forward", ) parser.add_argument( "--disable-disk-cache", action="store_true", help="Disable disk cache to avoid possible crashes related to file system or high concurrency.", ) parser.add_argument( "--api-key", type=str, default=ServerArgs.api_key, help="Set API Key", ) def from_cli_args(cls, args: argparse.Namespace): attrs = [attr.name for attr in dataclasses.fields(cls)] return cls(**{attr: getattr(args, attr) for attr in attrs}) def url(self): return f"http://{self.host}:{self.port}" def get_optional_modes_logging(self): return ( f"disable_radix_cache={self.disable_radix_cache}, " f"enable_flashinfer={self.enable_flashinfer}, " f"disable_regex_jump_forward={self.disable_regex_jump_forward}, " f"disable_disk_cache={self.disable_disk_cache}, " f"attention_reduce_in_fp32={self.attention_reduce_in_fp32}" ) def init_global_processor(server_args: ServerArgs): global global_processor transformers.logging.set_verbosity_error() global_processor = get_processor( server_args.tokenizer_path, tokenizer_mode=server_args.tokenizer_mode, trust_remote_code=server_args.trust_remote_code, )
null
7,184
import asyncio import concurrent.futures import dataclasses import multiprocessing as mp import os from typing import List import numpy as np import transformers import uvloop import zmq import zmq.asyncio from sglang.srt.hf_transformers_utils import ( get_config, get_context_length, get_processor, get_tokenizer, ) from sglang.srt.managers.io_struct import ( BatchStrOut, DetokenizeReqInput, FlushCacheReq, GenerateReqInput, TokenizedGenerateReqInput, ) from sglang.srt.mm_utils import expand2square, process_anyres_image from sglang.srt.sampling_params import SamplingParams from sglang.srt.server_args import PortArgs, ServerArgs from sglang.srt.utils import get_exception_traceback, is_multimodal_model, load_image global global_processor def process_anyres_image(image, processor, grid_pinpoints): """ Process an image with variable resolutions. Args: image (PIL.Image.Image): The input image to be processed. processor: The image processor object. grid_pinpoints (str): A string representation of a list of possible resolutions. Returns: np.array: An np array containing the processed image patches. """ if type(grid_pinpoints) is list: possible_resolutions = grid_pinpoints else: possible_resolutions = ast.literal_eval(grid_pinpoints) best_resolution = select_best_resolution(image.size, possible_resolutions) image_padded = resize_and_pad_image(image, best_resolution) patches = divide_to_patches(image_padded, processor.crop_size["height"]) image_original_resize = image.resize( (processor.size["shortest_edge"], processor.size["shortest_edge"]) ) image_patches = [image_original_resize] + patches image_patches = [ processor.preprocess(image_patch)["pixel_values"][0] for image_patch in image_patches ] return np.stack(image_patches, axis=0) def expand2square(pil_img, background_color): width, height = pil_img.size if width == height: return pil_img if pil_img.mode == "L": pil_img = pil_img.convert("RGB") if width > height: result = Image.new(pil_img.mode, (width, width), background_color) result.paste(pil_img, (0, (width - height) // 2)) return result else: result = Image.new(pil_img.mode, (height, height), background_color) result.paste(pil_img, ((height - width) // 2, 0)) return result def get_exception_traceback(): etype, value, tb = sys.exc_info() err_str = "".join(traceback.format_exception(etype, value, tb)) return err_str def load_image(image_file): from PIL import Image image = None if image_file.startswith("http://") or image_file.startswith("https://"): timeout = int(os.getenv("REQUEST_TIMEOUT", "3")) response = requests.get(image_file, timeout=timeout) image = Image.open(BytesIO(response.content)) elif image_file.lower().endswith(("png", "jpg", "jpeg", "webp", "gif")): image = Image.open(image_file) elif image_file.startswith("data:"): image_file = image_file.split(",")[1] image = Image.open(BytesIO(base64.b64decode(image_file))) else: image = Image.open(BytesIO(base64.b64decode(image_file))) return image def get_pixel_values( image_data, image_aspect_ratio=None, image_grid_pinpoints=None, processor=None ): try: processor = processor or global_processor image = load_image(image_data) image_hash = hash(image_data) if image_aspect_ratio == "pad": image = expand2square( image, tuple(int(x * 255) for x in processor.image_processor.image_mean) ) pixel_values = processor.image_processor(image)["pixel_values"][0] elif image_aspect_ratio == "anyres": pixel_values = process_anyres_image( image, processor.image_processor, image_grid_pinpoints ) else: pixel_values = processor.image_processor(image)["pixel_values"][0] pixel_values = pixel_values.astype(np.float16) return pixel_values, image_hash, image.size except Exception: print("Exception in TokenizerManager:\n" + get_exception_traceback())
null
7,185
import asyncio import logging import multiprocessing import time import warnings from concurrent.futures import ThreadPoolExecutor from typing import List import numpy as np import rpyc import torch from rpyc.utils.classic import obtain from rpyc.utils.server import ThreadedServer from sglang.srt.constrained.fsm_cache import FSMCache from sglang.srt.constrained.jump_forward import JumpForwardCache from sglang.srt.hf_transformers_utils import get_processor, get_tokenizer from sglang.srt.managers.io_struct import ( BatchTokenIDOut, FlushCacheReq, TokenizedGenerateReqInput, ) from sglang.srt.managers.router.infer_batch import Batch, ForwardMode, Req from sglang.srt.managers.router.model_runner import ModelRunner from sglang.srt.managers.router.radix_cache import RadixCache from sglang.srt.managers.router.scheduler import Scheduler from sglang.srt.model_config import ModelConfig from sglang.srt.server_args import PortArgs, ServerArgs from sglang.srt.utils import ( get_exception_traceback, get_int_token_logit_bias, is_multimodal_model, set_random_seed, ) from vllm.logger import _default_handler as vllm_default_handler def _init_service(port): t = ThreadedServer( ModelRpcServer(), port=port, protocol_config={"allow_pickle": True, "sync_request_timeout": 1800}, ) t.start() def start_model_process(port): proc = multiprocessing.Process(target=_init_service, args=(port,)) proc.start() time.sleep(1) repeat_count = 0 while repeat_count < 20: try: con = rpyc.connect( "localhost", port, config={"allow_pickle": True, "sync_request_timeout": 1800}, ) break except ConnectionRefusedError: time.sleep(1) repeat_count += 1 if repeat_count == 20: raise RuntimeError("init rpc env error!") assert proc.is_alive() return con.root, proc
null
7,186
from dataclasses import dataclass from enum import Enum, auto from typing import List import numpy as np import torch from sglang.srt.managers.router.radix_cache import RadixCache from sglang.srt.memory_pool import ReqToTokenPool, TokenToKVPool def _top_p_top_k(probs: torch.Tensor, top_ps: torch.Tensor, top_ks: torch.Tensor): probs_sort, probs_idx = probs.sort(dim=-1, descending=True) probs_sum = torch.cumsum(probs_sort, dim=-1) probs_sort[(probs_sum - probs_sort) > top_ps] = 0.0 probs_sort[ torch.arange(0, probs.shape[-1], device=probs.device).view(1, -1) >= top_ks ] = 0.0 probs_sort.div_(probs_sort.max(dim=-1, keepdim=True)[0]) return probs_sort, probs_idx
null
7,187
import importlib import logging import inspect from dataclasses import dataclass from functools import lru_cache from pathlib import Path import importlib.resources import numpy as np import torch from sglang.srt.managers.router.infer_batch import Batch, ForwardMode from sglang.srt.memory_pool import ReqToTokenPool, TokenToKVPool from sglang.srt.utils import is_multimodal_model from sglang.utils import get_available_gpu_memory from vllm.model_executor.layers.quantization.awq import AWQConfig from vllm.model_executor.layers.quantization.gptq import GPTQConfig from vllm.model_executor.layers.quantization.marlin import MarlinConfig from vllm.model_executor.model_loader import _set_default_torch_dtype from vllm.model_executor.parallel_utils.parallel_state import initialize_model_parallel import sglang def import_model_classes(): model_arch_name_to_cls = {} for f in importlib.resources.files("sglang.srt.models").iterdir(): if f.name.endswith(".py"): module_name = Path(f.name).with_suffix('') module = importlib.import_module(f"sglang.srt.models.{module_name}") if hasattr(module, "EntryClass"): model_arch_name_to_cls[module.EntryClass.__name__] = module.EntryClass return model_arch_name_to_cls def get_model_cls_by_arch_name(model_arch_names): model_arch_name_to_cls = import_model_classes() model_class = None for arch in model_arch_names: if arch in model_arch_name_to_cls: model_class = model_arch_name_to_cls[arch] break else: raise ValueError( f"Unsupported architectures: {arch}. " f"Supported list: {list(model_arch_name_to_cls.keys())}" ) return model_class
null
7,188
import heapq import time from collections import defaultdict from dataclasses import dataclass from typing import Tuple import torch def match(key, seq): i = 0 for k, w in zip(key, seq): if k != w: break i += 1 return i
null
7,189
import asyncio import sglang as sgl def multi_turn_question(s, question_1, question_2): s += sgl.system("You are a helpful assistant.") s += sgl.user(question_1) s += sgl.assistant(sgl.gen("answer_1", max_tokens=256)) s += sgl.user(question_2) s += sgl.assistant(sgl.gen("answer_2", max_tokens=256)) def stream_a_variable(): state = multi_turn_question.run( question_1="What is the capital of the United States?", question_2="List two local attractions.", stream=True ) for out in state.text_iter(var_name="answer_2"): print(out, end="", flush=True) print("\n")
null
7,190
import asyncio import sglang as sgl def multi_turn_question(s, question_1, question_2): s += sgl.system("You are a helpful assistant.") s += sgl.user(question_1) s += sgl.assistant(sgl.gen("answer_1", max_tokens=256)) s += sgl.user(question_2) s += sgl.assistant(sgl.gen("answer_2", max_tokens=256)) async def async_stream(): state = multi_turn_question.run( question_1="What is the capital of the United States?", question_2="List two local attractions.", stream=True ) async for out in state.text_async_iter(var_name="answer_2"): print(out, end="", flush=True) print("\n")
null
7,191
import sglang as sgl sgl.set_default_backend(sgl.OpenAI("gpt-3.5-turbo-instruct")) def parallel_sample(s, question, n): s += ( "Question: Compute 1 + 2 + 3\n" "Reasoning: I need to use a calculator.\n" "Tool: calculator\n" "Answer: 6\n" "Question: Compute 3 + 2 + 2\n" "Reasoning: I will try a calculator.\n" "Tool: calculator\n" "Answer: 7\n" ) s += "Question: " + question + "\n" forks = s.fork(n) forks += "Reasoning:" + sgl.gen("reasoning", stop="\n") + "\n" forks += "Tool:" + sgl.gen("tool", choices=["calculator", "browser"]) + "\n" forks += "Answer:" + sgl.gen("answer", stop="\n") + "\n" forks.join()
null
7,192
import sglang as sgl def tool_use(s, question): s += "To answer this question: " + question + ", " s += "I need to use a " + sgl.gen("tool", choices=["calculator", "search engine"])
null
7,193
import triton_python_backend_utils as pb_utils import numpy import sglang as sgl from sglang import function, set_default_backend from sglang.srt.constrained import build_regex_from_object from pydantic import BaseModel sgl.set_default_backend(sgl.RuntimeEndpoint("http://localhost:30000")) class Character(BaseModel): name: str eye_color: str house: str def character_gen(s, name): s += ( name + " is a character in Harry Potter. Please fill in the following information about this character.\n" ) s += sgl.gen("json_output", max_tokens=256, regex=build_regex_from_object(Character))
null
7,194
import sglang as sgl def tool_use(s, question): s += "To answer this question: " + question + ". " s += "I need to use a " + sgl.gen("tool", choices=["calculator", "search engine"]) + ". " if s["tool"] == "calculator": s += "The math expression is" + sgl.gen("expression") elif s["tool"] == "search engine": s += "The key word to search is" + sgl.gen("word") def driver_tool_use(): state = tool_use.run(question="What is the capital of the United States?") print(state.text()) print("\n")
null
7,195
import sglang as sgl def tip_suggestion(s): s += ( "Here are two tips for staying healthy: " "1. Balanced Diet. 2. Regular Exercise.\n\n" ) forks = s.fork(2) for i, f in enumerate(forks): f += f"Now, expand tip {i+1} into a paragraph:\n" f += sgl.gen(f"detailed_tip", max_tokens=256, stop="\n\n") s += "Tip 1:" + forks[0]["detailed_tip"] + "\n" s += "Tip 2:" + forks[1]["detailed_tip"] + "\n" s += "In summary" + sgl.gen("summary") def driver_tip_suggestion(): state = tip_suggestion.run() print(state.text()) print("\n")
null
7,196
import sglang as sgl def regular_expression_gen(s): s += "Q: What is the IP address of the Google DNS servers?\n" s += "A: " + sgl.gen( "answer", temperature=0, regex=r"((25[0-5]|2[0-4]\d|[01]?\d\d?).){3}(25[0-5]|2[0-4]\d|[01]?\d\d?)", ) def driver_regex(): state = regular_expression_gen.run() print(state.text()) print("\n")
null
7,197
import sglang as sgl def text_qa(s, question): s += "Q: " + question + "\n" s += "A:" + sgl.gen("answer", stop="\n") def driver_batching(): states = text_qa.run_batch( [ {"question": "What is the capital of the United Kingdom?"}, {"question": "What is the capital of France?"}, {"question": "What is the capital of Japan?"}, ], progress_bar=True ) for s in states: print(s.text()) print("\n")
null
7,198
import sglang as sgl def text_qa(s, question): s += "Q: " + question + "\n" s += "A:" + sgl.gen("answer", stop="\n") def driver_stream(): state = text_qa.run( question="What is the capital of France?", temperature=0.1, stream=True ) for out in state.text_iter(): print(out, end="", flush=True) print("\n")
null
7,199
from sglang import function, gen, set_default_backend, OpenAI def gen_character_spec(s): s += "Construct a character within the following format:\n" s += "Name: Steve Jobs.\nBirthday: February 24, 1955.\nJob: Apple CEO.\n" s += "\nPlease generate new Name, Birthday and Job.\n" s += "Name:" + gen("name", stop="\n") + "\nBirthday:" + gen("birthday", stop="\n") s += "\nJob:" + gen("job", stop="\n") + "\n"
null
7,200
import asyncio from sglang import Runtime async def generate( engine, prompt, sampling_params, ): tokenizer = engine.get_tokenizer() messages = [ {"role": "system", "content": "You will be given question answer tasks.",}, {"role": "user", "content": prompt}, ] prompt = tokenizer.apply_chat_template( messages, tokenize=False, add_generation_prompt=True ) stream = engine.add_request(prompt, sampling_params) async for output in stream: print(output, end="", flush=True) print()
null
7,201
from enum import Enum import sglang as sgl from pydantic import BaseModel from sglang.srt.constrained import build_regex_from_object def character_gen(s, name): s += ( name + " is a character in Harry Potter. Please fill in the following information about this character.\n" ) s += sgl.gen("json_output", max_tokens=256, regex=character_regex) def driver_character_gen(): state = character_gen.run(name="Hermione Granger") print(state.text())
null
7,202
from enum import Enum import sglang as sgl from pydantic import BaseModel from sglang.srt.constrained import build_regex_from_object def pydantic_wizard_gen(s): def driver_pydantic_wizard_gen(): state = pydantic_wizard_gen.run() print(state.text())
null
7,203
import sglang as sgl def image_qa(s, image_file1, image_file2, question): s += sgl.user(sgl.image(image_file1) + sgl.image(image_file2) + question) s += sgl.assistant(sgl.gen("answer", max_tokens=256))
null
7,204
import sglang as sgl import os def multi_turn_question(s, question_1, question_2): def single(): state = multi_turn_question.run( question_1="What is the capital of the United States?", question_2="List two local attractions.", ) for m in state.messages(): print(m["role"], ":", m["content"]) print("\n-- answer_1 --\n", state["answer_1"])
null
7,205
import sglang as sgl import os def multi_turn_question(s, question_1, question_2): def stream(): state = multi_turn_question.run( question_1="What is the capital of the United States?", question_2="List two local attractions.", stream=True ) for out in state.text_iter(): print(out, end="", flush=True) print()
null
7,206
import sglang as sgl import os def multi_turn_question(s, question_1, question_2): s += sgl.system("You are a helpful assistant.") s += sgl.user(question_1) s += sgl.assistant(sgl.gen("answer_1", max_tokens=256)) s += sgl.user(question_2) s += sgl.assistant(sgl.gen("answer_2", max_tokens=256)) def batch(): states = multi_turn_question.run_batch([ {"question_1": "What is the capital of the United States?", "question_2": "List two local attractions."}, {"question_1": "What is the capital of France?", "question_2": "What is the population of this city?"}, ]) for s in states: print(s.messages())
null
7,207
import sglang as sgl import os def multi_turn_question(s, question_1, question_2): s += sgl.system("You are a helpful assistant.") s += sgl.user(question_1) s += sgl.assistant(sgl.gen("answer_1", max_tokens=256)) s += sgl.user(question_2) s += sgl.assistant(sgl.gen("answer_2", max_tokens=256)) def single(): state = multi_turn_question.run( question_1="What is the capital of the United States?", question_2="List two local attractions.", ) for m in state.messages(): print(m["role"], ":", m["content"]) print("\n-- answer_1 --\n", state["answer_1"])
null
7,208
import sglang as sgl import os def multi_turn_question(s, question_1, question_2): s += sgl.system("You are a helpful assistant.") s += sgl.user(question_1) s += sgl.assistant(sgl.gen("answer_1", max_tokens=256)) s += sgl.user(question_2) s += sgl.assistant(sgl.gen("answer_2", max_tokens=256)) def stream(): state = multi_turn_question.run( question_1="What is the capital of the United States?", question_2="List two local attractions.", stream=True ) for out in state.text_iter(): print(out, end="", flush=True) print()
null
7,210
import sglang as sgl def few_shot_qa(s, question): s += ( """The following are questions with answers. Q: What is the capital of France? A: Paris Q: What is the capital of Germany? A: Berlin Q: What is the capital of Italy? A: Rome """) s += "Q: " + question + "\n" s += "A:" + sgl.gen("answer", stop="\n", temperature=0) def single(): state = few_shot_qa.run(question="What is the capital of the United States?") answer = state["answer"].strip().lower() assert "washington" in answer, f"answer: {state['answer']}" print(state.text())
null
7,211
import sglang as sgl def few_shot_qa(s, question): s += ( """The following are questions with answers. Q: What is the capital of France? A: Paris Q: What is the capital of Germany? A: Berlin Q: What is the capital of Italy? A: Rome """) s += "Q: " + question + "\n" s += "A:" + sgl.gen("answer", stop="\n", temperature=0) def stream(): state = few_shot_qa.run( question="What is the capital of the United States?", stream=True) for out in state.text_iter("answer"): print(out, end="", flush=True) print()
null
7,212
import sglang as sgl def few_shot_qa(s, question): def batch(): states = few_shot_qa.run_batch([ {"question": "What is the capital of the United States?"}, {"question": "What is the capital of China?"}, ]) for s in states: print(s["answer"])
null
7,213
import sglang as sgl def multi_turn_question(s, question_1, question_2): s += sgl.user(question_1) s += sgl.assistant(sgl.gen("answer_1", max_tokens=256)) s += sgl.user(question_2) s += sgl.assistant(sgl.gen("answer_2", max_tokens=256)) def single(): state = multi_turn_question.run( question_1="What is the capital of the United States?", question_2="List two local attractions.", ) for m in state.messages(): print(m["role"], ":", m["content"]) print("\n-- answer_1 --\n", state["answer_1"])
null
7,214
import sglang as sgl def multi_turn_question(s, question_1, question_2): s += sgl.user(question_1) s += sgl.assistant(sgl.gen("answer_1", max_tokens=256)) s += sgl.user(question_2) s += sgl.assistant(sgl.gen("answer_2", max_tokens=256)) def stream(): state = multi_turn_question.run( question_1="What is the capital of the United States?", question_2="List two local attractions.", stream=True ) for out in state.text_iter(): print(out, end="", flush=True) print()
null
7,215
import sglang as sgl def multi_turn_question(s, question_1, question_2): s += sgl.user(question_1) s += sgl.assistant(sgl.gen("answer_1", max_tokens=256)) s += sgl.user(question_2) s += sgl.assistant(sgl.gen("answer_2", max_tokens=256)) def batch(): states = multi_turn_question.run_batch([ {"question_1": "What is the capital of the United States?", "question_2": "List two local attractions."}, {"question_1": "What is the capital of France?", "question_2": "What is the population of this city?"}, ]) for s in states: print(s.messages())
null
7,219
import sglang as sgl def multi_turn_question(s, question_1, question_2): s += sgl.system("You are a helpful assistant.") s += sgl.user(question_1) s += sgl.assistant(sgl.gen("answer_1", max_tokens=256)) s += sgl.user(question_2) s += sgl.assistant(sgl.gen("answer_2", max_tokens=256)) def single(): state = multi_turn_question.run( question_1="What is the capital of the United States?", question_2="List two local attractions.", ) for m in state.messages(): print(m["role"], ":", m["content"]) print("\n-- answer_1 --\n", state["answer_1"])
null
7,220
import sglang as sgl def multi_turn_question(s, question_1, question_2): s += sgl.system("You are a helpful assistant.") s += sgl.user(question_1) s += sgl.assistant(sgl.gen("answer_1", max_tokens=256)) s += sgl.user(question_2) s += sgl.assistant(sgl.gen("answer_2", max_tokens=256)) def stream(): state = multi_turn_question.run( question_1="What is the capital of the United States?", question_2="List two local attractions.", stream=True ) for out in state.text_iter(): print(out, end="", flush=True) print()
null
7,221
import sglang as sgl def multi_turn_question(s, question_1, question_2): def batch(): states = multi_turn_question.run_batch([ {"question_1": "What is the capital of the United States?", "question_2": "List two local attractions."}, {"question_1": "What is the capital of France?", "question_2": "What is the population of this city?"}, ]) for s in states: print(s.messages())
null
7,222
import sglang as sgl import os def few_shot_qa(s, question): s += ( """The following are questions with answers. Q: What is the capital of France? A: Paris Q: What is the capital of Germany? A: Berlin Q: What is the capital of Italy? A: Rome """) s += "Q: " + question + "\n" s += "A:" + sgl.gen("answer", stop="\n", temperature=0) def single(): state = few_shot_qa.run(question="What is the capital of the United States?") answer = state["answer"].strip().lower() assert "washington" in answer, f"answer: {state['answer']}" print(state.text())
null
7,223
import sglang as sgl import os def few_shot_qa(s, question): s += ( """The following are questions with answers. Q: What is the capital of France? A: Paris Q: What is the capital of Germany? A: Berlin Q: What is the capital of Italy? A: Rome """) s += "Q: " + question + "\n" s += "A:" + sgl.gen("answer", stop="\n", temperature=0) def stream(): state = few_shot_qa.run( question="What is the capital of the United States?", stream=True) for out in state.text_iter("answer"): print(out, end="", flush=True) print()
null
7,224
import sglang as sgl import os def few_shot_qa(s, question): s += ( """The following are questions with answers. Q: What is the capital of France? A: Paris Q: What is the capital of Germany? A: Berlin Q: What is the capital of Italy? A: Rome """) s += "Q: " + question + "\n" s += "A:" + sgl.gen("answer", stop="\n", temperature=0) def batch(): states = few_shot_qa.run_batch([ {"question": "What is the capital of the United States?"}, {"question": "What is the capital of China?"}, ]) for s in states: print(s["answer"])
null
7,225
import sglang as sgl def image_qa(s, image_path, question): def single(): state = image_qa.run( image_path="images/cat.jpeg", question="What is this?", max_new_tokens=64) print(state["answer"], "\n")
null
7,226
import sglang as sgl def image_qa(s, image_path, question): s += sgl.user(sgl.image(image_path) + question) s += sgl.assistant(sgl.gen("answer")) def stream(): state = image_qa.run( image_path="images/cat.jpeg", question="What is this?", max_new_tokens=64, stream=True) for out in state.text_iter("answer"): print(out, end="", flush=True) print()
null
7,227
import sglang as sgl def image_qa(s, image_path, question): s += sgl.user(sgl.image(image_path) + question) s += sgl.assistant(sgl.gen("answer")) def batch(): states = image_qa.run_batch( [ {"image_path": "images/cat.jpeg", "question":"What is this?"}, {"image_path": "images/dog.jpeg", "question":"What is this?"}, ], max_new_tokens=64, ) for s in states: print(s["answer"], "\n")
null
7,233
import sglang as sgl def few_shot_qa(s, question): s += ( """The following are questions with answers. Q: What is the capital of France? A: Paris Q: What is the capital of Germany? A: Berlin Q: What is the capital of Italy? A: Rome """) s += "Q: " + question + "\n" s += "A:" + sgl.gen("answer", stop="\n", temperature=0) def batch(): states = few_shot_qa.run_batch([ {"question": "What is the capital of the United States?"}, {"question": "What is the capital of China?"}, ]) for s in states: print(s["answer"])
null
7,234
import sglang as sgl def image_qa(s, image_path, question): s += sgl.user(sgl.image(image_path) + question) s += sgl.assistant(sgl.gen("answer")) def single(): state = image_qa.run( image_path="images/cat.jpeg", question="What is this?", max_new_tokens=64, stop="###") print(state["answer"], "\n")
null
7,235
import sglang as sgl def image_qa(s, image_path, question): s += sgl.user(sgl.image(image_path) + question) s += sgl.assistant(sgl.gen("answer")) def stream(): state = image_qa.run( image_path="images/cat.jpeg", question="What is this?", max_new_tokens=64, stream=True, stop="###") for out in state.text_iter("answer"): print(out, end="", flush=True) print()
null
7,236
import sglang as sgl def image_qa(s, image_path, question): s += sgl.user(sgl.image(image_path) + question) s += sgl.assistant(sgl.gen("answer")) def batch(): states = image_qa.run_batch( [ {"image_path": "images/cat.jpeg", "question":"What is this?"}, {"image_path": "images/dog.jpeg", "question":"What is this?"}, ], max_new_tokens=64, stop="###" ) for s in states: print(s["answer"], "\n")
null
7,237
import sglang as sgl def few_shot_qa(s, question): s += ( """ \n\nHuman: What is the capital of France? \n\nAssistant: Paris \n\nHuman: What is the capital of Germany? \n\nAssistant: Berlin \n\nHuman: What is the capital of Italy? \n\nAssistant: Rome """) s += "\n\nHuman: " + question + "\n" s += "\n\nAssistant:" + sgl.gen("answer", stop="\n", temperature=0) def single(): state = few_shot_qa.run(question="What is the capital of the United States?") answer = state["answer"].strip().lower() assert "washington" in answer, f"answer: {state['answer']}" print(state.text())
null
7,238
import sglang as sgl def few_shot_qa(s, question): s += ( """ \n\nHuman: What is the capital of France? \n\nAssistant: Paris \n\nHuman: What is the capital of Germany? \n\nAssistant: Berlin \n\nHuman: What is the capital of Italy? \n\nAssistant: Rome """) s += "\n\nHuman: " + question + "\n" s += "\n\nAssistant:" + sgl.gen("answer", stop="\n", temperature=0) def stream(): state = few_shot_qa.run( question="What is the capital of the United States?", stream=True) for out in state.text_iter("answer"): print(out, end="", flush=True) print()
null
7,239
import sglang as sgl def few_shot_qa(s, question): s += ( """ \n\nHuman: What is the capital of France? \n\nAssistant: Paris \n\nHuman: What is the capital of Germany? \n\nAssistant: Berlin \n\nHuman: What is the capital of Italy? \n\nAssistant: Rome """) s += "\n\nHuman: " + question + "\n" s += "\n\nAssistant:" + sgl.gen("answer", stop="\n", temperature=0) def batch(): states = few_shot_qa.run_batch([ {"question": "What is the capital of the United States?"}, {"question": "What is the capital of China?"}, ]) for s in states: print(s["answer"])
null