| |
|
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from copy import deepcopy |
|
|
| from google.cloud.aiplatform import base |
| from google.cloud.aiplatform import initializer as aiplatform_initializer |
| from google.cloud.aiplatform.compat.services import dataset_service_client |
| from vertexai.generative_models import ( |
| Content, |
| Image, |
| Part, |
| GenerativeModel, |
| GenerationConfig, |
| SafetySetting, |
| Tool, |
| ToolConfig, |
| ) |
| from vertexai.generative_models._generative_models import ( |
| _to_content, |
| _validate_generate_content_parameters, |
| _reconcile_model_name, |
| _get_resource_name_from_model_name, |
| ContentsType, |
| GenerationConfigType, |
| GenerationResponse, |
| PartsType, |
| SafetySettingsType, |
| ) |
|
|
| import re |
| from typing import ( |
| Any, |
| Dict, |
| Iterable, |
| List, |
| Optional, |
| Union, |
| ) |
|
|
| _LOGGER = base.Logger(__name__) |
|
|
| DEFAULT_MODEL_NAME = "gemini-1.5-flash-002" |
| VARIABLE_NAME_REGEX = r"(\{[^\W0-9]\w*\})" |
|
|
|
|
| class Prompt: |
| """A prompt which may be a template with variables. |
| |
| The `Prompt` class allows users to define a template string with |
| variables represented in curly braces `{variable}`. The variable |
| name must be a valid Python variable name (no spaces, must start with a |
| letter). These placeholders can be replaced with specific values using the |
| `assemble_contents` method, providing flexibility in generating dynamic prompts. |
| |
| Usage: |
| Generate content from a single set of variables: |
| ``` |
| prompt = Prompt( |
| prompt_data="Hello, {name}! Today is {day}. How are you?", |
| variables=[{"name": "Alice", "day": "Monday"}] |
| generation_config=GenerationConfig( |
| temperature=0.1, |
| top_p=0.95, |
| top_k=20, |
| candidate_count=1, |
| max_output_tokens=100, |
| ), |
| model_name="gemini-1.0-pro-002", |
| safety_settings=[SafetySetting( |
| category=SafetySetting.HarmCategory.HARM_CATEGORY_DANGEROUS_CONTENT, |
| threshold=SafetySetting.HarmBlockThreshold.BLOCK_MEDIUM_AND_ABOVE, |
| method=SafetySetting.HarmBlockMethod.SEVERITY, |
| )], |
| system_instruction="Please answer in a short sentence.", |
| ) |
| |
| # Generate content using the assembled prompt. |
| prompt.generate_content( |
| contents=prompt.assemble_contents(**prompt.variables) |
| ) |
| ``` |
| Generate content with multiple sets of variables: |
| ``` |
| prompt = Prompt( |
| prompt_data="Hello, {name}! Today is {day}. How are you?", |
| variables=[ |
| {"name": "Alice", "day": "Monday"}, |
| {"name": "Bob", "day": "Tuesday"}, |
| ], |
| generation_config=GenerationConfig( |
| temperature=0.1, |
| top_p=0.95, |
| top_k=20, |
| candidate_count=1, |
| max_output_tokens=100, |
| ), |
| model_name="gemini-1.0-pro-002", |
| safety_settings=[SafetySetting( |
| category=SafetySetting.HarmCategory.HARM_CATEGORY_DANGEROUS_CONTENT, |
| threshold=SafetySetting.HarmBlockThreshold.BLOCK_MEDIUM_AND_ABOVE, |
| method=SafetySetting.HarmBlockMethod.SEVERITY, |
| )], |
| system_instruction="Please answer in a short sentence.", |
| ) |
| |
| # Generate content using the assembled prompt for each variable set. |
| for i in range(len(prompt.variables)): |
| prompt.generate_content( |
| contents=prompt.assemble_contents(**prompt.variables[i]) |
| ) |
| ``` |
| """ |
|
|
| def __init__( |
| self, |
| prompt_data: Optional[PartsType] = None, |
| *, |
| variables: Optional[List[Dict[str, PartsType]]] = None, |
| prompt_name: Optional[str] = None, |
| generation_config: Optional[GenerationConfig] = None, |
| model_name: Optional[str] = None, |
| safety_settings: Optional[SafetySetting] = None, |
| system_instruction: Optional[PartsType] = None, |
| tools: Optional[List[Tool]] = None, |
| tool_config: Optional[ToolConfig] = None, |
| ): |
| """Initializes the Prompt with a given prompt, and variables. |
| |
| Args: |
| prompt: A PartsType prompt which may be a template with variables or a prompt with no variables. |
| variables: A list of dictionaries containing the variable names and values. |
| prompt_name: The display name of the prompt, if stored in an online resource. |
| generation_config: A GenerationConfig object containing parameters for generation. |
| model_name: Model Garden model resource name. |
| Alternatively, a tuned model endpoint resource name can be provided. |
| If no model is provided, the default latest model will be used. |
| safety_settings: A SafetySetting object containing safety settings for generation. |
| system_instruction: A PartsType object representing the system instruction. |
| tools: A list of Tool objects for function calling. |
| tool_config: A ToolConfig object for function calling. |
| """ |
| self._prompt_data = None |
| self._variables = None |
| self._model_name = None |
| self._generation_config = None |
| self._safety_settings = None |
| self._system_instruction = None |
| self._tools = None |
| self._tool_config = None |
|
|
| |
| self._dataset_client_value = None |
| self._dataset = None |
| self._prompt_name = None |
| self._version_id = None |
| self._version_name = None |
|
|
| self.prompt_data = prompt_data |
| self.variables = variables if variables else [{}] |
| self.prompt_name = prompt_name |
| self.model_name = model_name |
| self.generation_config = generation_config |
| self.safety_settings = safety_settings |
| self.system_instruction = system_instruction |
| self.tools = tools |
| self.tool_config = tool_config |
|
|
| @property |
| def prompt_data(self) -> Optional[PartsType]: |
| return self._prompt_data |
|
|
| @property |
| def variables(self) -> Optional[List[Dict[str, PartsType]]]: |
| return self._variables |
|
|
| @property |
| def prompt_name(self) -> Optional[str]: |
| return self._prompt_name |
|
|
| @property |
| def generation_config(self) -> Optional[GenerationConfig]: |
| return self._generation_config |
|
|
| @property |
| def model_name(self) -> Optional[str]: |
| if self._model_name: |
| return self._model_name |
| else: |
| return Prompt._format_model_resource_name(DEFAULT_MODEL_NAME) |
|
|
| @property |
| def safety_settings(self) -> Optional[List[SafetySetting]]: |
| return self._safety_settings |
|
|
| @property |
| def system_instruction(self) -> Optional[PartsType]: |
| return self._system_instruction |
|
|
| @property |
| def tools(self) -> Optional[List[Tool]]: |
| return self._tools |
|
|
| @property |
| def tool_config(self) -> Optional[ToolConfig]: |
| return self._tool_config |
|
|
| @property |
| def prompt_id(self) -> Optional[str]: |
| if self._dataset: |
| return self._dataset.name.split("/")[-1] |
| return None |
|
|
| @property |
| def version_id(self) -> Optional[str]: |
| return self._version_id |
|
|
| @property |
| def version_name(self) -> Optional[str]: |
| return self._version_name |
|
|
| @prompt_data.setter |
| def prompt_data(self, prompt_data: Optional[PartsType]) -> None: |
| """Overwrites the existing saved local prompt_data. |
| |
| Args: |
| prompt_data: A PartsType prompt. |
| """ |
| if prompt_data is not None: |
| self._validate_parts_type_data(prompt_data) |
| self._prompt_data = prompt_data |
|
|
| @variables.setter |
| def variables(self, variables: List[Dict[str, PartsType]]) -> None: |
| """Overwrites the existing saved local variables. |
| |
| Args: |
| variables: A list of dictionaries containing the variable names and values. |
| """ |
| if isinstance(variables, list): |
| for i in range(len(variables)): |
| variables[i] = variables[i].copy() |
| Prompt._format_variable_value_to_parts(variables[i]) |
| self._variables = variables |
| else: |
| raise TypeError( |
| f"Variables must be a list of dictionaries, not {type(variables)}" |
| ) |
|
|
| @prompt_name.setter |
| def prompt_name(self, prompt_name: Optional[str]) -> None: |
| """Overwrites the existing saved local prompt_name.""" |
| if prompt_name: |
| self._prompt_name = prompt_name |
| else: |
| self._prompt_name = None |
|
|
| @model_name.setter |
| def model_name(self, model_name: Optional[str]) -> None: |
| """Overwrites the existing saved local model_name.""" |
| if model_name: |
| self._model_name = Prompt._format_model_resource_name(model_name) |
| else: |
| self._model_name = None |
|
|
| def _format_model_resource_name(model_name: Optional[str]) -> str: |
| """Formats the model resource name.""" |
| project = aiplatform_initializer.global_config.project |
| location = aiplatform_initializer.global_config.location |
| model_name = _reconcile_model_name(model_name, project, location) |
|
|
| prediction_resource_name = _get_resource_name_from_model_name( |
| model_name, project, location |
| ) |
| return prediction_resource_name |
|
|
| def _validate_configs( |
| self, |
| generation_config: Optional[GenerationConfig] = None, |
| safety_settings: Optional[SafetySetting] = None, |
| system_instruction: Optional[PartsType] = None, |
| tools: Optional[List[Tool]] = None, |
| tool_config: Optional[ToolConfig] = None, |
| ): |
| generation_config = generation_config or self._generation_config |
| safety_settings = safety_settings or self._safety_settings |
| tools = tools or self._tools |
| tool_config = tool_config or self._tool_config |
| system_instruction = system_instruction or self._system_instruction |
| return _validate_generate_content_parameters( |
| contents="test", |
| generation_config=generation_config, |
| safety_settings=safety_settings, |
| system_instruction=system_instruction, |
| tools=tools, |
| tool_config=tool_config, |
| ) |
|
|
| @generation_config.setter |
| def generation_config(self, generation_config: Optional[GenerationConfig]) -> None: |
| """Overwrites the existing saved local generation_config. |
| |
| Args: |
| generation_config: A GenerationConfig object containing parameters for generation. |
| """ |
| self._validate_configs(generation_config=generation_config) |
| self._generation_config = generation_config |
|
|
| @safety_settings.setter |
| def safety_settings(self, safety_settings: Optional[SafetySetting]) -> None: |
| """Overwrites the existing saved local safety_settings. |
| |
| Args: |
| safety_settings: A SafetySetting object containing safety settings for generation. |
| """ |
| self._validate_configs(safety_settings=safety_settings) |
| self._safety_settings = safety_settings |
|
|
| @system_instruction.setter |
| def system_instruction(self, system_instruction: Optional[PartsType]) -> None: |
| """Overwrites the existing saved local system_instruction. |
| |
| Args: |
| system_instruction: A PartsType object representing the system instruction. |
| """ |
| if system_instruction: |
| self._validate_parts_type_data(system_instruction) |
| self._system_instruction = system_instruction |
|
|
| @tools.setter |
| def tools(self, tools: Optional[List[Tool]]) -> None: |
| """Overwrites the existing saved local tools. |
| |
| Args: |
| tools: A list of Tool objects for function calling. |
| """ |
| self._validate_configs(tools=tools) |
| self._tools = tools |
|
|
| @tool_config.setter |
| def tool_config(self, tool_config: Optional[ToolConfig] = None) -> None: |
| """Overwrites the existing saved local tool_config. |
| |
| Args: |
| tool_config: A ToolConfig object for function calling. |
| """ |
| self._validate_configs(tool_config=tool_config) |
| self._tool_config = tool_config |
|
|
| def _format_variable_value_to_parts(variables_dict: Dict[str, PartsType]) -> None: |
| """Formats the variables values to be List[Part]. |
| |
| Args: |
| variables_dict: A single dictionary containing the variable names and values. |
| |
| Raises: |
| TypeError: If a variable value is not a PartsType Object. |
| """ |
| for key in variables_dict.keys(): |
| |
| if isinstance(variables_dict[key], Content): |
| raise TypeError( |
| "Variable values must be a PartsType object, not Content" |
| ) |
|
|
| |
| content = Content._from_gapic(_to_content(value=variables_dict[key])) |
| variables_dict[key] = content.parts |
|
|
| def _validate_parts_type_data(self, data: Any) -> None: |
| """ |
| Args: |
| prompt_data: The prompt input to validate |
| |
| Raises: |
| TypeError: If prompt_data is not a PartsType Object. |
| """ |
| |
| if isinstance(data, Content): |
| raise TypeError("Prompt data must be a PartsType object, not Content") |
|
|
| |
| _to_content(value=data) |
|
|
| def assemble_contents(self, **variables_dict: PartsType) -> List[Content]: |
| """Returns the prompt data, as a List[Content], assembled with variables if applicable. |
| Can be ingested into model.generate_content to make API calls. |
| |
| Returns: |
| A List[Content] prompt. |
| Usage: |
| ``` |
| prompt = Prompt( |
| prompt_data="Hello, {name}! Today is {day}. How are you?", |
| ) |
| |
| model.generate_content( |
| contents=prompt.assemble_contents(name="Alice", day="Monday") |
| ) |
| ``` |
| """ |
| |
| if self.prompt_data is None: |
| raise ValueError("prompt_data must not be empty.") |
|
|
| variables_dict = variables_dict.copy() |
|
|
| |
| if not variables_dict: |
| return [Content._from_gapic(_to_content(value=self.prompt_data))] |
|
|
| |
| Prompt._format_variable_value_to_parts(variables_dict) |
|
|
| |
| |
| assembled_prompt = [] |
| assembled_variables_cnt = {} |
| if isinstance(self.prompt_data, list): |
| |
| for part in self.prompt_data: |
| assembled_prompt.extend( |
| self._assemble_singular_part( |
| part, variables_dict, assembled_variables_cnt |
| ) |
| ) |
| else: |
| |
| assembled_prompt.extend( |
| self._assemble_singular_part( |
| self.prompt_data, variables_dict, assembled_variables_cnt |
| ) |
| ) |
|
|
| |
| simplified_assembled_prompt = [assembled_prompt[0]] |
| for i in range(1, len(assembled_prompt)): |
| |
| try: |
| prev_text = simplified_assembled_prompt[-1].text |
| curr_text = assembled_prompt[i].text |
| if isinstance(prev_text, str) and isinstance(curr_text, str): |
| simplified_assembled_prompt[-1] = Part.from_text( |
| prev_text + curr_text |
| ) |
| else: |
| simplified_assembled_prompt.append(assembled_prompt[i]) |
| except AttributeError: |
| simplified_assembled_prompt.append(assembled_prompt[i]) |
| continue |
|
|
| |
| for key in variables_dict: |
| if key not in assembled_variables_cnt: |
| raise ValueError(f"Variable {key} is not present in prompt_data.") |
|
|
| assemble_cnt_msg = "Assembled prompt replacing: " |
| for key in assembled_variables_cnt: |
| assemble_cnt_msg += ( |
| f"{assembled_variables_cnt[key]} instances of variable {key}, " |
| ) |
| if assemble_cnt_msg[-2:] == ", ": |
| assemble_cnt_msg = assemble_cnt_msg[:-2] |
| _LOGGER.info(assemble_cnt_msg) |
|
|
| |
| return [ |
| Content( |
| parts=simplified_assembled_prompt, |
| role="user", |
| ) |
| ] |
|
|
| def _assemble_singular_part( |
| self, |
| prompt_data_part: Union[str, Image, Part], |
| formatted_variables_set: Dict[str, List[Part]], |
| assembled_variables_cnt: Dict[str, int], |
| ) -> List[Part]: |
| """Assemble a str, Image, or Part.""" |
| if isinstance(prompt_data_part, Image): |
| |
| return [Part.from_image(prompt_data_part)] |
| elif isinstance(prompt_data_part, str): |
| |
| return self._assemble_single_str( |
| prompt_data_part, formatted_variables_set, assembled_variables_cnt |
| ) |
| elif isinstance(prompt_data_part, Part): |
| |
| try: |
| text = prompt_data_part.text |
| except AttributeError: |
| return [prompt_data_part] |
| return self._assemble_single_str( |
| text, formatted_variables_set, assembled_variables_cnt |
| ) |
|
|
| def _assemble_single_str( |
| self, |
| prompt_data_str: str, |
| formatted_variables_set: Dict[str, List[Part]], |
| assembled_variables_cnt: Dict[str, int], |
| ) -> List[Part]: |
| """Assemble a single string with 0 or more variables within the string.""" |
| |
| prompt_data_str_split = re.split(VARIABLE_NAME_REGEX, prompt_data_str) |
|
|
| assembled_data = [] |
| |
| for s in prompt_data_str_split: |
| if not s: |
| continue |
| variable_name = s[1:-1] |
| if ( |
| re.match(VARIABLE_NAME_REGEX, s) |
| and variable_name in formatted_variables_set |
| ): |
| assembled_data.extend(formatted_variables_set[variable_name]) |
| assembled_variables_cnt[variable_name] = ( |
| assembled_variables_cnt.get(variable_name, 0) + 1 |
| ) |
| else: |
| assembled_data.append(Part.from_text(s)) |
|
|
| return assembled_data |
|
|
| def generate_content( |
| self, |
| contents: ContentsType, |
| *, |
| generation_config: Optional[GenerationConfigType] = None, |
| safety_settings: Optional[SafetySettingsType] = None, |
| model_name: Optional[str] = None, |
| tools: Optional[List["Tool"]] = None, |
| tool_config: Optional["ToolConfig"] = None, |
| stream: bool = False, |
| system_instruction: Optional[PartsType] = None, |
| ) -> Union["GenerationResponse", Iterable["GenerationResponse"],]: |
| """Generates content using the saved Prompt configs. |
| |
| Args: |
| contents: Contents to send to the model. |
| Supports either a list of Content objects (passing a multi-turn conversation) |
| or a value that can be converted to a single Content object (passing a single message). |
| Supports |
| * str, Image, Part, |
| * List[Union[str, Image, Part]], |
| * List[Content] |
| generation_config: Parameters for the generation. |
| model_name: Prediction model resource name. |
| safety_settings: Safety settings as a mapping from HarmCategory to HarmBlockThreshold. |
| tools: A list of tools (functions) that the model can try calling. |
| tool_config: Config shared for all tools provided in the request. |
| stream: Whether to stream the response. |
| system_instruction: System instruction to pass to the model. |
| |
| Returns: |
| A single GenerationResponse object if stream == False |
| A stream of GenerationResponse objects if stream == True |
| |
| Usage: |
| ``` |
| prompt = Prompt( |
| prompt_data="Hello, {name}! Today is {day}. How are you?", |
| variables={"name": "Alice", "day": "Monday"}, |
| generation_config=GenerationConfig(temperature=0.1,), |
| system_instruction="Please answer in a short sentence.", |
| model_name="gemini-1.0-pro-002", |
| ) |
| |
| prompt.generate_content( |
| contents=prompt.assemble_contents(**prompt.variables) |
| ) |
| ``` |
| """ |
| if not (model_name or self._model_name): |
| _LOGGER.info( |
| "No model name specified, falling back to default model: %s", |
| self.model_name, |
| ) |
| model_name = model_name or self.model_name |
|
|
| generation_config = generation_config or self.generation_config |
| safety_settings = safety_settings or self.safety_settings |
| tools = tools or self.tools |
| tool_config = tool_config or self.tool_config |
| system_instruction = system_instruction or self.system_instruction |
|
|
| if not model_name: |
| raise ValueError( |
| "Model name must be specified to use Prompt.generate_content()" |
| ) |
| model_name = Prompt._format_model_resource_name(model_name) |
|
|
| model = GenerativeModel( |
| model_name=model_name, system_instruction=system_instruction |
| ) |
| return model.generate_content( |
| contents=contents, |
| generation_config=generation_config, |
| safety_settings=safety_settings, |
| tools=tools, |
| tool_config=tool_config, |
| stream=stream, |
| ) |
|
|
| @property |
| def _dataset_client(self) -> dataset_service_client.DatasetServiceClient: |
| if not getattr(self, "_dataset_client_value", None): |
| self._dataset_client_value = ( |
| aiplatform_initializer.global_config.create_client( |
| client_class=dataset_service_client.DatasetServiceClient, |
| ) |
| ) |
| return self._dataset_client_value |
|
|
| @classmethod |
| def _clone(cls, prompt: "Prompt") -> "Prompt": |
| """Returns a copy of the Prompt.""" |
| return Prompt( |
| prompt_data=deepcopy(prompt.prompt_data), |
| variables=deepcopy(prompt.variables), |
| generation_config=deepcopy(prompt.generation_config), |
| safety_settings=deepcopy(prompt.safety_settings), |
| tools=deepcopy(prompt.tools), |
| tool_config=deepcopy(prompt.tool_config), |
| system_instruction=deepcopy(prompt.system_instruction), |
| model_name=prompt.model_name, |
| ) |
|
|
| def get_unassembled_prompt_data(self) -> PartsType: |
| """Returns the prompt data, without any variables replaced.""" |
| return self.prompt_data |
|
|
| def __str__(self) -> str: |
| """Returns the prompt data as a string, without any variables replaced.""" |
| return str(self.prompt_data or "") |
|
|
| def __repr__(self) -> str: |
| """Returns a string representation of the unassembled prompt.""" |
| result = "Prompt(" |
| if self.prompt_data: |
| result += f"prompt_data='{self.prompt_data}', " |
| if self.variables and self.variables[0]: |
| result += f"variables={self.variables}), " |
| if self.system_instruction: |
| result += f"system_instruction={self.system_instruction}), " |
| if self._model_name: |
| |
| result += f"model_name={self._model_name}), " |
| if self.generation_config: |
| result += f"generation_config={self.generation_config}), " |
| if self.safety_settings: |
| result += f"safety_settings={self.safety_settings}), " |
| if self.tools: |
| result += f"tools={self.tools}), " |
| if self.tool_config: |
| result += f"tool_config={self.tool_config}, " |
| if self.prompt_id: |
| result += f"prompt_id={self.prompt_id}, " |
| if self.version_id: |
| result += f"version_id={self.version_id}, " |
| if self.prompt_name: |
| result += f"prompt_name={self.prompt_name}, " |
| if self.version_name: |
| result += f"version_name={self.version_name}, " |
|
|
| |
| if result[-2:] == ", ": |
| result = result[:-2] |
| result += ")" |
| return result |
|
|