| | from typing import Optional, Union |
| |
|
| | from google.cloud import aiplatform |
| | from google.cloud.aiplatform import initializer as aiplatform_initializer |
| | from vertexai.language_models import _language_models |
| | from vertexai.language_models import _language_models as tuning |
| |
|
| |
|
| | _DISTILLATION_PIPELINE_URI = ( |
| | "https://us-kfp.pkg.dev/ml-pipeline/distillation/distillation/v1.0.0" |
| | ) |
| |
|
| |
|
| | class DistillationMixin: |
| | def distill_from( |
| | self, |
| | *, |
| | dataset: str, |
| | teacher_model: Union[str, _language_models._TextGenerationModel], |
| | train_steps: Optional[int] = None, |
| | learning_rate_multiplier: Optional[float] = None, |
| | evaluation_spec: Optional[tuning.TuningEvaluationSpec] = None, |
| | accelerator_type: Optional[tuning._ACCELERATOR_TYPE_TYPE] = None, |
| | model_display_name: Optional[str] = None, |
| | max_context_length: Optional[str] = None, |
| | ): |
| | """Tunes a smaller model with help from another bigger model. |
| | |
| | Args: |
| | dataset: A URI pointing to data in JSON lines format. |
| | teacher_model: The teacher model to use for distillation. |
| | train_steps: Number of training batches to use (batch size is 8 samples). |
| | learning_rate_multiplier: Learning rate multiplier to use in tuning. |
| | evaluation_spec: Specification for the model evaluation during tuning. |
| | accelerator_type: Type of accelerator to use. Can be "TPU" or "GPU". |
| | model_display_name: Custom display name for the tuned model. |
| | max_context_length: The max context length used for tuning. |
| | Can be either '8k' or '32k' |
| | |
| | Returns: |
| | A tuning job for distillation. |
| | |
| | Raises: |
| | RuntimeError: If the model does not support distillation. |
| | """ |
| | if "/models/" not in self._endpoint_name: |
| | raise RuntimeError( |
| | f"Model does not support distillation: {self._endpoint_name}" |
| | ) |
| | student_short_model_id = self._endpoint_name.split("/")[-1] |
| |
|
| | if isinstance(teacher_model, str): |
| | teacher_short_model_id = teacher_model |
| | elif isinstance(teacher_model, _language_models._LanguageModel): |
| | if "/models/" not in teacher_model._endpoint_name: |
| | raise RuntimeError( |
| | f"Teacher model does not support distillation: {teacher_model._endpoint_name}" |
| | ) |
| | teacher_short_model_id = teacher_model._endpoint_name.split("/")[-1] |
| | else: |
| | raise RuntimeError(f"Unsupported teacher model type: {teacher_model}") |
| |
|
| | pipeline_job = submit_distillation_pipeline_job( |
| | teacher_model=teacher_short_model_id, |
| | student_model=student_short_model_id, |
| | dataset=dataset, |
| | train_steps=train_steps, |
| | learning_rate_multiplier=learning_rate_multiplier, |
| | evaluation_spec=evaluation_spec, |
| | accelerator_type=accelerator_type, |
| | model_display_name=model_display_name, |
| | max_context_length=max_context_length, |
| | ) |
| | tuning_job = tuning._LanguageModelTuningJob( |
| | base_model=self, |
| | job=pipeline_job, |
| | ) |
| | return tuning_job |
| |
|
| |
|
| | def submit_distillation_pipeline_job( |
| | *, |
| | teacher_model: str, |
| | student_model: str, |
| | dataset: str, |
| | train_steps: Optional[int] = None, |
| | learning_rate_multiplier: Optional[float] = None, |
| | evaluation_spec: Optional[tuning.TuningEvaluationSpec] = None, |
| | accelerator_type: Optional[tuning._ACCELERATOR_TYPE_TYPE] = None, |
| | model_display_name: Optional[str] = None, |
| | max_context_length: Optional[str] = None, |
| | ): |
| | teacher_short_model_id = teacher_model.split("/")[-1] |
| | student_short_model_id = student_model.split("/")[-1] |
| | pipeline_arguments = { |
| | "teacher_model_reference": teacher_model, |
| | "student_model_reference": student_model, |
| | "dataset_uri": dataset, |
| | "project": aiplatform_initializer.global_config.project, |
| | "location": aiplatform_initializer.global_config.location, |
| | } |
| | if train_steps is not None: |
| | pipeline_arguments["train_steps"] = train_steps |
| | if learning_rate_multiplier is not None: |
| | pipeline_arguments["learning_rate_multiplier"] = learning_rate_multiplier |
| | if evaluation_spec is not None: |
| | pipeline_arguments["evaluation_data_uri"] = evaluation_spec.evaluation_data |
| | pipeline_arguments["evaluation_interval"] = evaluation_spec.evaluation_interval |
| | pipeline_arguments[ |
| | "enable_early_stopping" |
| | ] = evaluation_spec.enable_early_stopping |
| | pipeline_arguments[ |
| | "enable_checkpoint_selection" |
| | ] = evaluation_spec.enable_checkpoint_selection |
| | pipeline_arguments["tensorboard_resource_id"] = evaluation_spec.tensorboard |
| | |
| | if accelerator_type is not None: |
| | pipeline_arguments["accelerator_type"] = accelerator_type |
| | if aiplatform_initializer.global_config.encryption_spec_key_name is not None: |
| | pipeline_arguments[ |
| | "encryption_spec_key_name" |
| | ] = aiplatform_initializer.global_config.encryption_spec_key_name |
| | if max_context_length is not None: |
| | pipeline_arguments["max_context_length"] = max_context_length |
| | if model_display_name is None: |
| | model_display_name = ( |
| | f"{student_short_model_id} distilled from {teacher_short_model_id}" |
| | ) |
| | pipeline_arguments["model_display_name"] = model_display_name |
| | |
| | |
| | |
| | |
| | |
| | pipeline_job = aiplatform.PipelineJob( |
| | template_path=_DISTILLATION_PIPELINE_URI, |
| | display_name=None, |
| | parameter_values=pipeline_arguments, |
| | ) |
| | pipeline_job.submit() |
| | return pipeline_job |
| |
|