TrieTran
Upload folder using huggingface_hub
58bf47f verified
"""
phi.py
Class definition for all LLMs derived from PhiForCausalLM.
"""
from typing import Optional, Type
import torch
from torch import nn as nn
from transformers import PhiForCausalLM
from transformers.models.phi.modeling_phi import PhiDecoderLayer
from prismatic.models.backbones.llm.base_llm import HFCausalLLMBackbone
from prismatic.models.backbones.llm.prompting import PhiPromptBuilder, PromptBuilder
# Registry ==> Support Phi Models (from HF Transformers)
# fmt: off
PHI_MODELS = {
# === Phi-2 ===
"phi-2-3b": {
"llm_family": "phi", "llm_cls": PhiForCausalLM, "hf_hub_path": "microsoft/phi-2"
}
}
# fmt: on
class PhiLLMBackbone(HFCausalLLMBackbone):
def __init__(
self,
llm_backbone_id: str,
llm_max_length: int = 2048,
hf_token: Optional[str] = None,
inference_mode: bool = False,
use_flash_attention_2: bool = True,
) -> None:
super().__init__(
llm_backbone_id,
llm_max_length=llm_max_length,
hf_token=hf_token,
inference_mode=inference_mode,
use_flash_attention_2=use_flash_attention_2,
**PHI_MODELS[llm_backbone_id],
)
# [Special Case] Phi PAD Token Handling --> for clarity, we add an extra token (and resize)
self.tokenizer.add_special_tokens({"pad_token": "<|pad|>"})
self.llm.config.pad_token_id = self.tokenizer.pad_token_id
self.llm.resize_token_embeddings(len(self.tokenizer), pad_to_multiple_of=64)
@property
def prompt_builder_fn(self) -> Type[PromptBuilder]:
if self.identifier.startswith("phi-2"):
return PhiPromptBuilder
raise ValueError(f"No PromptBuilder defined for LLM Backbone `{self.identifier}`")
@property
def transformer_layer_cls(self) -> Type[nn.Module]:
return PhiDecoderLayer
@property
def half_precision_dtype(self) -> torch.dtype:
return torch.bfloat16