|
"""MPT Blocks used for the MPT Model.""" |
|
import logging |
|
from copy import deepcopy |
|
from functools import partial |
|
from typing import Any, Callable, Optional, Union |
|
import torch |
|
import torch.nn as nn |
|
from .fc import FC_CLASS_REGISTRY |
|
try: |
|
import transformer_engine.pytorch as te |
|
except: |
|
te = None |
|
log = logging.getLogger(__name__) |
|
_FFN_ACT_FN_DEFAULT = {'name': 'gelu', 'approximate': 'none'} |
|
|
|
def resolve_ffn_act_fn(config: Optional[dict]=None) -> Callable[[torch.Tensor], torch.Tensor]: |
|
"""Resolve the activation function for the feed-forward network. |
|
Args: |
|
config (Optional[dict]): The configuration dictionary for the activation function. |
|
The dict config must specify the 'name' of a torch.nn.functional activation |
|
function. All of other key values pairs are bound to the function as a partial. |
|
Returns: |
|
Callable[[torch.Tensor], torch.Tensor]: The activation function. |
|
""" |
|
if config is None: |
|
config = _FFN_ACT_FN_DEFAULT |
|
config = deepcopy(config) |
|
name = config.pop('name') |
|
if not hasattr(torch.nn.functional, name): |
|
raise ValueError(f'Unrecognised activation function name ({name}).') |
|
act = getattr(torch.nn.functional, name) |
|
return partial(act, **config) |
|
_DEFAULT_ACT_FN = resolve_ffn_act_fn(_FFN_ACT_FN_DEFAULT) |
|
|
|
def resolve_ffn_hidden_size(d_model: int, expansion_ratio: Union[int, float], ffn_hidden_size: Optional[int]=None) -> int: |
|
"""Resolve the hidden size of the feed-forward network. |
|
Args: |
|
d_model (int): The dimension of the input and output of the feed-forward network. |
|
expansion_ratio (Union[int, float]): The expansion ratio of the feed-forward network. |
|
ffn_hidden_size (Optional[int]): The hidden size of the feed-forward network. |
|
Returns: |
|
int: The hidden size of the feed-forward network. |
|
""" |
|
if ffn_hidden_size is not None: |
|
log.info(f'`expansion_ratio` (={expansion_ratio}) ignored when `ffn_hidden_size` (={ffn_hidden_size}) is specified.') |
|
else: |
|
ffn_hidden_size = int(d_model * expansion_ratio) |
|
if ffn_hidden_size != d_model * expansion_ratio: |
|
raise ValueError(f'`d_model * expansion_ratio` must be an integer (d_model={d_model!r}; expansion_ratio={expansion_ratio!r}; d_model * expansion_ratio={d_model * expansion_ratio!r}).') |
|
return ffn_hidden_size |
|
|
|
class MPTMLP(nn.Module): |
|
|
|
def __init__(self, d_model: int, expansion_ratio: Union[int, float], fc_type: str='torch', ffn_hidden_size: Optional[int]=None, act_fn: Callable[[torch.Tensor], torch.Tensor]=_DEFAULT_ACT_FN, device: Optional[str]=None, bias: bool=True): |
|
super().__init__() |
|
ffn_hidden_size = resolve_ffn_hidden_size(d_model, expansion_ratio, ffn_hidden_size) |
|
self.fc_kwargs: dict[str, Any] = {'bias': bias} |
|
if fc_type != 'te': |
|
self.fc_kwargs['device'] = device |
|
self.up_proj = FC_CLASS_REGISTRY[fc_type](d_model, ffn_hidden_size, **self.fc_kwargs) |
|
self.act = act_fn |
|
self.down_proj = FC_CLASS_REGISTRY[fc_type](ffn_hidden_size, d_model, **self.fc_kwargs) |
|
self.down_proj._is_residual = True |
|
|
|
def forward(self, x: torch.Tensor) -> torch.Tensor: |
|
return self.down_proj(self.act(self.up_proj(x))) |
|
|
|
class MPTGLU(MPTMLP): |
|
|
|
def __init__(self, d_model: int, expansion_ratio: Union[int, float], fc_type: str='torch', ffn_hidden_size: Optional[int]=None, act_fn: Callable[[torch.Tensor], torch.Tensor]=_DEFAULT_ACT_FN, device: Optional[str]=None, bias: bool=True): |
|
super().__init__(d_model=d_model, expansion_ratio=expansion_ratio, fc_type=fc_type, ffn_hidden_size=ffn_hidden_size, act_fn=act_fn, device=device, bias=bias) |
|
self.gate_proj = FC_CLASS_REGISTRY[fc_type](d_model, self.up_proj.out_features, **self.fc_kwargs) |
|
|
|
def forward(self, x: torch.Tensor) -> torch.Tensor: |
|
return self.down_proj(self.act(self.gate_proj(x)) * self.up_proj(x)) |
|
FFN_CLASS_REGISTRY = {'mptmlp': MPTMLP, 'mptglu': MPTGLU} |
|
if te is not None: |
|
te.LayerNormMLP._has_norm = True |
|
FFN_CLASS_REGISTRY['te_ln_mlp'] = te.LayerNormMLP |
|
|
|
def build_ffn(d_model: int, expansion_ratio: Union[int, float], fc_type: str='torch', ffn_hidden_size: Optional[int]=None, ffn_act_fn: Optional[dict]=None, device: Optional[str]=None, bias: bool=True, **kwargs: Any) -> nn.Module: |
|
ffn_type = kwargs.pop('ffn_type') |
|
if ffn_type in ['mptmlp', 'mptglu']: |
|
if len(kwargs) > 0: |
|
raise ValueError(f'MPTMLP (or MPTGLU) got an unexpected keyword argument: {kwargs}') |
|
return FFN_CLASS_REGISTRY[ffn_type](d_model=d_model, expansion_ratio=expansion_ratio, fc_type=fc_type, act_fn=resolve_ffn_act_fn(ffn_act_fn), ffn_hidden_size=ffn_hidden_size, device=device, bias=bias) |
|
elif ffn_type == 'te_ln_mlp': |
|
assert te is not None |
|
ffn_hidden_size = resolve_ffn_hidden_size(d_model, expansion_ratio, ffn_hidden_size) |
|
if ffn_act_fn is not None: |
|
raise ValueError(f'Transformer Engine block does not support custom activation functions.') |
|
return te.LayerNormMLP(hidden_size=d_model, ffn_hidden_size=ffn_hidden_size, bias=bias, **kwargs) |
|
raise ValueError(f'ffn_type={ffn_type!r} not recognized.') |