|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""
|
|
|
The base class for reward model
|
|
|
"""
|
|
|
|
|
|
from abc import ABC, abstractmethod
|
|
|
|
|
|
from verl import DataProto
|
|
|
|
|
|
|
|
|
class BasePPORewardModel(ABC):
|
|
|
def __init__(self, config):
|
|
|
self.config = config
|
|
|
|
|
|
@abstractmethod
|
|
|
def compute_reward(self, data: DataProto) -> DataProto:
|
|
|
"""Computing reward given input_ids. The transformers should output a tensor with shape
|
|
|
[batch_size, sequence_length], and the value at [EOS] mask should be gathered.
|
|
|
|
|
|
Args:
|
|
|
data: must contain keys "input_ids", "attention_mask" and "position_ids".
|
|
|
- input_ids: [batch_size, sequence_length]
|
|
|
- attention_mask: [batch_size, sequence_length]
|
|
|
- position_ids: [batch_size, sequence_length]
|
|
|
|
|
|
Returns: a data pass protocol containing "reward". Only the [EOS] position contains the reward.
|
|
|
Other position should have zero reward. Note that this may change in the future if we use
|
|
|
dense reward. So, we leave the interface for general case.
|
|
|
- reward: [batch_size, sequence_length].
|
|
|
|
|
|
"""
|
|
|
pass
|
|
|
|