llm_blender_value_v4 / config.json
SiMajid's picture
Initial commit with checkpoint files
23f56d8
raw
history blame contribute delete
511 Bytes
{
"ranker_type": "pairranker",
"model_type": "roberta",
"model_name": "roberta-large",
"cache_dir": "./hf_models/roberta-large/",
"load_checkpoint": null,
"source_maxlength": 128,
"candidate_maxlength": 128,
"n_tasks": 1,
"num_pos": 5,
"num_neg": 5,
"sub_sampling_mode": "all_pair",
"sub_sampling_ratio": 0.4,
"loss_type": "instructgpt",
"reduce_type": "linear",
"inference_mode": "bubble",
"drop_out": 0.05,
"fp16": true,
"device": null
}