| /** @format */ | |
| function getPayloadDoc(prompt) { | |
| return { | |
| model: "", | |
| temperature: 0.3, | |
| max_tokens: 1024, | |
| stream: false, | |
| random_seed: 42, | |
| messages: [{ role: "user", content: prompt }], | |
| safe_prompt: false, | |
| }; | |
| } | |
| function getPayloadBuildContext(prompt) { | |
| return { | |
| model: "", | |
| messages: [{ role: "user", content: prompt }], | |
| temperature: 0.3, | |
| max_tokens: 2000, | |
| stream: false, | |
| safe_prompt: false, | |
| random_seed: 42, | |
| }; | |
| } | |
| function getPayloadWithContext(prompt) { | |
| return { | |
| model: "", | |
| messages: [{ role: "user", content: prompt }], | |
| temperature: 0.3, | |
| max_tokens: 2000, | |
| stream: false, | |
| safe_prompt: false, | |
| random_seed: 42, | |
| }; | |
| } | |
| function getPayloadThread(prompt) { | |
| return { | |
| model: "", | |
| messages: [{ role: "user", content: prompt }], | |
| temperature: 0.7, | |
| max_tokens: 2000, | |
| stream: false, | |
| safe_prompt: false, | |
| random_seed: 42, | |
| }; | |
| } | |
| /* | |
| function getPayloadDoc(prompt) { | |
| const payload = { | |
| inputs: prompt, | |
| parameters: { | |
| task: "text2text-generation", | |
| max_new_tokens: 2000, | |
| num_return_sequences: 1, | |
| temperature: 0.4, | |
| top_p: 0.85, | |
| top_k: 30, | |
| do_sample: false, | |
| no_repeat_ngram_size: 3, | |
| num_beams: 4, | |
| repetition_penalty: 1.2, | |
| return_full_text: false, | |
| details: false, | |
| max_time: 90.0, | |
| seed: 42, | |
| }, | |
| options: { | |
| use_cache: false, | |
| wait_for_model: true, | |
| }, | |
| }; | |
| return payload; | |
| } | |
| function getPayloadBuildContext(prompt) { | |
| const payload = { | |
| inputs: prompt, | |
| parameters: { | |
| task: "text2text-generation", | |
| max_new_tokens: 6000, | |
| num_return_sequences: 1, | |
| temperature: 0.7, | |
| top_p: 0.85, | |
| top_k: 30, | |
| do_sample: false, | |
| no_repeat_ngram_size: 4, | |
| num_beams: 6, | |
| repetition_penalty: 1.2, | |
| return_full_text: false, | |
| details: false, | |
| max_time: 180.0, | |
| seed: 42, | |
| }, | |
| options: { | |
| use_cache: false, | |
| wait_for_model: true, | |
| }, | |
| }; | |
| return payload; | |
| } | |
| function getPayloadWithContext(prompt) { | |
| const payload = { | |
| inputs: prompt, | |
| parameters: { | |
| task: "text2text-generation", | |
| max_new_tokens: 4000, | |
| num_return_sequences: 1, | |
| temperature: 0.7, | |
| top_p: 0.85, | |
| top_k: 30, | |
| do_sample: false, | |
| no_repeat_ngram_size: 4, | |
| num_beams: 5, | |
| repetition_penalty: 1.4, | |
| return_full_text: false, | |
| details: false, | |
| max_time: 120.0, | |
| seed: 42, | |
| }, | |
| options: { | |
| use_cache: false, | |
| wait_for_model: true, | |
| }, | |
| }; | |
| return payload; | |
| } | |
| function getPayloadThread(prompt) { | |
| const payload = { | |
| inputs: prompt, | |
| parameters: { | |
| task: "text2text-generation", | |
| max_new_tokens: 6048, | |
| num_return_sequences: 1, | |
| temperature: 0.7, | |
| top_p: 0.85, | |
| top_k: 30, | |
| do_sample: false, | |
| no_repeat_ngram_size: 4, | |
| num_beams: 5, | |
| repetition_penalty: 1.4, | |
| return_full_text: false, | |
| details: false, | |
| max_time: 120.0, | |
| seed: 42, | |
| }, | |
| options: { | |
| use_cache: false, | |
| wait_for_model: true, | |
| }, | |
| }; | |
| return payload; | |
| } | |
| */ | |