Update server.js
Browse files
server.js
CHANGED
|
@@ -137,17 +137,17 @@ app.post('/cr', async (req, res) => {
|
|
| 137 |
const apiKey = req.body.api || openai_key;
|
| 138 |
|
| 139 |
if (!prompt) {
|
| 140 |
-
return res.status(400).json(
|
| 141 |
}
|
| 142 |
|
| 143 |
try {
|
| 144 |
-
const response = await axios.post('
|
| 145 |
messages: [{'role': 'system', 'content': start}, {'role': 'user', 'content': prompt}],
|
| 146 |
max_tokens: 2000,
|
| 147 |
temperature: 0.3,
|
| 148 |
// presence_penalty: 0.0,
|
| 149 |
//frequency_penalty: -0.2,
|
| 150 |
-
|
| 151 |
//model: "gemini-1.5-flash-latest",
|
| 152 |
}, {
|
| 153 |
headers: {
|
|
|
|
| 137 |
const apiKey = req.body.api || openai_key;
|
| 138 |
|
| 139 |
if (!prompt) {
|
| 140 |
+
return res.status(400).json("wait"); // Не удалось принять данные
|
| 141 |
}
|
| 142 |
|
| 143 |
try {
|
| 144 |
+
const response = await axios.post('http://yangtautauaja.xp3.biz/v1/openchat.php', {
|
| 145 |
messages: [{'role': 'system', 'content': start}, {'role': 'user', 'content': prompt}],
|
| 146 |
max_tokens: 2000,
|
| 147 |
temperature: 0.3,
|
| 148 |
// presence_penalty: 0.0,
|
| 149 |
//frequency_penalty: -0.2,
|
| 150 |
+
// model: "gemini-1.0-ultra",
|
| 151 |
//model: "gemini-1.5-flash-latest",
|
| 152 |
}, {
|
| 153 |
headers: {
|