Spaces:
Paused
Paused
matt HOFFNER commited on
Commit ·
328552e
1
Parent(s): ac695ee
try fetch and post
Browse files- src/app/search/web/page.jsx +43 -19
- src/pages/api/llm.js +7 -3
src/app/search/web/page.jsx
CHANGED
|
@@ -1,31 +1,55 @@
|
|
| 1 |
-
"use client"
|
| 2 |
import { useEffect, useState } from "react";
|
| 3 |
|
| 4 |
export default function WebSearchPage({ searchParams }) {
|
| 5 |
-
|
| 6 |
-
const [aiResponse, setAiResponse] = useState(null);
|
| 7 |
-
const startIndex = searchParams.start || "1";
|
| 8 |
|
| 9 |
useEffect(() => {
|
|
|
|
|
|
|
|
|
|
| 10 |
const url = new URL('/api/llm', window.location.origin);
|
| 11 |
-
url.searchParams.append('question', searchParams.searchTerm);
|
| 12 |
-
url.searchParams.append('startIndex', startIndex);
|
| 13 |
-
|
| 14 |
-
const openaiRes = new EventSource(url);
|
| 15 |
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 26 |
|
| 27 |
console.log(aiResponse);
|
| 28 |
|
| 29 |
-
|
| 30 |
return <>{aiResponse ? JSON.stringify(aiResponse) : 'Loading...'}</>;
|
| 31 |
}
|
|
|
|
|
|
|
| 1 |
import { useEffect, useState } from "react";
|
| 2 |
|
| 3 |
export default function WebSearchPage({ searchParams }) {
|
| 4 |
+
const [aiResponse, setAiResponse] = useState("");
|
|
|
|
|
|
|
| 5 |
|
| 6 |
useEffect(() => {
|
| 7 |
+
if (!searchParams || !searchParams.searchTerm) return;
|
| 8 |
+
|
| 9 |
+
const { searchTerm, start = "1" } = searchParams;
|
| 10 |
const url = new URL('/api/llm', window.location.origin);
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
|
| 12 |
+
fetch(url, {
|
| 13 |
+
method: 'POST',
|
| 14 |
+
headers: {
|
| 15 |
+
'Content-Type': 'application/json'
|
| 16 |
+
},
|
| 17 |
+
body: JSON.stringify({
|
| 18 |
+
question: searchTerm,
|
| 19 |
+
startIndex: start
|
| 20 |
+
})
|
| 21 |
+
})
|
| 22 |
+
.then(response => {
|
| 23 |
+
if (!response.ok) {
|
| 24 |
+
throw new Error("HTTP error " + response.status);
|
| 25 |
+
}
|
| 26 |
+
// Create a reader to stream the response body
|
| 27 |
+
const reader = response.body.getReader();
|
| 28 |
+
|
| 29 |
+
// Read and process the response body chunks
|
| 30 |
+
return new ReadableStream({
|
| 31 |
+
start(controller) {
|
| 32 |
+
function push() {
|
| 33 |
+
reader.read().then(({ done, value }) => {
|
| 34 |
+
if (done) {
|
| 35 |
+
// Close the stream when done
|
| 36 |
+
controller.close();
|
| 37 |
+
return;
|
| 38 |
+
}
|
| 39 |
+
// Decode the response chunk and append it to the existing response
|
| 40 |
+
setAiResponse(prev => prev + new TextDecoder().decode(value));
|
| 41 |
+
// Get the next chunk
|
| 42 |
+
push();
|
| 43 |
+
});
|
| 44 |
+
}
|
| 45 |
+
push();
|
| 46 |
+
}
|
| 47 |
+
});
|
| 48 |
+
})
|
| 49 |
+
.catch(console.error);
|
| 50 |
+
}, [searchParams]);
|
| 51 |
|
| 52 |
console.log(aiResponse);
|
| 53 |
|
|
|
|
| 54 |
return <>{aiResponse ? JSON.stringify(aiResponse) : 'Loading...'}</>;
|
| 55 |
}
|
src/pages/api/llm.js
CHANGED
|
@@ -20,13 +20,17 @@ export default function handler(req, res) {
|
|
| 20 |
const openai = new OpenAIApi(configuration);
|
| 21 |
|
| 22 |
const messages = [
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
{
|
| 24 |
role: "user",
|
| 25 |
content: QUESTION,
|
| 26 |
},
|
| 27 |
];
|
| 28 |
|
| 29 |
-
const
|
| 30 |
apiKey: process.env.API_KEY,
|
| 31 |
googleCSEId: process.env.CONTEXT_KEY,
|
| 32 |
});
|
|
@@ -39,7 +43,7 @@ export default function handler(req, res) {
|
|
| 39 |
const response = await openai.createChatCompletion({
|
| 40 |
model: "gpt-3.5-turbo-0613",
|
| 41 |
messages,
|
| 42 |
-
functions: [
|
| 43 |
temperature: 0,
|
| 44 |
});
|
| 45 |
|
|
@@ -64,7 +68,7 @@ export default function handler(req, res) {
|
|
| 64 |
const args = response.data.choices[0].message.function_call.arguments;
|
| 65 |
|
| 66 |
const fn = functions[fnName];
|
| 67 |
-
const result = await fn
|
| 68 |
|
| 69 |
messages.push({
|
| 70 |
role: "assistant",
|
|
|
|
| 20 |
const openai = new OpenAIApi(configuration);
|
| 21 |
|
| 22 |
const messages = [
|
| 23 |
+
{
|
| 24 |
+
role: "system",
|
| 25 |
+
content: "You are a helpful assistant.",
|
| 26 |
+
},
|
| 27 |
{
|
| 28 |
role: "user",
|
| 29 |
content: QUESTION,
|
| 30 |
},
|
| 31 |
];
|
| 32 |
|
| 33 |
+
const googleCustomSearch = new GoogleCustomSearch({
|
| 34 |
apiKey: process.env.API_KEY,
|
| 35 |
googleCSEId: process.env.CONTEXT_KEY,
|
| 36 |
});
|
|
|
|
| 43 |
const response = await openai.createChatCompletion({
|
| 44 |
model: "gpt-3.5-turbo-0613",
|
| 45 |
messages,
|
| 46 |
+
functions: [googleCustomSearch.schema],
|
| 47 |
temperature: 0,
|
| 48 |
});
|
| 49 |
|
|
|
|
| 68 |
const args = response.data.choices[0].message.function_call.arguments;
|
| 69 |
|
| 70 |
const fn = functions[fnName];
|
| 71 |
+
const result = await fn.call(args);
|
| 72 |
|
| 73 |
messages.push({
|
| 74 |
role: "assistant",
|