code
stringlengths 619
138k
| apis
listlengths 1
8
| extract_api
stringlengths 79
7.3k
|
|---|---|---|
package com.api.chatgpt;
import com.theokanning.openai.edit.EditRequest;
import com.theokanning.openai.service.OpenAiService;
import org.springframework.boot.SpringApplication;
import org.springframework.boot.autoconfigure.SpringBootApplication;
@SpringBootApplication
public class ChatGptApplication {
public static void main(String[] args) {
SpringApplication.run(ChatGptApplication.class, args);
var service = new OpenAiService("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");
var request = EditRequest.builder()
.model("text-davinci-edit-001")
.input("i Jack, I like programn in java, how abt you")
.instruction("Fix the grammar and spelling mistakes")
.build();
service.createEdit(request).getChoices().forEach(System.out::println);
}
}
|
[
"com.theokanning.openai.edit.EditRequest.builder"
] |
[((534, 769), 'com.theokanning.openai.edit.EditRequest.builder'), ((534, 744), 'com.theokanning.openai.edit.EditRequest.builder'), ((534, 674), 'com.theokanning.openai.edit.EditRequest.builder'), ((534, 603), 'com.theokanning.openai.edit.EditRequest.builder')]
|
package com.hqy.cloud.chatgpt.config;
import cn.hutool.core.util.StrUtil;
import com.hqy.cloud.chatgpt.config.interceptor.AuthenticationInterceptor;
import com.hqy.cloud.chatgpt.config.interceptor.ProxyAuthenticator;
import com.hqy.cloud.chatgpt.core.UnofficialApi;
import com.hqy.cloud.chatgpt.service.OpenAiChatgptService;
import com.theokanning.openai.OpenAiApi;
import com.theokanning.openai.service.OpenAiService;
import lombok.RequiredArgsConstructor;
import lombok.extern.slf4j.Slf4j;
import okhttp3.ConnectionPool;
import okhttp3.OkHttpClient;
import org.apache.commons.lang3.StringUtils;
import org.jetbrains.annotations.NotNull;
import org.springframework.boot.autoconfigure.condition.ConditionalOnMissingBean;
import org.springframework.context.annotation.Bean;
import org.springframework.context.annotation.Configuration;
import retrofit2.Retrofit;
import retrofit2.adapter.rxjava2.RxJava2CallAdapterFactory;
import retrofit2.converter.jackson.JacksonConverterFactory;
import java.net.Authenticator;
import java.net.InetSocketAddress;
import java.net.Proxy;
import java.util.concurrent.TimeUnit;
import static com.hqy.cloud.chatgpt.common.lang.Constants.DEFAULT_UNOFFICIAL_PROXY_URL;
/**
* OpenAiChatGptAutoConfiguration.
* @author qiyuan.hong
* @version 1.0
* @date 2023/7/27 13:25
*/
@Slf4j
@Configuration
@RequiredArgsConstructor
public class OpenAiChatGptAutoConfiguration {
private final ChatGptConfigurationProperties properties;
@ConditionalOnMissingBean
@Bean(name = "apiOkHttpClient")
public OkHttpClient apiOkHttpClient() {
OkHttpClient.Builder builder = getBuilder();
builder.addInterceptor(new AuthenticationInterceptor(properties.getApiKey()));
//设置代理
settingProxy(builder);
return builder.build();
}
@ConditionalOnMissingBean
@Bean(name = "unofficialProxyOkHttpClient")
public OkHttpClient unofficialProxyOkHttpClient() {
OkHttpClient.Builder builder = getBuilder();
//设置代理
settingProxy(builder);
return builder.build();
}
@Bean
@ConditionalOnMissingBean
public OpenAiService openAiService(OkHttpClient apiOkHttpClient) {
Retrofit retrofit = OpenAiService.defaultRetrofit(apiOkHttpClient, OpenAiService.defaultObjectMapper()).newBuilder()
.baseUrl(properties.getApiBaseUrl()).build();
return new OpenAiService(retrofit.create(OpenAiApi.class), apiOkHttpClient.dispatcher().executorService());
}
/*@Bean
public OpenAiChatgptService openAiChatgptService(OpenAiService openAiService) {
return new OpenAiChatgptServiceImpl(openAiService, properties);
}*/
@Bean
@ConditionalOnMissingBean
public UnofficialApi unofficialApi(OkHttpClient unofficialProxyOkHttpClient) {
Retrofit retrofit = new Retrofit.Builder()
.baseUrl(DEFAULT_UNOFFICIAL_PROXY_URL + StrUtil.SLASH)
.client(unofficialProxyOkHttpClient)
.addConverterFactory(JacksonConverterFactory.create(OpenAiService.defaultObjectMapper()))
.addCallAdapterFactory(RxJava2CallAdapterFactory.create())
.build();
return retrofit.create(UnofficialApi.class);
}
@NotNull
private OkHttpClient.Builder getBuilder() {
ConnectionPool connectionPool = new ConnectionPool(Runtime.getRuntime().availableProcessors(), 1, TimeUnit.MINUTES);
return new OkHttpClient.Builder()
.connectionPool(connectionPool)
.readTimeout(properties.getApiRequestTimeout(), TimeUnit.MILLISECONDS);
}
private void settingProxy(OkHttpClient.Builder builder) {
Proxy proxy = null;
// HTTP代理
ChatGptConfigurationProperties.HttpProxy httpProxy = properties.getHttpProxy();
if (httpProxy != null && httpProxy.isAvailable()) {
proxy = new Proxy(Proxy.Type.HTTP, new InetSocketAddress(httpProxy.getHost(), httpProxy.getPort()));
}
// SOCKS代理
ChatGptConfigurationProperties.SocksProxy socksProxy = properties.getSocksProxy();
if (socksProxy != null && socksProxy.isAvailable()) {
proxy = new Proxy(Proxy.Type.SOCKS, new InetSocketAddress(socksProxy.getHost(), socksProxy.getPort()));
if (!StringUtils.isAllBlank(socksProxy.getPassword(), socksProxy.getUsername())) {
Authenticator.setDefault(new ProxyAuthenticator(socksProxy.getUsername(), socksProxy.getPassword()));
}
}
if (proxy != null) {
builder.proxy(proxy);
log.info("OkHttpClient using proxy: {}.", proxy);
}
}
}
|
[
"com.theokanning.openai.service.OpenAiService.defaultRetrofit"
] |
[((2226, 2383), 'com.theokanning.openai.service.OpenAiService.defaultRetrofit'), ((2226, 2375), 'com.theokanning.openai.service.OpenAiService.defaultRetrofit'), ((2226, 2322), 'com.theokanning.openai.service.OpenAiService.defaultRetrofit')]
|
package com.gltech.myai.service;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.service.OpenAiService;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.stereotype.Service;
import javax.annotation.PostConstruct;
import java.time.Duration;
@Service
public class OpenAiChatService {
private OpenAiService openAiService;
@Value("${openai.api.key}")
private String apiKey;
private String internalContext;
@PostConstruct
public void init() {
this.openAiService = new OpenAiService(apiKey, Duration.ofSeconds(30));
this.internalContext = "";
}
public String sendChatMessage(String prompt) {
internalContext += prompt + "\n";
System.out.println("Context: " + internalContext);
CompletionRequest completionRequest = CompletionRequest.builder()
.model("text-davinci-003")
.prompt(internalContext)
.maxTokens(150)
.build();
String responseText = openAiService.createCompletion(completionRequest).getChoices().get(0).getText();
System.out.println("OpenAI response: " + responseText);
return responseText;
}
public void shutdownOpenAiService() {
openAiService.shutdownExecutor();
}
public void resetInternalContext() {
internalContext = "";
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((867, 1019), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((867, 998), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((867, 970), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((867, 933), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
/**
* Copyright 2021 Rochester Institute of Technology (RIT). Developed with
* government support under contract 70RCSA22C00000008 awarded by the United
* States Department of Homeland Security for Cybersecurity and Infrastructure Security Agency.
*
* Permission is hereby granted, free of charge, to any person obtaining a copy
* of this software and associated documentation files (the “Software”), to deal
* in the Software without restriction, including without limitation the rights
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
* copies of the Software, and to permit persons to whom the Software is
* furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
package edu.rit.se.nvip.characterizer.cwe;
import com.theokanning.openai.OpenAiHttpException;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatCompletionResult;
import com.theokanning.openai.completion.chat.ChatMessage;
import edu.rit.se.nvip.db.model.CompositeVulnerability;
import edu.rit.se.nvip.reconciler.openai.OpenAIRequestHandler;
import edu.rit.se.nvip.reconciler.openai.RequestorIdentity;
import org.apache.logging.log4j.LogManager;
import org.apache.logging.log4j.Logger;
import java.util.ArrayList;
import java.util.HashSet;
import java.util.List;
import java.util.Set;
import java.util.concurrent.ExecutionException;
import java.util.concurrent.Future;
public class ChatGPTProcessor {
private final Logger logger = LogManager.getLogger(getClass().getSimpleName());
private OpenAIRequestHandler requestHandler;
private static final String MODEL = "gpt-3.5-turbo";
private static final double TEMP = 0.0;
private static final String SYS_MESSAGE = String.format("You will be presented with several CWE IDs and their corresponding names followed by a CVE description." +
" Your job is to provide a list, from the CWE IDs given to you, of CWE IDs that have a direct correlation to the CVE Description, based on the CWE ID's corresponding name" +
" if you believe that none of the CWEs have a direct correlation with the provided CVE description respond with \"NONE\" otherwise send " +
"ONLY a comma separated list of CWE Ids that match. If you ever send a CWE's name you have failed your job.");
private static final String SYS_ROLE = "system";
private static final String USER_ROLE = "user";
private Set<String> processedIds = new HashSet<>();
private Set<CWETree> out = new HashSet<>();
private Set<CWETree> matches = new HashSet<>();
private Set<Integer> matchedIds = new HashSet<>();
private int tokenCount = 0;
public ChatGPTProcessor() {
requestHandler = OpenAIRequestHandler.getInstance();
}
public String callModel(String arg) {
try {
ChatCompletionRequest request = formRequest(arg);
Future<ChatCompletionResult> futureRes = requestHandler.createChatCompletion(request, RequestorIdentity.FILTER);
ChatCompletionResult res = futureRes.get();
return res.getChoices().get(0).getMessage().getContent();// Return the obtained result
} catch (OpenAiHttpException | InterruptedException | ExecutionException ex) {
logger.error(ex);
return null;
}
}
private ChatCompletionRequest formRequest(String description) {
List<ChatMessage> messages = formMessages(description);
return ChatCompletionRequest.builder().model(MODEL).temperature(TEMP).n(1).messages(messages).maxTokens(1000).build();
}
private List<ChatMessage> formMessages(String description) {
List<ChatMessage> messages = new ArrayList<>();
messages.add(new ChatMessage(SYS_ROLE, SYS_MESSAGE));
messages.add(new ChatMessage(USER_ROLE, description));
return messages;
}
private Set<String> askChatGPT(Set<CWETree> candidates, CompositeVulnerability vuln){
StringBuilder cwes = new StringBuilder(); //String that will be sent to chat gpt
int count = 1; //count so we can ensure only 5 vulns get sent at a time (attempts to not overwhelm chatgpt)
Set<String> out = new HashSet<>(); //the output set
for (CWETree tree : candidates) {
cwes.append(tree.getRoot().getId()).append(": ").append(tree.getRoot().getName()).append("\n"); //append this string in the form{ 123: Cwe Name, 456: Cwe Name2, ...}
if (count % 5 == 0) { //when 5 vulns are added to the cwe string
String chatMessage = cwes + " \nCVE Description: \n" + vuln.getDescription(); //create the message to send to chat gpt
logger.info(chatMessage);
String msg = callModel(chatMessage); //call chatgpt
out.addAll(getIdsFromResponse(msg)); //add a set of ids from chat gpt to the output set
cwes = new StringBuilder(); //clear out previous cwes
}
count++;
}
if (cwes.length() > 0){ //case for if there are 4-1 vulns left... AKA cwes.length is only zero if there are no CWEs left
String chatMessage = cwes + " CVE Description: " + vuln.getDescription(); //create message to send to chat gpt
logger.info(chatMessage);
String finalRun = callModel(chatMessage); //send it
out.addAll(getIdsFromResponse(finalRun)); //add the response to the list of outputs
}
return out;
}
private Set<CWETree> parseResponse(Set<CWETree> candidates, Set<String> response){
Set<CWETree> set = new HashSet<>();
if(response.contains("NONE") || response.isEmpty()){
return set;
}
for (String id : response){ //for each id
for(CWETree cweTree : candidates){ //for each candidate id
try {
if(id.equals("NONE") || id.equals("")) continue;
if (cweTree.getRoot().getId() == Integer.parseInt(id)) { //if the root id matches the id present then add the tree to the set of trees
set.add(cweTree);
}
}catch(NumberFormatException e){
logger.error("Wrong format: {}", id); //in case chatgpt sends some weird format
break;
}
}
}
return set;
}
private Set<CWETree> whichMatchHelper(Set<CWETree> candidates, CompositeVulnerability vuln) {
if (candidates.isEmpty()) { //if candidates is empty return a new set
return new HashSet<>();
}
Set<String> response = askChatGPT(candidates, vuln); //ask chatgpt what candidates might be related to the cve
Set<String> filteredResponse = new HashSet<>();
if(!response.isEmpty()) {
for (String id : response) {
if(id.equals("NONE") || id.equals("")){
break;
}
if (!processedIds.contains(id)) { //keeps repeats from being sent
processedIds.add(id);
filteredResponse.add(id);
}
}
}
matches.addAll(parseResponse(candidates, filteredResponse)); //parse chatgpt's response
List<CWETree> treesToProcess = new ArrayList<>(matches);
for (CWETree match : treesToProcess) {
if(!matchedIds.contains(match.getRoot().getId())) {
matchedIds.add(match.getRoot().getId());
out.add(match);
whichMatchHelper(match.getSubtrees(), vuln);
}
}
return out;
}
public Set<CWE> assignCWEs(CompositeVulnerability vuln) {
CWEForest forest = new CWEForest(); // builds the forest
Set<CWETree> trees = whichMatchHelper(forest.getTrees(), vuln); //gets trees related to vuln
logger.info("trees size: " + trees.size());
Set<CWE> out = new HashSet<>();
for (CWETree tree : trees) {
out.add(tree.getRoot());
}
return out;
}
private Set<String> getIdsFromResponse(String response) {
String[] parts = response.split(","); //split the string by commas (the response string will look like{ 123,456,789 }
Set<String> out = new HashSet<>(); //output set
for (String part : parts) { //for each id
String[] finalParts = part.split("CWE-"); //split one more time (occasionally chatgpt will send {CWE-123,CWE-456} instead so this accounts for that)
for (String finalPart : finalParts){ //for finalPart or the final ID
String trimmedPart = finalPart.trim(); //trim it
if (trimmedPart.equals("")) continue;
out.add(trimmedPart); //add the trimmed part to the list
}
}
return out;
}
public static boolean isInt(String input) {
try {
Integer.parseInt(input);
return true;
} catch (NumberFormatException e) {
return false;
}
}
}
|
[
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((4103, 4213), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((4103, 4205), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((4103, 4189), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((4103, 4170), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((4103, 4165), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((4103, 4147), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
package com.qStivi.chatbot;
import com.qStivi.Main;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.messages.MessageRequest;
import com.theokanning.openai.runs.Run;
import com.theokanning.openai.runs.RunCreateRequest;
import com.theokanning.openai.service.OpenAiService;
import com.theokanning.openai.threads.ThreadRequest;
import net.dv8tion.jda.api.entities.ISnowflake;
import net.dv8tion.jda.api.entities.RichPresence;
import net.dv8tion.jda.api.entities.User;
import net.dv8tion.jda.internal.utils.JDALogger;
import org.jetbrains.annotations.NotNull;
import org.slf4j.Logger;
import java.io.File;
import java.io.FileNotFoundException;
import java.io.FileReader;
import java.io.IOException;
import java.util.List;
import java.util.Random;
import java.util.Set;
import java.util.concurrent.CopyOnWriteArrayList;
public class OpenAIChatBot implements IChatBot {
private static final Logger LOGGER = JDALogger.getLog(OpenAIChatBot.class);
private final static Random RANDOM = new Random();
private final OpenAiService service;
private final List<String> moods = List.of("(Bitte antworte auf eine freche Art und Weise)", "(Bitte antworte sarkastisch)", "(Führe mich mit deiner Antwort an der Nase herum)", "(Bitte antworte wütend)", "(Die Blähungen setzen ein!)");
private static final String ASSISTANT_ID = "asst_DY0gEKhlE9E4pdQvmQhF2Qxq";
private final String threadID;
public OpenAIChatBot(String apiKey) {
this.service = new OpenAiService(apiKey);
threadID = getThreadID();
// addToMemory(new ChatMessage(ChatMessageRole.SYSTEM.value(), INITIAL_MESSAGE));
}
/**
* Retrieves the thread ID from the file "threadID.txt" in the current working directory.
* @return The thread ID.
*/
private String getThreadID() {
try {
var fileReader = new FileReader("threadID.txt");
var bufferedReader = new java.io.BufferedReader(fileReader);
return bufferedReader.readLine();
} catch (FileNotFoundException ignored) {
LOGGER.warn("File threadID.txt not found!");
var id = service.createThread(ThreadRequest.builder().build()).getId();
LOGGER.info("Creating file threadID.txt with ID: {}", id);
try {
var fileWriter = new java.io.FileWriter("threadID.txt");
fileWriter.write(id);
fileWriter.close();
return id;
} catch (IOException e) {
e.printStackTrace();
}
} catch (IOException e) {
e.printStackTrace();
}
return null;
}
private void addMoodPrompt() {
if (RANDOM.nextInt(10) < 10) { // 100% chance to append a keyword
var message = moods.get(RANDOM.nextInt(moods.size()));
message += " (If you can't do anything with this message, just ignore it)";
// addToMemory(new ChatMessage(ChatMessageRole.SYSTEM.value(), message));
}
}
@Override
public String sendMessage(String message) {
return sendMessage(ChatMessageRole.USER, message);
}
@Override
public String sendMessage(String message, Set<User> speakers) {
// addToMemory(createChatMessage(ChatMessageRole.SYSTEM, String.format("Currently talking are %s", speakers.stream().map(User::getName).reduce((a, b) -> a + ", " + b).orElse("no one"))));
return sendMessage(ChatMessageRole.USER, message);
}
@Override
public String sendMessage(ChatMessageRole role, String message) {
addUserActivityInformation();
addMoodPrompt();
// addToMemory(createChatMessage(role, message));
var botMessage = getResponse(message);
// purgeMemory();
return botMessage;
}
private String getResponse(String message) {
service.createMessage(threadID, MessageRequest.builder().role("user").content(message).build());
var run = service.createRun(threadID, RunCreateRequest.builder().assistantId(ASSISTANT_ID).build());
waitForRunToComplete(run);
var response = getNewestMessageInThread();
LOGGER.info("Response: {}", response);
return response.toString();
}
@NotNull
private StringBuilder getNewestMessageInThread() {
var response = new StringBuilder();
var messages = service.listMessages(threadID);
var retrieveMessage = service.retrieveMessage(threadID, messages.getFirstId());
for (var messageRequest : retrieveMessage.getContent()) {
response.append(messageRequest.getText().getValue()).append("\n");
}
return response;
}
private void waitForRunToComplete(Run run) {
var lastRunStatus = "";
while (run.getStatus() == null || !run.getStatus().equals("completed")) {
if (run.getStatus() != null && !run.getStatus().equals(lastRunStatus)) {
LOGGER.info("Run status: {}", run.getStatus());
lastRunStatus = run.getStatus();
}
run = service.retrieveRun(threadID, run.getId());
}
}
private void addUserActivityInformation() {
// Add user activity of all users in channel to memory
StringBuilder userActivities = new StringBuilder();
Main.jda.getGuilds().forEach(guild -> guild.getVoiceChannels().forEach(voiceChannel -> {
if (voiceChannel.getMembers().stream().map(ISnowflake::getId).anyMatch(id -> id.equals(Main.jda.getSelfUser().getId()))) {
voiceChannel.getMembers().forEach(member -> {
if (!member.getActivities().isEmpty()) {
var activities = member.getActivities();
for (var activity : activities) {
RichPresence richPresence = activity.asRichPresence();
String details = richPresence != null ? richPresence.getDetails() : "";
String state = activity.getState();
if (state == null) {
state = "";
} else {
state = state + " on ";
}
userActivities.append("\n").append(member.getEffectiveName()) // qStivi
.append(" is ") // is
.append(activity.getType().name()).append(" ") // listening
.append(details).append(" ") // beat it
.append(state) // Michael Jackson on
.append(activity.getName()); // Spotify
}
}
});
}
}));
String activityMessage = """
(Du sieht, die Discord activity der Benutzer mit denen du chattest.
""" + userActivities + ")";
// addToMemory(createChatMessage(ChatMessageRole.SYSTEM, activityMessage));
}
}
|
[
"com.theokanning.openai.threads.ThreadRequest.builder",
"com.theokanning.openai.runs.RunCreateRequest.builder",
"com.theokanning.openai.messages.MessageRequest.builder"
] |
[((2238, 2269), 'com.theokanning.openai.threads.ThreadRequest.builder'), ((3980, 4042), 'com.theokanning.openai.messages.MessageRequest.builder'), ((3980, 4034), 'com.theokanning.openai.messages.MessageRequest.builder'), ((3980, 4017), 'com.theokanning.openai.messages.MessageRequest.builder'), ((4092, 4152), 'com.theokanning.openai.runs.RunCreateRequest.builder'), ((4092, 4144), 'com.theokanning.openai.runs.RunCreateRequest.builder'), ((5410, 6915), 'com.qStivi.Main.jda.getGuilds'), ((5410, 5430), 'com.qStivi.Main.jda.getGuilds'), ((5598, 5628), 'com.qStivi.Main.jda.getSelfUser'), ((5598, 5620), 'com.qStivi.Main.jda.getSelfUser')]
|
package com.example.wpfsboot.controller;
import com.example.wpfsboot.common.Constants;
import com.example.wpfsboot.common.Result;
import com.example.wpfsboot.entity.GPTParams;
import com.jcraft.jsch.Channel;
import com.jcraft.jsch.ChannelExec;
import com.jcraft.jsch.JSch;
import com.jcraft.jsch.Session;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.embedding.EmbeddingRequest;
import com.theokanning.openai.finetune.FineTuneRequest;
import com.theokanning.openai.image.CreateImageEditRequest;
import com.theokanning.openai.image.CreateImageVariationRequest;
import com.theokanning.openai.image.ImageResult;
import com.theokanning.openai.moderation.ModerationRequest;
import io.github.asleepyfish.config.ChatGPTProperties;
import io.github.asleepyfish.entity.billing.Billing;
import io.github.asleepyfish.entity.billing.Subscription;
import io.github.asleepyfish.enums.audio.AudioResponseFormatEnum;
import io.github.asleepyfish.enums.edit.EditModelEnum;
import io.github.asleepyfish.enums.embedding.EmbeddingModelEnum;
import io.github.asleepyfish.enums.image.ImageResponseFormatEnum;
import io.github.asleepyfish.enums.image.ImageSizeEnum;
import io.github.asleepyfish.service.OpenAiProxyService;
import io.github.asleepyfish.util.OpenAiUtils;
import io.swagger.annotations.ApiOperation;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.http.HttpHeaders;
import org.springframework.http.HttpStatus;
import org.springframework.http.MediaType;
import org.springframework.http.ResponseEntity;
import org.springframework.web.bind.annotation.*;
import javax.servlet.http.HttpServletResponse;
import java.io.*;
import java.net.MalformedURLException;
import java.nio.file.Files;
import java.nio.file.Path;
import java.nio.file.Paths;
import java.util.Arrays;
import java.util.List;
import org.springframework.core.io.Resource;
import org.springframework.core.io.UrlResource;
import org.springframework.http.HttpHeaders;
import org.springframework.http.MediaType;
import org.springframework.http.ResponseEntity;
import org.springframework.stereotype.Controller;
import org.springframework.web.bind.annotation.GetMapping;
import org.springframework.web.bind.annotation.PathVariable;
import java.io.IOException;
import java.net.MalformedURLException;
import java.nio.file.Path;
import java.nio.file.Paths;
/**
* @Author: 结束乐队
* @Date: 2023/7/18
*/
@RestController
@RequestMapping("/wpfgpt")
public class ChatGPTController {
@Value("${files.upload.path}")
private String fileUploadPath;
@Value("${server.ip}")
private String serverIp;
@Value("${server.port}")
private String serverPort;
@Value(("${server.password}"))
private String serverPassword;
@Value("${chatgpt.proxy-host}")
private String proxyPort;
@Autowired
private OpenAiUtils openAiUtils;
@Value("${cmd.start}")
private String cmdStart;
/**
* @param question
* @return
*/
public static int containsCode(String question) {
String[] keyword01 = {"图"};
String[] keyword02 = {"报表"};
for (String keyword : keyword01) {
if (question.contains(keyword)) {
return 1;
}
}
for (String keyword : keyword02) {
if (question.contains(keyword)) {
return 2;
}
}
return 3;
}
/**
* @param question
* @return
*/
public static boolean containsJudge(String question) {
String[] keywords = {"预处理后"};
for (String keyword : keywords) {
if (question.contains(keyword)) {
return true;
}
}
return false;
}
/**
* @param originalString
* @param startMarker
* @param endMarker
* @return
*/
public static String extractBetweenMarkers(String originalString, String startMarker, String endMarker) {
int startIndex = originalString.indexOf(startMarker);
int endIndex = originalString.lastIndexOf(endMarker);
if (startIndex != -1 && endIndex != -1 && startIndex < endIndex) {
// 使用substring方法提取两个标记之间的部分
return originalString.substring(startIndex + startMarker.length(), endIndex);
} else {
return ""; // 如果没有找到匹配的标记则返回空字符串
}
}
public static void writePythonCodeToFile(String pythonCode, String filePath) throws IOException {
File file = new File(filePath);
BufferedWriter writer = null;
try {
writer = new BufferedWriter(new FileWriter(file));
writer.write(pythonCode);
} finally {
if (writer != null) {
writer.close();
}
}
}
private String IMAGE_PATH = fileUploadPath + "/gpt_output/img/"; // 图片文件所在路径
@GetMapping("/api/images/{fileName}")
public ResponseEntity<byte[]> getImage(@PathVariable String fileName) {
try {
String imagePath = fileUploadPath + "/gpt_output/img/" + fileName.replace(".csv", ".png");
System.out.println("imagePath: " + imagePath);
byte[] imageData = readImageData(imagePath);
if (imageData != null) {
HttpHeaders headers = new HttpHeaders();
headers.setContentType(MediaType.IMAGE_JPEG); // 设置图片类型,可以根据实际情况调整
return new ResponseEntity<>(imageData, headers, HttpStatus.OK);
} else {
return new ResponseEntity<>(HttpStatus.NOT_FOUND); // 图片不存在
}
} catch (Exception e) {
// 处理异常
return new ResponseEntity<>(HttpStatus.INTERNAL_SERVER_ERROR);
}
}
// 根据图片路径读取图片数据的方法
private byte[] readImageData(String imagePath) {
try {
Path path = Paths.get(imagePath);
return Files.readAllBytes(path);
} catch (IOException e) {
// 处理读取文件异常
e.printStackTrace();
return null;
}
}
private String docxDirectory = fileUploadPath + "/gpt_output/docx"; // 指定存放 .docx 文件的目录
@GetMapping("/docx/{filename}")
public ResponseEntity<Resource> downloadDocx(@PathVariable String filename) throws MalformedURLException {
System.out.println("docxDirectory: " + fileUploadPath + "/gpt_output/docx");
Path filePath = Paths.get(fileUploadPath + "/gpt_output/docx").resolve(filename);
System.out.println("filePath: " + filePath);
Resource resource = new UrlResource(filePath.toUri());
return ResponseEntity.ok()
.header(HttpHeaders.CONTENT_DISPOSITION, "attachment; filename=\"" + resource.getFilename() + "\"")
.contentType(MediaType.APPLICATION_OCTET_STREAM)
.body(resource);
}
/**
* 问答
*
* @param gptParams 问题
* @return 答案
*/
@PostMapping("/postChat2")
public ResponseEntity<Result> postChat2(@RequestBody GPTParams gptParams) {
Result result = new Result();
//时间戳
String time = System.currentTimeMillis() + "";
result.setTime(time);
// 问题预处理
String question = gptParams.getQuestion();
question = question.replace("ROUND(A.WS,1)", "AWS").replace("ROUND(A.POWER,0)", "APOWER");
String fileName = gptParams.getFileName();
// 判断问题类型
int tag1 = containsCode(question);
boolean tag2 = containsJudge(question);
String filePath = tag2 ? fileUploadPath + "/outfile/" + fileName : fileUploadPath + "/pred/" + fileName;
if (tag1 == 3) {
result.setImage(false);
System.out.println(question);
result.setMsg(OpenAiUtils.createChatCompletion(question).get(0));
} else if (tag1 == 1) {
// 判断是否要展示图片
result.setImage(true);
if (tag2) {
question = question.replace("预处理后", "");
} else {
question = question.replace("预测后", "");
}
String text = "我有一个csv文件,位置在" + filePath + ",第一行是列名,第二行开始是数据,其中列名有:DATATIME,WINDSPEED,PREPOWER,WINDDIRECTION,TEMPERATURE,HUMIDITY,PRESSURE,AWS,APOWER,YD15,";
question = text + question
+ ", 图片输出至" + fileUploadPath + "/gpt_output/img/,"
+ "图片名为:" + time + ".png,"
+ "请给出完整的Python代码, 默认我已经安装了所有需要的包。plt.show()不需要,且我只需要你返回给我一个可以执行的完整代码段。整个代码段用markdown中的```包围";
// question = text + question
// + ", 图片输出至/home/wpfs/algorithm/submission75254/gpt_output/img/,"
// + "图片名为:" + time + ".png,"
// + "请给出完整的Python代码, 默认我已经安装了所有需要的包。且我只需要你返回给我一个可以执行的完整代码段。并且注释部分用4个#作为前缀";
System.out.println(question);
result.setMsg(OpenAiUtils.createChatCompletion(question).get(0));
System.out.println(result.getMsg());
String pythonCode = extractBetweenMarkers(result.getMsg(), "```python", "```");
// System.out.println(pythonCode);
String pyFileName = fileName.replace(".csv", ".py");
String pyFilePath = fileUploadPath + "/gpt_output/py/" + pyFileName;
try {
writePythonCodeToFile(pythonCode, pyFilePath);
System.out.println("Python file generated successfully at: " + pyFilePath);
} catch (IOException e) {
System.err.println("Error while generating Python file: " + e.getMessage());
}
String host = serverIp; // 远程服务器IP地址
String user = "root"; // 远程服务器用户名
String password = serverPassword; // 远程服务器密码
if (host.equals("localhost")) {
String command = "conda activate py37 && cd " + fileUploadPath + "/gpt_output/py/ && python ./" + pyFileName;
System.out.println(command);
try {
ProcessBuilder processBuilder = new ProcessBuilder("cmd.exe", "/c", command);
processBuilder.redirectErrorStream(true);
Process process = processBuilder.start();
InputStream inputStream = process.getInputStream();
BufferedReader reader = new BufferedReader(new InputStreamReader(inputStream));
String line;
while ((line = reader.readLine()) != null) {
System.out.println(line); // 输出结果到控制台
}
int exitCode = process.waitFor(); // 等待命令执行完成
System.out.println("exit-status: " + exitCode); // 输出退出状态
} catch (Exception e) {
e.printStackTrace(); // 输出错误信息
}
} else {
// 要执行的命令
StringBuilder command = new StringBuilder("conda activate py37;cd " + fileUploadPath + "/gpt_output/py/;python ./" + pyFileName + ";");
try {
JSch jsch = new JSch();
Session session = jsch.getSession(user, host, 22); // 创建一个SSH会话
session.setPassword(password); // 设置会话密码
session.setConfig("StrictHostKeyChecking", "no"); // 设置会话配置,不检查HostKey
session.connect(); // 连接会话
Channel channel = session.openChannel("exec"); // 打开一个exec通道
((ChannelExec) channel).setCommand(command.toString()); // 设置要执行的命令
channel.setInputStream(null);
((ChannelExec) channel).setErrStream(System.err); // 设置错误输出流
InputStream inputStream = channel.getInputStream();
channel.connect(); // 连接通道
byte[] buffer = new byte[1024];
while (true) {
while (inputStream.available() > 0) {
int i = inputStream.read(buffer, 0, 1024);
if (i < 0) {
break;
}
System.out.print(new String(buffer, 0, i)); // 输出结果到控制台
}
if (channel.isClosed()) {
if (inputStream.available() > 0) {
continue;
}
System.out.println("exit-status: " + channel.getExitStatus()); // 输出退出状态
break;
}
try {
Thread.sleep(1000);
} catch (Exception ee) {
} // 等待一秒钟
}
channel.disconnect(); // 断开通道
session.disconnect(); // 断开会话
} catch (Exception e) {
e.printStackTrace(); // 输出错误信息
}
}
} else {
// 判断是否要展示报表
result.setReport(true);
result.setMsg("报表生成完成");
String tag;
if (tag2) {
question = question.replace("预处理后", "");
tag = "outfile";
} else {
question = question.replace("预测后", "");
tag = "pred";
}
// TODO 数据报表造假
String host = serverIp; // 远程服务器IP地址
String user = "root"; // 远程服务器用户名
String password = serverPassword; // 远程服务器密码
String mdFilePath = fileUploadPath + "/gpt_output/markdown/" + fileName.replace(".csv", "_") + time + ".md";
String docxFilePath = fileUploadPath + "/gpt_output/docx/" + fileName.replace(".csv", "_") + time + ".docx";
String pdfFilePath = fileUploadPath + "/gpt_output/pdf/" + fileName.replace(".csv", "_") + time + ".pdf";
if (serverIp.equals("localhost")) {
String command = "conda activate py37 && cd " + fileUploadPath + "/gpt_output/ && python ./report_local.py --file_name " + fileName + " --time_stamp " + time + " --tag " + tag + " && " + "cd markdown && pandoc " + mdFilePath + " -o " + docxFilePath;
try {
ProcessBuilder processBuilder = new ProcessBuilder("cmd.exe", "/c", command);
processBuilder.redirectErrorStream(true);
Process process = processBuilder.start();
InputStream inputStream = process.getInputStream();
BufferedReader reader = new BufferedReader(new InputStreamReader(inputStream));
String line;
while ((line = reader.readLine()) != null) {
System.out.println(line); // 输出结果到控制台
}
int exitCode = process.waitFor(); // 等待命令执行完成
System.out.println("exit-status: " + exitCode); // 输出退出状态
} catch (Exception e) {
e.printStackTrace(); // 输出错误信息
}
} else {
// 要执行的命令
StringBuilder command = new StringBuilder("conda activate py37;cd " + fileUploadPath + "/gpt_output/;python ./report.py --file_name " + fileName + " --time_stamp " + time + " --tag " + tag + ";" + "cd markdown;pandoc " + mdFilePath + " -o " + docxFilePath + ";");
System.out.println("command: " + command.toString());
try {
JSch jsch = new JSch();
Session session = jsch.getSession(user, host, 22); // 创建一个SSH会话
session.setPassword(password); // 设置会话密码
session.setConfig("StrictHostKeyChecking", "no"); // 设置会话配置,不检查HostKey
session.connect(); // 连接会话
Channel channel = session.openChannel("exec"); // 打开一个exec通道
((ChannelExec) channel).setCommand(command.toString()); // 设置要执行的命令
channel.setInputStream(null);
((ChannelExec) channel).setErrStream(System.err); // 设置错误输出流
InputStream inputStream = channel.getInputStream();
channel.connect(); // 连接通道
byte[] buffer = new byte[1024];
while (true) {
while (inputStream.available() > 0) {
int i = inputStream.read(buffer, 0, 1024);
if (i < 0) {
break;
}
System.out.print(new String(buffer, 0, i)); // 输出结果到控制台
}
if (channel.isClosed()) {
if (inputStream.available() > 0) {
continue;
}
System.out.println("exit-status: " + channel.getExitStatus()); // 输出退出状态
break;
}
try {
Thread.sleep(1000);
} catch (Exception ee) {
} // 等待一秒钟
}
channel.disconnect(); // 断开通道
session.disconnect(); // 断开会话
} catch (Exception e) {
e.printStackTrace(); // 输出错误信息
}
}
}
result.setCode(Constants.CODE_200);
return new ResponseEntity<>(result, HttpStatus.OK);
}
/**
* 问答
*
* @param question 问题
* @return 答案
*/
@PostMapping("/postChat")
public ResponseEntity<Result> postChat(@RequestBody String question) {
Result result = new Result();
result.setMsg(OpenAiUtils.createChatCompletion(question).get(0));
result.setCode(Constants.CODE_200);
return new ResponseEntity<>(result, HttpStatus.OK);
}
/**
* 问答
*
* @param content 问题
* @return 答案
*/
@GetMapping("/getChat")
public List<String> getChat(String content) {
return OpenAiUtils.createChatCompletion(content);
}
/**
* 流式问答,返回到控制台
*/
@GetMapping("/streamChat")
public void streamChat(String content) {
// OpenAiUtils.createStreamChatCompletion(content, System.out);
// 下面的默认和上面这句代码一样,是输出结果到控制台
OpenAiUtils.createStreamChatCompletion(content);
}
/**
* 流式问答,输出结果到WEB浏览器端
*/
@GetMapping("/streamChatWithWeb")
public void streamChatWithWeb(String content, HttpServletResponse response) throws IOException, InterruptedException {
// 需要指定response的ContentType为流式输出,且字符编码为UTF-8
response.setContentType("text/event-stream");
response.setCharacterEncoding("UTF-8");
// 禁用缓存
response.setHeader("Cache-Control", "no-cache");
OpenAiUtils.createStreamChatCompletion(content, response.getOutputStream());
}
/**
* 生成图片
*
* @param prompt 图片描述
*/
@PostMapping("/createImage")
public void createImage(String prompt) {
System.out.println(OpenAiUtils.createImage(prompt));
}
/**
* 下载图片
*/
@GetMapping("/downloadImage")
public void downloadImage(String prompt, HttpServletResponse response) {
OpenAiUtils.downloadImage(prompt, response);
}
@PostMapping("/billing")
public void billing() {
String monthUsage = OpenAiUtils.billingUsage("2023-04-01", "2023-05-01");
System.out.println("四月使用:" + monthUsage + "美元");
String totalUsage = OpenAiUtils.billingUsage();
System.out.println("一共使用:" + totalUsage + "美元");
String stageUsage = OpenAiUtils.billingUsage("2023-01-31");
System.out.println("自从2023/01/31使用:" + stageUsage + "美元");
Subscription subscription = OpenAiUtils.subscription();
System.out.println("订阅信息(包含到期日期,账户总额度等信息):" + subscription);
// dueDate为到期日,total为总额度,usage为使用量,balance为余额
Billing totalBilling = OpenAiUtils.billing();
System.out.println("历史账单信息:" + totalBilling);
// 默认不传参的billing方法的使用量usage从2023-01-01开始,如果用户的账单使用早于该日期,可以传入开始日期startDate
Billing posibleStartBilling = OpenAiUtils.billing("2022-01-01");
System.out.println("可能的历史账单信息:" + posibleStartBilling);
}
/**
* 自定义Token使用(解决单个SpringBoot项目中只能指定唯一的Token[sk-xxxxxxxxxxxxx]的问题,现在可以自定义ChatGPTProperties内容,添加更多的Token实例)
*/
@PostMapping("/customToken")
public void customToken() {
ChatGPTProperties properties = ChatGPTProperties.builder().token("sk-002xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx")
.proxyHost(serverIp)
.proxyHost(proxyPort)
.build();
OpenAiProxyService openAiProxyService = new OpenAiProxyService(properties);
// 直接使用new出来的openAiProxyService来调用方法,每个OpenAiProxyService都拥有自己的Token。
// 这样在一个SpringBoot项目中,就可以有多个Token,可以有更多的免费额度供使用了
openAiProxyService.createStreamChatCompletion("Java的三大特性是什么");
}
@PostMapping("/models")
public void models() {
System.out.println("models列表:" + OpenAiUtils.listModels());
System.out.println("=============================================");
System.out.println("text-davinci-003信息:" + OpenAiUtils.getModel("text-davinci-003"));
}
/**
* 编辑
*/
@PostMapping("/edit")
public void edit() {
String input = "What day of the wek is it?";
String instruction = "Fix the spelling mistakes";
System.out.println("编辑前:" + input);
// 下面这句和OpenAiUtils.edit(input, instruction, EditModelEnum.TEXT_DAVINCI_EDIT_001);是一样的,默认使用模型TEXT_DAVINCI_EDIT_001
System.out.println("编辑后:" + OpenAiUtils.edit(input, instruction));
System.out.println("=============================================");
input = " public static void mian(String[] args) {\n" +
" system.in.println(\"hello world\");\n" +
" }";
instruction = "Fix the code mistakes";
System.out.println("修正代码前:\n" + input);
System.out.println("修正代码后:\n" + OpenAiUtils.edit(input, instruction, EditModelEnum.CODE_DAVINCI_EDIT_001));
}
@PostMapping("/embeddings")
public void embeddings() {
String text = "Once upon a time";
System.out.println("文本:" + text);
System.out.println("文本的嵌入向量:" + OpenAiUtils.embeddings(text));
System.out.println("=============================================");
String[] texts = {"Once upon a time", "There was a princess"};
System.out.println("文本数组:" + Arrays.toString(texts));
EmbeddingRequest embeddingRequest = EmbeddingRequest.builder()
.model(EmbeddingModelEnum.TEXT_EMBEDDING_ADA_002.getModelName()).input(Arrays.asList(texts)).build();
System.out.println("文本数组的嵌入向量:" + OpenAiUtils.embeddings(embeddingRequest));
}
@PostMapping("/transcription")
public void transcription() {
String filePath = "src/main/resources/audio/想象之中-许嵩.mp3";
System.out.println("语音文件转录后的text文本是:" + OpenAiUtils.transcription(filePath, AudioResponseFormatEnum.TEXT));
// File file = new File("src/main/resources/audio/想象之中-许嵩.mp3");
// System.out.println("语音文件转录后的text文本是:" + OpenAiUtils.transcription(file, AudioResponseFormatEnum.TEXT));
}
@PostMapping("/translation")
public void translation() {
String filePath = "src/main/resources/audio/想象之中-许嵩.mp3";
System.out.println("语音文件翻译成英文后的text文本是:" + OpenAiUtils.translation(filePath, AudioResponseFormatEnum.TEXT));
// File file = new File("src/main/resources/audio/想象之中-许嵩.mp3");
// System.out.println("语音文件翻译成英文后的text文本是:" + OpenAiUtils.translation(file, AudioResponseFormatEnum.TEXT));
}
@PostMapping("/createImageEdit")
public void createImageEdit() {
CreateImageEditRequest createImageEditRequest = CreateImageEditRequest.builder().prompt("Background changed to white")
.n(1).size(ImageSizeEnum.S512x512.getSize()).responseFormat(ImageResponseFormatEnum.URL.getResponseFormat()).build();
ImageResult imageEdit = OpenAiUtils.createImageEdit(createImageEditRequest, "src/main/resources/image/img.png", "src/main/resources/image/mask.png");
System.out.println("图片编辑结果:" + imageEdit);
}
@PostMapping("/createImageVariation")
public void createImageVariation() {
CreateImageVariationRequest createImageVariationRequest = CreateImageVariationRequest.builder()
.n(2).size(ImageSizeEnum.S512x512.getSize()).responseFormat(ImageResponseFormatEnum.URL.getResponseFormat()).build();
ImageResult imageVariation = OpenAiUtils.createImageVariation(createImageVariationRequest, "src/main/resources/image/img.png");
System.out.println("图片变体结果:" + imageVariation);
}
/**
* 文件操作(下面文件操作入参,用户可根据实际情况自行补全)
*/
@PostMapping("/files")
public void files() {
// 上传文件
System.out.println("上传文件信息:" + OpenAiUtils.uploadFile("", ""));
// 获取文件列表
System.out.println("文件列表:" + OpenAiUtils.listFiles());
// 获取文件信息
System.out.println("文件信息:" + OpenAiUtils.retrieveFile(""));
// 获取文件内容
System.out.println("文件内容:" + OpenAiUtils.retrieveFileContent(""));
// 删除文件
System.out.println("删除文件信息:" + OpenAiUtils.deleteFile(""));
}
@PostMapping("/fileTune")
public void fileTune() {
// 创建微调
FineTuneRequest fineTuneRequest = FineTuneRequest.builder().trainingFile("").build();
System.out.println("创建微调信息:" + OpenAiUtils.createFineTune(fineTuneRequest));
// 创建微调完成
CompletionRequest completionRequest = CompletionRequest.builder().build();
System.out.println("创建微调完成信息:" + OpenAiUtils.createFineTuneCompletion(completionRequest));
// 获取微调列表
System.out.println("获取微调列表:" + OpenAiUtils.listFineTunes());
// 获取微调信息
System.out.println("获取微调信息:" + OpenAiUtils.retrieveFineTune(""));
// 取消微调
System.out.println("取消微调信息:" + OpenAiUtils.cancelFineTune(""));
// 列出微调事件
System.out.println("列出微调事件:" + OpenAiUtils.listFineTuneEvents(""));
// 删除微调
System.out.println("删除微调信s息:" + OpenAiUtils.deleteFineTune(""));
}
@PostMapping("/moderation")
public void moderation() {
// 创建moderation
ModerationRequest moderationRequest = ModerationRequest.builder().input("I want to kill them.").build();
System.out.println("创建moderation信息:" + OpenAiUtils.createModeration(moderationRequest));
}
}
|
[
"com.theokanning.openai.moderation.ModerationRequest.builder",
"com.theokanning.openai.completion.CompletionRequest.builder",
"com.theokanning.openai.image.CreateImageVariationRequest.builder",
"com.theokanning.openai.finetune.FineTuneRequest.builder",
"com.theokanning.openai.embedding.EmbeddingRequest.builder",
"com.theokanning.openai.image.CreateImageEditRequest.builder"
] |
[((6691, 6755), 'java.nio.file.Paths.get'), ((6890, 7122), 'org.springframework.http.ResponseEntity.ok'), ((6890, 7090), 'org.springframework.http.ResponseEntity.ok'), ((6890, 7025), 'org.springframework.http.ResponseEntity.ok'), ((8072, 8121), 'io.github.asleepyfish.util.OpenAiUtils.createChatCompletion'), ((9556, 9605), 'io.github.asleepyfish.util.OpenAiUtils.createChatCompletion'), ((19097, 19146), 'io.github.asleepyfish.util.OpenAiUtils.createChatCompletion'), ((22380, 22559), 'io.github.asleepyfish.config.ChatGPTProperties.builder'), ((22380, 22534), 'io.github.asleepyfish.config.ChatGPTProperties.builder'), ((22380, 22496), 'io.github.asleepyfish.config.ChatGPTProperties.builder'), ((22380, 22459), 'io.github.asleepyfish.config.ChatGPTProperties.builder'), ((24754, 24897), 'com.theokanning.openai.embedding.EmbeddingRequest.builder'), ((24754, 24889), 'com.theokanning.openai.embedding.EmbeddingRequest.builder'), ((24754, 24861), 'com.theokanning.openai.embedding.EmbeddingRequest.builder'), ((24804, 24860), 'io.github.asleepyfish.enums.embedding.EmbeddingModelEnum.TEXT_EMBEDDING_ADA_002.getModelName'), ((26186, 26389), 'com.theokanning.openai.image.CreateImageEditRequest.builder'), ((26186, 26381), 'com.theokanning.openai.image.CreateImageEditRequest.builder'), ((26186, 26317), 'com.theokanning.openai.image.CreateImageEditRequest.builder'), ((26186, 26278), 'com.theokanning.openai.image.CreateImageEditRequest.builder'), ((26186, 26256), 'com.theokanning.openai.image.CreateImageEditRequest.builder'), ((26284, 26316), 'io.github.asleepyfish.enums.image.ImageSizeEnum.S512x512.getSize'), ((26333, 26380), 'io.github.asleepyfish.enums.image.ImageResponseFormatEnum.URL.getResponseFormat'), ((26770, 26940), 'com.theokanning.openai.image.CreateImageVariationRequest.builder'), ((26770, 26932), 'com.theokanning.openai.image.CreateImageVariationRequest.builder'), ((26770, 26868), 'com.theokanning.openai.image.CreateImageVariationRequest.builder'), ((26770, 26829), 'com.theokanning.openai.image.CreateImageVariationRequest.builder'), ((26835, 26867), 'io.github.asleepyfish.enums.image.ImageSizeEnum.S512x512.getSize'), ((26884, 26931), 'io.github.asleepyfish.enums.image.ImageResponseFormatEnum.URL.getResponseFormat'), ((27990, 28040), 'com.theokanning.openai.finetune.FineTuneRequest.builder'), ((27990, 28032), 'com.theokanning.openai.finetune.FineTuneRequest.builder'), ((28217, 28252), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((29087, 29152), 'com.theokanning.openai.moderation.ModerationRequest.builder'), ((29087, 29144), 'com.theokanning.openai.moderation.ModerationRequest.builder')]
|
package BackendProg.BookAppWeb.util;
import java.time.Duration;
import java.util.ArrayList;
import java.util.List;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
public class OpenAiComms {
private OpenAiService remoteAiService;
public OpenAiComms() {
// System.out.println("Openai key: " + System.getenv("BOOKAPP_OPENAI_KEY"));
// Fix timeouts
remoteAiService = new OpenAiService(System.getenv("BOOKAPP_OPENAI_KEY"), Duration.ZERO);
}
private OpenAiService getService() {
return remoteAiService;
}
public String getBookRecommendationText(String bookName, String bookAuthors) {
List<ChatMessage> messages = new ArrayList<>();
messages.add(new ChatMessage(
ChatMessageRole.SYSTEM.value(),
"You are a bot meant to summarize books in an engaging, entertaining and enticing way."
));
messages.add(new ChatMessage(
ChatMessageRole.USER.value(),
String.format("Please summarize the book %s by the author(s) %s for me. Make the summary interesting and entertaining. Separate the summary into paragraphs using two newlines.", bookName, bookAuthors)
));
// Could probably add something to customize the used model, maybe someone will want to shell out the cash for GPT-4 lol
ChatCompletionRequest req = ChatCompletionRequest
.builder()
.model("gpt-3.5-turbo")
.messages(messages)
.temperature(1.01)
.build();
ChatMessage resp = getService().createChatCompletion(req).getChoices().get(0).getMessage();
return resp.getContent();
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value"
] |
[((943, 973), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value'), ((1139, 1167), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value')]
|
package dev.danilobarreto.portalaluno.Controller;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.service.OpenAiService;
import dev.danilobarreto.portalaluno.Model.TextGenerate;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.ui.Model;
import org.springframework.web.bind.annotation.*;
import org.springframework.web.servlet.ModelAndView;
@RestController
public class ChatController {
@Value("${openai.token}")
private String TOKEN_OPEN_AI;
@GetMapping("/chat")
public ModelAndView telaInicio(){
ModelAndView mv = new ModelAndView();
mv.setViewName("chat");
return mv;
}
@ModelAttribute("textGenerate")
public TextGenerate textGenerate() {
return new TextGenerate();
}
@PostMapping("/text")
public Object generate(@RequestBody TextGenerate textGenerate, Model model){
try {
OpenAiService service = new OpenAiService(TOKEN_OPEN_AI);
CompletionRequest completionRequest = CompletionRequest.builder()
.model("text-davinci-003")
.prompt(textGenerate.getText())
.maxTokens(4000)
.build();
model.addAttribute("response", service.createCompletion(completionRequest).getChoices());
model.addAttribute("error", null); // Limpar qualquer mensagem de erro existente
return "response";
} catch (Exception e) {
model.addAttribute("response", null); // Limpar qualquer resposta existente
model.addAttribute("error", e.getMessage());
return "index"; // Redirecionar de volta para o formulário com mensagem de erro
}
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((1059, 1251), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1059, 1222), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1059, 1185), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1059, 1133), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
/*
* Copyright 2021 the original author or authors.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* https://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.springframework.cli.merger.ai.service;
import java.io.FileNotFoundException;
import java.io.IOException;
import java.nio.charset.StandardCharsets;
import java.time.Duration;
import java.time.temporal.ChronoUnit;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.Properties;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.service.OpenAiService;
import org.jetbrains.annotations.NotNull;
import org.springframework.cli.SpringCliException;
import org.springframework.cli.merger.ai.PromptRequest;
import org.springframework.cli.runtime.engine.templating.HandlebarsTemplateEngine;
import org.springframework.cli.util.PropertyFileUtils;
import org.springframework.cli.util.TerminalMessage;
import org.springframework.core.io.ClassPathResource;
import org.springframework.util.StreamUtils;
public abstract class AbstractOpenAiService implements org.springframework.cli.merger.ai.service.OpenAiService {
private final HandlebarsTemplateEngine handlebarsTemplateEngine = new HandlebarsTemplateEngine();
private com.theokanning.openai.service.OpenAiService openAiService;
private final TerminalMessage terminalMessage;
public AbstractOpenAiService(TerminalMessage terminalMessage) {
this.terminalMessage = terminalMessage;
}
protected ChatCompletionRequest getChatCompletionRequest(PromptRequest promptRequest) {
createOpenAiService();
ChatCompletionRequest chatCompletionRequest = ChatCompletionRequest.builder()
.model("gpt-3.5-turbo")
.temperature(0.3)
.messages(List.of(new ChatMessage("system", promptRequest.getSystemPrompt()),
new ChatMessage("user", promptRequest.getUserPrompt())))
.build();
return chatCompletionRequest;
}
private void createOpenAiService() {
if (this.openAiService == null) {
// get api token in file ~/.openai
Properties properties = PropertyFileUtils.getPropertyFile();
String apiKey = properties.getProperty("OPEN_AI_API_KEY");
this.openAiService = new OpenAiService(apiKey, Duration.of(5, ChronoUnit.MINUTES));
}
}
public OpenAiService getOpenAiService() {
return openAiService;
}
public HandlebarsTemplateEngine getHandlebarsTemplateEngine() {
return handlebarsTemplateEngine;
}
public TerminalMessage getTerminalMessage() {
return terminalMessage;
}
protected Map<String, String> getContext(String description) {
Map<String, String> context = new HashMap<>();
context.put("description", description);
return context;
}
protected String getPrompt(Map<String, String> context, String promptType) {
String resourceFileName = "/org/springframework/cli/merger/ai/openai-" + promptType + "-prompt.txt";
try {
ClassPathResource promptResource = new ClassPathResource(resourceFileName);
String promptRaw = StreamUtils.copyToString(promptResource.getInputStream(), StandardCharsets.UTF_8);
return getHandlebarsTemplateEngine().process(promptRaw, context);
}
catch (FileNotFoundException ex) {
throw new SpringCliException("Resource file note found:" + resourceFileName);
}
catch (IOException ex) {
throw new SpringCliException("Could read file " + resourceFileName, ex);
}
}
protected PromptRequest createPromptRequest(Map<String, String> context, String promptFamilyName) {
String systemPrompt = getPrompt(context, "system-" + promptFamilyName);
String userPrompt = getPrompt(context, "user-" + promptFamilyName);
return new PromptRequest(systemPrompt, userPrompt);
}
@NotNull
protected String getResponse(ChatCompletionRequest chatCompletionRequest) {
StringBuilder builder = new StringBuilder();
getOpenAiService().createChatCompletion(chatCompletionRequest).getChoices().forEach(choice -> {
builder.append(choice.getMessage().getContent());
});
String response = builder.toString();
return response;
}
}
|
[
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((2172, 2406), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2172, 2394), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2172, 2251), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2172, 2230), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
package com.touchbiz.chatgpt.controller;
import com.theokanning.openai.completion.CompletionRequest;
import com.touchbiz.chatgpt.application.ChatApplicationService;
import com.touchbiz.chatgpt.boot.config.OpenAiConfig;
import com.touchbiz.chatgpt.common.aspect.annotation.RequestLimit;
import com.touchbiz.chatgpt.common.dto.Result;
import com.touchbiz.chatgpt.common.proxy.OpenAiEventStreamService;
import com.touchbiz.chatgpt.database.domain.ChatSessionDetail;
import com.touchbiz.chatgpt.dto.Chat;
import com.touchbiz.chatgpt.dto.ChatResult;
import com.touchbiz.chatgpt.dto.request.ChatCompletionRequest;
import com.touchbiz.chatgpt.dto.request.ChatMessageRequest;
import com.touchbiz.chatgpt.dto.request.ValidChatRight;
import com.touchbiz.chatgpt.dto.response.ChatCompontionsResult;
import com.touchbiz.chatgpt.dto.response.ChatSessionDTO;
import com.touchbiz.chatgpt.infrastructure.constants.CommonConstant;
import com.touchbiz.chatgpt.infrastructure.converter.ChatSessionConverter;
import com.touchbiz.chatgpt.service.ChatSessionInfoService;
import com.touchbiz.common.entity.annotation.Auth;
import com.touchbiz.common.entity.result.MonoResult;
import com.touchbiz.common.utils.tools.JsonUtils;
import io.swagger.annotations.ApiOperation;
import lombok.SneakyThrows;
import lombok.extern.slf4j.Slf4j;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.http.MediaType;
import org.springframework.http.codec.ServerSentEvent;
import org.springframework.util.CollectionUtils;
import org.springframework.util.ObjectUtils;
import org.springframework.web.bind.annotation.*;
import reactor.core.publisher.Flux;
import reactor.core.publisher.Mono;
import reactor.core.scheduler.Schedulers;
import javax.validation.Valid;
import java.time.LocalTime;
import java.util.*;
import java.util.concurrent.atomic.AtomicReference;
import static com.touchbiz.chatgpt.infrastructure.constants.CacheConstant.*;
@Slf4j
@RequestMapping("/api/chatGpt/chatting")
@RestController
public class ChatController extends AbstractBaseController<ChatSessionDetail, ChatSessionInfoService> {
@Autowired
private OpenAiConfig config;
@Autowired
private OpenAiEventStreamService service;
@Autowired
private ChatApplicationService chatApplicationService;
@PostMapping
@RequestLimit()
public Mono<Result<?>> prompt(@RequestBody @Valid Chat chat) {
String sessionId = chat.getSessionId();
String prompt = chat.getPrompt();
//sessionId校验合法性
chatApplicationService.checkSessionId(sessionId);
var user = getCurrentUser();
log.info("chat:{}", chat);
String redisKey = CHAT_SESSION_CONTEXT_KEY + sessionId;
String question;
//拼接提问
if (getRedisTemplate().hasKey(redisKey)) {
question = JsonUtils.toJson(getRedisTemplate().get(redisKey)).trim().replace("\\", "").replace("\"", "").replace("n", "\\n") + CommonConstant.SPLICER + prompt;
} else {
question = prompt;
}
log.info("question:{}", question);
CompletionRequest completionRequest = generateRequest(prompt);
try {
long start = System.currentTimeMillis();
var result = service.createCompletion(completionRequest);
log.info("调用openAI接口耗时:{}", System.currentTimeMillis() - start + "ms");
String rt = JsonUtils.toJson(result);
ChatResult chatResult = JsonUtils.toObject(rt, ChatResult.class);
log.info("result:{}", chatResult);
if (!ObjectUtils.isEmpty(chatResult) && !CollectionUtils.isEmpty(chatResult.getChoices())) {
String answerContent = chatResult.getChoices().get(0).getText();
String answer = answerContent.replace("\\", "");
redisTemplate.set(CHAT_SESSION_CONTEXT_KEY + sessionId, question + answer, CHAT_SESSION_INFO_EXPIRE_SECONDS);
chatApplicationService.createSessionInfo(chat, answerContent, user);
return Mono.just(Result.ok(answerContent));
}
} catch (Exception ex) {
log.error("error:", ex);
return Mono.just(Result.error("系统超时,请联系管理员"));
}
return Mono.just(Result.error("请求失败,请重试"));
}
@SneakyThrows
@GetMapping(value = "/completion", produces = MediaType.TEXT_EVENT_STREAM_VALUE)
public Flux<ServerSentEvent<Result<String>>> completion(@RequestParam("sessionId") String sessionId,
@RequestParam("prompt") String prompt) {
log.info("sessionId:{},prompt:{}", sessionId, prompt);
if (sessionId.contains(" ")) {
sessionId = sessionId.replace(" ", "+");
}
//sessionId校验合法性
chatApplicationService.checkSessionId(sessionId);
var user = getCurrentUser();
String redisKey = CHAT_SESSION_CONTEXT_KEY + sessionId;
String question;
//拼接提问
if (getRedisTemplate().hasKey(redisKey)) {
question = JsonUtils.toJson(getRedisTemplate().get(redisKey)).replace("\"", "").replace(CommonConstant.CHARACTER,"\n") + CommonConstant.SPLICER + prompt;
} else {
question = prompt;
}
log.info("question:{}", question);
var eventStream = service.createCompletionFlux(this.generateRequest(question));
eventStream.doOnError(x -> log.error("doOnError SSE:", x));
String finalSessionId = sessionId;
List<ChatResult> list = new ArrayList<>();
AtomicReference<ChatResult> lastChatResult = null;
eventStream.subscribe(content -> {
String data = content.data();
if ("[DONE]".equals(data)) {
return;
}
ChatResult chatResult = JsonUtils.toObject(data, ChatResult.class);
lastChatResult.set(chatResult);
list.add(chatResult);
log.info("Time: {} - event: name[{}], id [{}], content[{}] ",
LocalTime.now(), content.event(), content.id(), data
);
}, error -> log.error("Error receiving SSE:", error),
() -> {
StringBuilder stringBuilder = new StringBuilder();
list.forEach(item -> {
List<ChatResult.Choice> choices = item.getChoices();
if (!CollectionUtils.isEmpty(choices)) {
String text = choices.get(0).getText();
stringBuilder.append(text);
}
});
String answerContent = stringBuilder.toString();
String qn = question.replace("\n", CommonConstant.CHARACTER);
String answer = answerContent.replace("\n", CommonConstant.CHARACTER);
redisTemplate.set(CHAT_SESSION_CONTEXT_KEY + finalSessionId, qn + answer, CHAT_SESSION_INFO_EXPIRE_SECONDS);
chatApplicationService.createSessionInfo(finalSessionId, prompt, answerContent, user);
}
);
return eventStream.map(x -> {
Result<String> result = Result.ok();
result.setResult(x.data());
return ServerSentEvent.builder(result).build();
}).subscribeOn(Schedulers.elastic());
}
@SneakyThrows
@GetMapping(value = "/completions", produces = MediaType.TEXT_EVENT_STREAM_VALUE)
public Flux<ServerSentEvent<Result<String>>> chatCompletion(@RequestParam("sessionId") String sessionId,
@RequestParam("prompt") String prompt) {
log.info("sessionId:{},prompt:{}", sessionId, prompt);
if (sessionId.contains(" ")) {
sessionId = sessionId.replace(" ", "+");
}
//sessionId校验合法性
chatApplicationService.checkSessionId(sessionId);
var user = getCurrentUser();
final String redisKey = CHAT_SESSION_CONTEXT_KEY + sessionId;
List<ChatMessageRequest> chatList = redisTemplate.getObjectList(redisKey, ChatMessageRequest.class);
if(chatList == null){
chatList = new ArrayList<>();
}
chatList.add(new ChatMessageRequest("user", prompt));
var eventStream = service.createChatCompletionFlux(this.generateChatRequest(chatList));
eventStream.doOnError(x -> log.error("doOnError SSE:", x));
String finalSessionId = sessionId;
List<ChatCompontionsResult> list = new ArrayList<>();
AtomicReference<ChatCompontionsResult> lastChatResult = null;
List<ChatMessageRequest> finalChatList = chatList;
eventStream.subscribe(content -> {
String data = content.data();
if ("[DONE]".equals(data)) {
//判断上一条的结束原因,如果是因为长度不足,则如何继续请求并拼接到目前的数据中去
return;
}
ChatCompontionsResult chatResult = JsonUtils.toObject(data, ChatCompontionsResult.class);
list.add(chatResult);
log.info("Time: {} - event: name[{}], id [{}], content[{}] ",
LocalTime.now(), content.event(), content.id(), data
);
}, error -> log.error("Error receiving SSE:", error),
() -> {
StringBuilder stringBuilder = new StringBuilder();
list.forEach(item -> {
List<ChatCompontionsResult.Choice> choices = item.getChoices();
if (!CollectionUtils.isEmpty(choices)) {
choices.forEach(choice->{
choice.getDelta().forEach(delta->{
stringBuilder.append(delta.getContent());
});
});
}
});
String answerContent = stringBuilder.toString();
ChatMessageRequest request = new ChatMessageRequest("system", answerContent);
finalChatList.add(request);
redisTemplate.setObjectList(CHAT_SESSION_CONTEXT_KEY + finalSessionId, finalChatList, CHAT_SESSION_INFO_EXPIRE_SECONDS);
chatApplicationService.createSessionInfo(finalSessionId, prompt, answerContent, user);
}
);
return eventStream.map(x -> {
Result<String> result = Result.ok();
result.setResult(x.data());
return ServerSentEvent.builder(result).build();
}).subscribeOn(Schedulers.elastic());
}
@SneakyThrows
@GetMapping(value = "/continueCompletion", produces = MediaType.TEXT_EVENT_STREAM_VALUE)
public Flux<ServerSentEvent<Result<String>>> continueCompletion(@RequestParam("sessionId") String sessionId) {
if (sessionId.contains(" ")) {
sessionId = sessionId.replace(" ", "+");
}
//sessionId校验合法性
chatApplicationService.checkSessionId(sessionId);
var user = getCurrentUser();
String redisKey = CHAT_SESSION_CONTEXT_KEY + sessionId;
String question;
//拼接提问
if (getRedisTemplate().hasKey(redisKey)) {
question = JsonUtils.toJson(getRedisTemplate().get(redisKey)).replace("\"", "").replace(CommonConstant.CHARACTER,"\n") + CommonConstant.SPLICER;
} else {
question = "";
}
log.info("question:{}", question);
var eventStream = service.createCompletionFlux(this.generateRequest(question));
eventStream.doOnError(x -> log.error("doOnError SSE:", x));
String finalSessionId = sessionId;
List<ChatResult> list = new ArrayList<>();
AtomicReference<ChatResult> lastChatResult = null;
eventStream.subscribe(content -> {
String data = content.data();
if ("[DONE]".equals(data)) {
return;
}
ChatResult chatResult = JsonUtils.toObject(data, ChatResult.class);
lastChatResult.set(chatResult);
list.add(chatResult);
log.info("Time: {} - event: name[{}], id [{}], content[{}] ",
LocalTime.now(), content.event(), content.id(), data
);
}, error -> log.error("Error receiving SSE:", error),
() -> {
StringBuilder stringBuilder = new StringBuilder();
list.forEach(item -> {
List<ChatResult.Choice> choices = item.getChoices();
if (!CollectionUtils.isEmpty(choices)) {
String text = choices.get(0).getText();
stringBuilder.append(text);
}
});
String answerContent = stringBuilder.toString();
String qn = question.replace("\n", CommonConstant.CHARACTER);
String answer = answerContent.replace("\n", CommonConstant.CHARACTER);
redisTemplate.set(CHAT_SESSION_CONTEXT_KEY + finalSessionId, qn + answer, CHAT_SESSION_INFO_EXPIRE_SECONDS);
chatApplicationService.createSessionInfo(finalSessionId, "prompt", answerContent, user);
}
);
return eventStream.map(x -> {
Result<String> result = Result.ok();
result.setResult(x.data());
return ServerSentEvent.builder(result).build();
}).subscribeOn(Schedulers.elastic());
}
@Auth
@ApiOperation("获取会话列表")
@GetMapping
public MonoResult<List<ChatSessionDTO>> getPageList(@RequestParam(name = "pageNo", defaultValue = "1") Integer pageNo,
@RequestParam(name = "pageSize", defaultValue = "10") Integer pageSize) {
var user = getCurrentUser();
var result = chatApplicationService.getChatSessionPageList(pageNo, pageSize, user);
var list = result.getRecords().stream().map(ChatSessionConverter.INSTANCE::transformOut)
.toList();
return MonoResult.ok(list);
}
@ApiOperation("新增会话id")
@PostMapping("/session")
public MonoResult<?> createSession() {
var user = getCurrentUser();
var session = chatApplicationService.createSession(user);
//添加缓存
String key = CHAT_SESSION_KEY + session.getSessionId();
getRedisTemplate().setObject(key, session, CHAT_SESSION_EXPIRE_SECONDS);
return MonoResult.OK(ChatSessionConverter.INSTANCE.transformOut(session));
}
/**
* 判断是否允许进行聊天,如果没有相应的次数,则不能进行后续的聊天,并返回相应的提示内容
*
* @return
*/
@PostMapping("/validRight")
public MonoResult<Object> validChatRight(@RequestBody ValidChatRight validChatRight) {
return MonoResult.ok("");
}
@Auth
@ApiOperation(value = "删除会话")
@DeleteMapping("/{id}")
public MonoResult<?> delete(@PathVariable String id) {
var user = getCurrentUser();
chatApplicationService.deleteSession(id, user);
return MonoResult.ok("删除成功!");
}
private CompletionRequest generateRequest(String prompt) {
return CompletionRequest.builder()
.prompt(prompt)
.model(config.getModel())
.stop(Arrays.asList(" Human:", " AI:"))
.maxTokens(1280)
.presencePenalty(0.6d)
.frequencyPenalty(0d)
.temperature(0.9D)
.bestOf(1)
.topP(1d)
.build();
}
private ChatCompletionRequest generateChatRequest(List<ChatMessageRequest> list) {
var request = ChatCompletionRequest.builder().temperature(0d)
.messages(list)
.build();
request.setFrequencyPenalty(0d);
request.setMaxTokens(4096);
return request;
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((2841, 2954), 'com.touchbiz.common.utils.tools.JsonUtils.toJson'), ((2841, 2934), 'com.touchbiz.common.utils.tools.JsonUtils.toJson'), ((2841, 2916), 'com.touchbiz.common.utils.tools.JsonUtils.toJson'), ((2841, 2898), 'com.touchbiz.common.utils.tools.JsonUtils.toJson'), ((5131, 5238), 'com.touchbiz.common.utils.tools.JsonUtils.toJson'), ((5131, 5199), 'com.touchbiz.common.utils.tools.JsonUtils.toJson'), ((7426, 7465), 'org.springframework.http.codec.ServerSentEvent.builder'), ((10847, 10886), 'org.springframework.http.codec.ServerSentEvent.builder'), ((11585, 11692), 'com.touchbiz.common.utils.tools.JsonUtils.toJson'), ((11585, 11653), 'com.touchbiz.common.utils.tools.JsonUtils.toJson'), ((13869, 13908), 'org.springframework.http.codec.ServerSentEvent.builder'), ((14986, 15037), 'com.touchbiz.chatgpt.infrastructure.converter.ChatSessionConverter.INSTANCE.transformOut'), ((15749, 16129), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((15749, 16104), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((15749, 16078), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((15749, 16051), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((15749, 16016), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((15749, 15978), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((15749, 15939), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((15749, 15906), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((15749, 15850), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((15749, 15808), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((16247, 16351), 'com.touchbiz.chatgpt.dto.request.ChatCompletionRequest.builder'), ((16247, 16326), 'com.touchbiz.chatgpt.dto.request.ChatCompletionRequest.builder'), ((16247, 16294), 'com.touchbiz.chatgpt.dto.request.ChatCompletionRequest.builder')]
|
package com.notorious;
import com.notorious.models.Data;
import com.notorious.models.User;
import com.notorious.repositorys.DataRepository;
import com.notorious.repositorys.UserRepository;
import com.notorious.request.AssociationRequest;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.image.CreateImageRequest;
import com.theokanning.openai.service.OpenAiService;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.http.ResponseEntity;
import org.springframework.web.bind.annotation.*;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
@CrossOrigin(origins = {"http://localhost:3000", "https://notorious-799tu.ondigitalocean.app"})
@RestController("/")
public class OpenAIController {
Object response;
OpenAiService service;
final List<ChatMessage> messages;
ChatMessage systemMessage;
public OpenAIController(@Value("${openai.apikey}") String token) {
this.service = new OpenAiService(token);
this.systemMessage = new ChatMessage();
this.response = "";
messages = new ArrayList<>();
}
@GetMapping("hello")
public String hello() {
return "Hola!";
}
@GetMapping("words")
public Object GenerateWords (String word) {
/*Streaming chat completion...*/
systemMessage = new ChatMessage(ChatMessageRole.USER.value(), "I am memorizing, and I need to create associations between words." + "Give me 5 Spanish words that exist, with a 90% similarity in writing to: " + word +
". Like these similarities: shower/chofer, snake/esnife, get/jet. Give me only words that exist in real life. May your answer only be the words, eliminate numbers and signs from your answer.");
return getObject();
}
@GetMapping("idea")
public Object GenerateWords (String wordOne, String wordTwo) {
/*Streaming chat completion...*/
systemMessage = new ChatMessage(ChatMessageRole.USER.value(), "I am memorizing Give me a single short implausible idea in Spanish with a maximum of 30 words that includes the words: " + wordOne + " and " + wordTwo + ". Let my 5 senses be involved in the implausible idea. that your answer is in spanish");
return getObject();
}
private Object getObject() {
messages.add(systemMessage);
ChatCompletionRequest chatCompletionRequest = ChatCompletionRequest
.builder()
.model("gpt-3.5-turbo")
.messages(messages)
.n(1)
.logitBias(new HashMap<>())
.build();
Object response = service.createChatCompletion(chatCompletionRequest).getChoices().get(0).getMessage();
service.shutdownExecutor();
return response;
}
@GetMapping("image")
public Object ImageGenerate(String history) {
//Creating Image
CreateImageRequest request = CreateImageRequest.builder()
.prompt(history)
.size("256x256")
.build();
//Image is located at:
response = service.createImage(request).getData().get(0);
service.shutdownExecutor();
return response;
}
/*______________________*/
@Autowired
private UserRepository userRepository;
@Autowired
private DataRepository dataRepository;
//Añadir Usuario
@PostMapping(path = "add")
public @ResponseBody User addNewUser(@RequestParam String username, @RequestParam String password) {
User user = new User();
user.setUsername(username);
user.setPassword(password);
userRepository.save(user);
return user;
}
//Obtener Usuario
@GetMapping(path = "getUser")
public ResponseEntity<Object> getUser(@RequestParam String username) {
User user = userRepository.findByUsername(username);
if(user != null) {
return ResponseEntity.ok(user);
} else {
return ResponseEntity.ok(new HashMap<>());
}
}
//Obtener todos los usuarios
@GetMapping(path = "all")
public @ResponseBody Iterable<User> getAllUsers() {
return userRepository.findAll();
}
//Añadir nueva Asociacion
@PostMapping(path = "addNewAssociation")
public @ResponseBody String addNewAssociation (@RequestBody
AssociationRequest request) {
User user = userRepository.findByUsername(request.getUsernameFK());
if(user != null) {
Data data = new Data();
data.setUser(user);
data.setWordEnglish(request.getWordEnglish());
data.setWordSimilar(request.getWordSimilar());
data.setIdea(request.getIdea());
data.setImage(request.getImage());
dataRepository.save(data);
return "Data added successfully";
} else {
return "User not found";
}
}
@GetMapping(path = "getAssociation")
public @ResponseBody Data getAssociation (@RequestParam String wordEnglish) {
Data data = dataRepository.findByWordEnglish(wordEnglish);
return data;
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value",
"com.theokanning.openai.image.CreateImageRequest.builder"
] |
[((1585, 1613), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((2188, 2216), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((3159, 3278), 'com.theokanning.openai.image.CreateImageRequest.builder'), ((3159, 3253), 'com.theokanning.openai.image.CreateImageRequest.builder'), ((3159, 3220), 'com.theokanning.openai.image.CreateImageRequest.builder')]
|
package edu.uniceub.calendar_man.chatworkflowmanager.open_ai.functions;
import com.theokanning.openai.completion.chat.ChatFunction;
import edu.uniceub.calendar_man.chatworkflowmanager.models.Event;
import edu.uniceub.calendar_man.chatworkflowmanager.open_ai.functions.contexts.GetCurrentDateResponse;
import edu.uniceub.calendar_man.chatworkflowmanager.open_ai.functions.contexts.GetEventsRequest;
import java.time.LocalDateTime;
import java.util.List;
import java.util.function.Function;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
public final class FunctionUtils {
final Logger LOGGER = LoggerFactory.getLogger(FunctionUtils.class);
private FunctionUtils() {}
public static final List<String> ASSISTANT_ACTION_FUNCTIONS =
List.of("get_calendar_context", "schedule_event", "get_current_date");
public static ChatFunction getCalendarEventsFunction(
final Function<GetEventsRequest, Object> getEvents) {
return ChatFunction.builder()
.name("get_calendar_context")
.description("Accesses the user's calendar and gets the events for the given date.")
.executor(GetEventsRequest.class, getEvents)
.build();
}
public static ChatFunction scheduleNewEvent(final Function<Event, Object> scheduleEvent) {
return ChatFunction.builder()
.name("schedule_event")
.description("Schedules a new event to the user's calendar.")
.executor(Event.class, scheduleEvent)
.build();
}
public static ChatFunction getCurrentDate() {
return ChatFunction.builder()
.name("get_current_date")
.description("Returns the current date.")
.executor(
Object.class, empty -> GetCurrentDateResponse.fromDate(LocalDateTime.now().toString()))
.build();
}
}
|
[
"com.theokanning.openai.completion.chat.ChatFunction.builder"
] |
[((954, 1177), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((954, 1160), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((954, 1107), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((954, 1014), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1289, 1476), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1289, 1459), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1289, 1413), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1289, 1343), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1542, 1784), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1542, 1767), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1542, 1648), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1542, 1598), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1735, 1765), 'java.time.LocalDateTime.now')]
|
package br.com.alura.screenmatch.service;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.service.OpenAiService;
public class ConsultaChatGPT {
public static String obterTraducao(String texto) {
OpenAiService service = new OpenAiService("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");
CompletionRequest requisicao = CompletionRequest.builder()
.model("text-davinci-003")
.prompt("traduza para o português o texto: " + texto)
.maxTokens(1000)
.temperature(0.7)
.build();
var resposta = service.createCompletion(requisicao);
return resposta.getChoices().get(0).getText();
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((390, 623), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((390, 598), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((390, 564), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((390, 531), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((390, 460), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
package learn.scraibe.controllers;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatCompletionResult;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
import learn.scraibe.models.Note;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.http.HttpStatus;
import org.springframework.http.ResponseEntity;
import org.springframework.web.bind.annotation.PostMapping;
import org.springframework.web.bind.annotation.RequestBody;
import org.springframework.web.bind.annotation.RequestMapping;
import org.springframework.web.bind.annotation.RestController;
import java.time.Duration;
import java.util.ArrayList;
import java.util.List;
@RestController
@RequestMapping("/generate-completion")
public class OpenAIController {
@Value("${openai.api.key}")
private String openaiApiKey;
@PostMapping
public ResponseEntity<Object> generateCompletion(@RequestBody Note note) {
if(note.getContent() == null || note.getContent().isBlank()){
return new ResponseEntity<>("Cannot have blank notes", HttpStatus.BAD_REQUEST);
}
//create service that will route to OpenAI endpoint, provide key and timeout value incase openai takes a long time
OpenAiService service = new OpenAiService(openaiApiKey, Duration.ofSeconds(60));
//set up messages and Roles
List<ChatMessage> messages = new ArrayList<>();
ChatMessage userMessage = new ChatMessage(ChatMessageRole.USER.value(), "organize with bullet points, only respond with bullet points "+ note.getContent());
ChatMessage systemMessage = new ChatMessage(ChatMessageRole.ASSISTANT.value(), "you are a helpful assistant");
messages.add(userMessage);
messages.add((systemMessage));
// configure chatCompletionRequest object that will be sent over via the api
ChatCompletionRequest chatCompletionRequest = ChatCompletionRequest
.builder()
.model("gpt-3.5-turbo-0613")
.messages(messages)
.build();
//use service to make the request to OpenAI and then get the specific message to send back to the frontend.
ChatMessage responseMessage = service.createChatCompletion(chatCompletionRequest).getChoices().get(0).getMessage();
note.setContent(responseMessage.getContent());
return new ResponseEntity<>(note, HttpStatus.OK);
//TODO make a conditional statement based on the success of a response message,
//one previous error occurred because the request timed out(openai took too long to send back a request)
// but extending the duration seemed to solved the issue, just wondering what other issues to anticipate.
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.ASSISTANT.value"
] |
[((1638, 1666), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((1805, 1838), 'com.theokanning.openai.completion.chat.ChatMessageRole.ASSISTANT.value')]
|
package Server;
import Protocol.ChatServerInterface;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
import java.io.FileInputStream;
import java.io.IOException;
import java.io.InputStream;
import java.net.ServerSocket;
import java.net.Socket;
import java.util.*;
public class ChatServer {
private final ServerSocket serverSocket;
private final List<ClientHandler> clients;
private final List<String> chatLog;
private final OpenAiService openapi;
List<ChatMessage> messages = new ArrayList<>();
private int nextUserId = 0;
public ChatServer(int port) throws IOException {
clients = new ArrayList<>();
chatLog = new ArrayList<>();
serverSocket = new ServerSocket(port);
System.out.println("Server is now open on port " + port); // 서버가 열린 포트 번호를 로그로 찍는 코드 추가
Properties prop = new Properties();
String openaiKey = "";
try (InputStream input = new FileInputStream("config.properties")) {
// .properties 파일 로드
prop.load(input);
// 키를 사용하여 값을 검색
openaiKey = prop.getProperty("OPENAI_KEY");
} catch (IOException ex) {
ex.printStackTrace();
}
openapi = new OpenAiService(openaiKey);
ChatMessage customInstruction = new ChatMessage(ChatMessageRole.SYSTEM.value(), "'지피티' is designed for group chat interactions, understanding and responding to individual users based on their names in the chat. It has a rough, friendly speaking style, similar to that of a close friend. The GPT is knowledgeable in computer science, especially Java, AI, and gaming, with a particular interest in League of Legends and the latest computer hardware. Its responses are brief, typically no more than two sentences, and mirror the user's speech style. The GPT seamlessly integrates into group conversations, offering tech and gaming insights in a casual, engaging manner.");
messages.add(customInstruction);
}
public void start() throws IOException {
while (true) {
Socket socket = serverSocket.accept();
ClientHandler clientHandler = new ClientHandler(socket, this, nextUserId++);
clients.add(clientHandler);
clientHandler.start();
}
}
public synchronized void broadcastMessage(String message, int userId) {
chatLog.add(message);
System.out.println(message);
for (ClientHandler client : clients) {
client.sendMessage(message, userId);
}
}
public List<String> getChatLog() {
return chatLog;
}
public static class ClientHandler extends Thread {
private final Socket socket;
private final ChatServer server;
private final int userId;
private final ChatServerInterface chatServerInterface;
private static final Map<Integer, String> nameMap = new HashMap<>();
public ClientHandler(Socket socket, ChatServer server, int userId) throws IOException {
this.socket = socket;
this.server = server;
this.userId = userId;
this.chatServerInterface = new ChatServerInterface(socket.getInputStream(), socket.getOutputStream());
chatServerInterface.setClientHandler(new ChatServerInterface.ClientHandler() {
@Override
public void onNameSet(String name) {
nameMap.put(userId, name);
}
@Override
public void onMessageReceived(String message) {
String userName = nameMap.getOrDefault(userId, "Unknown");
server.broadcastMessage(userName + ": " + message, userId); // 사용자 이름과 메시지를 모든 클라이언트에게 전송
try {
ChatMessage userMessage = new ChatMessage(ChatMessageRole.USER.value(), userName + ": " + message);
server.messages.add(userMessage);
ChatCompletionRequest chatCompletionRequest = ChatCompletionRequest
.builder()
.model("gpt-3.5-turbo-0613")
.messages(server.messages)
// .functionCall(new ChatCompletionRequest.ChatCompletionRequestFunctionCall("auto"))
.maxTokens(256)
.build();
ChatMessage responseMessage = server.openapi.createChatCompletion(chatCompletionRequest).getChoices().get(0).getMessage();
server.messages.add(responseMessage);
server.broadcastMessage(responseMessage.getContent(), userId);
} catch (Exception e) {
e.printStackTrace();
}
}
@Override
public void onMessageEditRequest(int messageId, String newMessage) {
// 메시지 수정 요청 처리
}
@Override
public void onMessageDeleteRequest(int messageId) {
// 메시지 삭제 요청 처리
}
@Override
public void onInvalidRequest(String[] messages) {
// 잘못된 요청 처리
}
});
}
public void run() {
try {
while (!interrupted() && chatServerInterface.readCommand()) ;
} catch (IOException e) {
e.printStackTrace();
} finally {
closeConnection();
}
}
private void closeConnection() {
try {
System.out.println("Client " + userId + " disconnected.");
socket.close();
server.clients.remove(this);
} catch (IOException e) {
e.printStackTrace();
}
}
public void sendMessage(String message, int userId) {
try {
chatServerInterface.sendMessageToClient(server.getChatLog().size() - 1, userId, message);
} catch (IOException e) {
e.printStackTrace();
}
}
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value"
] |
[((1440, 1470), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value'), ((3681, 3709), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value')]
|
package com.odde.doughnut.services.ai.tools;
import com.odde.doughnut.services.ai.*;
import com.theokanning.openai.completion.chat.ChatFunction;
import java.util.List;
public class AiToolFactory {
public static final String COMPLETE_NOTE_DETAILS = "complete_note_details";
public static AiToolList mcqWithAnswerAiTool() {
return new AiToolList(
"""
Please assume the role of a Memory Assistant, which involves helping me review, recall, and reinforce information from my notes. As a Memory Assistant, focus on creating exercises that stimulate memory and comprehension. Please adhere to the following guidelines:
1. Generate a MCQ based on the note in the current context path
2. Only the top-level of the context path is visible to the user.
3. Provide 2 to 4 choices with only 1 correct answer.
4. Vary the lengths of the choice texts so that the correct answer isn't consistently the longest.
5. If there's insufficient information in the note to create a question, leave the 'stem' field empty.
Note: The specific note of focus and its more detailed contexts are not known. Focus on memory reinforcement and recall across various subjects.
""",
List.of(
ChatFunction.builder()
.name("ask_single_answer_multiple_choice_question")
.description("Ask a single-answer multiple-choice question to the user")
.executor(MCQWithAnswer.class, null)
.build()));
}
public static AiToolList questionEvaluationAiTool(MCQWithAnswer question) {
MultipleChoicesQuestion clone = question.getMultipleChoicesQuestion();
String messageBody =
"""
Please assume the role of a learner, who has learned the note of focus as well as many other notes.
Only the top-level of the context path is visible to you.
Without the specific note of focus and its more detailed contexts revealed to you,
please critically check if the following question makes sense and is possible to you:
%s
"""
.formatted(clone.toJsonString());
return new AiToolList(
messageBody,
List.of(
ChatFunction.builder()
.name("evaluate_question")
.description("answer and evaluate the feasibility of the question")
.executor(QuestionEvaluation.class, null)
.build()));
}
}
|
[
"com.theokanning.openai.completion.chat.ChatFunction.builder"
] |
[((1248, 1505), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1248, 1480), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1248, 1427), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((1248, 1338), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((2165, 2397), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((2165, 2372), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((2165, 2314), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((2165, 2230), 'com.theokanning.openai.completion.chat.ChatFunction.builder')]
|
package com.github.starrygaze.midjourney.service.translate.impl;
import cn.hutool.core.text.CharSequenceUtil;
import com.github.starrygaze.midjourney.ProxyProperties;
import com.github.starrygaze.midjourney.service.translate.TranslateService;
import com.theokanning.openai.completion.chat.ChatCompletionChoice;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.service.OpenAiService;
import lombok.extern.slf4j.Slf4j;
import org.springframework.beans.factory.support.BeanDefinitionValidationException;
import java.util.List;
/**
* 这个类叫做 GPTTranslateServiceImpl,它实现了 TranslateService 接口。根据类名和代码内容,我们可以看出这个类是用来提供基于 OpenAI GPT (Generative Pre-training Transformer) 的翻译服务。
* 这个类提供了一个基于 OpenAI GPT 的翻译服务,可以将中文文本翻译成英语文本。尽管 OpenAI GPT 不是专门的翻译模型,但是由于其强大的语言生成和理解能力,也可以用来做一些简单的翻译任务。
*/
@Slf4j
public class GPTTranslateServiceImpl implements TranslateService {
/**
* 初始化(构造函数):在构造函数中,接收一个类型为 ProxyProperties.OpenaiConfig 的参数 openaiConfig,并从这个参数中获取了 OpenAI GPT 所需要的 API 密钥。
* 如果这个密钥为空,将会抛出一个 BeanDefinitionValidationException 异常。然后,使用这个 API 密钥和超时时间创建一个 OpenAiService 实例。
*/
private final OpenAiService openAiService;
private final ProxyProperties.OpenaiConfig openaiConfig;
public GPTTranslateServiceImpl(ProxyProperties.OpenaiConfig openaiConfig) {
if (CharSequenceUtil.isBlank(openaiConfig.getGptApiKey())) {
throw new BeanDefinitionValidationException("mj-proxy.openai.gpt-api-key未配置");
}
this.openaiConfig = openaiConfig;
this.openAiService = new OpenAiService(openaiConfig.getGptApiKey(), openaiConfig.getTimeout());
}
/**
* translateToEnglish(String prompt):这个方法用来将输入的文本翻译成英语。首先,它会检查输入的文本是否包含中文,如果不包含,直接返回原来的文本,否则会进行翻译。
* 翻译的过程中,它会创建两个 ChatMessage 实例,一个是系统消息,告诉 GPT 需要将中文翻译成英文,另一个是用户消息,包含了需要翻译的中文。然后,使用这两个消息创建一个
* ChatCompletionRequest 实例。接下来,使用 OpenAiService 来执行这个请求,并获取返回的结果。从结果中提取出翻译后的英文并返回。如果在调用过程中出现任何异常,
* 都会被捕获并打印警告信息,然后返回原来的文本。
* @param prompt
* @return
*/
@Override
public String translateToEnglish(String prompt) {
if (!containsChinese(prompt)) {
return prompt;
}
ChatMessage m1 = new ChatMessage("system", "把中文翻译成英文");
ChatMessage m2 = new ChatMessage("user", prompt);
ChatCompletionRequest request = ChatCompletionRequest.builder()
.model(this.openaiConfig.getModel())
.temperature(this.openaiConfig.getTemperature())
.maxTokens(this.openaiConfig.getMaxTokens())
.messages(List.of(m1, m2))
.build();
try {
List<ChatCompletionChoice> choices = this.openAiService.createChatCompletion(request).getChoices();
if (!choices.isEmpty()) {
return choices.get(0).getMessage().getContent();
}
} catch (Exception e) {
log.warn("调用chat-gpt接口翻译中文失败: {}", e.getMessage());
}
return prompt;
}
}
|
[
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((3138, 3356), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3138, 3343), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3138, 3312), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3138, 3263), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3138, 3210), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
package alura.jmilhas.api.infra.integration;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.service.OpenAiService;
import org.springframework.stereotype.Service;
@Service
public class ChatGptIntegrationService {
private static final String API_KEY = System.getenv("MY_API_KEY");
private final OpenAiService service = new OpenAiService(API_KEY);
public String destinationText(String destination) {
String prompt =
String.format("Aja como um redator para um site de venda de viagens. " +
"Faça um resumo sobre o local %s. Enfatize os pontos positivos da cidade." +
"Utilize uma linguagem informal. " +
"Cite ideias de passeios neste lugar. " +
"Crie 2 parágrafos neste resumo.", destination);
CompletionRequest request = CompletionRequest.builder()
.model("text-davinci-003")
.prompt(prompt)
.maxTokens(2048)
.temperature(0.6)
.build();
return service.createCompletion(request)
.getChoices()
.get(0)
.getText();
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((910, 1104), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((910, 1079), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((910, 1045), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((910, 1012), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((910, 980), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
package com.jwtly10.aicontentgenerator.service.OpenAI;
import com.jwtly10.aicontentgenerator.model.Gender;
import com.theokanning.openai.OpenAiHttpException;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
import lombok.extern.slf4j.Slf4j;
import org.springframework.stereotype.Service;
import java.util.ArrayList;
import java.util.List;
@Service
@Slf4j
public class OpenAPIService {
private final OpenAiService service;
public OpenAPIService(OpenAiService service) {
this.service = service;
}
/**
* Improve content by correcting grammar mistakes
*
* @param content Content to improve
* @return Improved content
*/
public String improveContent(String content) throws OpenAiHttpException {
log.info("Improving content via OpenAPI");
String instructions = "Given the following reddit post, correct grammar mistakes. Don't alter curse words or swearing. Replace slashes and " +
"dashes with the appropriate word.Remove dashes between words like high-end. Add punctuation as necessary for smooth speech flow. Only respond " +
"with the modified (or unmodified if no changes were made) text. Do not include any other information. ";
ChatMessage responseMessage = getResponseMessage(List.of(instructions, content));
log.debug("Original content: {}", content);
log.debug("Improved content: {}", responseMessage.getContent());
log.info("Content improved successfully");
return responseMessage.getContent();
}
/**
* Determine gender from given content
*
* @param content Content to determine gender
* @return Gender of content
*/
public Gender determineGender(String content) {
log.info("Determining gender");
String instructions = "From the given text, determine the poster's gender. Use the context provided by the text. " +
"If the gender is ambiguous, reply with the most probable gender. Respond with a single letter: 'M' for Male or 'F' for Female. Again. ONLY REPLY WITH M OR F. NOTHING ELSE.";
ChatMessage responseMessage = getResponseMessage(List.of(instructions, content));
log.info("Determined gender: {}", responseMessage.getContent());
if (responseMessage.getContent().equalsIgnoreCase("M")) {
return Gender.MALE;
} else if (responseMessage.getContent().equalsIgnoreCase("F")) {
return Gender.FEMALE;
} else {
log.error("OPEN API responded with something else: '{}'. Defaulting to male", responseMessage.getContent());
return Gender.MALE;
}
}
/**
* Get response message from given messages
*
* @param messages Messages to get response from
* @return Response message
*/
private ChatMessage getResponseMessage(List<String> messages) throws OpenAiHttpException {
List<ChatMessage> chatMessages = new ArrayList<>();
for (String message : messages) {
chatMessages.add(new ChatMessage(ChatMessageRole.USER.value(), message));
}
ChatCompletionRequest completionRequest = ChatCompletionRequest.builder()
.messages(chatMessages)
.model("gpt-3.5-turbo")
.build();
return service.createChatCompletion(completionRequest).getChoices().get(0).getMessage();
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value",
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((3261, 3289), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((3363, 3499), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3363, 3474), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3363, 3434), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
package br.com.alura.screenmatch.service;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.service.OpenAiService;
public class ConsultaChatGPT {
public static String obterTraducao(String texto) {
OpenAiService service = new OpenAiService(System.getenv("OPENAI_APIKEY"));
CompletionRequest requisicao = CompletionRequest.builder()
.model("gpt-3.5-turbo-instruct")
.prompt("traduza para o português o texto: " + texto)
.maxTokens(1000)
.temperature(0.7)
.build();
var resposta = service.createCompletion(requisicao);
return resposta.getChoices().get(0).getText();
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((366, 605), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((366, 580), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((366, 546), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((366, 513), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((366, 442), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
package school.sptech.zup.service;
import com.theokanning.openai.OpenAiHttpException;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.service.OpenAiService;
import lombok.RequiredArgsConstructor;
import org.springframework.http.HttpStatus;
import org.springframework.stereotype.Service;
import org.springframework.web.server.ResponseStatusException;
import school.sptech.zup.domain.ChaveKey;
import school.sptech.zup.domain.Gpt;
import school.sptech.zup.domain.ZupLog;
import school.sptech.zup.dto.response.GptResponse;
import school.sptech.zup.repository.ChaveKeyRepository;
import school.sptech.zup.repository.ZupLogRepository;
import school.sptech.zup.util.DateUtil;
import java.time.LocalDateTime;
@Service
@RequiredArgsConstructor
public class GptService {
private final ZupLogRepository _zupLog;
private final ChaveKeyRepository _chaveKey;
private final DateUtil _dateutil;
public ChaveKey InserirChave(String idChave){
ChaveKey chave = new ChaveKey();
chave.setIdChave(idChave);
_chaveKey.save(chave);
return chave;
}
public GptResponse gptNoticia(Gpt gpt) {
GptResponse gptResponse = new GptResponse();
var buscaChave = _chaveKey.UltimaChaveInserida();
try{
OpenAiService service = new OpenAiService(buscaChave.get(0).getIdChave());
CompletionRequest request = CompletionRequest.builder()
.model("text-davinci-003")
.prompt(gpt.getTitulo() + gpt.getPergunta())
.maxTokens(1000)
.build();
gptResponse.setResposta(service.createCompletion(request).getChoices().get(0).getText());
gptResponse.setId(gpt.getId());
return gptResponse;
}
catch (OpenAiHttpException ex){
if (ex.statusCode == 401){
ZupLog log = new ZupLog();
log.setDescricao("GPT: ERRO 401, Token expirado => " + ex.getMessage());
log.setDt_entrada(
_dateutil.formLocalDate(LocalDateTime.now())
);
_zupLog.save(log);
} else if (ex.statusCode == 404) {
ZupLog log = new ZupLog();
log.setDescricao("GPT: ERRO 404 => " + ex.getMessage());
log.setDt_entrada(
_dateutil.formLocalDate(LocalDateTime.now())
);
_zupLog.save(log);
}
}
throw new ResponseStatusException(HttpStatus.NOT_FOUND, "Erro, acesse o Log!");
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((1422, 1627), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1422, 1598), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1422, 1561), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1422, 1496), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
package com.challenge.challengeapi.controller;
import com.challenge.challengeapi.entity.Message;
import com.challenge.challengeapi.entity.UserIdRequest;
import com.challenge.challengeapi.service.MessageService;
import com.theokanning.openai.OpenAiService;
import com.theokanning.openai.completion.CompletionRequest;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.http.HttpStatus;
import org.springframework.http.ResponseEntity;
import org.springframework.web.bind.annotation.*;
import java.util.List;
@CrossOrigin(origins = "http://127.0.0.1:5173/", maxAge = 5173)
@RestController
@RequestMapping("/hearmeout/messages")
public class MessageController {
private final MessageService messageService;
@Autowired
public MessageController(MessageService messageService) {
this.messageService = messageService;
}
@PostMapping("/send")
@CrossOrigin
public ResponseEntity<String> sendMessage(@RequestBody Message message) {
try {
messageService.sendMessage(message.getUserId(), message.getText());
OpenAiService service = new OpenAiService("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");
CompletionRequest completionRequest = CompletionRequest.builder()
.prompt(message.getText())
.model("ada")
.build();
String response;
try {
response = service.createCompletion(completionRequest).getChoices().get(0).getText();
} catch (Exception e) {
System.err.println("Erro ao chamar a API do GPT: " + e.getMessage());
return ResponseEntity.status(HttpStatus.INTERNAL_SERVER_ERROR).body("Erro ao chamar a API do GPT");
}
System.out.println("Resposta do GPT: " + response);
message.setResponse(response);
messageService.saveMessage(message);
// Retorna a mensagem retornada pelo GPT-3 no corpo da resposta
return ResponseEntity.ok(response);
} catch (Exception e) {
return ResponseEntity.status(HttpStatus.INTERNAL_SERVER_ERROR).body("Erro na mensagem");
}
}
@GetMapping("/history")
@CrossOrigin
public ResponseEntity<List<Message>> getMessageHistory(@RequestBody UserIdRequest request) {
try {
List<Message> history = messageService.getMessageHistory(request.getUserId());
return ResponseEntity.ok(history);
} catch (Exception e) {
return ResponseEntity.status(HttpStatus.INTERNAL_SERVER_ERROR).body(null);
}
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((1255, 1392), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1255, 1363), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1255, 1329), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1689, 1780), 'org.springframework.http.ResponseEntity.status'), ((2130, 2210), 'org.springframework.http.ResponseEntity.status'), ((2574, 2640), 'org.springframework.http.ResponseEntity.status')]
|
package alura.api.challenge.service;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.service.OpenAiService;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.stereotype.Service;
@Service
public class GeraTextoDescritivoService {
@Value("${chatgpt.api.key}")
private String API_KEY;
public String geraTextoDescritivo(String destino){
OpenAiService service = new OpenAiService(API_KEY);
var pergunta = String.format(
"Gere um resumo de até 200 caracteres sobre %s enfatizanfo características do local e argumentando porque é um bom local para viajar."
, destino);
CompletionRequest request = CompletionRequest.builder()
.model("text-davinci-003")
.prompt(pergunta)
.maxTokens(500)
.build();
var response = service.createCompletion(request);
return response.getChoices().get(0).getText().replace("\n", "").trim();
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((744, 905), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((744, 880), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((744, 848), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((744, 814), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
package br.com.alura.screenmatch.service;
import br.com.alura.screenmatch.config.OpenAiConfig;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.service.OpenAiService;
public class ConsultaChatGPT {
public static String obterTraducao(String texto) {
String apiKey = OpenAiConfig.getApiKey();
OpenAiService service = new OpenAiService(apiKey);
CompletionRequest requisicao = CompletionRequest.builder()
.model("gpt-3.5-turbo-instruct")
.prompt("traduza para o português o texto: " + texto)
.maxTokens(1000)
.temperature(0.7)
.build();
var resposta = service.createCompletion(requisicao);
return resposta.getChoices().get(0).getText();
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((446, 685), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((446, 660), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((446, 626), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((446, 593), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((446, 522), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
package br.com.fiap.gsjava.controllers;
import br.com.fiap.gsjava.models.ChatGPT;
import br.com.fiap.gsjava.repositories.ChatGPTRepository;
import jakarta.validation.ConstraintViolationException;
import jakarta.validation.Valid;
import org.slf4j.LoggerFactory;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.data.web.PageableDefault;
import org.springframework.data.web.PagedResourcesAssembler;
import org.springframework.hateoas.EntityModel;
import org.springframework.hateoas.PagedModel;
import org.springframework.http.HttpStatus;
import org.springframework.http.ResponseEntity;
import org.springframework.web.bind.annotation.*;
import org.springframework.web.server.ResponseStatusException;
import com.theokanning.openai.OpenAiService;
import com.theokanning.openai.completion.CompletionRequest;
import org.springframework.data.domain.Pageable;
import org.slf4j.Logger;
@RestController
@RequestMapping("/chatbot")
public class ChatGPTController {
@Autowired
ChatGPTRepository repo;
@Autowired
PagedResourcesAssembler<ChatGPT> assembler;
Logger log = LoggerFactory.getLogger(ChatGPTController.class);
private static final String API_KEY = "Sua Chave da API aqui";
@GetMapping
public PagedModel<EntityModel<ChatGPT>> index(@PageableDefault(size = 5) Pageable pageable) {
return assembler.toModel(repo.findAll(pageable));
}
@GetMapping("/busca/{id}")
public EntityModel<ChatGPT> show(@PathVariable Long id) {
log.info("buscar chat com id: " + id);
ChatGPT chatGPT = repo.findById(id).orElseThrow(() ->
new ResponseStatusException(HttpStatus.NOT_FOUND, "Cliente não encontrado"));
return chatGPT.toModel();
}
@PostMapping("/api")
public ResponseEntity<ChatGPT> create(@RequestBody @Valid ChatGPT input) {
OpenAiService service = new OpenAiService(API_KEY);
CompletionRequest request = CompletionRequest.builder()
.model("text-davinci-003")
.prompt(input.getPergunta())
.maxTokens(100)
.build();
String resposta = service.createCompletion(request).getChoices().get(0).getText();
ChatGPT chatGPT = new ChatGPT(input.getPergunta(), resposta);
log.info("Saída do chatbot: " + chatGPT);
repo.save(chatGPT);
return ResponseEntity.status(HttpStatus.CREATED).body(chatGPT);
}
@DeleteMapping("/{id}")
public ResponseEntity<ChatGPT>destroy(@PathVariable Long id) {
log.info("deletar chat com o id: " + id);
ChatGPT chatgpt = repo.findById(id).orElseThrow(() ->
new ResponseStatusException(HttpStatus.NOT_FOUND, "Chat não encontrado"));;
repo.delete(chatgpt);
return ResponseEntity.noContent().build();
}
@ResponseStatus(HttpStatus.BAD_REQUEST)
@ExceptionHandler(ConstraintViolationException.class)
public ResponseEntity<String> handleValidationExceptions(ConstraintViolationException ex) {
log.error("Erro de validação: ", ex);
return ResponseEntity.badRequest().body(ex.getMessage());
}
@ResponseStatus(HttpStatus.INTERNAL_SERVER_ERROR)
@ExceptionHandler(Exception.class)
public ResponseEntity<String> handleAllExceptions(Exception ex) {
log.error("Erro não esperado: ", ex);
return ResponseEntity.status(HttpStatus.INTERNAL_SERVER_ERROR).body("Ocorreu um erro inesperado. Tente novamente mais tarde.");
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((2009, 2185), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2009, 2159), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2009, 2126), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2009, 2080), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2451, 2506), 'org.springframework.http.ResponseEntity.status'), ((2871, 2905), 'org.springframework.http.ResponseEntity.noContent'), ((3182, 3231), 'org.springframework.http.ResponseEntity.badRequest'), ((3472, 3591), 'org.springframework.http.ResponseEntity.status')]
|
package com.d_d.aifoodideageneratord_d.services;
import com.d_d.aifoodideageneratord_d.model.RecipeType;
import com.theokanning.openai.completion.chat.ChatCompletionChoice;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatCompletionResult;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.service.OpenAiService;
import java.time.Duration;
import java.util.List;
public class AiRecommendationService {
private static final String MODEL = "gpt-4-turbo-preview";
private static final String OPENAI_API_KEY_ENV = "OPENAI_API_KEY";
private final OpenAiService openAiService;
public AiRecommendationService() {
String apiToken = System.getenv(OPENAI_API_KEY_ENV);
validApiToken(apiToken);
this.openAiService = new OpenAiService(apiToken, Duration.ofSeconds(60));
}
public String getRecommendation(List<String> products, String choice) {
RecipeType recipeType = RecipeType.fromChoice(choice);
if (recipeType == null) {
throw new IllegalArgumentException("Invalid choice");
}
String question = createQuestion(products, recipeType);
ChatCompletionRequest request = buildChatCompletionRequest(question);
return getChatCompletionResult(request);
}
private String createQuestion(List<String> products, RecipeType recipeType) {
String productList = String.join(", ", products);
return String.format("I have the following products in my fridge: %s. " +
"Create a recipe using only the items in this list. " +
"You do not have to use all the products, but make sure that the recipe does not contain anything that is not on the list. " +
"Present the result in a clear, organised way, always starting with 'Recipe Title:', including the name of the recipe, a list of ingredients used including grams, preparation time, calories, step-by-step instructions. " +
"If it is not possible to make something with these products, please let me know and let the answer be in Polish..", productList, recipeType.getDescription());
}
private ChatCompletionRequest buildChatCompletionRequest(String question) {
return ChatCompletionRequest.builder()
.messages(List.of(new ChatMessage("user", question)))
.model(MODEL)
.build();
}
private String getChatCompletionResult(ChatCompletionRequest request) {
ChatCompletionResult result = openAiService.createChatCompletion(request);
return result.getChoices().stream()
.map(ChatCompletionChoice::getMessage)
.map(ChatMessage::getContent)
.findFirst()
.orElse("Failed to get a response");
}
private void validApiToken(String apiToken) {
if (apiToken == null || apiToken.isEmpty()) {
throw new IllegalArgumentException("Please set the OPENAI_API_KEY in environment variable");
}
}
}
|
[
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((2320, 2476), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2320, 2451), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2320, 2421), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
package io.github.asleepyfish.service;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.google.common.base.Strings;
import com.google.common.cache.Cache;
import com.google.common.cache.CacheBuilder;
import com.google.common.collect.Lists;
import com.knuddels.jtokkit.Encodings;
import com.knuddels.jtokkit.api.Encoding;
import com.knuddels.jtokkit.api.EncodingRegistry;
import com.knuddels.jtokkit.api.ModelType;
import com.theokanning.openai.DeleteResult;
import com.theokanning.openai.completion.CompletionChoice;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.completion.CompletionResult;
import com.theokanning.openai.completion.chat.ChatCompletionChoice;
import com.theokanning.openai.completion.chat.ChatCompletionChunk;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.edit.EditRequest;
import com.theokanning.openai.edit.EditResult;
import com.theokanning.openai.embedding.EmbeddingRequest;
import com.theokanning.openai.embedding.EmbeddingResult;
import com.theokanning.openai.finetune.FineTuneEvent;
import com.theokanning.openai.finetune.FineTuneRequest;
import com.theokanning.openai.finetune.FineTuneResult;
import com.theokanning.openai.image.Image;
import com.theokanning.openai.image.*;
import com.theokanning.openai.moderation.ModerationRequest;
import com.theokanning.openai.moderation.ModerationResult;
import io.github.asleepyfish.client.OpenAiApi;
import io.github.asleepyfish.config.ChatGPTProperties;
import io.github.asleepyfish.entity.audio.TranscriptionRequest;
import io.github.asleepyfish.entity.audio.TranslationRequest;
import io.github.asleepyfish.entity.billing.Billing;
import io.github.asleepyfish.entity.billing.Subscription;
import io.github.asleepyfish.enums.audio.AudioModelEnum;
import io.github.asleepyfish.enums.audio.AudioResponseFormatEnum;
import io.github.asleepyfish.enums.chat.FinishReasonEnum;
import io.github.asleepyfish.enums.chat.RoleEnum;
import io.github.asleepyfish.enums.edit.EditModelEnum;
import io.github.asleepyfish.enums.embedding.EmbeddingModelEnum;
import io.github.asleepyfish.enums.exception.ChatGPTErrorEnum;
import io.github.asleepyfish.enums.image.ImageResponseFormatEnum;
import io.github.asleepyfish.enums.image.ImageSizeEnum;
import io.github.asleepyfish.enums.model.ModelEnum;
import io.github.asleepyfish.exception.ChatGPTException;
import io.github.asleepyfish.service.openai.OpenAiService;
import lombok.NonNull;
import lombok.extern.slf4j.Slf4j;
import okhttp3.*;
import org.springframework.util.CollectionUtils;
import retrofit2.Retrofit;
import javax.imageio.ImageIO;
import java.awt.*;
import java.awt.image.BufferedImage;
import java.awt.image.ComponentColorModel;
import java.io.ByteArrayInputStream;
import java.io.File;
import java.io.IOException;
import java.io.OutputStream;
import java.math.BigDecimal;
import java.net.InetSocketAddress;
import java.net.Proxy;
import java.nio.charset.Charset;
import java.text.SimpleDateFormat;
import java.time.Duration;
import java.time.LocalDate;
import java.time.Period;
import java.time.format.DateTimeFormatter;
import java.util.List;
import java.util.*;
import java.util.concurrent.ConcurrentMap;
import java.util.concurrent.ExecutionException;
import java.util.concurrent.TimeUnit;
import java.util.stream.Collectors;
import java.util.zip.ZipEntry;
import java.util.zip.ZipOutputStream;
/**
* @Author: asleepyfish
* @Date: 2023/3/3 14:00
* @Description: OpenAiProxyService
*/
@Slf4j
public class OpenAiProxyService extends OpenAiService {
private final Random random = new Random();
private final ChatGPTProperties chatGPTProperties;
private final OkHttpClient client;
private Cache<String, LinkedList<ChatMessage>> cache;
private static final EncodingRegistry REGISTRY = Encodings.newDefaultEncodingRegistry();
public OpenAiProxyService(ChatGPTProperties chatGPTProperties) {
this(chatGPTProperties, Duration.ZERO);
}
public OpenAiProxyService(ChatGPTProperties chatGPTProperties, OkHttpClient okHttpClient) {
this(chatGPTProperties, Duration.ZERO, okHttpClient);
}
public OpenAiProxyService(ChatGPTProperties chatGPTProperties, Duration timeout) {
super(buildApi(chatGPTProperties, timeout, null), defaultClient(chatGPTProperties, timeout).dispatcher().executorService(), chatGPTProperties.getBaseUrl());
this.chatGPTProperties = chatGPTProperties;
this.cache = chatGPTProperties.getSessionExpirationTime() == null ? CacheBuilder.newBuilder().build() :
CacheBuilder.newBuilder().expireAfterAccess(chatGPTProperties.getSessionExpirationTime(), TimeUnit.MINUTES).build();
this.client = OpenAiProxyService.defaultClient(chatGPTProperties, timeout);
}
public OpenAiProxyService(ChatGPTProperties chatGPTProperties, Duration timeout, OkHttpClient client) {
super(buildApi(chatGPTProperties, timeout, client), client.dispatcher().executorService(), chatGPTProperties.getBaseUrl());
this.chatGPTProperties = chatGPTProperties;
this.cache = chatGPTProperties.getSessionExpirationTime() == null ? CacheBuilder.newBuilder().build() :
CacheBuilder.newBuilder().expireAfterAccess(chatGPTProperties.getSessionExpirationTime(), TimeUnit.MINUTES).build();
this.client = client;
}
public static OpenAiApi buildApi(ChatGPTProperties properties, Duration timeout, OkHttpClient okHttpClient) {
ObjectMapper mapper = defaultObjectMapper();
OkHttpClient client = okHttpClient == null ? defaultClient(properties, timeout) : okHttpClient;
Retrofit retrofit = defaultRetrofit(client, mapper, properties.getBaseUrl());
return retrofit.create(OpenAiApi.class);
}
public static OkHttpClient defaultClient(ChatGPTProperties properties, Duration timeout) {
if (Strings.isNullOrEmpty(properties.getProxyHost())) {
return OpenAiService.defaultClient(properties, timeout);
}
// Create proxy object
Proxy proxy = new Proxy(Proxy.Type.SOCKS, new InetSocketAddress(properties.getProxyHost(), properties.getProxyPort()));
return OpenAiService.defaultClient(properties, timeout).newBuilder()
.proxy(proxy)
.build();
}
/**
* createStreamChatCompletion
*
* @param content content
*/
public void createStreamChatCompletion(String content) {
createStreamChatCompletion(content, "DEFAULT USER", System.out);
}
/**
* createStreamChatCompletion
*
* @param content content
* @param os os
*/
public void createStreamChatCompletion(String content, OutputStream os) {
createStreamChatCompletion(content, "DEFAULT USER", os);
}
/**
* createStreamChatCompletion
*
* @param content content
* @param user user
* @param os os
*/
public void createStreamChatCompletion(String content, String user, OutputStream os) {
createStreamChatCompletion(content, user, chatGPTProperties.getChatModel(), os);
}
/**
* createStreamChatCompletion
*
* @param content content
* @param user user
* @param model model
* @param os os
*/
public void createStreamChatCompletion(String content, String user, String model, OutputStream os) {
createStreamChatCompletion(RoleEnum.USER.getRoleName(), content, user, model, 1.0D, 1.0D, os);
}
/**
* createStreamChatCompletion
*
* @param role role
* @param content content
* @param user user
* @param model model
* @param temperature temperature
* @param topP topP
* @param os os
*/
public void createStreamChatCompletion(String role, String content, String user, String model, Double temperature, Double topP, OutputStream os) {
createStreamChatCompletion(ChatCompletionRequest.builder()
.model(model)
.messages(Collections.singletonList(new ChatMessage(role, content)))
.user(user)
.temperature(temperature)
.topP(topP)
.stream(true)
.build(), os);
}
/**
* createStreamChatCompletion
*
* @param chatCompletionRequest chatCompletionRequest
* @param os os
*/
public void createStreamChatCompletion(ChatCompletionRequest chatCompletionRequest, OutputStream os) {
chatCompletionRequest.setStream(true);
chatCompletionRequest.setN(1);
String user = chatCompletionRequest.getUser();
LinkedList<ChatMessage> contextInfo = new LinkedList<>();
try {
contextInfo = cache.get(user, LinkedList::new);
} catch (ExecutionException e) {
e.printStackTrace();
}
// if the contextInfo is empty, add system prompt
addSystemPrompt(contextInfo);
contextInfo.addAll(chatCompletionRequest.getMessages());
chatCompletionRequest.setMessages(contextInfo);
List<ChatCompletionChunk> chunks = new ArrayList<>();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
// avoid frequently request, random sleep 0.5s~0.7s
if (i > 0) {
randomSleep();
}
super.streamChatCompletion(chatCompletionRequest).doOnError(Throwable::printStackTrace).blockingForEach(chunk -> {
chunk.getChoices().stream().map(choice -> choice.getMessage().getContent())
.filter(Objects::nonNull).findFirst().ifPresent(o -> {
try {
os.write(o.getBytes(Charset.defaultCharset()));
os.flush();
} catch (Exception e) {
throw new RuntimeException(e);
}
});
chunks.add(chunk);
});
os.close();
// if the last line code is correct, we can simply break the circle
break;
} catch (Exception e) {
String message = e.getMessage();
boolean overload = checkTokenUsage(message);
if (overload) {
int size = Objects.requireNonNull(cache.getIfPresent(user)).size();
Iterator<ChatMessage> iterator = Objects.requireNonNull(cache.getIfPresent(user)).iterator();
for (int j = 0; j < size / 2; j++) {
ChatMessage chatMessage = iterator.next();
if (!RoleEnum.SYSTEM.getRoleName().equals(chatMessage.getRole())) {
iterator.remove();
}
}
chatCompletionRequest.setMessages(cache.getIfPresent(user));
}
log.info("answer failed " + (i + 1) + " times, the error message is: " + message);
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
// when the call fails, remove the last item in the list
Objects.requireNonNull(cache.getIfPresent(user)).removeLast();
throw new ChatGPTException(ChatGPTErrorEnum.FAILED_TO_GENERATE_ANSWER, message);
}
}
}
LinkedList<ChatMessage> chatMessages = new LinkedList<>();
try {
chatMessages = cache.get(user, LinkedList::new);
} catch (ExecutionException e) {
e.printStackTrace();
}
chatMessages.add(new ChatMessage(RoleEnum.ASSISTANT.getRoleName(), chunks.stream()
.flatMap(chunk -> chunk.getChoices().stream())
.map(ChatCompletionChoice::getMessage)
.map(ChatMessage::getContent)
.filter(Objects::nonNull)
.collect(Collectors.joining())));
}
/**
* chatCompletion
*
* @param content content
* @return List
*/
public List<String> chatCompletion(String content) {
return chatCompletion(content, "DEFAULT USER");
}
/**
* chatCompletion
*
* @param content content
* @param user user
* @return List
*/
public List<String> chatCompletion(String content, String user) {
return chatCompletion(content, user, chatGPTProperties.getChatModel());
}
/**
* chatCompletion
*
* @param content content
* @param user user
* @param model model
* @return List
*/
public List<String> chatCompletion(String content, String user, String model) {
return chatCompletion(RoleEnum.USER.getRoleName(), content, user, model, 1.0D, 1.0D);
}
/**
* chatCompletion
*
* @param role role
* @param content content
* @param user user
* @param model model
* @param temperature temperature
* @param topP topP
* @return List
*/
public List<String> chatCompletion(String role, String content, String user, String model, Double temperature, Double topP) {
return chatCompletion(ChatCompletionRequest.builder()
.model(model)
.messages(Collections.singletonList(new ChatMessage(role, content)))
.user(user)
.temperature(temperature)
.topP(topP)
.build());
}
/**
* chatCompletion
*
* @param chatCompletionRequest chatCompletionRequest
* @return List
*/
public List<String> chatCompletion(ChatCompletionRequest chatCompletionRequest) {
String user = chatCompletionRequest.getUser();
LinkedList<ChatMessage> contextInfo = new LinkedList<>();
try {
contextInfo = cache.get(user, LinkedList::new);
} catch (ExecutionException e) {
e.printStackTrace();
}
addSystemPrompt(contextInfo);
contextInfo.addAll(chatCompletionRequest.getMessages());
chatCompletionRequest.setMessages(contextInfo);
List<ChatCompletionChoice> choices = new ArrayList<>();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
// avoid frequently request, random sleep 0.5s~0.7s
if (i > 0) {
randomSleep();
}
choices = super.createChatCompletion(chatCompletionRequest).getChoices();
// if the last line code is correct, we can simply break the circle
break;
} catch (Exception e) {
String message = e.getMessage();
boolean overload = checkTokenUsage(message);
if (overload) {
int size = Objects.requireNonNull(cache.getIfPresent(user)).size();
Iterator<ChatMessage> iterator = Objects.requireNonNull(cache.getIfPresent(user)).iterator();
for (int j = 0; j < size / 2; j++) {
ChatMessage chatMessage = iterator.next();
if (!RoleEnum.SYSTEM.getRoleName().equals(chatMessage.getRole())) {
iterator.remove();
}
}
chatCompletionRequest.setMessages(cache.getIfPresent(user));
}
log.info("answer failed " + (i + 1) + " times, the error message is: " + message);
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
// when the call fails, remove the last item in the list
Objects.requireNonNull(cache.getIfPresent(user)).removeLast();
throw new ChatGPTException(ChatGPTErrorEnum.FAILED_TO_GENERATE_ANSWER, message);
}
}
}
List<String> results = new ArrayList<>();
LinkedList<ChatMessage> chatMessages = new LinkedList<>();
try {
chatMessages = cache.get(user, LinkedList::new);
} catch (ExecutionException e) {
e.printStackTrace();
}
for (ChatCompletionChoice choice : choices) {
String text = choice.getMessage().getContent();
results.add(text);
if (FinishReasonEnum.LENGTH.getMessage().equals(choice.getFinishReason())) {
results.add("答案过长,请输入继续~");
}
chatMessages.add(choice.getMessage());
}
return results;
}
/**
* please use ChatCompletion instead
*
* @param prompt prompt
* @return List<String>
*/
@Deprecated
public List<String> completion(String prompt) {
return completion(prompt, "DEFAULT USER");
}
@Deprecated
public List<String> completion(String prompt, String user) {
return completion(prompt, user, chatGPTProperties.getModel());
}
@Deprecated
public List<String> completion(String prompt, String user, String model) {
return completion(prompt, user, model, 0D, 1D);
}
@Deprecated
public List<String> completion(String prompt, String user, String model, Double temperature, Double topP) {
return completion(CompletionRequest.builder()
.model(model)
.prompt(prompt)
.user(user)
.temperature(temperature)
.topP(topP)
.maxTokens(ModelEnum.getMaxTokens(model))
.build());
}
@Deprecated
public List<String> completion(CompletionRequest completionRequest) {
List<CompletionChoice> choices = new ArrayList<>();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
// avoid frequently request, random sleep 0.5s~0.7s
if (i > 0) {
randomSleep();
}
choices = super.createCompletion(completionRequest).getChoices();
// if the last line code is correct, we can simply break the circle
break;
} catch (Exception e) {
log.info("answer failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.FAILED_TO_GENERATE_ANSWER, e.getMessage());
}
}
}
List<String> results = new ArrayList<>();
choices.forEach(choice -> {
String text = choice.getText();
if (FinishReasonEnum.LENGTH.getMessage().equals(choice.getFinish_reason())) {
text = text + System.lineSeparator() + "The answer is too long, Please disassemble the above problems into several minor problems.";
}
results.add(text);
});
return results;
}
/**
* createImages
*
* @param prompt prompt
* @return List
*/
public List<String> createImages(String prompt) {
return createImages(prompt, "DEFAULT USER");
}
/**
* createImages
*
* @param prompt prompt
* @param user user
* @return List
*/
public List<String> createImages(String prompt, String user) {
return createImages(prompt, user, ImageResponseFormatEnum.URL);
}
/**
* createImages
*
* @param prompt prompt
* @param user user
* @param responseFormat responseFormat
* @return List
*/
public List<String> createImages(String prompt, String user, ImageResponseFormatEnum responseFormat) {
ImageResult imageResult = createImages(CreateImageRequest.builder()
.prompt(prompt)
.user(user)
.responseFormat(responseFormat.getResponseFormat())
.build());
String format = responseFormat.getResponseFormat();
return imageResult.getData().stream().map(image -> format == null ||
ImageResponseFormatEnum.URL.getResponseFormat().equals(format) ?
image.getUrl() : image.getB64Json()).collect(Collectors.toList());
}
/**
* createImages
*
* @param prompt prompt
* @param user user
* @param responseFormat responseFormat
* @param imageSizeEnum imageSizeEnum
* @return List
*/
public List<String> createImages(String prompt, String user, ImageResponseFormatEnum responseFormat, ImageSizeEnum imageSizeEnum) {
ImageResult imageResult = createImages(CreateImageRequest.builder()
.prompt(prompt)
.user(user)
.responseFormat(responseFormat.getResponseFormat())
.size(imageSizeEnum.getSize())
.build());
String format = responseFormat.getResponseFormat();
return imageResult.getData().stream().map(image -> format == null ||
ImageResponseFormatEnum.URL.getResponseFormat().equals(format) ?
image.getUrl() : image.getB64Json()).collect(Collectors.toList());
}
/**
* createImages
*
* @param createImageRequest createImageRequest
* @return ImageResult
*/
public ImageResult createImages(CreateImageRequest createImageRequest) {
ImageResult imageResult = new ImageResult();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
imageResult = super.createImage(createImageRequest);
break;
} catch (Exception e) {
log.info("image generate failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.FAILED_TO_GENERATE_IMAGE, e.getMessage());
}
}
}
return imageResult;
}
/**
* downloadImage
*
* @param prompt prompt
* @param os os
*/
public void downloadImage(String prompt, OutputStream os) {
downloadImage(prompt, ImageSizeEnum.S1024x1024.getSize(), os);
}
/**
* downloadImage
*
* @param prompt prompt
* @param n n
* @param os os
*/
public void downloadImage(String prompt, Integer n, OutputStream os) {
downloadImage(prompt, n, ImageSizeEnum.S1024x1024.getSize(), os);
}
/**
* downloadImage
*
* @param prompt prompt
* @param size size
* @param os os
*/
public void downloadImage(String prompt, String size, OutputStream os) {
downloadImage(prompt, 1, size, os);
}
/**
* downloadImage
*
* @param prompt prompt
* @param n size
* @param size size
* @param os os
*/
public void downloadImage(String prompt, Integer n, String size, OutputStream os) {
downloadImage(CreateImageRequest.builder()
.prompt(prompt)
.n(n)
.size(size)
.user("DEFAULT USER").build(), os);
}
/**
* downloadImage
*
* @param createImageRequest createImageRequest
* @param os os
*/
public void downloadImage(CreateImageRequest createImageRequest, OutputStream os) {
createImageRequest.setResponseFormat(ImageResponseFormatEnum.B64_JSON.getResponseFormat());
if (!ImageResponseFormatEnum.B64_JSON.getResponseFormat().equals(createImageRequest.getResponseFormat())) {
throw new ChatGPTException(ChatGPTErrorEnum.ERROR_RESPONSE_FORMAT);
}
List<String> imageList = createImages(createImageRequest).getData().stream()
.map(Image::getB64Json).collect(Collectors.toList());
try {
if (imageList.size() == 1) {
BufferedImage bufferedImage = getImageFromBase64(imageList.get(0));
ImageIO.write(bufferedImage, "png", os);
} else {
try (ZipOutputStream zipOut = new ZipOutputStream(os)) {
for (int i = 0; i < imageList.size(); i++) {
BufferedImage bufferedImage = getImageFromBase64(imageList.get(i));
ZipEntry zipEntry = new ZipEntry("image" + (i + 1) + ".png");
zipOut.putNextEntry(zipEntry);
ImageIO.write(bufferedImage, "png", zipOut);
zipOut.closeEntry();
}
}
}
} catch (Exception e) {
throw new ChatGPTException(ChatGPTErrorEnum.DOWNLOAD_IMAGE_ERROR);
}
}
/**
* Get Bill Since startDate
*
* @param startDate startDate (yyyy-MM-dd)
* @return bill
*/
@Deprecated
public String billingUsage(String... startDate) {
String start = startDate.length == 0 ? "2023-01-01" : startDate[0];
BigDecimal totalUsage = BigDecimal.ZERO;
DateTimeFormatter formatter = DateTimeFormatter.ofPattern("yyyy-MM-dd");
try {
LocalDate endDate = LocalDate.now();
// the max query bills scope up to 100 days. The interval for each query is defined as 3 months.
Period threeMonth = Period.ofMonths(3);
LocalDate nextDate = LocalDate.parse(start, formatter);
while (nextDate.isBefore(endDate)) {
String left = nextDate.format(formatter);
nextDate = nextDate.plus(threeMonth);
String right = nextDate.format(formatter);
totalUsage = totalUsage.add(new BigDecimal(billingUsage(left, right)));
}
} catch (Exception e) {
e.printStackTrace();
}
return totalUsage.toPlainString();
}
/**
* You can query bills for up to 100 days at a time.
*
* @param startDate startDate (yyyy-MM-dd)
* @param endDate endDate (yyyy-MM-dd)
* @return Unit: (USD)
*/
@Deprecated
public String billingUsage(@NonNull String startDate, @NonNull String endDate) {
String billingUsage = "0";
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
String cents = execute(api.billingUsage(startDate, endDate)).getTotalUsage();
billingUsage = new BigDecimal(cents).divide(new BigDecimal("100")).toPlainString();
break;
} catch (Exception e) {
log.info("query billingUsage failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.QUERY_BILLINGUSAGE_ERROR, e.getMessage());
}
}
}
return billingUsage;
}
/**
* You can query all the available billing for a given date range.
*
* @param startDate startDate
* @return billing
*/
@Deprecated
public Billing billing(String... startDate) {
String start = startDate.length == 0 ? "2023-01-01" : startDate[0];
Subscription subscription = subscription();
String usage = billingUsage(start);
String dueDate = new SimpleDateFormat("yyyy-MM-dd").format(new Date(subscription.getAccessUntil() * 1000));
String total = subscription.getSystemHardLimitUsd();
Billing billing = new Billing();
billing.setDueDate(dueDate);
billing.setTotal(total);
billing.setUsage(usage);
billing.setBalance(new BigDecimal(total).subtract(new BigDecimal(usage)).toPlainString());
return billing;
}
/**
* Obtain subscription information
*
* @return subscription information
*/
public Subscription subscription() {
Subscription subscription = null;
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
subscription = execute(api.subscription());
break;
} catch (Exception e) {
log.info("query billingUsage failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.QUERY_BILLINGUSAGE_ERROR, e.getMessage());
}
}
}
return subscription;
}
/**
* Edit
*
* @param input input
* @param instruction instruction
* @return {@link String}
*/
public String edit(String input, String instruction) {
return edit(input, instruction, EditModelEnum.TEXT_DAVINCI_EDIT_001);
}
/**
* Edit
*
* @param input input
* @param instruction instruction
* @param editModelEnum editModelEnum
* @return {@link String}
*/
public String edit(String input, String instruction, EditModelEnum editModelEnum) {
return edit(input, instruction, 1D, 1D, editModelEnum);
}
/**
* Edit
*
* @param input input
* @param instruction instruction
* @param temperature temperature
* @param topP topP
* @param editModelEnum editModelEnum
* @return {@link String}
*/
public String edit(String input, String instruction, Double temperature, Double topP, EditModelEnum editModelEnum) {
EditResult editResult = edit(EditRequest.builder()
.model(editModelEnum.getModelName())
.input(input)
.instruction(instruction)
.temperature(temperature)
.topP(topP)
.build());
List<String> results = Lists.newArrayList();
editResult.getChoices().forEach(choice -> results.add(choice.getText()));
return results.get(0);
}
/**
* edit
*
* @param editRequest editRequest
* @return results
*/
public EditResult edit(EditRequest editRequest) {
EditResult editResult = new EditResult();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
editResult = super.createEdit(editRequest);
break;
} catch (Exception e) {
log.info("edit failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.EDIT_ERROR, e.getMessage());
}
}
}
return editResult;
}
/**
* embeddings
*
* @param input input
* @return results
*/
public List<Double> embeddings(String input) {
return embeddings(input, EmbeddingModelEnum.TEXT_EMBEDDING_ADA_002);
}
/**
* embeddings
*
* @param input input
* @param embeddingModelEnum embeddingModelEnum
* @return results
*/
public List<Double> embeddings(String input, EmbeddingModelEnum embeddingModelEnum) {
return embeddings(EmbeddingRequest.builder()
.input(Collections.singletonList(input))
.model(embeddingModelEnum.getModelName())
.build()).getData().get(0).getEmbedding();
}
/**
* edit
*
* @param embeddingRequest embeddingRequest
* @return results
*/
public EmbeddingResult embeddings(EmbeddingRequest embeddingRequest) {
EmbeddingResult embeddingResult = new EmbeddingResult();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
embeddingResult = super.createEmbeddings(embeddingRequest);
} catch (Exception e) {
log.info("embeddings failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.EMBEDDINGS_ERROR, e.getMessage());
}
}
}
return embeddingResult;
}
/**
* Transcribes audio into the input language.
*
* @param filePath filePath
* @param audioResponseFormatEnum audioResponseFormatEnum
* @return text
*/
public String transcription(String filePath, AudioResponseFormatEnum audioResponseFormatEnum) {
File file = new File(filePath);
return transcription(file, audioResponseFormatEnum);
}
/**
* Transcribes audio into the input language.
*
* @param file file
* @param audioResponseFormatEnum audioResponseFormatEnum
* @return text
*/
public String transcription(File file, AudioResponseFormatEnum audioResponseFormatEnum) {
TranscriptionRequest transcriptionRequest = TranscriptionRequest.builder()
.file(file).model(AudioModelEnum.WHISPER_1.getModelName())
.responseFormat(audioResponseFormatEnum.getFormat()).build();
return transcription(transcriptionRequest);
}
/**
* Transcribes audio into the input language.
*
* @param transcriptionRequest transcriptionRequest
* @return text
*/
public String transcription(TranscriptionRequest transcriptionRequest) {
// Create Request Body
MultipartBody.Builder builder = new MultipartBody.Builder()
.setType(MultipartBody.FORM)
.addFormDataPart("model", transcriptionRequest.getModel())
.addFormDataPart("file", transcriptionRequest.getFile().getName(),
RequestBody.Companion.create(transcriptionRequest.getFile(), MediaType.parse("application/octet-stream")));
if (transcriptionRequest.getPrompt() != null) {
builder.addFormDataPart("prompt", transcriptionRequest.getPrompt());
}
if (transcriptionRequest.getResponseFormat() != null) {
builder.addFormDataPart("response_format", transcriptionRequest.getResponseFormat());
}
if (transcriptionRequest.getTemperature() != null) {
builder.addFormDataPart("temperature", String.valueOf(transcriptionRequest.getTemperature()));
}
if (transcriptionRequest.getLanguage() != null) {
builder.addFormDataPart("language", transcriptionRequest.getLanguage());
}
RequestBody requestBody = builder.build();
Request request = new Request.Builder()
.post(requestBody)
.url(baseUrl + "v1/audio/transcriptions")
.build();
String text = null;
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try (Response response = client.newCall(request).execute()) {
if (i > 0) {
randomSleep();
}
text = response.body().string();
break;
} catch (Exception e) {
log.info("transcription failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.TRANSCRIPTION_ERROR, e.getMessage());
}
}
}
return text;
}
/**
* Translates audio into English.
*
* @param filePath filePath
* @param audioResponseFormatEnum audioResponseFormatEnum
* @return text
*/
public String translation(String filePath, AudioResponseFormatEnum audioResponseFormatEnum) {
File file = new File(filePath);
return translation(file, audioResponseFormatEnum);
}
/**
* Translates audio into English.
*
* @param file file
* @param audioResponseFormatEnum audioResponseFormatEnum
* @return text
*/
public String translation(File file, AudioResponseFormatEnum audioResponseFormatEnum) {
TranslationRequest translationRequest = TranslationRequest.builder()
.file(file).model(AudioModelEnum.WHISPER_1.getModelName())
.responseFormat(audioResponseFormatEnum.getFormat()).build();
return translation(translationRequest);
}
/**
* Translates audio into English.
*
* @param translationRequest translationRequest
* @return text
*/
public String translation(TranslationRequest translationRequest) {
// Create Request Body
MultipartBody.Builder builder = new MultipartBody.Builder()
.setType(MultipartBody.FORM)
.addFormDataPart("model", translationRequest.getModel())
.addFormDataPart("file", translationRequest.getFile().getName(),
RequestBody.Companion.create(translationRequest.getFile(), MediaType.parse("application/octet-stream")));
if (translationRequest.getPrompt() != null) {
builder.addFormDataPart("prompt", translationRequest.getPrompt());
}
if (translationRequest.getResponseFormat() != null) {
builder.addFormDataPart("response_format", translationRequest.getResponseFormat());
}
if (translationRequest.getTemperature() != null) {
builder.addFormDataPart("temperature", String.valueOf(translationRequest.getTemperature()));
}
RequestBody requestBody = builder.build();
Request request = new Request.Builder()
.post(requestBody)
.url(baseUrl + "v1/audio/translations")
.build();
String text = null;
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try (Response response = client.newCall(request).execute()) {
if (i > 0) {
randomSleep();
}
text = response.body().string();
break;
} catch (Exception e) {
log.info("translation failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.TRANSLATION_ERROR, e.getMessage());
}
}
}
return text;
}
/**
* create Image Edit
*
* @param createImageEditRequest createImageEditRequest
* @param imagePath imagePath
* @param maskPath maskPath
* @return imageResult
*/
public ImageResult createImageEdit(CreateImageEditRequest createImageEditRequest, String imagePath, String maskPath) {
File image = new File(imagePath);
File mask = null;
if (maskPath != null) {
mask = new File(maskPath);
}
return createImageEdit(createImageEditRequest, image, mask);
}
/**
* create Image Edit
*
* @param createImageEditRequest createImageEditRequest
* @param image image
* @param mask mask
* @return imageResult
*/
public ImageResult createImageEdit(CreateImageEditRequest createImageEditRequest, File image, File mask) {
try {
convertColorFormats(image);
if (mask != null) {
convertColorFormats(mask);
}
} catch (Exception e) {
log.error(e.getMessage());
}
// Create Request Body
MultipartBody.Builder builder = new MultipartBody.Builder()
.setType(MultipartBody.FORM)
.addFormDataPart("prompt", createImageEditRequest.getPrompt())
.addFormDataPart("image", image.getName(),
RequestBody.Companion.create(image, MediaType.parse("application/octet-stream")));
if (mask != null) {
builder.addFormDataPart("mask", mask.getName(),
RequestBody.Companion.create(mask, MediaType.parse("application/octet-stream")));
}
if (createImageEditRequest.getN() != null) {
builder.addFormDataPart("n", String.valueOf(createImageEditRequest.getN()));
}
if (createImageEditRequest.getResponseFormat() != null) {
builder.addFormDataPart("response_format", createImageEditRequest.getResponseFormat());
}
if (createImageEditRequest.getSize() != null) {
builder.addFormDataPart("size", createImageEditRequest.getSize());
}
if (createImageEditRequest.getUser() != null) {
builder.addFormDataPart("user", createImageEditRequest.getUser());
}
ImageResult imageResult = new ImageResult();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
imageResult = execute(api.createImageEdit(builder.build()));
break;
} catch (Exception e) {
log.info("create image edit failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.CREATE_IMAGE_EDIT_ERROR, e.getMessage());
}
}
}
return imageResult;
}
/**
* create Image Edit
*
* @param createImageVariationRequest createImageVariationRequest
* @param imagePath imagePath
* @return imageResult
*/
public ImageResult createImageVariation(CreateImageVariationRequest createImageVariationRequest, String imagePath) {
File image = new File(imagePath);
return createImageVariation(createImageVariationRequest, image);
}
/**
* create Image Variation
*
* @param createImageVariationRequest createImageVariationRequest
* @param image image
* @return imageResult
*/
public ImageResult createImageVariation(CreateImageVariationRequest createImageVariationRequest, File image) {
try {
convertColorFormats(image);
} catch (Exception e) {
log.error(e.getMessage());
}
// Create Request Body
MultipartBody.Builder builder = new MultipartBody.Builder()
.setType(MultipartBody.FORM)
.addFormDataPart("image", image.getName(),
RequestBody.Companion.create(image, MediaType.parse("application/octet-stream")));
if (createImageVariationRequest.getN() != null) {
builder.addFormDataPart("n", String.valueOf(createImageVariationRequest.getN()));
}
if (createImageVariationRequest.getResponseFormat() != null) {
builder.addFormDataPart("response_format", createImageVariationRequest.getResponseFormat());
}
if (createImageVariationRequest.getSize() != null) {
builder.addFormDataPart("size", createImageVariationRequest.getSize());
}
if (createImageVariationRequest.getUser() != null) {
builder.addFormDataPart("user", createImageVariationRequest.getUser());
}
ImageResult imageResult = new ImageResult();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
imageResult = execute(api.createImageVariation(builder.build()));
break;
} catch (Exception e) {
log.info("create image variation failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.CREATE_IMAGE_VARIATION_ERROR, e.getMessage());
}
}
}
return imageResult;
}
/**
* list files
*
* @return files
*/
public List<com.theokanning.openai.file.File> listFiles() {
List<com.theokanning.openai.file.File> files = new ArrayList<>();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
files = super.listFiles();
break;
} catch (Exception e) {
log.info("list files failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.LIST_FILES_ERROR, e.getMessage());
}
}
}
return files;
}
/**
* upload file
*
* @param purpose purpose
* @param filepath filepath
* @return file
*/
public com.theokanning.openai.file.File uploadFile(@NonNull String purpose, @NonNull String filepath) {
com.theokanning.openai.file.File file = null;
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
file = super.uploadFile(purpose, filepath);
break;
} catch (Exception e) {
log.info("upload file failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.UPLOAD_FILE_ERROR, e.getMessage());
}
}
}
return file;
}
/**
* delete file
*
* @param fileId fileId
* @return deleteResult
*/
public DeleteResult deleteFile(@NonNull String fileId) {
DeleteResult deleteResult = null;
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
deleteResult = super.deleteFile(fileId);
break;
} catch (Exception e) {
log.info("delete file failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.DELETE_FILE_ERROR, e.getMessage());
}
}
}
return deleteResult;
}
/**
* retrieve file
*
* @param fileId fileId
* @return file
*/
public com.theokanning.openai.file.File retrieveFile(@NonNull String fileId) {
com.theokanning.openai.file.File file = null;
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
file = super.retrieveFile(fileId);
break;
} catch (Exception e) {
log.info("retrieve file failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.RETRIEVE_FILE_ERROR, e.getMessage());
}
}
}
return file;
}
/**
* retrieve file content
*
* @param fileId fileId
* @return file content
*/
public String retrieveFileContent(@NonNull String fileId) {
Request request = new Request.Builder()
.url(baseUrl + "v1/files/{" + fileId + "}/content")
.build();
String fileContent = null;
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try (Response response = client.newCall(request).execute()) {
if (i > 0) {
randomSleep();
}
fileContent = response.body().string();
break;
} catch (Exception e) {
log.info("retrieve file content failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.RETRIEVE_FILE_CONTENT_ERROR, e.getMessage());
}
}
}
return fileContent;
}
/**
* list fine-tunes
*
* @param fineTuneRequest fineTuneRequest
* @return fineTunes
*/
public FineTuneResult createFineTune(FineTuneRequest fineTuneRequest) {
FineTuneResult fineTuneResult = new FineTuneResult();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
fineTuneResult = super.createFineTune(fineTuneRequest);
break;
} catch (Exception e) {
log.info("create fine tune failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.CREATE_FINE_TUNE_ERROR, e.getMessage());
}
}
}
return fineTuneResult;
}
/**
* createFineTuneCompletion
*
* @param completionRequest completionRequest
* @return completionResult
*/
public CompletionResult createFineTuneCompletion(CompletionRequest completionRequest) {
CompletionResult completionResult = new CompletionResult();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
completionResult = super.createFineTuneCompletion(completionRequest);
break;
} catch (Exception e) {
log.info("create fine tune completion failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.CREATE_FINE_TUNE_COMPLETION_ERROR, e.getMessage());
}
}
}
return completionResult;
}
/**
* list fine-tunes
*
* @return fineTunes
*/
public List<FineTuneResult> listFineTunes() {
List<FineTuneResult> fineTunes = new ArrayList<>();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
fineTunes = super.listFineTunes();
break;
} catch (Exception e) {
log.info("list fine tunes failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.LIST_FINE_TUNES_ERROR, e.getMessage());
}
}
}
return fineTunes;
}
/**
* retrieve fine-tune
*
* @param fineTuneId fineTuneId
* @return fineTune
*/
public FineTuneResult retrieveFineTune(String fineTuneId) {
FineTuneResult fineTuneResult = new FineTuneResult();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
fineTuneResult = super.retrieveFineTune(fineTuneId);
break;
} catch (Exception e) {
log.info("retrieve fine tune failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.RETRIEVE_FINE_TUNE_ERROR, e.getMessage());
}
}
}
return fineTuneResult;
}
/**
* cancel fine-tune
*
* @param fineTuneId fineTuneId
* @return fineTune
*/
public FineTuneResult cancelFineTune(String fineTuneId) {
FineTuneResult fineTuneResult = new FineTuneResult();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
fineTuneResult = super.cancelFineTune(fineTuneId);
break;
} catch (Exception e) {
log.info("cancel fine tune failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.CANCEL_FINE_TUNE_ERROR, e.getMessage());
}
}
}
return fineTuneResult;
}
/**
* list fine-tune events
*
* @param fineTuneId fineTuneId
* @return fineTuneEvents
*/
public List<FineTuneEvent> listFineTuneEvents(String fineTuneId) {
List<FineTuneEvent> fineTuneEvents = new ArrayList<>();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
fineTuneEvents = super.listFineTuneEvents(fineTuneId);
break;
} catch (Exception e) {
log.info("list fine tune events failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.LIST_FINE_TUNE_EVENTS_ERROR, e.getMessage());
}
}
}
return fineTuneEvents;
}
/**
* delete fine-tune
*
* @param fineTuneId fineTuneId
* @return deleteResult
*/
public DeleteResult deleteFineTune(String fineTuneId) {
DeleteResult deleteResult = new DeleteResult();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
deleteResult = super.deleteFineTune(fineTuneId);
break;
} catch (Exception e) {
log.info("delete fine tune failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.DELETE_FINE_TUNE_ERROR, e.getMessage());
}
}
}
return deleteResult;
}
/**
* create moderation
*
* @param moderationRequest moderationRequest
* @return moderationResult
*/
public ModerationResult createModeration(ModerationRequest moderationRequest) {
ModerationResult moderationResult = new ModerationResult();
for (int i = 0; i < chatGPTProperties.getRetries(); i++) {
try {
if (i > 0) {
randomSleep();
}
moderationResult = super.createModeration(moderationRequest);
break;
} catch (Exception e) {
log.info("create moderation failed " + (i + 1) + " times, the error message is: " + e.getMessage());
if (i == chatGPTProperties.getRetries() - 1) {
e.printStackTrace();
throw new ChatGPTException(ChatGPTErrorEnum.CREATE_MODERATION_ERROR, e.getMessage());
}
}
}
return moderationResult;
}
/**
* force clear cache
*
* @param cacheName cacheName
*/
public void forceClearCache(String cacheName) {
this.cache.invalidate(cacheName);
}
/**
* retrieveCache
*
* @return cache
*/
public Cache<String, LinkedList<ChatMessage>> retrieveCache() {
return this.cache;
}
/**
* retrieveChatMessage
*
* @param key key
* @return chatMessage
*/
public LinkedList<ChatMessage> retrieveChatMessage(String key) {
return this.cache.getIfPresent(key);
}
/**
* setCache
*
* @param cache cache
*/
public void setCache(Cache<String, LinkedList<ChatMessage>> cache) {
String systemPrompt = getSystemPrompt();
if (Strings.isNullOrEmpty(getSystemPrompt())) {
this.cache = cache;
return;
}
ConcurrentMap<String, LinkedList<ChatMessage>> map = cache.asMap();
for (LinkedList<ChatMessage> chatMessages : map.values()) {
boolean findSystemFlag = false;
for (ChatMessage chatMessage : chatMessages) {
if (chatMessage.getRole().equals(RoleEnum.SYSTEM.getRoleName())) {
chatMessage.setContent(systemPrompt);
findSystemFlag = true;
break;
}
}
if (!findSystemFlag) {
chatMessages.addFirst(new ChatMessage(RoleEnum.SYSTEM.getRoleName(), systemPrompt));
}
}
this.cache = cache;
}
/**
* addCache
*
* @param key key
* @param chatMessages chatMessages
*/
public void addCache(String key, LinkedList<ChatMessage> chatMessages) {
if (Strings.isNullOrEmpty(getSystemPrompt())) {
this.cache.put(key, chatMessages);
return;
}
boolean findSystemFlag = false;
for (ChatMessage chatMessage : chatMessages) {
if (chatMessage.getRole().equals(RoleEnum.SYSTEM.getRoleName())) {
chatMessage.setContent(getSystemPrompt());
findSystemFlag = true;
break;
}
}
if (!findSystemFlag) {
chatMessages.addFirst(new ChatMessage(RoleEnum.SYSTEM.getRoleName(), getSystemPrompt()));
}
this.cache.put(key, chatMessages);
}
/**
* 设置系统提示
*
* @param systemPrompt 系统提示
*/
public void setSystemPrompt(String systemPrompt) {
super.setSystemPrompt(systemPrompt);
ConcurrentMap<String, LinkedList<ChatMessage>> map = cache.asMap();
for (LinkedList<ChatMessage> chatMessages : map.values()) {
for (ChatMessage chatMessage : chatMessages) {
if (chatMessage.getRole().equals(RoleEnum.SYSTEM.getRoleName())) {
chatMessage.setContent(systemPrompt);
break;
}
}
}
}
/**
* 获取系统提示
*
* @return {@link String}
*/
public String getSystemPrompt() {
return super.getSystemPrompt();
}
/**
* 清理系统提示
*/
public void cleanUpSystemPrompt() {
super.setSystemPrompt(null);
ConcurrentMap<String, LinkedList<ChatMessage>> map = cache.asMap();
for (LinkedList<ChatMessage> chatMessages : map.values()) {
Iterator<ChatMessage> iterator = chatMessages.iterator();
while (iterator.hasNext()) {
ChatMessage chatMessage = iterator.next();
if (chatMessage.getRole().equals(RoleEnum.SYSTEM.getRoleName())) {
iterator.remove();
break;
}
}
}
}
/**
* 计数Token
*
* @param text 文本
* @return int
*/
public int countTokens(String text) {
return countTokens(text, ModelType.GPT_3_5_TURBO);
}
/**
* 计数Token
*
* @param text 文本
* @param modelType 型号类型
* @return int
*/
public int countTokens(String text, ModelType modelType) {
return getEncodingForModel(modelType).countTokens(text);
}
/**
* 获取模型编码
*
* @param modelType 型号类型
* @return {@link Encoding}
*/
public Encoding getEncodingForModel(ModelType modelType) {
return REGISTRY.getEncodingForModel(modelType);
}
private void randomSleep() throws InterruptedException {
Thread.sleep(500 + random.nextInt(200));
}
private static boolean checkTokenUsage(String message) {
return message != null && message.contains("This model's maximum context length is");
}
private BufferedImage getImageFromBase64(String base64) throws IOException {
byte[] imageBytes = Base64.getDecoder().decode(base64.getBytes());
try (ByteArrayInputStream bis = new ByteArrayInputStream(imageBytes)) {
return ImageIO.read(bis);
}
}
/**
* convert color formats
*
* @param image image
* @throws IOException IOException
*/
private void convertColorFormats(File image) throws IOException {
BufferedImage inputImage = ImageIO.read(image);
// Get the color model of the image
ComponentColorModel componentColorModel = (ComponentColorModel) inputImage.getColorModel();
// Check the pixel format of the image
int pixelSize = componentColorModel.getPixelSize();
int numComponents = componentColorModel.getNumComponents();
boolean isRGBA = pixelSize == 32 && numComponents == 4;
boolean isL = pixelSize == 8 && numComponents == 1;
boolean isLA = pixelSize == 16 && numComponents == 2;
if (!isRGBA && !isL && !isLA) {
// Create a new RGBA image
BufferedImage outputImage = new BufferedImage(inputImage.getWidth(), inputImage.getHeight(),
BufferedImage.TYPE_INT_ARGB);
// Draw the original image to the new image
Graphics2D g2d = outputImage.createGraphics();
g2d.drawImage(inputImage, 0, 0, null);
g2d.dispose();
// Save New Image
ImageIO.write(outputImage, "png", image);
}
}
private void addSystemPrompt(LinkedList<ChatMessage> contextInfo) {
if (CollectionUtils.isEmpty(contextInfo) && !Strings.isNullOrEmpty(getSystemPrompt())) {
ChatMessage chatMessage = new ChatMessage();
chatMessage.setRole(RoleEnum.SYSTEM.getRoleName());
chatMessage.setContent(getSystemPrompt());
contextInfo.addFirst(chatMessage);
}
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder",
"com.theokanning.openai.edit.EditRequest.builder",
"com.theokanning.openai.embedding.EmbeddingRequest.builder",
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((4608, 4641), 'com.google.common.cache.CacheBuilder.newBuilder'), ((4660, 4775), 'com.google.common.cache.CacheBuilder.newBuilder'), ((4660, 4767), 'com.google.common.cache.CacheBuilder.newBuilder'), ((5236, 5269), 'com.google.common.cache.CacheBuilder.newBuilder'), ((5288, 5403), 'com.google.common.cache.CacheBuilder.newBuilder'), ((5288, 5395), 'com.google.common.cache.CacheBuilder.newBuilder'), ((6267, 6383), 'io.github.asleepyfish.service.openai.OpenAiService.defaultClient'), ((6267, 6358), 'io.github.asleepyfish.service.openai.OpenAiService.defaultClient'), ((6267, 6328), 'io.github.asleepyfish.service.openai.OpenAiService.defaultClient'), ((7515, 7542), 'io.github.asleepyfish.enums.chat.RoleEnum.USER.getRoleName'), ((8059, 8358), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8059, 8333), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8059, 8303), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8059, 8275), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8059, 8233), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8059, 8205), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8059, 8120), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((10908, 10967), 'io.github.asleepyfish.enums.chat.RoleEnum.SYSTEM.getRoleName'), ((10908, 10937), 'io.github.asleepyfish.enums.chat.RoleEnum.SYSTEM.getRoleName'), ((11938, 11970), 'io.github.asleepyfish.enums.chat.RoleEnum.ASSISTANT.getRoleName'), ((13009, 13036), 'io.github.asleepyfish.enums.chat.RoleEnum.USER.getRoleName'), ((13502, 13771), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((13502, 13746), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((13502, 13718), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((13502, 13676), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((13502, 13648), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((13502, 13563), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((15459, 15518), 'io.github.asleepyfish.enums.chat.RoleEnum.SYSTEM.getRoleName'), ((15459, 15488), 'io.github.asleepyfish.enums.chat.RoleEnum.SYSTEM.getRoleName'), ((16659, 16728), 'io.github.asleepyfish.enums.chat.FinishReasonEnum.LENGTH.getMessage'), ((16659, 16695), 'io.github.asleepyfish.enums.chat.FinishReasonEnum.LENGTH.getMessage'), ((17619, 17889), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((17619, 17864), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((17619, 17806), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((17619, 17778), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((17619, 17736), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((17619, 17708), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((17619, 17676), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((19016, 19086), 'io.github.asleepyfish.enums.chat.FinishReasonEnum.LENGTH.getMessage'), ((19016, 19052), 'io.github.asleepyfish.enums.chat.FinishReasonEnum.LENGTH.getMessage'), ((20463, 20525), 'io.github.asleepyfish.enums.image.ImageResponseFormatEnum.URL.getResponseFormat'), ((20463, 20510), 'io.github.asleepyfish.enums.image.ImageResponseFormatEnum.URL.getResponseFormat'), ((21405, 21467), 'io.github.asleepyfish.enums.image.ImageResponseFormatEnum.URL.getResponseFormat'), ((21405, 21452), 'io.github.asleepyfish.enums.image.ImageResponseFormatEnum.URL.getResponseFormat'), ((22699, 22733), 'io.github.asleepyfish.enums.image.ImageSizeEnum.S1024x1024.getSize'), ((22974, 23008), 'io.github.asleepyfish.enums.image.ImageSizeEnum.S1024x1024.getSize'), ((23965, 24017), 'io.github.asleepyfish.enums.image.ImageResponseFormatEnum.B64_JSON.getResponseFormat'), ((24033, 24132), 'io.github.asleepyfish.enums.image.ImageResponseFormatEnum.B64_JSON.getResponseFormat'), ((24033, 24085), 'io.github.asleepyfish.enums.image.ImageResponseFormatEnum.B64_JSON.getResponseFormat'), ((30314, 30555), 'com.theokanning.openai.edit.EditRequest.builder'), ((30314, 30530), 'com.theokanning.openai.edit.EditRequest.builder'), ((30314, 30502), 'com.theokanning.openai.edit.EditRequest.builder'), ((30314, 30460), 'com.theokanning.openai.edit.EditRequest.builder'), ((30314, 30418), 'com.theokanning.openai.edit.EditRequest.builder'), ((30314, 30388), 'com.theokanning.openai.edit.EditRequest.builder'), ((32090, 32256), 'com.theokanning.openai.embedding.EmbeddingRequest.builder'), ((32090, 32231), 'com.theokanning.openai.embedding.EmbeddingRequest.builder'), ((32090, 32173), 'com.theokanning.openai.embedding.EmbeddingRequest.builder'), ((33971, 34153), 'io.github.asleepyfish.entity.audio.TranscriptionRequest.builder'), ((33971, 34145), 'io.github.asleepyfish.entity.audio.TranscriptionRequest.builder'), ((33971, 34076), 'io.github.asleepyfish.entity.audio.TranscriptionRequest.builder'), ((33971, 34029), 'io.github.asleepyfish.entity.audio.TranscriptionRequest.builder'), ((34036, 34075), 'io.github.asleepyfish.enums.audio.AudioModelEnum.WHISPER_1.getModelName'), ((37211, 37391), 'io.github.asleepyfish.entity.audio.TranslationRequest.builder'), ((37211, 37383), 'io.github.asleepyfish.entity.audio.TranslationRequest.builder'), ((37211, 37314), 'io.github.asleepyfish.entity.audio.TranslationRequest.builder'), ((37211, 37267), 'io.github.asleepyfish.entity.audio.TranslationRequest.builder'), ((37274, 37313), 'io.github.asleepyfish.enums.audio.AudioModelEnum.WHISPER_1.getModelName'), ((58713, 58742), 'io.github.asleepyfish.enums.chat.RoleEnum.SYSTEM.getRoleName'), ((58996, 59025), 'io.github.asleepyfish.enums.chat.RoleEnum.SYSTEM.getRoleName'), ((59562, 59591), 'io.github.asleepyfish.enums.chat.RoleEnum.SYSTEM.getRoleName'), ((59822, 59851), 'io.github.asleepyfish.enums.chat.RoleEnum.SYSTEM.getRoleName'), ((60376, 60405), 'io.github.asleepyfish.enums.chat.RoleEnum.SYSTEM.getRoleName'), ((61190, 61219), 'io.github.asleepyfish.enums.chat.RoleEnum.SYSTEM.getRoleName'), ((64151, 64180), 'io.github.asleepyfish.enums.chat.RoleEnum.SYSTEM.getRoleName')]
|
/**
* Copyright 2021 Rochester Institute of Technology (RIT). Developed with
* government support under contract 70RCSA22C00000008 awarded by the United
* States Department of Homeland Security for Cybersecurity and Infrastructure Security Agency.
*
* Permission is hereby granted, free of charge, to any person obtaining a copy
* of this software and associated documentation files (the “Software”), to deal
* in the Software without restriction, including without limitation the rights
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
* copies of the Software, and to permit persons to whom the Software is
* furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
package edu.rit.se.nvip.reconciler.openai;
import com.theokanning.openai.OpenAiHttpException;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatCompletionResult;
import com.theokanning.openai.completion.chat.ChatMessage;
import org.apache.logging.log4j.LogManager;
import org.apache.logging.log4j.Logger;
import java.util.ArrayList;
import java.util.List;
import java.util.concurrent.*;
public class GPTFilterModel {
private final Logger logger = LogManager.getLogger(getClass().getSimpleName());
private static final String MODEL = "gpt-3.5-turbo";
private static final double TEMP = 0.0;
private static final String PASS = "0";
private static final String FAIL = "1";
private static final String SYS_MESSAGE = String.format("You are a validation engine for vulnerability data scraped from the web." +
" If a user's message looks like a CVE description without errors, respond with \"%s\" or else \"%s\"", PASS, FAIL);
private static final String SYS_ROLE = "system";
private static final String USER_ROLE = "user";
private OpenAIRequestHandler requestHandler;
public GPTFilterModel() {
requestHandler = OpenAIRequestHandler.getInstance();
}
public void setRequestHandler(OpenAIRequestHandler handler) {
this.requestHandler = handler;
}
public boolean callModel(String arg) throws OpenAiInvalidReturnException{
try {
ChatCompletionRequest request = formRequest(arg);
Future<ChatCompletionResult> futureRes = requestHandler.createChatCompletion(request, RequestorIdentity.FILTER);
ChatCompletionResult res = futureRes.get();
return getAnswer(res);
} catch (OpenAiHttpException | InterruptedException | ExecutionException ex) {
logger.error(ex);
return true; // need a default answer
}
}
public int tokenCount(String description) {
return requestHandler.chatCompletionTokenCount(formRequest(description));
}
private ChatCompletionRequest formRequest(String description) {
List<ChatMessage> messages = formMessages(description);
return ChatCompletionRequest.builder().model(MODEL).temperature(TEMP).n(1).messages(messages).maxTokens(1).build();
}
private List<ChatMessage> formMessages(String description) {
List<ChatMessage> messages = new ArrayList<>();
messages.add(new ChatMessage(SYS_ROLE, SYS_MESSAGE));
messages.add(new ChatMessage(USER_ROLE, description));
return messages;
}
private boolean getAnswer(ChatCompletionResult res) throws OpenAiInvalidReturnException {
String answer = res.getChoices().get(0).getMessage().getContent();
switch (answer) {
case PASS:
return true;
case FAIL:
return false;
default:
throw new OpenAiInvalidReturnException("OpenAi responded with \"" + answer + "\"");
}
}
public static class OpenAiInvalidReturnException extends Exception {
public OpenAiInvalidReturnException(String errorMessage) {
super(errorMessage);
}
}
public static void main(String[] args) throws OpenAiInvalidReturnException, InterruptedException {
GPTFilterModel model = new GPTFilterModel();
ExecutorService executor = Executors.newFixedThreadPool(Runtime.getRuntime().availableProcessors());
int a = 0;
for (int i = 0; i < 5; i++) {
int finalI = i;
executor.submit(() -> {
try {
boolean result = model.callModel("testing # " + finalI);
System.out.println("trial # " + finalI + " evaluated as " + result);
} catch (OpenAiInvalidReturnException e) {
System.out.println(e.toString());
}
});
}
executor.shutdown();
boolean res = executor.awaitTermination(10, TimeUnit.SECONDS);
OpenAIRequestHandler.getInstance().shutdown();
}
}
|
[
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((3549, 3656), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3549, 3648), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3549, 3635), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3549, 3616), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3549, 3611), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3549, 3593), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
package com.example.gpt3javaexample.services;
import com.example.gpt3javaexample.aop.SaveToLogs;
import com.theokanning.openai.OpenAiService;
import com.theokanning.openai.completion.CompletionChoice;
import com.theokanning.openai.completion.CompletionRequest;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.stereotype.Service;
@Service
public class GPTService {
@Value("${openai.max_tokens}")
private int MAX_TOKENS;
@Value("${openai.model}")
private String MODEL;
private final OpenAiService service;
private final StringBuilder chatHistory;
@Autowired
public GPTService(OpenAiService service) {
this.service = service;
this.chatHistory = new StringBuilder();
}
@SaveToLogs
public String doRequest(String prompt, Boolean newChat){
if (newChat){
clearHistory();
}
chatHistory.append("Input: ").append(prompt).append("\nOutput: ");
CompletionRequest request = CompletionRequest.builder()
.prompt(chatHistory.toString())
.model(MODEL)
.maxTokens(MAX_TOKENS)
.build();
String response = service.createCompletion(request).getChoices().stream()
.map(CompletionChoice::getText)
.reduce(String::concat)
.orElse("I don't know what to say");
chatHistory.append(response).append("\n");
return response;
}
public void clearHistory(){
chatHistory.delete(0, chatHistory.length());
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((1077, 1246), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1077, 1221), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1077, 1182), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1077, 1152), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
package br.com.alura.screenmatch.service;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.service.OpenAiService;
public class ConsumoChatGPT {
public static String obterTraducao(String texto) {
OpenAiService service = new OpenAiService("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");
CompletionRequest requisicao = CompletionRequest.builder()
.model("text-davinci-003")
.prompt("traduza para o português o texto: " + texto)
.maxTokens(1000)
.temperature(0.7)
.build();
var resposta = service.createCompletion(requisicao);
return resposta.getChoices().get(0).getText();
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((389, 622), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((389, 597), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((389, 563), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((389, 530), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((389, 459), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
package me.bowon.springbootdeveloper.controller;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.service.OpenAiService;
import lombok.RequiredArgsConstructor;
import me.bowon.springbootdeveloper.domain.Song;
import me.bowon.springbootdeveloper.domain.YoutubeData;
import me.bowon.springbootdeveloper.service.BlogService;
import me.bowon.springbootdeveloper.service.GptService;
import me.bowon.springbootdeveloper.service.YoutubeService;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.http.ResponseEntity;
import org.springframework.web.bind.annotation.PostMapping;
import org.springframework.web.bind.annotation.RequestBody;
import org.springframework.web.bind.annotation.RequestMapping;
import org.springframework.web.bind.annotation.RestController;
import me.bowon.springbootdeveloper.service.BlogService;
import java.io.IOException;
import java.security.GeneralSecurityException;
import java.util.ArrayList;
import java.util.List;
@RequiredArgsConstructor
@RestController
@RequestMapping(value = "/gpt")
public class GptTest {
@Value("${openai.api-key}")
private String apiKey;
private final GptService gptService;
private final YoutubeService youtubeService;
private final String promptFormat = // 프롬프트 양식
"Desired Format: 1. song-singer, \n Input: 다음 일기를 보고 노래 3가지를 추천해줘 \n";
private String data;
@PostMapping("/post")
public List<YoutubeData> sendQuestion(@RequestBody String request) throws GeneralSecurityException, IOException {
OpenAiService service = new OpenAiService(apiKey);
CompletionRequest completionRequest = CompletionRequest.builder()
.prompt(promptFormat + request)
.model("text-davinci-003")
.echo(false)
.maxTokens(100)
.temperature(0.7)
.build();
data = service.createCompletion(completionRequest).getChoices().toString();
List<Song> songs = gptService.parseSong(data);
System.out.println(songs);
List<YoutubeData> youtubeDataList = youtubeService.youtubeApi(songs);
return youtubeDataList;
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((1721, 1959), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1721, 1934), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1721, 1900), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1721, 1868), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1721, 1839), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1721, 1796), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
package com.github.pablwoaraujo;
import java.util.Arrays;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
public class Main {
public static void main(String[] args) {
var user = "Gere 5 produtos";
var system = "Você é um gerador de produtos fictícios para um ecommerce e deve gerar apenas o nome dos produtos solicitados pelo usuário";
var apiKey = System.getenv("OPENAI_API_KEY");
OpenAiService service = new OpenAiService(apiKey);
ChatCompletionRequest completionRequest = ChatCompletionRequest
.builder()
.model("gpt-3.5-turbo")
.messages(Arrays.asList(
new ChatMessage(ChatMessageRole.USER.value(), user),
new ChatMessage(ChatMessageRole.SYSTEM.value(), system)))
.build();
service
.createChatCompletion(completionRequest)
.getChoices()
.forEach(c -> System.out.println(c.getMessage().getContent()));
System.out.println("Hello world!");
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value"
] |
[((893, 921), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((970, 1000), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value')]
|
package br.com.fiap.gsjava.controllers;
import br.com.fiap.gsjava.models.ChatGPT;
import br.com.fiap.gsjava.repositories.ChatGPTRepository;
import br.com.fiap.gsjava.service.OpenAiService;
import jakarta.validation.ConstraintViolationException;
import jakarta.validation.Valid;
import org.slf4j.LoggerFactory;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.data.web.PageableDefault;
import org.springframework.data.web.PagedResourcesAssembler;
import org.springframework.hateoas.EntityModel;
import org.springframework.hateoas.PagedModel;
import org.springframework.http.HttpStatus;
import org.springframework.http.ResponseEntity;
import org.springframework.web.bind.annotation.*;
import org.springframework.web.server.ResponseStatusException;
import com.theokanning.openai.completion.CompletionRequest;
import org.springframework.data.domain.Pageable;
import org.slf4j.Logger;
@RestController
@RequestMapping("/chatbot")
public class ChatGPTController {
@Autowired
ChatGPTRepository repo;
@Autowired
PagedResourcesAssembler<ChatGPT> assembler;
Logger log = LoggerFactory.getLogger(ChatGPTController.class);
private static final String API_KEY = "Sua Chave Aqui";
@GetMapping
public PagedModel<EntityModel<ChatGPT>> index(@PageableDefault(size = 5) Pageable pageable) {
return assembler.toModel(repo.findAll(pageable));
}
@GetMapping("/busca/{id}")
public EntityModel<ChatGPT> show(@PathVariable Long id) {
log.info("buscar chat com id: " + id);
ChatGPT chatGPT = repo.findById(id).orElseThrow(() ->
new ResponseStatusException(HttpStatus.NOT_FOUND, "Cliente não encontrado"));
return chatGPT.toModel();
}
@PostMapping("/api")
public ResponseEntity<ChatGPT> create(@RequestBody @Valid ChatGPT input) {
OpenAiService service = new OpenAiService(API_KEY);
CompletionRequest request = CompletionRequest.builder()
.model("text-davinci-003")
.prompt(input.getPergunta())
.maxTokens(400)
.build();
String resposta = service.createCompletion(request).getChoices().get(0).getText();
ChatGPT chatGPT = new ChatGPT(input.getPergunta(), resposta);
log.info("Saída do chatbot: " + chatGPT);
repo.save(chatGPT);
return ResponseEntity.status(HttpStatus.CREATED).body(chatGPT);
}
@DeleteMapping("/{id}")
public ResponseEntity<ChatGPT>destroy(@PathVariable Long id) {
log.info("deletar chat com o id: " + id);
ChatGPT chatgpt = repo.findById(id).orElseThrow(() ->
new ResponseStatusException(HttpStatus.NOT_FOUND, "Chat não encontrado"));;
repo.delete(chatgpt);
return ResponseEntity.noContent().build();
}
@ResponseStatus(HttpStatus.BAD_REQUEST)
@ExceptionHandler(ConstraintViolationException.class)
public ResponseEntity<String> handleValidationExceptions(ConstraintViolationException ex) {
log.error("Erro de validação: ", ex);
return ResponseEntity.badRequest().body(ex.getMessage());
}
@ResponseStatus(HttpStatus.INTERNAL_SERVER_ERROR)
@ExceptionHandler(Exception.class)
public ResponseEntity<String> handleAllExceptions(Exception ex) {
log.error("Erro não esperado: ", ex);
return ResponseEntity.status(HttpStatus.INTERNAL_SERVER_ERROR).body("Ocorreu um erro inesperado. Tente novamente mais tarde.");
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((2006, 2182), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2006, 2156), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2006, 2123), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2006, 2077), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2448, 2503), 'org.springframework.http.ResponseEntity.status'), ((2868, 2902), 'org.springframework.http.ResponseEntity.noContent'), ((3179, 3228), 'org.springframework.http.ResponseEntity.badRequest'), ((3469, 3588), 'org.springframework.http.ResponseEntity.status')]
|
package com.technoguyfication.admingpt;
import java.io.InputStream;
import java.io.InputStreamReader;
import java.time.Duration;
import java.util.LinkedList;
import java.util.List;
import java.util.regex.Matcher;
import java.util.regex.Pattern;
import java.util.stream.Stream;
import org.bstats.bukkit.Metrics;
import org.bstats.charts.SimplePie;
import org.bstats.charts.SingleLineChart;
import org.bukkit.Bukkit;
import org.bukkit.ChatColor;
import org.bukkit.configuration.file.FileConfiguration;
import org.bukkit.configuration.file.YamlConfiguration;
import org.bukkit.event.EventException;
import org.bukkit.event.EventHandler;
import org.bukkit.event.Listener;
import org.bukkit.event.player.AsyncPlayerChatEvent;
import org.bukkit.plugin.java.JavaPlugin;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatCompletionResult;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
public class AdminGPT extends JavaPlugin implements Listener {
Pattern responsePattern = Pattern.compile("<([ctp])>\\/?(.*)<\\/[ctp]>");
OpenAiService service;
LinkedList<ChatMessage> messageHistory = new LinkedList<ChatMessage>();
String systemPrompt;
String languageModel;
int historyLength;
long timeoutSeconds;
Double temperature;
List<String> commandBlacklist;
// metrics
int totalMessages = 0;
int totalCommands = 0;
int totalResponses = 0;
@Override
public void onEnable() {
// bStats
int pluginId = 18196;
Metrics metrics = new Metrics(this, pluginId);
FileConfiguration config = this.getConfig();
InputStream langStream = this.getResource("lang.yml");
// Load lang.yml
YamlConfiguration langConfig = new YamlConfiguration();
try {
langConfig.load(new InputStreamReader(langStream));
// Load system prompt from lang.yml
systemPrompt = langConfig.getString("openai-system-prompt");
} catch (Exception e) {
getLogger().severe("Failed to load lang.yml file.");
e.printStackTrace();
// Disable plugin
this.setEnabled(false);
return;
}
// Load config
String apiKey = config.getString("openai-api-key");
if (apiKey == null || apiKey.isBlank() || apiKey.equals("your-api-key-here")) {
getLogger().severe("No OpenAI API key found in config.yml. Please add one and restart the server.");
// Save default config
this.saveDefaultConfig();
// Disable plugin
this.setEnabled(false);
return;
}
languageModel = config.getString("openai-language-model");
temperature = config.getDouble("openai-model-temperature");
timeoutSeconds = config.getLong("openai-timeout-secs");
historyLength = config.getInt("history-length");
commandBlacklist = config.getStringList("command-blacklist");
// Add bStats charts
metrics.addCustomChart(new SimplePie("language-model", () -> languageModel));
metrics.addCustomChart(new SingleLineChart("messages-sent", () -> {
var total = totalMessages;
totalMessages = 0;
return total;
}));
metrics.addCustomChart(new SingleLineChart("commands-run", () -> {
var total = totalCommands;
totalCommands = 0;
return total;
}));
metrics.addCustomChart(new SingleLineChart("responses-received", () -> {
var total = totalResponses;
totalResponses = 0;
return total;
}));
// Create OpenAI service
service = new OpenAiService(apiKey, Duration.ofSeconds(timeoutSeconds)); // set response timeout
// Register event listeners
getServer().getPluginManager().registerEvents(this, this);
// Startup messages
getLogger().info("Command blacklist: " + String.join(", ", commandBlacklist));
}
@Override
public void onDisable() {
// Plugin disabled
}
@EventHandler
public void onChat(AsyncPlayerChatEvent event) throws EventException {
// Increment total messages counter
totalMessages++;
// Add new message to list
addChatMessage(new ChatMessage(ChatMessageRole.USER.value(),
String.format("%s: %s", event.getPlayer().getName(), event.getMessage())));
// Replace placeholders in the system prompt
String templatedSystemPrompt = systemPrompt
.replace("{plugins}",
String.join(", ",
Stream.of(Bukkit.getPluginManager().getPlugins()).map(p -> p.getName())
.toArray(String[]::new)))
.replace("{players}",
String.join(", ",
Bukkit.getOnlinePlayers().stream().map(p -> p.getName()).toArray(String[]::new)))
.replace("{version}", Bukkit.getVersion());
// Make a new list with the system prompt and all messages
List<ChatMessage> messages = new LinkedList<ChatMessage>();
messages.add(new ChatMessage(ChatMessageRole.SYSTEM.value(), templatedSystemPrompt));
messages.addAll(messageHistory);
// Create a chat completion request
ChatCompletionRequest request = ChatCompletionRequest
.builder()
.model(languageModel)
.messages(messages)
.user(event.getPlayer().getUniqueId().toString())
.temperature(temperature)
.build();
getLogger().fine("Sending chat completion request to OpenAI...");
Bukkit.getScheduler().runTaskAsynchronously(this, () -> {
ChatCompletionResult result = service.createChatCompletion(request);
ChatMessage responseMessage = result.getChoices().get(0).getMessage();
getLogger().fine("Received chat completion result from OpenAI.");
List<String> commands = new LinkedList<String>();
List<String> responses = new LinkedList<String>();
// Run regex on each line of the result
for (String line : responseMessage.getContent().split("\\r?\\n")) {
Matcher matcher = responsePattern.matcher(line);
if (matcher.find()) {
switch (matcher.group(1)) {
case "c":
String command = matcher.group(2);
getLogger().info(String.format("Command: %s", command));
commands.add(command);
break;
case "t":
String thought = matcher.group(2);
getLogger().info(String.format("Thought: %s", thought));
break;
case "p":
String response = matcher.group(2);
getLogger().info(String.format("Response: %s", response));
responses.add(response);
break;
default:
getLogger().warning(String.format("Invalid response pattern: %s", line));
break;
}
}
}
// Run the rest of the code on the main thread
Bukkit.getScheduler().runTask(this, () -> {
// Add commands and responses to total counters
totalCommands += commands.size();
totalResponses += responses.size();
// add the result to the list of messages
addChatMessage(responseMessage);
// Run the commands
for (String command : commands) {
// Check if command is blacklisted
String rootCommand = command.split(" ")[0];
if (commandBlacklist.contains(rootCommand.toLowerCase())) {
getLogger().warning(String.format("Command %s is blacklisted.", command));
continue;
}
Bukkit.dispatchCommand(Bukkit.getConsoleSender(), command);
}
// Broadcast response lines
for (String response : responses) {
Bukkit.broadcastMessage(ChatColor.AQUA + String.format("<AdminGPT> %s", response));
}
});
});
}
private void addChatMessage(ChatMessage message) {
// Remove oldest message if list is full
if (messageHistory.size() >= historyLength) {
messageHistory.removeFirst();
}
// Add new message to list
messageHistory.add(message);
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value"
] |
[((4511, 4539), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((4851, 4986), 'java.util.stream.Stream.of'), ((4851, 4922), 'java.util.stream.Stream.of'), ((4861, 4899), 'org.bukkit.Bukkit.getPluginManager'), ((5101, 5180), 'org.bukkit.Bukkit.getOnlinePlayers'), ((5101, 5157), 'org.bukkit.Bukkit.getOnlinePlayers'), ((5101, 5135), 'org.bukkit.Bukkit.getOnlinePlayers'), ((5416, 5446), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value'), ((5940, 8809), 'org.bukkit.Bukkit.getScheduler'), ((7705, 8797), 'org.bukkit.Bukkit.getScheduler')]
|
package com.vission.chatGPT.service;
import com.google.common.collect.Lists;
import com.theokanning.openai.completion.chat.ChatCompletionChoice;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatCompletionResult;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
import com.vission.chatGPT.properties.ChatGPTProperties;
import com.vission.chatGPT.utils.BeanUtils;
import com.vission.chatGPT.utils.JsonUtils;
import com.vission.chatGPT.utils.RedisUtils;
import java.util.List;
import lombok.RequiredArgsConstructor;
import lombok.extern.slf4j.Slf4j;
import org.apache.commons.lang3.StringUtils;
import org.springframework.stereotype.Service;
@Service
@Slf4j
@RequiredArgsConstructor
public class ChatGPTService {
private final ChatGPTProperties properties;
private final OpenAiService openAiService;
private final RedisUtils redisUtils;
/**
* 翻译助手
*
* @param original 原文
* @return 翻译结果
*/
public String translation(String original) {
StringBuilder completion = new StringBuilder();
ChatMessage newQuestionMessage = new ChatMessage(ChatMessageRole.USER.value(), original);
ChatMessage system = new ChatMessage(ChatMessageRole.SYSTEM.value(),
"你是一个翻译助手,将我说的所有话翻译成中文");
ChatCompletionRequest request = ChatCompletionRequest.builder()
.model("gpt-3.5-turbo")
.messages(Lists.newArrayList(system, newQuestionMessage))
.build();
ChatCompletionResult chatCompletion = openAiService.createChatCompletion(request);
List<ChatCompletionChoice> choices = chatCompletion.getChoices();
for (ChatCompletionChoice choice : choices) {
completion.append(choice.getMessage().getContent());
}
return completion.toString();
}
/**
* 聊天 不会保存上下文聊天
*
* @param original 原文
* @return 翻译结果
*/
public String chatCompletion(String original) {
StringBuilder completion = new StringBuilder();
ChatMessage newQuestionMessage = new ChatMessage(ChatMessageRole.USER.value(), original);
ChatCompletionRequest request = ChatCompletionRequest.builder()
.model("gpt-3.5-turbo")
.messages(Lists.newArrayList(newQuestionMessage))
.build();
ChatCompletionResult chatCompletion = openAiService.createChatCompletion(request);
List<ChatCompletionChoice> choices = chatCompletion.getChoices();
for (ChatCompletionChoice choice : choices) {
completion.append(choice.getMessage().getContent());
}
return completion.toString();
}
/**
* 聊天 会保存上下文聊天
*
* @param original 原文
* @param userUuid 用户唯一标识
* @return 翻译结果
*/
public String chatCompletionByContext(String original, String userUuid) {
List<ChatMessage> messages = findChatMessagesByUuid(userUuid);
int messageCount = (int) messages.stream().map(ChatMessage::getRole)
.filter(t -> StringUtils.equals(t, ChatMessageRole.USER.value())).count();
if (messageCount > properties.getChatGptFlowNum()) {
redisUtils.del(userUuid);
return "您的连续对话已超过上限,系统已自动清空上下文";
}
StringBuilder result = new StringBuilder();
ChatMessage newMessage = new ChatMessage(ChatMessageRole.USER.value(), original);
messages.add(newMessage);
ChatCompletionRequest request = ChatCompletionRequest.builder()
.model("gpt-3.5-turbo").messages(messages).build();
ChatGPTService.log.info("request:{}", JsonUtils.toJson(request));
ChatCompletionResult chatCompletion = openAiService.createChatCompletion(request);
List<ChatCompletionChoice> choices = chatCompletion.getChoices();
for (ChatCompletionChoice choice : choices) {
messages.add(choice.getMessage());
result.append(choice.getMessage().getContent());
}
redisUtils.set(userUuid, messages, 1800);
return result.toString();
}
private List<ChatMessage> findChatMessagesByUuid(String userUuid) {
List result = redisUtils.getList(userUuid);
return BeanUtils.deepCopyList(result, ChatMessage.class);
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value",
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((1310, 1338), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((1396, 1426), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value'), ((1552, 1722), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((1552, 1697), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((1552, 1623), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2351, 2379), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((2432, 2594), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2432, 2569), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2432, 2503), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3374, 3402), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((3711, 3739), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((3826, 3924), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3826, 3916), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3826, 3897), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
package cos.peerna.domain.gpt.service;
import com.amazonaws.services.kms.model.NotFoundException;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.theokanning.openai.completion.chat.ChatCompletionChunk;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.service.OpenAiService;
import cos.peerna.domain.gpt.dto.request.SendMessageRequest;
import cos.peerna.domain.gpt.event.ReviewReplyEvent;
import cos.peerna.domain.gpt.model.GPT;
import cos.peerna.domain.history.model.History;
import cos.peerna.domain.history.repository.HistoryRepository;
import cos.peerna.domain.reply.model.Reply;
import cos.peerna.domain.reply.repository.ReplyRepository;
import cos.peerna.domain.room.model.Chat;
import cos.peerna.domain.room.repository.ChatRepository;
import cos.peerna.global.security.dto.SessionUser;
import java.util.ArrayList;
import java.util.List;
import lombok.RequiredArgsConstructor;
import lombok.extern.slf4j.Slf4j;
import org.springframework.data.redis.core.RedisTemplate;
import org.springframework.messaging.simp.SimpMessagingTemplate;
import org.springframework.stereotype.Service;
@Slf4j
@Service
@RequiredArgsConstructor
public class GPTService {
private final ReplyRepository replyRepository;
private final SimpMessagingTemplate template;
private final RedisTemplate<String, Object> redisTemplate;
private final ObjectMapper objectMapper;
private final OpenAiService openAIService;
private final ChatRepository chatRepository;
private final HistoryRepository historyRepository;
/*
TODO: Async 로 변경
*/
public void reviewReply(ReviewReplyEvent event) {
/*
TODO: 사용자의 권한에 따른 gpt 모델 선택
*/
ChatMessage systemMessage = new ChatMessage("system", GPT.getConcept(event.question()));
ChatMessage userMessage = new ChatMessage("user", event.answer());
StringBuilder assistantMessageBuilder = new StringBuilder();
openAIService.streamChatCompletion(ChatCompletionRequest.builder()
.model(GPT.getModel())
.messages(List.of(
systemMessage,
userMessage
))
.build())
.doOnError(throwable -> sendErrorMessage(event.userId()))
.blockingForEach(chunk -> sendChatMessage(chunk, event.userId(), assistantMessageBuilder));
ChatMessage assistantMessage = new ChatMessage("assistant", assistantMessageBuilder.toString());
redisTemplate.opsForList().rightPush(String.valueOf(event.historyId()), systemMessage);
redisTemplate.opsForList().rightPush(String.valueOf(event.historyId()), userMessage);
redisTemplate.opsForList().rightPush(String.valueOf(event.historyId()), assistantMessage);
History history = historyRepository.findById(event.historyId())
.orElseThrow(() -> new NotFoundException("history not found"));
chatRepository.save(Chat.builder()
.writerId(0L)
.content(assistantMessageBuilder.toString())
.history(history)
.build());
}
/*
TODO: Async 로 변경
*/
public void sendMessage(SessionUser user, SendMessageRequest request) {
Reply lastReply = replyRepository.findFirstByUserIdOrderByIdDesc(user.getId())
.orElseThrow(() -> new NotFoundException("reply not found"));
List<ChatMessage> messages = getChatMessages(lastReply.getHistory().getId());
ChatMessage userMessage = new ChatMessage("user", request.message());
messages.add(userMessage);
StringBuilder assistantMessageBuilder = new StringBuilder();
openAIService.streamChatCompletion(ChatCompletionRequest.builder()
.model(GPT.getModel())
.messages(messages)
.build())
.doOnError(throwable -> sendErrorMessage(user.getId()))
.blockingForEach(chunk -> sendChatMessage(chunk, user.getId(), assistantMessageBuilder));
ChatMessage assistantMessage = new ChatMessage("assistant", assistantMessageBuilder.toString());
redisTemplate.opsForList().rightPush(String.valueOf(lastReply.getHistory().getId()), userMessage);
redisTemplate.opsForList().rightPush(String.valueOf(lastReply.getHistory().getId()), assistantMessage);
chatRepository.save(Chat.builder()
.writerId(user.getId())
.content(request.message())
.history(lastReply.getHistory())
.build());
chatRepository.save(Chat.builder()
.writerId(0L)
.content(assistantMessageBuilder.toString())
.history(lastReply.getHistory())
.build());
}
private List<ChatMessage> getChatMessages(Long historyId) {
List<Object> messageObjects = redisTemplate.opsForList().range(String.valueOf(historyId), 0, -1);
List<ChatMessage> messages = new ArrayList<>();
if (messageObjects == null) {
throw new NotFoundException("messageObjects is null");
}
for (Object messageObject : messageObjects) {
ChatMessage chatMessage = objectMapper.convertValue(messageObject, ChatMessage.class);
messages.add(chatMessage);
}
return messages;
}
private void sendChatMessage(ChatCompletionChunk chunk, Long userId, StringBuilder assistantMessageBuilder) {
/*
TODO: stream 이 끝나면, gpt 답변 전체를 저장
TODO: gpt에게서 오는 chunk의 순서가 보장되지 않음
*/
String message = chunk.getChoices().get(0).getMessage().getContent();
if (message == null) {
template.convertAndSend("/user/" + userId + "/gpt", GPT.getENDMessage());
return;
}
template.convertAndSend("/user/" + userId + "/gpt", message);
assistantMessageBuilder.append(message);
}
private void sendErrorMessage(Long userId) {
template.convertAndSend("/user/" + userId + "/gpt", GPT.getErrorMessage());
}
}
|
[
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((2107, 2379), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2107, 2346), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2107, 2185), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3139, 3303), 'cos.peerna.domain.room.model.Chat.builder'), ((3139, 3278), 'cos.peerna.domain.room.model.Chat.builder'), ((3139, 3244), 'cos.peerna.domain.room.model.Chat.builder'), ((3139, 3183), 'cos.peerna.domain.room.model.Chat.builder'), ((3909, 4064), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3909, 4031), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3909, 3987), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((4598, 4770), 'cos.peerna.domain.room.model.Chat.builder'), ((4598, 4745), 'cos.peerna.domain.room.model.Chat.builder'), ((4598, 4696), 'cos.peerna.domain.room.model.Chat.builder'), ((4598, 4652), 'cos.peerna.domain.room.model.Chat.builder'), ((4801, 4980), 'cos.peerna.domain.room.model.Chat.builder'), ((4801, 4955), 'cos.peerna.domain.room.model.Chat.builder'), ((4801, 4906), 'cos.peerna.domain.room.model.Chat.builder'), ((4801, 4845), 'cos.peerna.domain.room.model.Chat.builder')]
|
package link.locutus.discord.gpt.imps;
import com.knuddels.jtokkit.api.Encoding;
import com.knuddels.jtokkit.api.EncodingRegistry;
import com.knuddels.jtokkit.api.ModelType;
import com.theokanning.openai.service.OpenAiService;
import com.theokanning.openai.embedding.Embedding;
import com.theokanning.openai.embedding.EmbeddingRequest;
import com.theokanning.openai.embedding.EmbeddingResult;
import link.locutus.discord.db.AEmbeddingDatabase;
import link.locutus.discord.gpt.pw.GptDatabase;
import java.sql.SQLException;
import java.util.List;
public class AdaEmbedding extends AEmbeddingDatabase {
private final EncodingRegistry registry;
private final Encoding embeddingEncoder;
private final OpenAiService service;
public AdaEmbedding(EncodingRegistry registry, OpenAiService service, GptDatabase database) throws SQLException, ClassNotFoundException {
super("ada", database);
this.registry = registry;
this.service = service;
this.embeddingEncoder = registry.getEncodingForModel(ModelType.TEXT_EMBEDDING_ADA_002);
}
public int getEmbeddingTokenSize(String text) {
return embeddingEncoder.encode(text).size();
}
@Override
public float[] fetchEmbedding(String text) {
EmbeddingRequest request = EmbeddingRequest.builder()
.model("text-embedding-ada-002")
.input(List.of(text))
.build();
EmbeddingResult embedResult = service.createEmbeddings(request);
List<Embedding> data = embedResult.getData();
if (data.size() != 1) {
throw new RuntimeException("Expected 1 embedding, got " + data.size());
}
List<Double> result = data.get(0).getEmbedding();
float[] target = new float[result.size()];
for (int i = 0; i < target.length; i++) {
target[i] = result.get(i).floatValue();
}
return target;
}
}
|
[
"com.theokanning.openai.embedding.EmbeddingRequest.builder"
] |
[((1288, 1426), 'com.theokanning.openai.embedding.EmbeddingRequest.builder'), ((1288, 1401), 'com.theokanning.openai.embedding.EmbeddingRequest.builder'), ((1288, 1363), 'com.theokanning.openai.embedding.EmbeddingRequest.builder')]
|
package com.redis.vss;
import redis.clients.jedis.JedisPooled;
import redis.clients.jedis.Protocol;
import redis.clients.jedis.search.Document;
import redis.clients.jedis.search.IndexDefinition;
import redis.clients.jedis.search.IndexOptions;
import redis.clients.jedis.search.Query;
import redis.clients.jedis.search.Schema;
import redis.clients.jedis.search.SearchResult;
import redis.clients.jedis.util.SafeEncoder;
import java.io.FileInputStream;
import java.io.IOException;
import java.io.InputStream;
import java.io.InputStreamReader;
import java.nio.ByteBuffer;
import java.nio.ByteOrder;
import java.util.Collections;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.Properties;
import java.util.regex.Pattern;
import java.util.stream.Collectors;
import com.opencsv.CSVReader;
import com.opencsv.CSVReaderBuilder;
import com.theokanning.openai.embedding.EmbeddingRequest;
import com.theokanning.openai.service.OpenAiService;
/**
* Java VSS Wiki Articles Example
*
* @author Michael Yuan
*/
public class JavaVSSWikiArticlesExample {
// Redis client connection
private static JedisPooled client = null;
// OpenAI connection
private static OpenAiService service = null;
// Model
private static String MODEL = "text-embedding-ada-002";
private static int VECTOR_DIM = 1536; // length of the vectors
private static int VECTOR_NUMBER = 25000; // initial number of vectors
private static String INDEX_NAME = "idx_wiki"; // name of the search index
private static String INDEX_NAME_HNSW = "idx_wiki_hnsw"; // name of the search index
private static String PREFIX = "wiki"; // prefix for the document keys
private static String DISTANCE_METRIC = "COSINE"; // prefix for the document keys
private JavaVSSWikiArticlesExample() {
try {
// Initialize Redis connection
InputStream input = ClassLoader.getSystemResourceAsStream("config.properties");
Properties prop = new Properties();
prop.load(input);
client = new JedisPooled(prop.getProperty("redis.host"),
Integer.parseInt(prop.getProperty("redis.port")));
// Initialize OpenAI service connection
String token = System.getenv("OPENAI_API_KEY");
service = new OpenAiService(token);
// client = new JedisPooled(prop.getProperty("redis.host"),
// Integer.parseInt(prop.getProperty("redis.port")),
// prop.getProperty("redis.user"),
// prop.getProperty("redis.password"));
Object result = client.sendCommand(Protocol.Command.PING, "Connected to Redis...");
System.out.println(SafeEncoder.encode((byte[]) result));
} catch (Exception ex) {
ex.printStackTrace();
}
}
private void createFlatIndex() {
try {
// Drop index if exists
try {
client.ftDropIndex(INDEX_NAME);
} catch (Exception e) {
}
;
System.out.println("Creating Flat index...");
HashMap<String, Object> attr = new HashMap<String, Object>();
attr.put("TYPE", "FLOAT64");
attr.put("DIM", VECTOR_DIM);
attr.put("DISTANCE_METRIC", DISTANCE_METRIC);
attr.put("INITIAL_CAP", VECTOR_NUMBER);
// Define index schema
Schema schema = new Schema().addNumericField("id")
.addTextField("title", 3.0).as("title")
.addTextField("url", 1.0).as("url")
.addTextField("text", 2.0).as("text")
.addVectorField("title_vector", Schema.VectorField.VectorAlgo.FLAT, attr).as("title_vector")
.addVectorField("content_vector", Schema.VectorField.VectorAlgo.FLAT, attr).as("content_vector");
IndexDefinition rule = new IndexDefinition(IndexDefinition.Type.HASH)
.setPrefixes(new String[] { "wiki:" });
client.ftCreate(INDEX_NAME, IndexOptions.defaultOptions().setDefinition(rule), schema);
} catch (Exception ex) {
ex.printStackTrace();
}
}
private void createHNSWIndex() {
try {
// Drop index if exists
try {
client.ftDropIndex(INDEX_NAME_HNSW);
} catch (Exception e) {
}
;
System.out.println("Creating HNSW index...");
HashMap<String, Object> attr = new HashMap<String, Object>();
attr.put("TYPE", "FLOAT64");
attr.put("DIM", VECTOR_DIM);
attr.put("DISTANCE_METRIC", DISTANCE_METRIC);
attr.put("INITIAL_CAP", VECTOR_NUMBER);
// Define index schema
Schema schema = new Schema().addNumericField("id")
.addTextField("title", 3.0).as("title")
.addTextField("url", 1.0).as("url")
.addTextField("text", 2.0).as("text")
.addVectorField("title_vector", Schema.VectorField.VectorAlgo.HNSW, attr).as("title_vector")
.addVectorField("content_vector", Schema.VectorField.VectorAlgo.HNSW, attr).as("content_vector");
IndexDefinition rule = new IndexDefinition(IndexDefinition.Type.HASH)
.setPrefixes(new String[] { "wiki:" });
client.ftCreate(INDEX_NAME_HNSW, IndexOptions.defaultOptions().setDefinition(rule), schema);
} catch (Exception ex) {
ex.printStackTrace();
}
}
/**
* @param csvFile
* Load data from csv file to Redis hashes
*/
private void loadData(String csvFile) {
System.out.println("Loading data in Redis...");
try {
FileInputStream input = new FileInputStream(csvFile);
String[] record = null;
String key;
try (CSVReader reader = new CSVReaderBuilder(new InputStreamReader(input)).withSkipLines(1).build()) {
while ((record = reader.readNext()) != null) {
key = PREFIX + ":" + record[0];
double[] title_vector = Pattern.compile(", ")
.splitAsStream(record[4].replaceAll("\\[", "").replaceAll("\\]", ""))
.map(elem -> Double.parseDouble(elem))
.collect(Collectors.toList())
.stream().mapToDouble(Double::doubleValue).toArray();
double[] content_vector = Pattern.compile(", ")
.splitAsStream(record[5].replaceAll("\\[", "").replaceAll("\\]", ""))
.map(elem -> Double.parseDouble(elem))
.collect(Collectors.toList())
.stream().mapToDouble(Double::doubleValue).toArray();
Map<byte[], byte[]> map = new HashMap<>();
map.put("id".getBytes(), record[0].getBytes());
map.put("url".getBytes(), record[1].getBytes());
map.put("title".getBytes(), record[2].getBytes());
map.put("text".getBytes(), record[3].getBytes());
map.put("title_vector".getBytes(), doubleToByte(title_vector));
map.put("content_vector".getBytes(), doubleToByte(content_vector));
map.put("vector_id".getBytes(), record[6].getBytes());
client.hset(key.getBytes(), map);
}
}
} catch (Exception ex) {
ex.printStackTrace();
}
}
/**
* @param input
* @return byte[]
*/
public byte[] doubleToByte(double[] input) {
ByteBuffer buffer = ByteBuffer.allocate(input.length * Double.BYTES);
buffer.order(ByteOrder.LITTLE_ENDIAN);
buffer.asDoubleBuffer().put(input);
return buffer.array();
}
public void searchRedis(String indexName, String queryString, String vector_field, int k) {
// Build OpenAI embedding request
EmbeddingRequest embeddingRequest = EmbeddingRequest.builder()
.model(MODEL)
.input(Collections.singletonList(queryString))
.build();
// Get vector embeddings from Open AI service
double[] embedding = service.createEmbeddings(embeddingRequest).getData().get(0).getEmbedding()
.stream().mapToDouble(Double::doubleValue).toArray();
// Build query
Query q = new Query("*=>[KNN $k @" + vector_field + "$vec AS vector_score]")
.setSortBy("vector_score", true)
.addParam("k", k)
.addParam("vec", doubleToByte(embedding))
.limit(0, k)
.dialect(2);
// Get and iterate over search results
SearchResult res = client.ftSearch(indexName, q);
List<Document> wikis = res.getDocuments();
int i = 1;
for (Document wiki : wikis) {
float score = Float.parseFloat((String) wiki.get("vector_score"));
System.out.println(i + ". " + wiki.get("title") + " (Score: " + (1 - score) + ")");
i++;
}
}
/**
* Run Redis VSS search examples using wiki articles.
*
* @param args The arguments of the program.
*/
public static void main(String[] args) {
// Zip archive of wiki articles with OpenAI embeddings
String fileUrl = "https://cdn.openai.com/API/examples/data/vector_database_wikipedia_articles_embedded.zip";
String saveAt = "/tmp/vector_database_wikipedia_articles_embedded.zip";
// CSV file of wiki articles with OpenAI embeddings
String csvFile = "/tmp/vector_database_wikipedia_articles_embedded.csv";
// Download and unzip csv file of wiki articles with OpenAI embeddings
try {
System.out.println("Downloading and unzipping csv file...");
LoadOpenAIData.downloadUsingNIO(fileUrl, saveAt);
LoadOpenAIData.unzipZip4j(saveAt, "/tmp");
} catch (IOException e) {
e.printStackTrace();
}
JavaVSSWikiArticlesExample vssArticles = new JavaVSSWikiArticlesExample();
vssArticles.createFlatIndex();
vssArticles.createHNSWIndex();
vssArticles.loadData(csvFile);
System.out.println("### VSS query: 'modern art in Europe' in 'title_vector'");
vssArticles.searchRedis(INDEX_NAME, "modern art in Europe", "title_vector", 10);
System.out.println("### VSS query: 'modern art in Europe' in 'title_vector'");
vssArticles.searchRedis(INDEX_NAME_HNSW, "modern art in Europe", "title_vector", 10);
System.out.println("### VSS query: 'Famous battles in Scottish history' in 'content_vector'");
vssArticles.searchRedis(INDEX_NAME, "Famous battles in Scottish history", "content_vector", 10);
}
}
|
[
"com.theokanning.openai.embedding.EmbeddingRequest.builder"
] |
[((4075, 4124), 'redis.clients.jedis.search.IndexOptions.defaultOptions'), ((5457, 5506), 'redis.clients.jedis.search.IndexOptions.defaultOptions'), ((6208, 6533), 'java.util.regex.Pattern.compile'), ((6208, 6523), 'java.util.regex.Pattern.compile'), ((6208, 6490), 'java.util.regex.Pattern.compile'), ((6208, 6452), 'java.util.regex.Pattern.compile'), ((6208, 6394), 'java.util.regex.Pattern.compile'), ((6208, 6327), 'java.util.regex.Pattern.compile'), ((6582, 6907), 'java.util.regex.Pattern.compile'), ((6582, 6897), 'java.util.regex.Pattern.compile'), ((6582, 6864), 'java.util.regex.Pattern.compile'), ((6582, 6826), 'java.util.regex.Pattern.compile'), ((6582, 6768), 'java.util.regex.Pattern.compile'), ((6582, 6701), 'java.util.regex.Pattern.compile'), ((8173, 8317), 'com.theokanning.openai.embedding.EmbeddingRequest.builder'), ((8173, 8292), 'com.theokanning.openai.embedding.EmbeddingRequest.builder'), ((8173, 8229), 'com.theokanning.openai.embedding.EmbeddingRequest.builder')]
|
package com.asleepyfish.strategy.event;
import com.alibaba.fastjson2.JSONObject;
import com.asleepyfish.dto.AiQa;
import com.asleepyfish.enums.WxMessageType;
import com.asleepyfish.repository.AiQaRepository;
import com.asleepyfish.strategy.WxEventStrategy;
import com.google.common.collect.Lists;
import com.theokanning.openai.image.CreateImageRequest;
import io.github.asleepyfish.enums.ImageResponseFormatEnum;
import io.github.asleepyfish.enums.ImageSizeEnum;
import io.github.asleepyfish.util.OpenAiUtils;
import lombok.extern.slf4j.Slf4j;
import me.chanjar.weixin.common.api.WxConsts;
import me.chanjar.weixin.common.bean.result.WxMediaUploadResult;
import me.chanjar.weixin.mp.api.WxMpService;
import me.chanjar.weixin.mp.bean.kefu.WxMpKefuMessage;
import org.springframework.stereotype.Service;
import javax.annotation.Resource;
import javax.servlet.http.HttpServletResponse;
import java.io.ByteArrayInputStream;
import java.util.Base64;
import java.util.List;
import java.util.Map;
/**
* @Author: asleepyfish
* @Date: 2022/8/31 19:55
* @Description: 消息策略
*/
@Service("text")
@Slf4j
public class TextStrategy implements WxEventStrategy {
@Resource
private AiQaRepository aiQaRepository;
@Resource
private WxMpService wxMpService;
@Override
public void execute(Map<String, String> requestMap, HttpServletResponse response) throws Exception {
// 发送方账号
String openId = requestMap.get("FromUserName");
String acceptContent = requestMap.get("Content");
log.info(">>> 用户输入:{}", acceptContent);
// 关闭输出流,避免微信服务端重复发送信息
response.getOutputStream().close();
if (acceptContent.charAt(0) == '/') {
createImage(acceptContent, openId);
} else {
createCompletion(acceptContent, openId);
}
}
private void createCompletion(String acceptContent, String openId) throws Exception {
WxMpKefuMessage wxMpKefuMessage = new WxMpKefuMessage();
wxMpKefuMessage.setToUser(openId);
wxMpKefuMessage.setMsgType(WxMessageType.TEXT.getType());
List<String> results = Lists.newArrayList();
// 初始化标记status = 0,表示解答成功
int status = 0;
try {
results = OpenAiUtils.createChatCompletion(acceptContent, openId);
} catch (Exception e) {
status = -1;
log.error(e.getMessage());
results.add(e.getMessage());
}
for (String result : results) {
if (result.startsWith("?") || result.startsWith("?")) {
result = result.substring(1);
}
result = result.trim();
wxMpKefuMessage.setContent(result);
log.info(">>> ChatGPT:{}", result);
AiQa aiQa = new AiQa();
aiQa.setUser(openId);
aiQa.setQuestion(acceptContent);
aiQa.setAnswer(result);
aiQa.setStatus(status);
aiQaRepository.save(aiQa);
// 客服接口发送信息
wxMpService.getKefuService().sendKefuMessage(wxMpKefuMessage);
}
}
private void createImage(String acceptContent, String openId) throws Exception {
WxMpKefuMessage wxMpKefuMessage = new WxMpKefuMessage();
wxMpKefuMessage.setToUser(openId);
wxMpKefuMessage.setMsgType(WxMessageType.IMAGE.getType());
List<String> results = Lists.newArrayList();
// 初始化标记status = 0,表示解答成功
int status = 0;
try {
acceptContent = acceptContent.substring(1);
results = OpenAiUtils.createImage(CreateImageRequest.builder()
.prompt(acceptContent)
.size(ImageSizeEnum.S512x512.getSize())
.user(openId)
.responseFormat(ImageResponseFormatEnum.B64_JSON.getResponseFormat())
.build());
} catch (Exception e) {
status = -1;
log.error(e.getMessage());
results.add(e.getMessage());
}
for (String result : results) {
AiQa aiQa = new AiQa();
aiQa.setUser(openId);
aiQa.setQuestion(acceptContent);
aiQa.setAnswer(result);
aiQa.setStatus(status);
aiQaRepository.save(aiQa);
if (status == -1) {
wxMpKefuMessage.setMsgType(WxMessageType.TEXT.getType());
wxMpKefuMessage.setContent("生成图片失败!原因:" + result);
wxMpService.getKefuService().sendKefuMessage(wxMpKefuMessage);
return;
}
WxMediaUploadResult wxMediaUploadResult = getMediaUploadResult(result);
log.info(">>> 图片上传结果:{}", JSONObject.toJSONString(wxMediaUploadResult));
wxMpKefuMessage.setMediaId(wxMediaUploadResult.getMediaId());
// 客服接口发送信息
wxMpService.getKefuService().sendKefuMessage(wxMpKefuMessage);
}
}
private WxMediaUploadResult getMediaUploadResult(String base64) throws Exception {
byte[] imageBytes = Base64.getDecoder().decode(base64);
try (ByteArrayInputStream bis = new ByteArrayInputStream(imageBytes)) {
return wxMpService.getMaterialService().mediaUpload(WxConsts.MediaFileType.IMAGE, "PNG", bis);
}
}
}
|
[
"com.theokanning.openai.image.CreateImageRequest.builder"
] |
[((2115, 2143), 'com.asleepyfish.enums.WxMessageType.TEXT.getType'), ((3411, 3440), 'com.asleepyfish.enums.WxMessageType.IMAGE.getType'), ((3694, 3978), 'com.theokanning.openai.image.CreateImageRequest.builder'), ((3694, 3949), 'com.theokanning.openai.image.CreateImageRequest.builder'), ((3694, 3859), 'com.theokanning.openai.image.CreateImageRequest.builder'), ((3694, 3825), 'com.theokanning.openai.image.CreateImageRequest.builder'), ((3694, 3765), 'com.theokanning.openai.image.CreateImageRequest.builder'), ((3792, 3824), 'io.github.asleepyfish.enums.ImageSizeEnum.S512x512.getSize'), ((3896, 3948), 'io.github.asleepyfish.enums.ImageResponseFormatEnum.B64_JSON.getResponseFormat'), ((4469, 4497), 'com.asleepyfish.enums.WxMessageType.TEXT.getType'), ((5208, 5242), 'java.util.Base64.getDecoder')]
|
package com.odde.doughnut.services.ai.tools;
import static com.theokanning.openai.service.OpenAiService.defaultObjectMapper;
import com.fasterxml.jackson.core.JsonProcessingException;
import com.fasterxml.jackson.databind.JsonNode;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.kjetland.jackson.jsonSchema.JsonSchemaGenerator;
import com.odde.doughnut.controllers.dto.AiCompletionRequiredAction;
import com.theokanning.openai.assistants.AssistantFunction;
import com.theokanning.openai.assistants.AssistantToolsEnum;
import com.theokanning.openai.assistants.Tool;
import com.theokanning.openai.runs.ToolCall;
import com.theokanning.openai.runs.ToolCallFunction;
import java.util.Map;
import java.util.function.Function;
import java.util.stream.Stream;
public record AiTool(
String name,
String description,
Class<?> parameterClass,
Function<Object, AiCompletionRequiredAction> executor) {
public static <T> AiTool build(
String name,
String description,
Class<T> parameterClass,
Function<T, AiCompletionRequiredAction> executor) {
return new AiTool(
name, description, parameterClass, (arguments) -> executor.apply((T) arguments));
}
public Tool getTool() {
return new Tool(
AssistantToolsEnum.FUNCTION,
AssistantFunction.builder()
.name(name)
.description(description)
.parameters(serializeClassSchema(parameterClass))
.build());
}
private static Map<String, Object> serializeClassSchema(Class<?> value) {
ObjectMapper objectMapper = new ObjectMapper();
JsonSchemaGenerator jsonSchemaGenerator = new JsonSchemaGenerator(objectMapper);
JsonNode jsonSchema = jsonSchemaGenerator.generateJsonSchema(value);
JsonNode jsonNode = objectMapper.valueToTree(jsonSchema);
return objectMapper.convertValue(jsonNode, Map.class);
}
public Stream<AiCompletionRequiredAction> tryConsume(ToolCall toolCall) {
ToolCallFunction function = toolCall.getFunction();
if (name.equals(function.getName())) {
return Stream.of(executor.apply(convertArguments(function)));
}
return Stream.empty();
}
private Object convertArguments(ToolCallFunction function) {
String arguments = function.getArguments();
try {
JsonNode jsonNode = defaultObjectMapper().readTree(arguments);
return defaultObjectMapper().treeToValue(jsonNode, parameterClass);
} catch (JsonProcessingException e) {
throw new RuntimeException(e);
}
}
}
|
[
"com.theokanning.openai.assistants.AssistantFunction.builder"
] |
[((1303, 1475), 'com.theokanning.openai.assistants.AssistantFunction.builder'), ((1303, 1454), 'com.theokanning.openai.assistants.AssistantFunction.builder'), ((1303, 1392), 'com.theokanning.openai.assistants.AssistantFunction.builder'), ((1303, 1354), 'com.theokanning.openai.assistants.AssistantFunction.builder')]
|
/*
* Copyright 2008-2009 the original author or authors.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package egovframework.example.sample.web;
import java.awt.Choice;
import java.io.File;
import java.io.IOException;
import java.nio.file.Files;
import java.nio.file.Path;
import java.nio.file.Paths;
import java.time.Duration;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import egovframework.example.API.Keys;
import egovframework.example.sample.service.EgovSampleService;
import egovframework.example.sample.service.SampleDefaultVO;
import egovframework.example.sample.service.SampleVO;
import egovframework.rte.fdl.property.EgovPropertyService;
import egovframework.rte.ptl.mvc.tags.ui.pagination.PaginationInfo;
import javax.annotation.Resource;
import javax.servlet.ServletContext;
import javax.servlet.annotation.MultipartConfig;
import javax.servlet.http.HttpServletRequest;
import org.springframework.stereotype.Controller;
import org.springframework.ui.Model;
import org.springframework.ui.ModelMap;
import org.springframework.validation.BindingResult;
import org.springframework.web.bind.annotation.ModelAttribute;
import org.springframework.web.bind.annotation.RequestMapping;
import org.springframework.web.bind.annotation.RequestMethod;
import org.springframework.web.bind.annotation.RequestParam;
import org.springframework.web.bind.annotation.RequestPart;
import org.springframework.web.bind.support.SessionStatus;
import org.springframework.web.multipart.MultipartFile;
import org.springmodules.validation.commons.DefaultBeanValidator;
import org.springframework.http.ResponseEntity;
import org.springframework.stereotype.Controller;
import org.springframework.web.bind.annotation.GetMapping;
import org.springframework.web.bind.annotation.PathVariable;
import org.springframework.web.bind.annotation.PostMapping;
import org.springframework.web.bind.annotation.RequestBody;
import org.springframework.web.bind.annotation.RequestMapping;
import org.springframework.web.bind.annotation.RequestMethod;
import org.springframework.web.bind.annotation.RestController;
import com.theokanning.openai.audio.CreateTranscriptionRequest;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.service.OpenAiService;
import org.apache.logging.log4j.LogManager;
import org.apache.logging.log4j.Logger;
/**
* @Class Name : EgovSampleController.java
* @Description : EgovSample Controller Class
* @Modification Information
* @
* @ 수정일 수정자 수정내용
* @ --------- --------- -------------------------------
* @ 2009.03.16 최초생성
*
* @author 개발프레임웍크 실행환경 개발팀
* @since 2009. 03.16
* @version 1.0
* @see
*
* Copyright (C) by MOPAS All right reserved.
*/
@Controller
@MultipartConfig(
maxFileSize = 1024 * 1024 * 25, // 최대 25MB 파일 크기
maxRequestSize = 1024 * 1024 * 25, // 최대 25MB 요청 크기
fileSizeThreshold = 1024 * 1024 // 1MB 이상부터 디스크에 저장
)
public class EgovSampleController {
private static final Logger logger = LogManager.getLogger(EgovSampleController.class);
private final String UPLOAD_DIR = "uploads";
/** EgovSampleService */
@Resource(name = "sampleService")
private EgovSampleService sampleService;
/** EgovPropertyService */
@Resource(name = "propertiesService")
protected EgovPropertyService propertiesService;
/** Validator */
@Resource(name = "beanValidator")
protected DefaultBeanValidator beanValidator;
/**
* 글 목록을 조회한다. (pageing)
* @param searchVO - 조회할 정보가 담긴 SampleDefaultVO
* @param model
* @return "egovSampleList"
* @exception Exception
*/
@RequestMapping(value = "/egovSampleList.do")
public String selectSampleList(@ModelAttribute("searchVO") SampleDefaultVO searchVO, ModelMap model) throws Exception {
/** EgovPropertyService.sample */
searchVO.setPageUnit(propertiesService.getInt("pageUnit"));
searchVO.setPageSize(propertiesService.getInt("pageSize"));
/** pageing setting */
PaginationInfo paginationInfo = new PaginationInfo();
paginationInfo.setCurrentPageNo(searchVO.getPageIndex());
paginationInfo.setRecordCountPerPage(searchVO.getPageUnit());
paginationInfo.setPageSize(searchVO.getPageSize());
searchVO.setFirstIndex(paginationInfo.getFirstRecordIndex());
searchVO.setLastIndex(paginationInfo.getLastRecordIndex());
searchVO.setRecordCountPerPage(paginationInfo.getRecordCountPerPage());
List<?> sampleList = sampleService.selectSampleList(searchVO);
model.addAttribute("resultList", sampleList);
int totCnt = sampleService.selectSampleListTotCnt(searchVO);
paginationInfo.setTotalRecordCount(totCnt);
model.addAttribute("paginationInfo", paginationInfo);
return "sample/egovSampleList";
}
/**
* 글 등록 화면을 조회한다.
* @param searchVO - 목록 조회조건 정보가 담긴 VO
* @param model
* @return "egovSampleRegister"
* @exception Exception
*/
@RequestMapping(value = "/addSample.do", method = RequestMethod.GET)
public String addSampleView(@ModelAttribute("searchVO") SampleDefaultVO searchVO, Model model) throws Exception {
model.addAttribute("sampleVO", new SampleVO());
return "sample/egovSampleRegister";
}
/**
* 글을 등록한다.
* @param sampleVO - 등록할 정보가 담긴 VO
* @param searchVO - 목록 조회조건 정보가 담긴 VO
* @param status
* @return "forward:/egovSampleList.do"
* @exception Exception
*/
@RequestMapping(value = "/addSample.do", method = RequestMethod.POST)
public String addSample(@ModelAttribute("searchVO") SampleDefaultVO searchVO, SampleVO sampleVO, BindingResult bindingResult, Model model, SessionStatus status)
throws Exception {
// Server-Side Validation
beanValidator.validate(sampleVO, bindingResult);
if (bindingResult.hasErrors()) {
model.addAttribute("sampleVO", sampleVO);
return "sample/egovSampleRegister";
}
sampleService.insertSample(sampleVO);
status.setComplete();
return "forward:/egovSampleList.do";
}
/**
* 글 수정화면을 조회한다.
* @param id - 수정할 글 id
* @param searchVO - 목록 조회조건 정보가 담긴 VO
* @param model
* @return "egovSampleRegister"
* @exception Exception
*/
@RequestMapping("/updateSampleView.do")
public String updateSampleView(@RequestParam("selectedId") String id, @ModelAttribute("searchVO") SampleDefaultVO searchVO, Model model) throws Exception {
SampleVO sampleVO = new SampleVO();
sampleVO.setId(id);
// 변수명은 CoC 에 따라 sampleVO
model.addAttribute(selectSample(sampleVO, searchVO));
return "sample/egovSampleRegister";
}
/**
* 글을 조회한다.
* @param sampleVO - 조회할 정보가 담긴 VO
* @param searchVO - 목록 조회조건 정보가 담긴 VO
* @param status
* @return @ModelAttribute("sampleVO") - 조회한 정보
* @exception Exception
*/
public SampleVO selectSample(SampleVO sampleVO, @ModelAttribute("searchVO") SampleDefaultVO searchVO) throws Exception {
return sampleService.selectSample(sampleVO);
}
/**
* 글을 수정한다.
* @param sampleVO - 수정할 정보가 담긴 VO
* @param searchVO - 목록 조회조건 정보가 담긴 VO
* @param status
* @return "forward:/egovSampleList.do"
* @exception Exception
*/
@RequestMapping("/updateSample.do")
public String updateSample(@ModelAttribute("searchVO") SampleDefaultVO searchVO, SampleVO sampleVO, BindingResult bindingResult, Model model, SessionStatus status)
throws Exception {
beanValidator.validate(sampleVO, bindingResult);
if (bindingResult.hasErrors()) {
model.addAttribute("sampleVO", sampleVO);
return "sample/egovSampleRegister";
}
sampleService.updateSample(sampleVO);
status.setComplete();
return "forward:/egovSampleList.do";
}
/**
* 글을 삭제한다.
* @param sampleVO - 삭제할 정보가 담긴 VO
* @param searchVO - 목록 조회조건 정보가 담긴 VO
* @param status
* @return "forward:/egovSampleList.do"
* @exception Exception
*/
@RequestMapping("/deleteSample.do")
public String deleteSample(SampleVO sampleVO, @ModelAttribute("searchVO") SampleDefaultVO searchVO, SessionStatus status) throws Exception {
sampleService.deleteSample(sampleVO);
status.setComplete();
return "forward:/egovSampleList.do";
}
@RequestMapping("/file.do")
public String fileReg() throws Exception {
return "sample/file";
}
//static String englishAudioFilePath = "/Users/jiuhyeong/Documents/Handong/capstone1/Dani_california.mp3";
//static String englishAudioFilePath = "/Users/jiuhyeong/Documents/Handong/capstone1/interview.mp4";
//requestparam으로 임시로 저장한 파일의 위치를 string으로 받은 후 whisper에게 전사를 맡김, 임시 파일 삭제?
@RequestMapping(value = "/file.do", method = RequestMethod.POST)
public String createTranscription(@RequestParam String absolutePath, Model model) {
OpenAiService service = new OpenAiService(Keys.OPENAPI_KEY,Duration.ofMinutes(9999));
CreateTranscriptionRequest createTranscriptionRequest = CreateTranscriptionRequest.builder()
.model("whisper-1")
.build();
String text = service.createTranscription(createTranscriptionRequest, absolutePath).getText();
logger.debug(text);
model.addAttribute("result", text);
model.addAttribute("absolutePath", absolutePath);
File fileToDelete = new File(absolutePath);
if (fileToDelete.exists()) {
if (fileToDelete.delete()) {
logger.debug("temp File deleted successfully.");
} else {
logger.error("Failed to delete the file.");
}
} else {
logger.debug("temp File not found");
}
return "sample/file";
}
//jsp에 저장버튼 추가 후 restapi로 보내기
@RequestMapping(value = "/summarize.do", method = RequestMethod.POST)
public String showSummaryResult(@RequestParam String transcription_result, Model model) {
OpenAiService service = new OpenAiService(Keys.OPENAPI_KEY,Duration.ofMinutes(9999));
List<ChatMessage> message = new ArrayList<ChatMessage>();
message.add(new ChatMessage("user", "텍스트의 주제를 파악해서 해당 언어로 다섯줄 내외 요약해줘 \""+transcription_result+"\""));
ChatCompletionRequest completionRequest = ChatCompletionRequest.builder()
.messages(message)
.model("gpt-3.5-turbo")
.maxTokens(1500)
.temperature((double) 0.5f)
.build();
String summary_restult=service.createChatCompletion(completionRequest).getChoices().get(0).getMessage().getContent();
model.addAttribute("summary_result",summary_restult);
return "sample/summarize";
}
//파일을 임시저장 후 file.do에 경로를 보냄.
@RequestMapping(value = "/postfile.do", method = RequestMethod.POST)
public String handleFile(@RequestParam(value = "file", required = false) MultipartFile file, Model model, HttpServletRequest request) throws IOException{
ServletContext context = request.getSession().getServletContext();
String projectPath = context.getRealPath("/");
System.out.println("Project Path: " + projectPath);
if (file.isEmpty()) {
return "redirect:/file.do"; // 파일이 선택되지 않았을 경우 폼으로 리다이렉트
}
try {
byte[] bytes = file.getBytes();
Path directoryPath = Paths.get(projectPath+UPLOAD_DIR);
// 디렉토리가 존재하지 않으면 생성
if (!Files.exists(directoryPath)) {
Files.createDirectories(directoryPath);
}
Path filePath = directoryPath.resolve(file.getOriginalFilename());
Files.write(filePath, bytes);
Path absolutePath = filePath.toAbsolutePath();
String absolutePathString = absolutePath.toString();
logger.debug("AbsolutePathString received"+absolutePathString);
model.addAttribute("absolutePath", absolutePathString);
} catch (IOException e) {
e.printStackTrace();
}
model.addAttribute("inputFile", file.getOriginalFilename());
return "sample/file";
}
@RequestMapping(value = "/save-result.do", method = RequestMethod.POST)
public String saveFile(@RequestParam(value = "dir", required = false) MultipartFile dir, @RequestParam String summ_result, Model model, HttpServletRequest request) throws IOException{
return "redirect:/summary.do";
}
}
|
[
"com.theokanning.openai.audio.CreateTranscriptionRequest.builder",
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((10123, 10222), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((10123, 10196), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((11541, 11746), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((11541, 11724), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((11541, 11683), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((11541, 11638), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((11541, 11601), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
/*
* Click nbfs://nbhost/SystemFileSystem/Templates/Licenses/license-default.txt to change this license
* Click nbfs://nbhost/SystemFileSystem/Templates/Classes/Class.java to edit this template
*/
package cloud.cleo.connectgpt;
import cloud.cleo.connectgpt.lang.LangUtil;
import static cloud.cleo.connectgpt.lang.LangUtil.LanguageIds.*;
import com.amazonaws.services.lambda.runtime.Context;
import com.amazonaws.services.lambda.runtime.RequestHandler;
import com.amazonaws.services.lambda.runtime.events.LexV2Event;
import com.amazonaws.services.lambda.runtime.events.LexV2Event.DialogAction;
import com.amazonaws.services.lambda.runtime.events.LexV2Event.Intent;
import com.amazonaws.services.lambda.runtime.events.LexV2Event.SessionState;
import com.amazonaws.services.lambda.runtime.events.LexV2Response;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.service.OpenAiService;
import java.net.SocketTimeoutException;
import java.time.Duration;
import java.time.LocalDate;
import java.time.ZoneId;
import org.apache.logging.log4j.LogManager;
import org.apache.logging.log4j.Logger;
import software.amazon.awssdk.enhanced.dynamodb.DynamoDbEnhancedClient;
import software.amazon.awssdk.enhanced.dynamodb.DynamoDbTable;
import software.amazon.awssdk.enhanced.dynamodb.Key;
import software.amazon.awssdk.enhanced.dynamodb.TableSchema;
import software.amazon.awssdk.enhanced.dynamodb.extensions.AutoGeneratedTimestampRecordExtension;
/**
*
* @author sjensen
*/
public class ChatGPTLambda implements RequestHandler<LexV2Event, LexV2Response> {
// Initialize the Log4j logger.
final static Logger log = LogManager.getLogger(ChatGPTLambda.class);
final static ObjectMapper mapper = new ObjectMapper();
final static TableSchema<ChatGPTSessionState> schema = TableSchema.fromBean(ChatGPTSessionState.class);
final static DynamoDbEnhancedClient enhancedClient = DynamoDbEnhancedClient.builder()
.extensions(AutoGeneratedTimestampRecordExtension.create()).build();
final static DynamoDbTable<ChatGPTSessionState> sessionState = enhancedClient.table(System.getenv("SESSION_TABLE_NAME"), schema);
final static OpenAiService open_ai_service = new OpenAiService(System.getenv("OPENAI_API_KEY"), Duration.ofSeconds(20));
final static String OPENAI_MODEL = System.getenv("OPENAI_MODEL");
@Override
public LexV2Response handleRequest(LexV2Event lexRequest, Context cntxt) {
try {
log.debug(mapper.valueToTree(lexRequest).toString());
final var intentName = lexRequest.getSessionState().getIntent().getName();
log.debug("Intent: " + intentName);
return processGPT(lexRequest);
} catch (Exception e) {
log.error(e);
// Unhandled Exception
return buildResponse(lexRequest, new LangUtil(lexRequest.getBot().getLocaleId()).getString(UNHANDLED_EXCEPTION));
}
}
private LexV2Response processGPT(LexV2Event lexRequest) {
final var input = lexRequest.getInputTranscript();
final var localId = lexRequest.getBot().getLocaleId();
final var lang = new LangUtil(localId);
log.debug("Java Locale is " + lang.getLocale());
if (input == null || input.isBlank()) {
log.debug("Got blank input, so just silent or nothing");
final var attrs = lexRequest.getSessionState().getSessionAttributes();
var count = Integer.valueOf(attrs.getOrDefault("blankCounter", "0"));
count++;
if (count > 2) {
log.debug("Two blank responses, sending to Quit Intent");
// Hang up on caller after 2 silience requests
return buildQuitResponse(lexRequest);
} else {
attrs.put("blankCounter", count.toString());
// If we get slience (timeout without speech), then we get empty string on the transcript
return buildResponse(lexRequest, lang.getString(BLANK_RESPONSE));
}
}
// When testing in lex console input will be text, so use session ID, for speech we shoud have a phone via Connect
final var user_id = lexRequest.getSessionId();
// Key to record in Dynamo
final var key = Key.builder().partitionValue(user_id).sortValue(LocalDate.now(ZoneId.of("America/Chicago")).toString()).build();
// load session state if it exists
log.debug("Start Retreiving Session State");
var session = sessionState.getItem(key);
log.debug("End Retreiving Session State");
if (session == null) {
session = new ChatGPTSessionState(user_id);
}
// Since we can call and change language during session, always specifiy how we want responses
session.addSystemMessage(lang.getString(CHATGPT_RESPONSE_LANGUAGE));
// add this request to the session
session.addUserMessage(input);
String botResponse;
try {
ChatCompletionRequest request = ChatCompletionRequest.builder()
.messages(session.getChatMessages())
.model(OPENAI_MODEL)
.maxTokens(500)
.temperature(0.2) // More focused
.n(1) // Only return 1 completion
.build();
log.debug("Start API Call to ChatGPT");
final var completion = open_ai_service.createChatCompletion(request);
log.debug("End API Call to ChatGPT");
log.debug(completion);
botResponse = completion.getChoices().get(0).getMessage().getContent();
// Add response to session
session.addAssistantMessage(botResponse);
// Since we have a valid response, add message asking if there is anything else
if ( ! "Text".equalsIgnoreCase(lexRequest.getInputMode()) ) {
// Only add if not text (added to voice response)
botResponse = botResponse + lang.getString(ANYTHING_ELSE);
}
// Save the session to dynamo
log.debug("Start Saving Session State");
session.incrementCounter();
sessionState.putItem(session);
log.debug("End Saving Session State");
} catch (RuntimeException rte) {
if (rte.getCause() != null && rte.getCause() instanceof SocketTimeoutException) {
log.error("Response timed out", rte);
botResponse = lang.getString(OPERATION_TIMED_OUT);
} else {
throw rte;
}
}
return buildResponse(lexRequest, botResponse);
}
/**
* Response that sends you to the Quit intent so the call can be ended
*
* @param lexRequest
* @param response
* @return
*/
private LexV2Response buildQuitResponse(LexV2Event lexRequest) {
// State to return
final var ss = SessionState.builder()
// Retain the current session attributes
.withSessionAttributes(lexRequest.getSessionState().getSessionAttributes())
// Send back Quit Intent
.withIntent(Intent.builder().withName("Quit").withState("ReadyForFulfillment").build())
// Indicate the state is Delegate
.withDialogAction(DialogAction.builder().withType("Delegate").build())
.build();
final var lexV2Res = LexV2Response.builder()
.withSessionState(ss)
.build();
log.debug("Response is " + mapper.valueToTree(lexV2Res));
return lexV2Res;
}
/**
* General Response used to send back a message and Elicit Intent again at LEX
*
* @param lexRequest
* @param response
* @return
*/
private LexV2Response buildResponse(LexV2Event lexRequest, String response) {
// State to return
final var ss = SessionState.builder()
// Retain the current session attributes
.withSessionAttributes(lexRequest.getSessionState().getSessionAttributes())
// Always ElictIntent, so you're back at the LEX Bot looking for more input
.withDialogAction(DialogAction.builder().withType("ElicitIntent").build())
.build();
final var lexV2Res = LexV2Response.builder()
.withSessionState(ss)
// We are using plain text responses
.withMessages(new LexV2Response.Message[]{new LexV2Response.Message("PlainText", response, null)})
.build();
log.debug("Response is " + mapper.valueToTree(lexV2Res));
return lexV2Res;
}
}
|
[
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((1987, 2099), 'software.amazon.awssdk.enhanced.dynamodb.DynamoDbEnhancedClient.builder'), ((1987, 2091), 'software.amazon.awssdk.enhanced.dynamodb.DynamoDbEnhancedClient.builder'), ((4370, 4481), 'software.amazon.awssdk.enhanced.dynamodb.Key.builder'), ((4370, 4473), 'software.amazon.awssdk.enhanced.dynamodb.Key.builder'), ((4370, 4407), 'software.amazon.awssdk.enhanced.dynamodb.Key.builder'), ((4418, 4472), 'java.time.LocalDate.now'), ((5138, 5440), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5138, 5383), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5138, 5341), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5138, 5303), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5138, 5267), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5138, 5226), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((7069, 7547), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.SessionState.builder'), ((7069, 7522), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.SessionState.builder'), ((7069, 7385), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.SessionState.builder'), ((7069, 7240), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.SessionState.builder'), ((7310, 7384), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.Intent.builder'), ((7310, 7376), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.Intent.builder'), ((7310, 7343), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.Intent.builder'), ((7470, 7521), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.DialogAction.builder'), ((7470, 7513), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.DialogAction.builder'), ((7579, 7665), 'com.amazonaws.services.lambda.runtime.events.LexV2Response.builder'), ((7579, 7640), 'com.amazonaws.services.lambda.runtime.events.LexV2Response.builder'), ((8067, 8446), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.SessionState.builder'), ((8067, 8421), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.SessionState.builder'), ((8067, 8238), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.SessionState.builder'), ((8365, 8420), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.DialogAction.builder'), ((8365, 8412), 'com.amazonaws.services.lambda.runtime.events.LexV2Event.DialogAction.builder'), ((8478, 8732), 'com.amazonaws.services.lambda.runtime.events.LexV2Response.builder'), ((8478, 8707), 'com.amazonaws.services.lambda.runtime.events.LexV2Response.builder'), ((8478, 8539), 'com.amazonaws.services.lambda.runtime.events.LexV2Response.builder')]
|
package de.throughput.ircbot.handler;
import com.fasterxml.jackson.annotation.JsonIgnore;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatCompletionResult;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
import de.throughput.ircbot.api.Command;
import de.throughput.ircbot.api.CommandEvent;
import de.throughput.ircbot.api.CommandHandler;
import de.throughput.ircbot.api.MessageHandler;
import org.apache.commons.lang3.exception.ExceptionUtils;
import org.pircbotx.hooks.events.MessageEvent;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.stereotype.Component;
import java.io.IOException;
import java.io.UncheckedIOException;
import java.nio.file.Files;
import java.nio.file.Path;
import java.text.SimpleDateFormat;
import java.time.LocalDateTime;
import java.util.ArrayList;
import java.util.Date;
import java.util.LinkedList;
import java.util.List;
import java.util.Locale;
import java.util.Map;
import java.util.Set;
import java.util.TimeZone;
import java.util.concurrent.ConcurrentHashMap;
@Component
public class OpenAiChatMessageHandler implements MessageHandler, CommandHandler {
private static final Logger LOG = LoggerFactory.getLogger(OpenAiChatMessageHandler.class);
public static final Command CMD_RESET_CONTEXT = new Command("aireset",
"aireset - deletes the current context for the channel and reloads the system prompt from the file system.");
private static final String MODEL_GPT_3_5_TURBO = "gpt-3.5-turbo";
private static final int MAX_CONTEXT_MESSAGES = 10;
private static final int MAX_TOKENS = 100;
private static final int MAX_IRC_MESSAGE_LENGTH = 420;
private static final String SHORT_ANSWER_HINT = " (Antwort auf 200 Zeichen begrenzen)";
private final Map<String, LinkedList<TimedChatMessage>> contextMessagesPerChannel = new ConcurrentHashMap<>();
private final OpenAiService openAiService;
private final Path systemPromptPath;
private String systemPrompt;
public OpenAiChatMessageHandler(OpenAiService openAiService, @Value("${openai.systemPrompt.path}") Path systemPromptPath) {
this.openAiService = openAiService;
this.systemPromptPath = systemPromptPath;
readSystemPromptFromFile();
}
@Override
public Set<Command> getCommands() {
return Set.of(CMD_RESET_CONTEXT);
}
@Override
public boolean onMessage(MessageEvent event) {
String message = event.getMessage().trim();
String botNick = event.getBot().getNick();
if (message.startsWith(botNick + ":") || message.startsWith(botNick + ",")) {
message = message.substring(event.getBot().getNick().length() + 1).trim();
generateResponse(event, message);
return true;
}
return false;
}
@Override
public boolean onCommand(CommandEvent command) {
// handles the aireset command
var contextMessages = contextMessagesPerChannel.get(command.getEvent().getChannel().getName());
if (contextMessages != null) {
synchronized (contextMessages) {
contextMessages.clear();
}
}
readSystemPromptFromFile();
command.respond("system prompt reloaded. context reset complete.");
return true;
}
/**
* Generates a response to the given (trimmed) message using the OpenAI API.
*/
private void generateResponse(MessageEvent event, String message) {
var contextMessages = contextMessagesPerChannel.computeIfAbsent(event.getChannel().getName(), k -> new LinkedList<>());
synchronized (contextMessages) {
try {
String channel = event.getChannel().getName();
var request = ChatCompletionRequest.builder()
.model(MODEL_GPT_3_5_TURBO)
.maxTokens(MAX_TOKENS)
.messages(createPromptMessages(contextMessages, channel, event.getUser().getNick(), message))
.build();
ChatCompletionResult completionResult = openAiService.createChatCompletion(request);
ChatMessage responseMessage = completionResult.getChoices().get(0).getMessage();
contextMessages.add(new TimedChatMessage(responseMessage));
event.respond(sanitizeResponse(responseMessage.getContent()));
} catch (Exception e) {
LOG.error(e.getMessage(), e);
event.respond("Tja. (" + ExceptionUtils.getRootCauseMessage(e) + ")");
}
}
}
/**
* Sanitizes the response by removing excessive whitespace and limiting the length.
*/
private static String sanitizeResponse(String content) {
String trim = content.replaceAll("\\s+", " ").trim();
return trim.length() > MAX_IRC_MESSAGE_LENGTH ? trim.substring(0, MAX_IRC_MESSAGE_LENGTH) : trim;
}
/**
* Creates the list of prompt messages for the OpenAI API call.
*/
private List<ChatMessage> createPromptMessages(LinkedList<TimedChatMessage> contextMessages, String channel, String nick, String message) {
message += SHORT_ANSWER_HINT;
contextMessages.add(new TimedChatMessage(new ChatMessage(ChatMessageRole.USER.value(), message, nick)));
pruneOldMessages(contextMessages);
List<ChatMessage> promptMessages = new ArrayList<>();
promptMessages.add(new ChatMessage(ChatMessageRole.SYSTEM.value(), systemPrompt));
promptMessages.add(new ChatMessage(ChatMessageRole.SYSTEM.value(), getDatePrompt()));
promptMessages.addAll(contextMessages);
return promptMessages;
}
/**
* Generates a system prompt containing the current date and time.
*/
private String getDatePrompt() {
TimeZone timeZone = TimeZone.getTimeZone("Europe/Berlin");
SimpleDateFormat dateFormat = new SimpleDateFormat("EEEE, 'der' dd. MMMM yyyy", Locale.GERMAN);
dateFormat.setTimeZone(timeZone);
SimpleDateFormat timeFormat = new SimpleDateFormat("HH:mm", Locale.GERMAN);
timeFormat.setTimeZone(timeZone);
Date now = new Date();
return "Heute ist " + dateFormat.format(now) + ", und es ist " + timeFormat.format(now) + " Uhr in Deutschland.";
}
/**
* Removes old messages from the context.
*/
private void pruneOldMessages(LinkedList<TimedChatMessage> contextMessages) {
LocalDateTime twoHoursAgo = LocalDateTime.now().minusHours(2);
contextMessages.removeIf(message -> message.getTimestamp().isBefore(twoHoursAgo));
while (contextMessages.size() > MAX_CONTEXT_MESSAGES) {
contextMessages.removeFirst();
}
}
/**
* Reads the system prompt from the file system.
*/
private void readSystemPromptFromFile() {
try {
systemPrompt = Files.readString(systemPromptPath);
} catch (IOException e) {
throw new UncheckedIOException(e);
}
}
@Override
public boolean isOnlyTalkChannels() {
return true;
}
/**
* Adds a timestamp to ChatMessage, allowing us to drop old messages from the context.
*/
private static class TimedChatMessage extends ChatMessage {
private final LocalDateTime timestamp;
public TimedChatMessage(ChatMessage chatMessage) {
super(chatMessage.getRole(), chatMessage.getContent(), chatMessage.getName());
this.timestamp = LocalDateTime.now();
}
@JsonIgnore
public LocalDateTime getTimestamp() {
return timestamp;
}
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value",
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((4011, 4292), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((4011, 4259), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((4011, 4141), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((4011, 4094), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5521, 5549), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((5718, 5748), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value'), ((5809, 5839), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value'), ((6750, 6783), 'java.time.LocalDateTime.now')]
|
package com.cvcopilot.resumebuilding.service;
import com.cvcopilot.resumebuilding.models.Modification;
import com.cvcopilot.resumebuilding.repository.ModificationRepository;
import com.cvcopilot.resumebuilding.repository.ProfileRepository;
import com.theokanning.openai.completion.chat.ChatCompletionChoice;
import java.time.Duration;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
import javax.annotation.PostConstruct;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.data.redis.core.HashOperations;
import org.springframework.data.redis.core.RedisTemplate;
import org.springframework.data.redis.core.ZSetOperations;
import org.springframework.kafka.annotation.KafkaListener;
import org.springframework.messaging.handler.annotation.Payload;
import org.springframework.stereotype.Service;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
@Service
public class ResumeService {
@Autowired
private ProfileRepository profileRepository;
@Autowired
private StateService stateService;
@Autowired
private RedisTemplate<String, String> redisTemplate;
private HashOperations<String, String, String> hashOperations;
private ZSetOperations<String, String> zSetOperations;
@Autowired
private ModificationRepository modificationRepository;
@Value("${openai.api-key}")
private String openAIKey;
@Value("${openai.model}")
private String openAIModel;
@PostConstruct
private void init() {
hashOperations = redisTemplate.opsForHash();
zSetOperations = redisTemplate.opsForZSet();
}
private String prompt = "Based on the user's experiences, write a optimized resume according to the job description. Emit the personal information.";
private static final Logger logger = LoggerFactory.getLogger(ResumeService.class);
@KafkaListener(topics = "resume", groupId = "test-group", containerFactory = "kafkaListenerContainerFactory")
public void consume(@Payload String message) {
String userId = message.substring(0, 19);
String modificationId = message.substring(19, 55);
List<ChatCompletionChoice> res;
try {
stateService.addOrUpdateState(userId, modificationId, "in_progress");
OpenAiService service = new OpenAiService(openAIKey, Duration.ofSeconds(120));
List<ChatMessage> messages = new ArrayList<>();
final ChatMessage systemMessage = new ChatMessage(ChatMessageRole.SYSTEM.value(), "You are a hr from big tech company.");
final ChatMessage userMessage = new ChatMessage(ChatMessageRole.USER.value(), message.substring(56) + prompt);
messages.add(systemMessage);
messages.add(userMessage);
ChatCompletionRequest chatCompletionRequest = ChatCompletionRequest
.builder()
.model("gpt-3.5-turbo")
.messages(messages)
.n(1)
.maxTokens(600)
.logitBias(new HashMap<>())
.build();
res = service.createChatCompletion(chatCompletionRequest).getChoices();
service.shutdownExecutor();
} catch (RuntimeException e) {
logger.error("RuntimeException: " + e.getMessage());
stateService.addOrUpdateState(userId, modificationId, "failed");
return;
}
try {
// write to postgres
modificationRepository.save(new Modification(modificationId, res.get(0).getMessage().getContent(), Long.valueOf(userId), System.currentTimeMillis()));
} catch (RuntimeException e) {
logger.error("Failed to write to Postgres: " + e.getMessage());
stateService.addOrUpdateState(userId, modificationId, "failed_db_error");
return;
}
// write state to redis
stateService.addOrUpdateState(userId, modificationId, "finished");
// invalidate cache of all results of this user
zSetOperations.remove(userId);
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value"
] |
[((2811, 2841), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value'), ((2943, 2971), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value')]
|
package podsofkon;
import com.theokanning.openai.image.CreateImageRequest;
import com.theokanning.openai.service.OpenAiService;
import org.springframework.core.io.ByteArrayResource;
import org.springframework.http.*;
import org.springframework.util.LinkedMultiValueMap;
import org.springframework.util.MultiValueMap;
import org.springframework.web.bind.annotation.*;
import org.springframework.web.client.RestTemplate;
import javax.servlet.http.HttpServletRequest;
import javax.sound.sampled.*;
import java.io.*;
import java.time.Duration;
import java.util.*;
@RestController
@RequestMapping("/picturestory")
public class GenerateAPictureStoryUsingOnlySpeech {
static List<String> storyImages = new ArrayList();
@GetMapping("/form")
public String newstory(
HttpServletRequest request) throws Exception {
storyImages = new ArrayList();
return getHtmlString("");
}
@GetMapping("/picturestory")
public String picturestory(@RequestParam("genopts") String genopts) throws Exception {
AudioFormat format =
new AudioFormat(AudioFormat.Encoding.PCM_SIGNED, 44100.0f, 16, 1,
(16 / 8) * 1, 44100.0f, true);
SoundRecorder soundRecorder = new SoundRecorder();
soundRecorder.build(format);
System.out.println("Start recording ....");
soundRecorder.start();
Thread.sleep(8000);
soundRecorder.stop();
System.out.println("Stopped recording ....");
Thread.sleep(3000); //give the process time
String name = "AISoundClip";
AudioFileFormat.Type fileType = AudioFileFormat.Type.WAVE;
AudioInputStream audioInputStream = soundRecorder.audioInputStream;
System.out.println("Saving...");
File file = new File(name + "." + fileType.getExtension());
audioInputStream.reset();
AudioSystem.write(audioInputStream, fileType, file);
System.out.println("Saved " + file.getAbsolutePath());
String transcription = transcribe(file) + genopts;
System.out.println("transcription " + transcription);
String imageLocation = imagegeneration(transcription);
System.out.println("imageLocation " + imageLocation);
storyImages.add(imageLocation);
String htmlStoryFrames = "";
Iterator<String> iterator = storyImages.iterator();
while(iterator.hasNext()) {
htmlStoryFrames += "<td><img src=\"" + iterator.next() +"\" width=\"400\" height=\"400\"></td>";
}
return getHtmlString(htmlStoryFrames);
}
private static String getHtmlString(String htmlStoryFrames) {
return "<html><table>" +
" <tr>" +
htmlStoryFrames +
" </tr>" +
"</table><br><br>" +
"<form action=\"/picturestory/picturestory\">" +
" <input type=\"submit\" value=\"Click here and record (up to 10 seconds of audio) describing next scene.\">" +
"<br> Some additional options..." +
"<br><input type=\"radio\" id=\"genopts\" name=\"genopts\" value=\", using only one line\" checked >using only one line" +
"<br><input type=\"radio\" id=\"genopts\" name=\"genopts\" value=\", photo taken on a Pentax k1000\">photo taken on a Pentax k1000" +
"<br><input type=\"radio\" id=\"genopts\" name=\"genopts\" value=\", pixel art\">pixel art" +
"<br><input type=\"radio\" id=\"genopts\" name=\"genopts\" value=\", digital art\">digital art" +
"<br><input type=\"radio\" id=\"genopts\" name=\"genopts\" value=\", 3d render\">3d render" +
"</form><br><br>" +
"<form action=\"/picturestory/form\">" +
" <input type=\"submit\" value=\"Or click here to start a new story\">\n" +
"</form>" +
"</html>";
}
public String imagegeneration(String imagedescription) throws Exception {
OpenAiService service =
new OpenAiService("sk-sdf3HSWvb2HgV", Duration.ofSeconds(60));
CreateImageRequest openairequest = CreateImageRequest.builder()
.prompt(imagedescription)
.build();
System.out.println("\nImage is located at:");
String imageLocation = service.createImage(openairequest).getData().get(0).getUrl();
service.shutdownExecutor();
return imageLocation;
}
public String transcribe(File file) throws Exception {
OpenAiService service =
new OpenAiService("sk-nMVoZmUsOBjRasdfvb2HgV", Duration.ofSeconds(60));
String audioTranscription = transcribeFile(file, service);
service.shutdownExecutor();
return audioTranscription;
}
private String transcribeFile(File file, OpenAiService service) throws Exception
{
String endpoint = "https://api.openai.com/v1/audio/transcriptions";
String modelName = "whisper-1";
HttpHeaders headers = new HttpHeaders();
headers.setContentType(MediaType.MULTIPART_FORM_DATA);
headers.setBearerAuth(System.getenv("OPENAI_KEY"));
MultiValueMap<String, Object> body = new LinkedMultiValueMap<>();
byte[] fileBytes = new byte[0];
try (FileInputStream fis = new FileInputStream(file);
ByteArrayOutputStream bos = new ByteArrayOutputStream()) {
byte[] buffer = new byte[1024];
int bytesRead;
while ((bytesRead = fis.read(buffer)) != -1) {
bos.write(buffer, 0, bytesRead);
}
fileBytes = bos.toByteArray();
} catch (IOException e) {
e.printStackTrace();
}
body.add("file", new ByteArrayResource(fileBytes) {
@Override
public String getFilename() {
return file.getName();
}
});
body.add("model", modelName);
HttpEntity<MultiValueMap<String, Object>> requestEntity = new HttpEntity<>(body, headers);
RestTemplate restTemplate = new RestTemplate();
ResponseEntity<String> response = restTemplate.exchange(endpoint, HttpMethod.POST, requestEntity, String.class);
return response.getBody();
}
public class SoundRecorder implements Runnable {
AudioInputStream audioInputStream;
private AudioFormat format;
public Thread thread;
public SoundRecorder build(AudioFormat format) {
this.format = format;
return this;
}
public void start() {
thread = new Thread(this);
thread.start();
}
public void stop() {
thread = null;
}
@Override
public void run() {
try (final ByteArrayOutputStream out = new ByteArrayOutputStream(); final TargetDataLine line = getTargetDataLineForRecord();) {
int frameSizeInBytes = format.getFrameSize();
int bufferLengthInFrames = line.getBufferSize() / 8;
final int bufferLengthInBytes = bufferLengthInFrames * frameSizeInBytes;
buildByteOutputStream(out, line, frameSizeInBytes, bufferLengthInBytes);
this.audioInputStream = new AudioInputStream(line);
setAudioInputStream(convertToAudioIStream(out, frameSizeInBytes));
audioInputStream.reset();
} catch (IOException ex) {
ex.printStackTrace();
} catch (Exception ex) {
ex.printStackTrace();
}
}
public void buildByteOutputStream(final ByteArrayOutputStream out, final TargetDataLine line, int frameSizeInBytes, final int bufferLengthInBytes) throws IOException {
final byte[] data = new byte[bufferLengthInBytes];
int numBytesRead;
line.start();
while (thread != null) {
if ((numBytesRead = line.read(data, 0, bufferLengthInBytes)) == -1) {
break;
}
out.write(data, 0, numBytesRead);
}
}
private void setAudioInputStream(AudioInputStream aStream) {
this.audioInputStream = aStream;
}
public AudioInputStream convertToAudioIStream(final ByteArrayOutputStream out, int frameSizeInBytes) {
byte[] audioBytes = out.toByteArray();
AudioInputStream audioStream =
new AudioInputStream(new ByteArrayInputStream(audioBytes), format,
audioBytes.length / frameSizeInBytes);
System.out.println("Recording finished");
return audioStream;
}
public TargetDataLine getTargetDataLineForRecord() {
TargetDataLine line;
DataLine.Info info = new DataLine.Info(TargetDataLine.class, format);
if (!AudioSystem.isLineSupported(info)) {
return null;
}
try {
line = (TargetDataLine) AudioSystem.getLine(info);
line.open(format, line.getBufferSize());
} catch (final Exception ex) {
return null;
}
return line;
}
}
}
|
[
"com.theokanning.openai.image.CreateImageRequest.builder"
] |
[((4160, 4255), 'com.theokanning.openai.image.CreateImageRequest.builder'), ((4160, 4230), 'com.theokanning.openai.image.CreateImageRequest.builder')]
|
package de.garrafao.phitag.computationalannotator.usepair.service;
import com.theokanning.openai.OpenAiHttpException;
import com.theokanning.openai.completion.chat.ChatCompletionChoice;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.service.OpenAiService;
import de.garrafao.phitag.computationalannotator.common.error.WrongApiKeyException;
import de.garrafao.phitag.computationalannotator.common.error.WrongModelException;
import de.garrafao.phitag.computationalannotator.common.function.CommonFunction;
import de.garrafao.phitag.computationalannotator.usepair.data.UsePairPrompt;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.stereotype.Service;
import java.util.List;
@Service
public class UsePairOpenAIService {
private final UsePairPrompt usePairPrompt;
private final CommonFunction commonFunction;
@Autowired
public UsePairOpenAIService(UsePairPrompt usePairPrompt, CommonFunction commonFunction) {
this.usePairPrompt = usePairPrompt;
this.commonFunction = commonFunction;
}
public String chat(final String apiKey, final String model, final String prompt, final String firstUsage,
final String secondUsage, final String lemma) {
try {
List<ChatMessage> messages = this.usePairPrompt.getChatMessages(prompt, firstUsage, secondUsage, lemma);
OpenAiService service = new OpenAiService(apiKey);
ChatCompletionRequest completionRequest = ChatCompletionRequest.builder()
.messages(messages)
.model(model)
.temperature(0.9)
.topP(0.9)
.n(1)
.build();
List<ChatCompletionChoice> choices = service.createChatCompletion(completionRequest).getChoices();
StringBuilder returnString = new StringBuilder();
for (ChatCompletionChoice choice : choices) {
ChatMessage message = choice.getMessage();
if (message != null) {
System.out.println(message.getContent());
returnString.append(message.getContent()).append(System.lineSeparator());
}
}
System.out.println("response "+ returnString);
int result = this.commonFunction.extractInteger(returnString.toString());
System.out.println("integer " + result);
return String.valueOf(result);
}catch (OpenAiHttpException e) {
if (e.getMessage().contains("The model")) {
throw new WrongModelException(model);
}
if (e.getMessage().contains("Incorrect API key provided")) {
throw new WrongApiKeyException();
}
throw e;
}
}
}
|
[
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((1606, 1835), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((1606, 1806), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((1606, 1780), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((1606, 1749), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((1606, 1711), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((1606, 1677), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
package idatt2106v231.backend.service;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.theokanning.openai.OpenAiApi;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.service.OpenAiService;
import idatt2106v231.backend.model.OpenAiKey;
import idatt2106v231.backend.repository.OpenAiKeyRepository;
import io.github.cdimascio.dotenv.Dotenv;
import okhttp3.OkHttpClient;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.stereotype.Service;
import retrofit2.Retrofit;
import java.time.Duration;
import java.util.ArrayList;
import java.util.List;
import java.util.Optional;
import static com.theokanning.openai.service.OpenAiService.*;
/**
* Class to manage Ai.
*/
@Service
public class AiServices {
private final OpenAiKeyRepository openAiKeyRepo;
/**
* Constructor which sets the Open AI key repository.
*/
@Autowired
public AiServices(OpenAiKeyRepository openAiKeyRepo) {
this.openAiKeyRepo = openAiKeyRepo;
}
/**
* Gets a chat completion using OpenAI GPT-3.
*
* @param content the content of the query
* @return the answer produced by the AI
*/
public String getChatCompletion(String content) {
try {
String token = getOpenAiApiKey();
if (token.startsWith("ERROR :")) throw new Exception(token);
ObjectMapper mapper = defaultObjectMapper();
Duration timeout = Duration.ofSeconds(300);
OkHttpClient client = defaultClient(token, timeout)
.newBuilder()
.build();
Retrofit retrofit = defaultRetrofit(client, mapper);
OpenAiApi api = retrofit.create(OpenAiApi.class);
OpenAiService service = new OpenAiService(api);
List<ChatMessage> messages = new ArrayList<>();
messages.add(new ChatMessage("user", content));
ChatCompletionRequest chatCompletionRequest = ChatCompletionRequest.builder()
.messages(messages)
.model("gpt-3.5-turbo")
.temperature(0.0)
.build();
return String.valueOf(service.createChatCompletion(chatCompletionRequest)
.getChoices().get(0).getMessage().getContent());
} catch (Exception e) {
return "ERROR: " + e.getMessage();
}
}
/**
* Gets the OpenAi API key.
* This must either be stored in the table 'open_ai_key' in the database,
* or in a .env file in the root of the project folder as OPENAI_TOKEN=your_token.
*
* @return the key
*/
public String getOpenAiApiKey() {
try {
String token = null;
Optional<OpenAiKey> openAiKey = openAiKeyRepo.findFirstByOrderByIdDesc();
if (openAiKey.isPresent()) token = openAiKey.get().getApiKey();
if (token == null) {
Dotenv dotenv = Dotenv.configure().load();
token = dotenv.get("OPENAI_TOKEN");
if (token == null) {
return "Token is missing. " +
"Make sure a valid OpenAI API key is stored in the database " +
"or in a .env file in the root of the project";
}
}
return token;
} catch (Exception e) {
return "ERROR: " + e.getMessage();
}
}
}
|
[
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((2086, 2268), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2086, 2239), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2086, 2201), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((2086, 2157), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3079, 3104), 'io.github.cdimascio.dotenv.Dotenv.configure')]
|
package com.ramesh.openai;
import java.time.Duration;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
/***
* This project demonstrates the Chain of Thought (CoT) prompting technique which is useful when there is need
* for analytical, reasoning, deriving etc. kind of problems
***/
class ChainOfThoughtPrompting {
public static void main(String... args) {
// Set the Open AI Token & Model
String token = "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";
String model = "gpt-3.5-turbo";
// service handle for calling OpenAI APIs
OpenAiService service = new OpenAiService(token, Duration.ofSeconds(30));
System.out.println("-----------------------------------------------------------");
// prompt - change this and run again and again. Mostly ChatGPT will not give the right response for complex prompt like puzzle.
// that's where Chain of thought comes to help (next prompt with COT is given below)
String prompt="I went to the market and bought 10 apples. I gave 2 apples to the neighbor and 2 to the repairman. I then went and bought 5 more apples and ate 1. How many apples did I remain with?";
System.out.println(prompt);
// create the Chat message object
final List<ChatMessage> messages = new ArrayList<>();
final ChatMessage userMessage = new ChatMessage(ChatMessageRole.USER.value(), prompt);
messages.add(userMessage);
// call ChatGPT ChatCompletion API and get the response
ChatCompletionRequest chatCompletionRequest = ChatCompletionRequest
.builder()
.model(model)
.messages(messages)
.n(1)
.temperature(.1)
.maxTokens(200)
.logitBias(new HashMap<>())
.build();
System.out.println("------------");
System.out.print("ChatGPT response=");
service.createChatCompletion(chatCompletionRequest).getChoices().forEach((c) -> {
System.out.println(c.getMessage().getContent());
});
System.out.println("\n-----------------------------------------------------------");
// Call ChatGPT Chat Completion with a CoT (Chain of THought) prompting technique
// You will see that ChatGPT most likely will give the right answer. This is because in the prompt
// the thinking process is given in the form of examples
String[] prompts = new String[10];
prompts[0] = "The odd numbers in this group add up to an even number: 4, 8, 9, 15, 12, 2, 1.";
prompts[1] = "A: The answer is False.";
prompts[2] = "The odd numbers in this group add up to an even number: 17, 10, 19, 4, 8, 12, 24.";
prompts[3] = "A: The answer is True.";
prompts[4] = "The odd numbers in this group add up to an even number: 16, 11, 14, 4, 8, 13, 24.";
prompts[5] = "A: The answer is True.";
prompts[6] = "The odd numbers in this group add up to an even number: 17, 9, 10, 12, 13, 4, 2.";
prompts[7] = "A: The answer is False.";
prompts[8] = "The odd numbers in this group add up to an even number: 15, 32, 5, 13, 82, 7, 1. ";
prompts[9] = "A: ";
final List<ChatMessage> messages_cot = new ArrayList<>();
for (int i = 0; i < 10; i++) {
System.out.println(prompts[i]);
final ChatMessage assistantMessage = new ChatMessage(ChatMessageRole.ASSISTANT.value(), prompts[i]);
messages_cot.add(assistantMessage);
}
ChatCompletionRequest chatCompletionRequest2 = ChatCompletionRequest
.builder()
.model(model)
.messages(messages_cot)
.n(1)
.temperature(.1)
.maxTokens(50)
.logitBias(new HashMap<>())
.build();
System.out.println("------------");
System.out.print("ChatGPT response=");
service.createChatCompletion(chatCompletionRequest2).getChoices().forEach((c) -> {
System.out.println(c.getMessage().getContent());
});
service.shutdownExecutor();
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.ASSISTANT.value"
] |
[((1626, 1654), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((3533, 3566), 'com.theokanning.openai.completion.chat.ChatMessageRole.ASSISTANT.value')]
|
package com.bambooleanlogic.ai;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.service.OpenAiService;
import java.io.IOException;
import java.nio.file.Files;
import java.nio.file.Path;
import java.util.List;
public class Main {
public static void main(String[] args) throws IOException {
SqlCode sql = generateSql(
"MySQL",
"Get all students who has at least one class where their grade is above average"
);
if (sql.code != null) {
System.out.println("--- CODE -----------------------");
System.out.println(sql.code);
System.out.println("--- COMMENT --------------------");
System.out.println(sql.comment);
System.out.println("--------------------------------");
} else {
System.out.println("--------------------------------");
System.out.println(sql.comment);
System.out.println("--------------------------------");
}
}
private static SqlCode generateSql(String dialect, String prompt) throws IOException {
String apiToken = Files.readString(Path.of("P:\\oapi.txt"));
OpenAiService service = new OpenAiService(apiToken);
ChatCompletionRequest request = ChatCompletionRequest.builder()
.model("gpt-3.5-turbo")
.messages(List.of(
new ChatMessage("system",
"You are a helpful assistant who produces " + dialect + " code."
),
new ChatMessage("user", prompt)
))
.build();
String response = service.createChatCompletion(request).getChoices().get(0).getMessage().getContent();
int start = response.indexOf("```");
if (start != -1) {
start += 3;
int end = response.indexOf("```", start);
if (end != -1) {
String code = response.substring(start, end).trim();
String comment = response.substring(end + 3).trim();
return new SqlCode(code, comment);
}
}
return new SqlCode(null, response);
}
private static final class SqlCode {
public final String code;
public final String comment;
public SqlCode(String code, String comment) {
this.code = code;
this.comment = comment;
}
}
}
|
[
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((1375, 1755), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((1375, 1730), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((1375, 1446), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
package com.chat.base.controller;
import com.chat.base.bean.annotation.VisitLimit;
import com.chat.base.bean.common.BaseCodeEnum;
import com.chat.base.bean.constants.*;
import com.chat.base.bean.entity.GptModelConfig;
import com.chat.base.bean.vo.*;
import com.chat.base.bean.entity.PromptModel;
import com.chat.base.bean.gpt.ApiChatReq;
import com.chat.base.bean.gpt.ChatReq;
import com.chat.base.bean.req.CompletionReq;
import com.chat.base.handler.*;
import com.chat.base.handler.gpt.OpenAiProxyServiceFactory;
import com.chat.base.service.ChatBaseOpenAiProxyService;
import com.chat.base.utils.*;
import com.google.common.cache.Cache;
import com.google.common.cache.CacheBuilder;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import io.github.asleepyfish.enums.RoleEnum;
import io.github.asleepyfish.exception.ChatGPTException;
import org.springframework.beans.factory.annotation.Value;
import lombok.extern.slf4j.Slf4j;
import org.apache.commons.lang3.StringUtils;
import org.springframework.beans.factory.annotation.Autowired;
import org.springframework.validation.annotation.Validated;
import org.springframework.web.bind.annotation.*;
import javax.servlet.http.HttpServletResponse;
import javax.validation.Valid;
import java.io.*;
import java.util.*;
import java.util.concurrent.TimeUnit;
/**
* @author huyd
* @date 2023/5/5 11:19 PM
*/
@Slf4j
@RestController
public class AIChatController extends BaseController {
@Autowired
private UserLogManager userLogManager;
@Autowired
private AIChatManger AIChatManger;
@Autowired
private PromptModelManager promptModelManager;
@Autowired
private DrawTaskInfoManager drawTaskInfoManager;
@Autowired
private WeightAlgorithmManager weightAlgorithmManager;
@Value("${file-token-path}")
private String mjTokenPath;
private static Cache<String, ChatBaseOpenAiProxyService> cache = CacheBuilder.newBuilder().initialCapacity(10).maximumSize(1000).expireAfterWrite(1000, TimeUnit.SECONDS).build();
@VisitLimit(value = {LimitEnum.IP}, scope = CommonConstant.NO_LOGIN_SCOPE)
@PostMapping("/chat/streamChatWithWeb/V3")
public void streamChatWithWebV3(@RequestBody @Valid ChatReq chatReq, HttpServletResponse response) throws Exception {
String ip = HttpUtil.getIpAddress();
String browserName = HttpUtil.browserName();
Long id = SessionUser.getUserId();
String conversationId = chatReq.getConversationId();
String userId = id == null ? conversationId : String.valueOf(id);
ModelPriceEnum modelPriceEnum = ModelPriceEnum.modelPriceMap.get(chatReq.getModel());
if (modelPriceEnum == null) {
response.getOutputStream().write(BaseCodeEnum.MODEL_NO_OPEN.getMsg().getBytes());
return;
}
CacheUserInfoVo cacheUserInfoVo = SessionUser.get();
try {
if (Objects.nonNull(cacheUserInfoVo) && Objects.nonNull(cacheUserInfoVo.getGptApiTokenVo())) {
AIChatManger.chatStream(chatReq, cacheUserInfoVo, response);
} else {
AIChatManger.streamChatWithWebV3NoStatus(chatReq, response);
}
} catch (ChatGPTException e) {
// 用户主动停掉回答
log.error("streamChatWithWebV3 user error chatReq={} ", chatReq, e);
} catch (Exception e) {
log.error("streamChatWithWebV3 error chatReq={} ", chatReq, e);
userLogManager.addUserLog(chatReq.getAppName(), userId, OpEnum.GPT3.getOp(), ip, browserName);
response.getOutputStream().write(BaseCodeEnum.SERVER_BUSY.getMsg().getBytes());
} finally {
response.getOutputStream().close();
}
}
/**
* 验证gpt的token效果
*
* @param chatReq
* @param response
* @throws Exception
*/
@PostMapping("/chat/streamChatWithWeb/api/chat")
public void streamChatWithApiChatWeb(@RequestBody @Valid ApiChatReq chatReq, HttpServletResponse response) throws Exception {
String ip = HttpUtil.getIpAddress();
String browserName = HttpUtil.browserName();
String uid = chatReq.getToken();
try {
response.setContentType("text/event-stream");
response.setCharacterEncoding("UTF-8");
response.setHeader("Cache-Control", "no-cache");
String model = StringUtils.isNoneEmpty(chatReq.getModel()) ? chatReq.getModel() : "gpt-3.5-turbo";
ChatBaseOpenAiProxyService proxyService = cache.get(chatReq.getToken() + model, () ->
OpenAiProxyServiceFactory.getService(chatReq.getToken(), chatReq.getProxyUrl(), model));
Integer contentNumber = CommonConstant.CONTENT_NUMBER;
String user = chatReq.getConversationId();
LinkedList<ChatMessage> userChatMessages = ChatMessageCacheUtil.getUserChatMessages(user, contentNumber);
userChatMessages.add(new ChatMessage(RoleEnum.USER.getRoleName(), chatReq.getPrompt()));
ChatMessageCacheUtil.getOkUserChatMessages(userChatMessages, model);
if (userChatMessages.size() <= 0) {
response.getOutputStream().write(BaseCodeEnum.TOKEN_OVER.getMsg().getBytes());
response.getOutputStream().close();
return;
}
ChatMessageResultVo streamChatCompletion = proxyService.createStreamChatCompletion(ChatCompletionRequest.builder()
.model(model)
.messages(userChatMessages)
.user(user)
.temperature(chatReq.getTemperature())
.topP(chatReq.getTop_p())
.stream(true)
.build(), response.getOutputStream(), uid);
if(streamChatCompletion!=null){
ChatMessageCacheUtil.saveChatMessage(user,streamChatCompletion.getChatMessage());
}
} catch (ChatGPTException e) {
// 用户主动停掉回答
log.error("streamChatWithWebV3 user error chatReq={} ", chatReq, e);
response.getOutputStream().write(BaseCodeEnum.TERMINATE.getMsg().getBytes());
} catch (Exception e) {
log.error("streamChatWithWebV3 error chatReq={} ", chatReq, e);
userLogManager.addUserLog("BlueCatApiChat", uid, OpEnum.GPT3.getOp(), ip, browserName);
response.getOutputStream().write(BaseCodeEnum.SERVER_BUSY.getMsg().getBytes());
} finally {
response.getOutputStream().close();
}
}
@PostMapping("/chat/streamChatWithWeb/completion")
public void completion(@RequestBody @Validated CompletionReq completionReq, HttpServletResponse response) throws IOException {
CacheUserInfoVo cacheUserInfoVo = SessionUser.get();
if (cacheUserInfoVo == null) {
response.getOutputStream().write("请登录之后再使用!".getBytes());
return;
}
response.setContentType("text/event-stream");
response.setCharacterEncoding("UTF-8");
response.setHeader("Cache-Control", "no-cache");
StringBuilder builder = new StringBuilder();
PromptModel prompt = promptModelManager.getPromptById(Long.parseLong(completionReq.getModelId()));
if (prompt == null || StringUtils.isBlank(prompt.getContent())) {
response.getOutputStream().write("模板已过期,请联系管理员".getBytes());
return;
}
builder.append(prompt.getContent()).append("\n");
builder.append(completionReq.getContent());
String uid = UUID.randomUUID().toString();
String model = StringUtils.isNoneEmpty(completionReq.getModel()) ? completionReq.getModel() : "gpt-3.5-turbo";
Optional<GptModelConfig> modelConfig = weightAlgorithmManager.round(cacheUserInfoVo, model);
if (!modelConfig.isPresent()) {
response.getOutputStream().write(BaseCodeEnum.NO_MODEL_ROLE.getMsg().getBytes());
return;
}
GptModelConfig gptModelConfig = modelConfig.get();
ChatBaseOpenAiProxyService proxyService = OpenAiProxyServiceFactory.createProxyService(gptModelConfig.getId().toString());
if (proxyService == null) {
response.getOutputStream().write(BaseCodeEnum.NO_MODEL.getMsg().getBytes());
response.getOutputStream().close();
return;
}
LinkedList<ChatMessage> userChatMessages = new LinkedList<>();
userChatMessages.add(new ChatMessage(RoleEnum.USER.getRoleName(), builder.toString()));
proxyService.createStreamChatCompletion(ChatCompletionRequest.builder()
.model(model)
.messages(userChatMessages)
.user(uid)
.temperature(1.0)
.topP(1.0)
.stream(true)
.build(), response.getOutputStream(), cacheUserInfoVo.getGptApiTokenVo().getToken());
}
}
|
[
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((1974, 2086), 'com.google.common.cache.CacheBuilder.newBuilder'), ((1974, 2078), 'com.google.common.cache.CacheBuilder.newBuilder'), ((1974, 2037), 'com.google.common.cache.CacheBuilder.newBuilder'), ((1974, 2019), 'com.google.common.cache.CacheBuilder.newBuilder'), ((2792, 2838), 'com.chat.base.bean.common.BaseCodeEnum.MODEL_NO_OPEN.getMsg'), ((2792, 2827), 'com.chat.base.bean.common.BaseCodeEnum.MODEL_NO_OPEN.getMsg'), ((3663, 3707), 'com.chat.base.bean.common.BaseCodeEnum.SERVER_BUSY.getMsg'), ((3663, 3696), 'com.chat.base.bean.common.BaseCodeEnum.SERVER_BUSY.getMsg'), ((5036, 5063), 'io.github.asleepyfish.enums.RoleEnum.USER.getRoleName'), ((5266, 5309), 'com.chat.base.bean.common.BaseCodeEnum.TOKEN_OVER.getMsg'), ((5266, 5298), 'com.chat.base.bean.common.BaseCodeEnum.TOKEN_OVER.getMsg'), ((5498, 5811), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5498, 5782), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5498, 5748), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5498, 5702), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5498, 5643), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5498, 5611), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((5498, 5563), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((6209, 6251), 'com.chat.base.bean.common.BaseCodeEnum.TERMINATE.getMsg'), ((6209, 6240), 'com.chat.base.bean.common.BaseCodeEnum.TERMINATE.getMsg'), ((6507, 6551), 'com.chat.base.bean.common.BaseCodeEnum.SERVER_BUSY.getMsg'), ((6507, 6540), 'com.chat.base.bean.common.BaseCodeEnum.SERVER_BUSY.getMsg'), ((8032, 8078), 'com.chat.base.bean.common.BaseCodeEnum.NO_MODEL_ROLE.getMsg'), ((8032, 8067), 'com.chat.base.bean.common.BaseCodeEnum.NO_MODEL_ROLE.getMsg'), ((8383, 8424), 'com.chat.base.bean.common.BaseCodeEnum.NO_MODEL.getMsg'), ((8383, 8413), 'com.chat.base.bean.common.BaseCodeEnum.NO_MODEL.getMsg'), ((8622, 8649), 'io.github.asleepyfish.enums.RoleEnum.USER.getRoleName'), ((8722, 8970), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8722, 8945), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8722, 8915), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8722, 8888), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8722, 8854), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8722, 8827), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((8722, 8783), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
package com.chunxia.chatgpt.chatapi;
import android.util.Log;
import com.blankj.utilcode.util.ThreadUtils;
import com.chunxia.chatgpt.model.review.SentenceCard;
import com.theokanning.openai.completion.chat.ChatCompletionChoice;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
import java.util.ArrayList;
import java.util.List;
public class MultiRoundChatAgent {
private static final String TAG = "MultiRoundChatAiApi";
private final List<ChatMessage> oldMessages = new ArrayList<>();
private String model = "gpt-3.5-turbo";
private int responseN = 1;
private int maxTokenN = 512;
private final ChatMessage systemMessage;
private final String systemCommand;
private final List<ThreadUtils.Task<String>> threadTasks = new ArrayList<>();
public MultiRoundChatAgent(String systemCommand, String model, int responseN, int maxTokenN) {
this.systemCommand = systemCommand;
this.model = model;
this.responseN = responseN;
this.maxTokenN = maxTokenN;
this.systemMessage = new ChatMessage(ChatMessageRole.SYSTEM.value(), this.systemCommand);
oldMessages.add(systemMessage);
}
public MultiRoundChatAgent() {
this.systemCommand = "";
this.systemMessage = new ChatMessage(ChatMessageRole.SYSTEM.value(), this.systemCommand);
oldMessages.add(systemMessage);
}
public MultiRoundChatAgent(String systemCommand) {
this.systemCommand = systemCommand;
this.systemMessage = new ChatMessage(ChatMessageRole.SYSTEM.value(), this.systemCommand);
oldMessages.add(systemMessage);
}
public void sendMessageInThread(String message, ReceiveOpenAiReply onReceiveOpenAiReply) {
ThreadUtils.Task<String> tTask = new ThreadUtils.SimpleTask<String>() {
@Override
public String doInBackground() throws Throwable {
return sendToChatAi(message);
}
@Override
public void onSuccess(String result) {
Log.i(TAG, "receive reply from chatgpt");
onReceiveOpenAiReply.onSuccess(result);
}
};
threadTasks.add(tTask);
ThreadUtils.getIoPool().execute(tTask);
}
public String sendMessage(String message) {
return sendToChatAi(message);
}
public void cancelAllCurrentThread() {
// todo 只取消当前正在执行的
threadTasks.forEach(ThreadUtils::cancel);
}
public SentenceCard getOneRoundSentenceCard() {
if (oldMessages.size() < 3) {
return null;
}
SentenceCard sentenceCard = new SentenceCard(oldMessages.get(2).getContent(), oldMessages.get(1).getContent());
return sentenceCard;
}
public interface ReceiveOpenAiReply {
void onSuccess(String reply);
}
private void insertUserMessage(String message) {
final ChatMessage userMessage = new ChatMessage(ChatMessageRole.USER.value(), message);
oldMessages.add(userMessage);
}
private String sendToChatAi(String message) {
Log.i(TAG, "User: " + message);
insertUserMessage(message);
ChatCompletionRequest chatCompletionRequest = ChatCompletionRequest
.builder()
.model(model)
.messages(oldMessages)
.n(responseN)
.maxTokens(maxTokenN)
.build();
OpenAiService openAiService = OpenAIServiceManager.getOpenAiService();
if (openAiService == null) {
return null;
} else {
List<ChatCompletionChoice> choices = openAiService.createChatCompletion(chatCompletionRequest).getChoices();
if (!choices.isEmpty()) {
String content = choices.get(0).getMessage().getContent();
Log.i(TAG, "ChatGpt: " + content);
addChatGptReplyToMessage(choices.get(0).getMessage());
return content;
}
}
return null;
}
public void clearOldMessage() {
oldMessages.clear();
oldMessages.add(systemMessage);
}
public void addChatGptReplyToMessage(ChatMessage message) {
oldMessages.add(message);
}
public int getMaxTokenN() {
return maxTokenN;
}
public void setMaxTokenN(int maxTokenN) {
this.maxTokenN = maxTokenN;
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value"
] |
[((1259, 1289), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value'), ((1473, 1503), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value'), ((1717, 1747), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value'), ((2390, 2428), 'com.blankj.utilcode.util.ThreadUtils.getIoPool'), ((3155, 3183), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value')]
|
package com.theokanning.openai.service;
import com.theokanning.openai.moderation.Moderation;
import com.theokanning.openai.moderation.ModerationRequest;
import org.junit.jupiter.api.Test;
import static org.junit.jupiter.api.Assertions.assertTrue;
public class ModerationTest {
String token = System.getenv("OPENAI_TOKEN");
com.theokanning.openai.service.OpenAiService service = new OpenAiService(token);
@Test
void createModeration() {
ModerationRequest moderationRequest = ModerationRequest.builder()
.input("I want to kill them")
.model("text-moderation-latest")
.build();
Moderation moderationScore = service.createModeration(moderationRequest).getResults().get(0);
assertTrue(moderationScore.isFlagged());
}
}
|
[
"com.theokanning.openai.moderation.ModerationRequest.builder"
] |
[((504, 651), 'com.theokanning.openai.moderation.ModerationRequest.builder'), ((504, 626), 'com.theokanning.openai.moderation.ModerationRequest.builder'), ((504, 577), 'com.theokanning.openai.moderation.ModerationRequest.builder')]
|
package com.theokanning.openai.service;
import com.theokanning.openai.completion.chat.ChatCompletionChoice;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import org.junit.jupiter.api.Test;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
import static org.junit.jupiter.api.Assertions.assertEquals;
class ChatCompletionTest {
String token = System.getenv("OPENAI_TOKEN");
OpenAiService service = new OpenAiService(token);
@Test
void createChatCompletion() {
final List<ChatMessage> messages = new ArrayList<>();
final ChatMessage systemMessage = new ChatMessage(ChatMessageRole.SYSTEM.value(), "You are a dog and will speak as such.");
messages.add(systemMessage);
ChatCompletionRequest chatCompletionRequest = ChatCompletionRequest
.builder()
.model("gpt-3.5-turbo")
.messages(messages)
.n(5)
.maxTokens(50)
.logitBias(new HashMap<>())
.build();
List<ChatCompletionChoice> choices = service.createChatCompletion(chatCompletionRequest).getChoices();
assertEquals(5, choices.size());
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value"
] |
[((772, 802), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value')]
|
package com.couchbase.intellij.tree.iq.intents;
import com.couchbase.client.java.json.JsonArray;
import com.couchbase.client.java.json.JsonObject;
import com.couchbase.intellij.tree.iq.IQWindowContent;
import com.couchbase.intellij.tree.iq.chat.ChatExchangeAbortException;
import com.couchbase.intellij.tree.iq.chat.ChatGptHandler;
import com.couchbase.intellij.tree.iq.chat.ChatLink;
import com.couchbase.intellij.tree.iq.chat.ChatLinkService;
import com.couchbase.intellij.tree.iq.chat.ChatLinkState;
import com.couchbase.intellij.tree.iq.chat.ChatMessageEvent;
import com.couchbase.intellij.tree.iq.chat.ChatMessageListener;
import com.couchbase.intellij.tree.iq.chat.ConfigurationPage;
import com.couchbase.intellij.tree.iq.chat.ConversationContext;
import com.couchbase.intellij.tree.iq.core.IQCredentials;
import com.couchbase.intellij.tree.iq.intents.actions.ActionInterface;
import com.couchbase.intellij.tree.iq.settings.OpenAISettingsState;
import com.couchbase.intellij.workbench.Log;
import com.intellij.testFramework.fixtures.BasePlatformTestCase;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import java.util.ArrayList;
import java.util.List;
import java.util.function.Consumer;
public abstract class AbstractIQTest extends BasePlatformTestCase {
private static final String IQ_URL = System.getenv("CAPELLA_DOMAIN") + "/v2/organizations/%s/integrations/iq/";
private static final ChatGptHandler handler = new ChatGptHandler();
private static ConversationContext ctx;
private static ChatLink link;
@Override
protected void setUp() throws Exception {
super.setUp();
IQCredentials credentials = new IQCredentials(System.getenv("IQ_ORG_LOGIN"), System.getenv("IQ_ORG_PASSWD"));
assertTrue("Please set capella domain and IQ credentials using `CAPELLA_DOMAIN`, `IQ_ORG_ID`, `IQ_ORG_LOGIN`, and `IQ_ORG_PASSWD` envvars", credentials.doLogin());
String orgId = System.getenv("IQ_ORG_ID");
final String iqUrl = String.format(IQ_URL, orgId);
OpenAISettingsState.OpenAIConfig iqGptConfig = new OpenAISettingsState.OpenAIConfig();
OpenAISettingsState.getInstance().setGpt4Config(iqGptConfig);
OpenAISettingsState.getInstance().setEnableInitialMessage(false);
iqGptConfig.setApiKey(credentials.getAuth().getJwt());
iqGptConfig.setEnableStreamResponse(false);
iqGptConfig.setModelName("gpt-4");
iqGptConfig.setApiEndpointUrl(iqUrl);
iqGptConfig.setEnableCustomApiEndpointUrl(true);
ConfigurationPage cp = iqGptConfig.withSystemPrompt(IQWindowContent::systemPrompt);
Log.setLevel(3);
Log.setPrinter(new Log.StdoutPrinter());
link = new ChatLinkService(getProject(), null, cp);
ctx = new ChatLinkState(cp);
}
protected void send(String message, Consumer<ChatMessageEvent.ResponseArrived> listener) {
send(message, false, listener);
}
protected void send(String message, boolean isSystem, Consumer<ChatMessageEvent.ResponseArrived> listener) {
ChatMessage chatMessage = new ChatMessage(
isSystem ? ChatMessageRole.SYSTEM.value() : ChatMessageRole.USER.value(),
message
);
ChatMessageEvent.Starting event = ChatMessageEvent.starting(AbstractIQTest.link, chatMessage);
ctx.addChatMessage(chatMessage);
List<ChatMessage> messages = ctx.getChatMessages(ctx.getModelType(), chatMessage);
if (isSystem) {
messages.add(chatMessage);
}
ChatCompletionRequest request = ChatCompletionRequest.builder()
.messages(messages)
.build();
handler.handle(AbstractIQTest.ctx, event.initiating(request), new ChatMessageListener() {
@Override
public void exchangeStarting(ChatMessageEvent.Starting event) throws ChatExchangeAbortException {
}
@Override
public void exchangeStarted(ChatMessageEvent.Started event) {
}
@Override
public void responseArriving(ChatMessageEvent.ResponseArriving event) {
}
@Override
public void responseArrived(ChatMessageEvent.ResponseArrived event) {
listener.accept(event);
}
@Override
public void responseCompleted(ChatMessageEvent.ResponseArrived event) {
}
@Override
public void exchangeFailed(ChatMessageEvent.Failed event) {
throw new RuntimeException("IQ Exchange failed", event.getCause());
}
@Override
public void exchangeCancelled(ChatMessageEvent.Cancelled event) {
}
}).blockingLast();
}
protected String getResponse(ChatMessageEvent.ResponseArrived response) {
assertEquals(1, response.getResponseChoices().size());
return response.getResponseChoices().get(0).getContent();
}
protected JsonObject getJson(ChatMessageEvent.ResponseArrived response) {
return JsonObject.fromJson(getResponse(response));
}
protected void assertJsonResponse(ChatMessageEvent.ResponseArrived response) {
String message = getResponse(response);
assertTrue(message.startsWith("{"));
}
protected void assertNotJson(ChatMessageEvent.ResponseArrived response) {
assertFalse(getResponse(response).trim().charAt(0) == '{');
}
protected List<JsonObject> getIntents(ChatMessageEvent.ResponseArrived response, Class<? extends ActionInterface> action) {
List<JsonObject> results = new ArrayList<>();
JsonObject json = getJson(response);
assertInstanceOf(json.get("actions"), JsonArray.class);
JsonArray actions = json.getArray("actions");
for (int i = 0; i < actions.size(); i++) {
assertInstanceOf(actions.get(i), JsonObject.class);
JsonObject intent = actions.getObject(i);
assertInstanceOf(intent.get("action"), String.class);
if (intent.getString("action").equals(action.getSimpleName())) {
results.add(intent);
}
}
return results;
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value",
"com.theokanning.openai.completion.chat.ChatCompletionRequest.builder"
] |
[((2263, 2323), 'com.couchbase.intellij.tree.iq.settings.OpenAISettingsState.getInstance'), ((2333, 2397), 'com.couchbase.intellij.tree.iq.settings.OpenAISettingsState.getInstance'), ((3263, 3293), 'com.theokanning.openai.completion.chat.ChatMessageRole.SYSTEM.value'), ((3296, 3324), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((3709, 3801), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder'), ((3709, 3776), 'com.theokanning.openai.completion.chat.ChatCompletionRequest.builder')]
|
package com.theokanning.openai.service;
import com.fasterxml.jackson.annotation.JsonInclude;
import com.fasterxml.jackson.core.JsonProcessingException;
import com.fasterxml.jackson.core.type.TypeReference;
import com.fasterxml.jackson.databind.DeserializationFeature;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.fasterxml.jackson.databind.PropertyNamingStrategy;
import com.theokanning.openai.ListSearchParameters;
import com.theokanning.openai.OpenAiResponse;
import com.theokanning.openai.assistants.Assistant;
import com.theokanning.openai.assistants.AssistantFunction;
import com.theokanning.openai.assistants.AssistantRequest;
import com.theokanning.openai.assistants.AssistantToolsEnum;
import com.theokanning.openai.assistants.Tool;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatFunction;
import com.theokanning.openai.completion.chat.ChatFunctionCall;
import com.theokanning.openai.messages.Message;
import com.theokanning.openai.messages.MessageRequest;
import com.theokanning.openai.runs.RequiredAction;
import com.theokanning.openai.runs.Run;
import com.theokanning.openai.runs.RunCreateRequest;
import com.theokanning.openai.runs.RunStep;
import com.theokanning.openai.runs.SubmitToolOutputRequestItem;
import com.theokanning.openai.runs.SubmitToolOutputs;
import com.theokanning.openai.runs.SubmitToolOutputsRequest;
import com.theokanning.openai.runs.ToolCall;
import com.theokanning.openai.threads.Thread;
import com.theokanning.openai.threads.ThreadRequest;
import com.theokanning.openai.utils.TikTokensUtil;
import org.junit.jupiter.api.Test;
import java.time.Duration;
import java.util.ArrayList;
import java.util.List;
import java.util.Map;
import java.util.Objects;
import static org.junit.jupiter.api.Assertions.assertEquals;
import static org.junit.jupiter.api.Assertions.assertNotNull;
class AssistantFunctionTest {
String token = System.getenv("OPENAI_TOKEN");
OpenAiService service = new OpenAiService(token, Duration.ofMinutes(1));
@Test
void createRetrieveRun() throws JsonProcessingException {
ObjectMapper mapper = new ObjectMapper();
mapper.configure(DeserializationFeature.FAIL_ON_UNKNOWN_PROPERTIES, false);
mapper.setSerializationInclusion(JsonInclude.Include.NON_NULL);
mapper.setPropertyNamingStrategy(PropertyNamingStrategy.SNAKE_CASE);
mapper.addMixIn(ChatFunction.class, ChatFunctionMixIn.class);
mapper.addMixIn(ChatCompletionRequest.class, ChatCompletionRequestMixIn.class);
mapper.addMixIn(ChatFunctionCall.class, ChatFunctionCallMixIn.class);
String funcDef = "{\n" +
" \"type\": \"object\",\n" +
" \"properties\": {\n" +
" \"location\": {\n" +
" \"type\": \"string\",\n" +
" \"description\": \"The city and state, e.g. San Francisco, CA\"\n" +
" },\n" +
" \"unit\": {\n" +
" \"type\": \"string\",\n" +
" \"enum\": [\"celsius\", \"fahrenheit\"]\n" +
" }\n" +
" },\n" +
" \"required\": [\"location\"]\n" +
"}";
Map<String, Object> funcParameters = mapper.readValue(funcDef, new TypeReference<Map<String, Object>>() {});
AssistantFunction function = AssistantFunction.builder()
.name("weather_reporter")
.description("Get the current weather of a location")
.parameters(funcParameters)
.build();
List<Tool> toolList = new ArrayList<>();
Tool funcTool = new Tool(AssistantToolsEnum.FUNCTION, function);
toolList.add(funcTool);
AssistantRequest assistantRequest = AssistantRequest.builder()
.model(TikTokensUtil.ModelEnum.GPT_4_1106_preview.getName())
.name("MATH_TUTOR")
.instructions("You are a personal Math Tutor.")
.tools(toolList)
.build();
Assistant assistant = service.createAssistant(assistantRequest);
ThreadRequest threadRequest = ThreadRequest.builder()
.build();
Thread thread = service.createThread(threadRequest);
MessageRequest messageRequest = MessageRequest.builder()
.content("What's the weather of Xiamen?")
.build();
Message message = service.createMessage(thread.getId(), messageRequest);
RunCreateRequest runCreateRequest = RunCreateRequest.builder()
.assistantId(assistant.getId())
.build();
Run run = service.createRun(thread.getId(), runCreateRequest);
assertNotNull(run);
Run retrievedRun = service.retrieveRun(thread.getId(), run.getId());
while (!(retrievedRun.getStatus().equals("completed"))
&& !(retrievedRun.getStatus().equals("failed"))
&& !(retrievedRun.getStatus().equals("requires_action"))){
retrievedRun = service.retrieveRun(thread.getId(), run.getId());
}
if (retrievedRun.getStatus().equals("requires_action")) {
RequiredAction requiredAction = retrievedRun.getRequiredAction();
System.out.println("requiredAction");
System.out.println(mapper.writeValueAsString(requiredAction));
List<ToolCall> toolCalls = requiredAction.getSubmitToolOutputs().getToolCalls();
ToolCall toolCall = toolCalls.get(0);
String toolCallId = toolCall.getId();
SubmitToolOutputRequestItem toolOutputRequestItem = SubmitToolOutputRequestItem.builder()
.toolCallId(toolCallId)
.output("sunny")
.build();
List<SubmitToolOutputRequestItem> toolOutputRequestItems = new ArrayList<>();
toolOutputRequestItems.add(toolOutputRequestItem);
SubmitToolOutputsRequest submitToolOutputsRequest = SubmitToolOutputsRequest.builder()
.toolOutputs(toolOutputRequestItems)
.build();
retrievedRun = service.submitToolOutputs(retrievedRun.getThreadId(), retrievedRun.getId(), submitToolOutputsRequest);
while (!(retrievedRun.getStatus().equals("completed"))
&& !(retrievedRun.getStatus().equals("failed"))
&& !(retrievedRun.getStatus().equals("requires_action"))){
retrievedRun = service.retrieveRun(thread.getId(), run.getId());
}
OpenAiResponse<Message> response = service.listMessages(thread.getId());
List<Message> messages = response.getData();
System.out.println(mapper.writeValueAsString(messages));
}
}
}
|
[
"com.theokanning.openai.utils.TikTokensUtil.ModelEnum.GPT_4_1106_preview.getName",
"com.theokanning.openai.assistants.AssistantRequest.builder",
"com.theokanning.openai.messages.MessageRequest.builder",
"com.theokanning.openai.assistants.AssistantFunction.builder",
"com.theokanning.openai.runs.SubmitToolOutputsRequest.builder",
"com.theokanning.openai.runs.SubmitToolOutputRequestItem.builder",
"com.theokanning.openai.threads.ThreadRequest.builder",
"com.theokanning.openai.runs.RunCreateRequest.builder"
] |
[((3437, 3645), 'com.theokanning.openai.assistants.AssistantFunction.builder'), ((3437, 3620), 'com.theokanning.openai.assistants.AssistantFunction.builder'), ((3437, 3576), 'com.theokanning.openai.assistants.AssistantFunction.builder'), ((3437, 3506), 'com.theokanning.openai.assistants.AssistantFunction.builder'), ((3864, 4125), 'com.theokanning.openai.assistants.AssistantRequest.builder'), ((3864, 4100), 'com.theokanning.openai.assistants.AssistantRequest.builder'), ((3864, 4067), 'com.theokanning.openai.assistants.AssistantRequest.builder'), ((3864, 4003), 'com.theokanning.openai.assistants.AssistantRequest.builder'), ((3864, 3967), 'com.theokanning.openai.assistants.AssistantRequest.builder'), ((3914, 3966), 'com.theokanning.openai.utils.TikTokensUtil.ModelEnum.GPT_4_1106_preview.getName'), ((4239, 4287), 'com.theokanning.openai.threads.ThreadRequest.builder'), ((4391, 4498), 'com.theokanning.openai.messages.MessageRequest.builder'), ((4391, 4473), 'com.theokanning.openai.messages.MessageRequest.builder'), ((4627, 4726), 'com.theokanning.openai.runs.RunCreateRequest.builder'), ((4627, 4701), 'com.theokanning.openai.runs.RunCreateRequest.builder'), ((5724, 5871), 'com.theokanning.openai.runs.SubmitToolOutputRequestItem.builder'), ((5724, 5842), 'com.theokanning.openai.runs.SubmitToolOutputRequestItem.builder'), ((5724, 5805), 'com.theokanning.openai.runs.SubmitToolOutputRequestItem.builder'), ((6090, 6210), 'com.theokanning.openai.runs.SubmitToolOutputsRequest.builder'), ((6090, 6181), 'com.theokanning.openai.runs.SubmitToolOutputsRequest.builder')]
|
package com.theokanning.openai.service;
import com.theokanning.openai.audio.CreateSpeechRequest;
import com.theokanning.openai.audio.CreateTranscriptionRequest;
import com.theokanning.openai.audio.CreateTranslationRequest;
import com.theokanning.openai.audio.TranscriptionResult;
import com.theokanning.openai.audio.TranslationResult;
import org.junit.jupiter.api.Test;
import java.io.IOException;
import java.time.Duration;
import okhttp3.MediaType;
import okhttp3.ResponseBody;
import static org.junit.jupiter.api.Assertions.*;
public class AudioTest {
static String englishAudioFilePath = "src/test/resources/hello-world.mp3";
static String koreanAudioFilePath = "src/test/resources/korean-hello.mp3";
String token = System.getenv("OPENAI_TOKEN");
OpenAiService service = new OpenAiService(token, Duration.ofSeconds(30));
@Test
void createTranscription() {
CreateTranscriptionRequest createTranscriptionRequest = CreateTranscriptionRequest.builder()
.model("whisper-1")
.build();
String text = service.createTranscription(createTranscriptionRequest, englishAudioFilePath).getText();
assertEquals("Hello World.", text);
}
@Test
void createTranscriptionVerbose() {
CreateTranscriptionRequest createTranscriptionRequest = CreateTranscriptionRequest.builder()
.model("whisper-1")
.responseFormat("verbose_json")
.build();
TranscriptionResult result = service.createTranscription(createTranscriptionRequest, englishAudioFilePath);
assertEquals("Hello World.", result.getText());
assertEquals("transcribe", result.getTask());
assertEquals("english", result.getLanguage());
assertTrue(result.getDuration() > 0);
assertEquals(1, result.getSegments().size());
}
@Test
void createTranslation() {
CreateTranslationRequest createTranslationRequest = CreateTranslationRequest.builder()
.model("whisper-1")
.build();
String text = service.createTranslation(createTranslationRequest, koreanAudioFilePath).getText();
assertEquals("Hello, my name is Yoona. I am a Korean native speaker.", text);
}
@Test
void createTranslationVerbose() {
CreateTranslationRequest createTranslationRequest = CreateTranslationRequest.builder()
.model("whisper-1")
.responseFormat("verbose_json")
.build();
TranslationResult result = service.createTranslation(createTranslationRequest, koreanAudioFilePath);
assertEquals("Hello, my name is Yoona. I am a Korean native speaker.", result.getText());
assertEquals("translate", result.getTask());
assertEquals("english", result.getLanguage());
assertTrue(result.getDuration() > 0);
assertEquals(1, result.getSegments().size());
}
@Test
void createSpeech() throws IOException {
CreateSpeechRequest createSpeechRequest = CreateSpeechRequest.builder()
.model("tts-1")
.input("Hello World.")
.voice("alloy")
.build();
final ResponseBody speech = service.createSpeech(createSpeechRequest);
assertNotNull(speech);
assertEquals(MediaType.get("audio/mpeg"), speech.contentType());
assertTrue(speech.bytes().length > 0);
}
}
|
[
"com.theokanning.openai.audio.CreateTranslationRequest.builder",
"com.theokanning.openai.audio.CreateSpeechRequest.builder",
"com.theokanning.openai.audio.CreateTranscriptionRequest.builder"
] |
[((958, 1055), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((958, 1030), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((1334, 1479), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((1334, 1454), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((1334, 1406), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((1971, 2066), 'com.theokanning.openai.audio.CreateTranslationRequest.builder'), ((1971, 2041), 'com.theokanning.openai.audio.CreateTranslationRequest.builder'), ((2376, 2519), 'com.theokanning.openai.audio.CreateTranslationRequest.builder'), ((2376, 2494), 'com.theokanning.openai.audio.CreateTranslationRequest.builder'), ((2376, 2446), 'com.theokanning.openai.audio.CreateTranslationRequest.builder'), ((3049, 3206), 'com.theokanning.openai.audio.CreateSpeechRequest.builder'), ((3049, 3181), 'com.theokanning.openai.audio.CreateSpeechRequest.builder'), ((3049, 3149), 'com.theokanning.openai.audio.CreateSpeechRequest.builder'), ((3049, 3110), 'com.theokanning.openai.audio.CreateSpeechRequest.builder')]
|
package org.zhong.chatgpt.wechat.bot.chatgptwechatbot.test;
import java.time.Duration;
import java.util.List;
import org.apache.http.client.CookieStore;
import org.apache.http.cookie.Cookie;
import org.apache.http.impl.client.BasicCookieStore;
import org.apache.http.impl.client.CloseableHttpClient;
import org.apache.http.impl.client.HttpClients;
import org.junit.jupiter.api.Test;
import org.zhong.chatgpt.wechat.bot.config.BotConfig;
import com.theokanning.openai.completion.CompletionRequest;
import cn.zhouyafeng.itchat4j.utils.MyHttpClient;
import com.theokanning.openai.OpenAiService;
public class TestOpenAI {
private static CloseableHttpClient httpClient;
private static MyHttpClient instance = null;
private static CookieStore cookieStore;
static {
cookieStore = new BasicCookieStore();
// 将CookieStore设置到httpClient中
httpClient = HttpClients.custom().setDefaultCookieStore(cookieStore).build();
}
public static String getCookie(String name) {
List<Cookie> cookies = cookieStore.getCookies();
for (Cookie cookie : cookies) {
if (cookie.getName().equalsIgnoreCase(name)) {
return cookie.getValue();
}
}
return null;
}
@Test
public void test() {
OpenAiService service = new OpenAiService(BotConfig.getAppKey(),"https://api.openai.com/", Duration.ofSeconds(300));
CompletionRequest completionRequest = CompletionRequest.builder()
.prompt("你好")
.model("text-davinci-003")
.maxTokens(2000)
.temperature(0.8)
.topP(1.0)
.frequencyPenalty(0.55)
.presencePenalty(0.19)
.echo(true)
.user("1234213213")
.build();
String text = service.createCompletion(completionRequest).getChoices().get(0).getText();
System.out.print(text);
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((872, 935), 'org.apache.http.impl.client.HttpClients.custom'), ((872, 927), 'org.apache.http.impl.client.HttpClients.custom'), ((1374, 1638), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1374, 1625), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1374, 1601), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1374, 1585), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1374, 1558), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1374, 1530), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1374, 1515), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1374, 1493), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1374, 1466), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1374, 1429), 'com.theokanning.openai.completion.CompletionRequest.builder')]
|
package com.odde.doughnut.testability;
import static com.odde.doughnut.services.ai.builder.OpenAIChatRequestBuilder.askClarificationQuestion;
import static com.odde.doughnut.services.ai.tools.AiToolFactory.COMPLETE_NOTE_DETAILS;
import static org.mockito.ArgumentMatchers.any;
import static org.mockito.Mockito.when;
import com.fasterxml.jackson.databind.JsonNode;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.theokanning.openai.OpenAiResponse;
import com.theokanning.openai.client.OpenAiApi;
import com.theokanning.openai.messages.Message;
import com.theokanning.openai.messages.MessageContent;
import com.theokanning.openai.messages.content.Text;
import com.theokanning.openai.runs.*;
import io.reactivex.Single;
import java.util.List;
import org.mockito.ArgumentMatchers;
import org.mockito.Mockito;
public record OpenAIAssistantMock(OpenAiApi openAiApi) {
public void mockThreadRunCompletionToolCalled(Object result, String runId) {
mockCreateRunInProcess(runId);
Run retrievedRun = getRunThatCallCompletionTool(runId, result);
Mockito.doReturn(Single.just(retrievedRun))
.when(openAiApi)
.retrieveRun(ArgumentMatchers.any(), ArgumentMatchers.any());
}
public void mockThreadRunCompletedAndListMessage(String msg, String runId) {
mockCreateRunInProcess(runId);
Run retrievedRun = getRunThatCompleted(runId);
Mockito.doReturn(Single.just(retrievedRun))
.when(openAiApi)
.retrieveRun(ArgumentMatchers.any(), ArgumentMatchers.any());
Text txt = new Text(msg, List.of());
MessageContent cnt = new MessageContent();
cnt.setText(txt);
List<MessageContent> contentList = List.of(cnt);
OpenAiResponse<Message> msgs = new OpenAiResponse<>();
msgs.setData(List.of(Message.builder().content(contentList).build()));
Mockito.doReturn(Single.just(msgs)).when(openAiApi).listMessages(retrievedRun.getThreadId());
}
public void mockSubmitOutputAndCompletion(Object result, String runId) {
Run run = getRunThatCallCompletionTool(runId, result);
when(openAiApi.submitToolOutputs(any(), any(), any())).thenReturn(Single.just(run));
}
public void mockSubmitOutputAndRequiredMoreAction(Object result, String runId) {
Run run =
getRunThatRequiresAction(
new ObjectMapper().valueToTree(result).toString(), runId, askClarificationQuestion);
when(openAiApi.submitToolOutputs(any(), any(), any())).thenReturn(Single.just(run));
}
private static Run getRunThatCallCompletionTool(String runId, Object result) {
JsonNode arguments = new ObjectMapper().valueToTree(result);
return getRunThatRequiresAction(arguments.toString(), runId, COMPLETE_NOTE_DETAILS);
}
private void mockCreateRunInProcess(String runId) {
Run run = new Run();
run.setId(runId);
run.setStatus("processing");
Mockito.doReturn(Single.just(run))
.when(openAiApi)
.createRun(ArgumentMatchers.any(), ArgumentMatchers.any());
}
private static Run getRunThatCompleted(String runId) {
Run retrievedRun = new Run();
retrievedRun.setId(runId);
retrievedRun.setStatus("completed");
return retrievedRun;
}
private static Run getRunThatRequiresAction(
String arguments, String runId, String function_name) {
Run retrievedRun = new Run();
retrievedRun.setId(runId);
retrievedRun.setStatus("requires_action");
retrievedRun.setRequiredAction(
RequiredAction.builder()
.submitToolOutputs(
SubmitToolOutputs.builder()
.toolCalls(
List.of(
ToolCall.builder()
.id("mocked-tool-call-id")
.function(
ToolCallFunction.builder()
.name(function_name)
.arguments(arguments)
.build())
.build()))
.build())
.build());
return retrievedRun;
}
}
|
[
"com.theokanning.openai.messages.Message.builder"
] |
[((1068, 1205), 'org.mockito.Mockito.doReturn'), ((1068, 1136), 'org.mockito.Mockito.doReturn'), ((1381, 1518), 'org.mockito.Mockito.doReturn'), ((1381, 1449), 'org.mockito.Mockito.doReturn'), ((1767, 1813), 'com.theokanning.openai.messages.Message.builder'), ((1767, 1805), 'com.theokanning.openai.messages.Message.builder'), ((1821, 1913), 'org.mockito.Mockito.doReturn'), ((1821, 1872), 'org.mockito.Mockito.doReturn'), ((2848, 2974), 'org.mockito.Mockito.doReturn'), ((2848, 2907), 'org.mockito.Mockito.doReturn')]
|
package com.touchbiz.chatgpt.simple;
import com.fasterxml.jackson.annotation.JsonInclude;
import com.fasterxml.jackson.databind.DeserializationFeature;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.fasterxml.jackson.databind.PropertyNamingStrategy;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.service.OpenAiService;
import com.touchbiz.common.utils.tools.JsonUtils;
import lombok.AllArgsConstructor;
import lombok.Builder;
import lombok.Data;
import lombok.SneakyThrows;
import lombok.extern.slf4j.Slf4j;
import org.junit.Test;
import org.springframework.core.ParameterizedTypeReference;
import org.springframework.http.MediaType;
import org.springframework.http.codec.ServerSentEvent;
import org.springframework.web.reactive.function.BodyInserters;
import org.springframework.web.reactive.function.client.WebClient;
import reactor.core.publisher.Flux;
import java.net.URI;
import java.net.http.HttpClient;
import java.net.http.HttpRequest;
import java.net.http.HttpResponse;
import java.time.LocalTime;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.List;
import java.util.function.Consumer;
@Slf4j
public class EventStreamTest {
String token = "";
@Test
public void testRetrofit(){
CompletionRequest completionRequest = CompletionRequest.builder()
// .prompt("Human:" + chat.prompt +"\nAI:")
.prompt("胡寅恺帅嘛")
.model("text-davinci-003")
// .echo(true)
// .stop(Arrays.asList(" Human:"," AI:"))
.maxTokens(128)
.presencePenalty(0d)
.frequencyPenalty(0d)
.temperature(0.7D)
.bestOf(1)
.topP(1d)
// .stream(true)
.build();
OpenAiService service = new OpenAiService(token);
var result = service.createCompletion(completionRequest);
log.info("result:{}", JsonUtils.toJson(result));
}
@SneakyThrows
@Test
public void testHttp() {
HttpClient client = HttpClient.newBuilder().build();
CompletionRequest completionRequest = CompletionRequest.builder()
// .prompt("Human:" + chat.prompt +"\nAI:")
.prompt("给我推荐10本小说")
.model("text-davinci-001")
// .echo(true)
.stop(Arrays.asList(" Human:"," AI:"))
.maxTokens(1024)
.presencePenalty(0d)
.frequencyPenalty(0d)
.temperature(0.7D)
.bestOf(1)
.topP(1d)
.stream(true)
.build();
ObjectMapper mapper = new ObjectMapper();
mapper.configure(DeserializationFeature.FAIL_ON_UNKNOWN_PROPERTIES, false);
mapper.setSerializationInclusion(JsonInclude.Include.NON_NULL);
mapper.setPropertyNamingStrategy(PropertyNamingStrategy.SNAKE_CASE);
var json = mapper.writeValueAsString(completionRequest);
log.info("json:{}", json);
HttpRequest request = HttpRequest.newBuilder()
.header("Authorization", "Bearer " + this.token)
.header( "Content-Type", "application/json")
.POST(HttpRequest.BodyPublishers.ofString(json))
.uri(URI.create("https://api.openai.com/v1/completions"))
.build();
client.sendAsync(request, HttpResponse.BodyHandlers.ofLines())
.thenApply(HttpResponse::body).get()
.forEach(System.out::println);
}
@SneakyThrows
@Test
public void testFlux(){
WebClient client = WebClient.create("https://api.openai.com/v1/completions");
ParameterizedTypeReference<ServerSentEvent<String>> type
= new ParameterizedTypeReference<>() {
};
CompletionRequest completionRequest = CompletionRequest.builder()
// .prompt("Human:" + chat.prompt +"\nAI:")
.prompt("给我推荐10本小说")
.model("text-davinci-001")
// .echo(true)
.stop(Arrays.asList(" Human:"," AI:"))
.maxTokens(1024)
.presencePenalty(0d)
.frequencyPenalty(0d)
.temperature(0.7D)
.bestOf(1)
.topP(1d)
.stream(true)
.build();
ObjectMapper mapper = new ObjectMapper();
mapper.configure(DeserializationFeature.FAIL_ON_UNKNOWN_PROPERTIES, false);
mapper.setSerializationInclusion(JsonInclude.Include.NON_NULL);
mapper.setPropertyNamingStrategy(PropertyNamingStrategy.SNAKE_CASE);
Flux<ServerSentEvent<String>> eventStream = client.post()
.accept(MediaType.APPLICATION_JSON)
.contentType(MediaType.APPLICATION_JSON)
.header("Authorization", "Bearer ")
.body(BodyInserters.fromValue(mapper.writeValueAsString(completionRequest)))
.retrieve()
.bodyToFlux(type);
eventStream.doOnError(x-> log.error("doOnError SSE:", x));
eventStream.subscribe(consumer
,
error -> log.error("Error receiving SSE:", error),
() -> log.info("Completed!!!"));
Thread.sleep(10*1000);
}
private Consumer<ServerSentEvent<String>> consumer = content -> log.info("Time: {} - event: name[{}], id [{}], content[{}] ",
LocalTime.now(), content.event(), content.id(), content.data());
@SneakyThrows
@Test
public void testModels() {
HttpClient client = HttpClient.newBuilder().build();
ObjectMapper mapper = new ObjectMapper();
mapper.configure(DeserializationFeature.FAIL_ON_UNKNOWN_PROPERTIES, false);
mapper.setSerializationInclusion(JsonInclude.Include.NON_NULL);
mapper.setPropertyNamingStrategy(PropertyNamingStrategy.SNAKE_CASE);
HttpRequest request = HttpRequest.newBuilder()
.header("Authorization", "Bearer " + this.token)
.header( "Content-Type", "application/json")
.GET()
.uri(URI.create("https://api.openai.com/v1/models"))
.build();
var response = client.sendAsync(request, HttpResponse.BodyHandlers.ofString())
.thenApply(HttpResponse::body).get();
log.info("response:{}", response);
}
@SneakyThrows
@Test
public void testChatGptModelHttp() {
HttpClient client = HttpClient.newBuilder().build();
List<ChatMessage> message = new ArrayList<>();
message.add(new ChatMessage("user","请给我推荐10本书"));
ChatCompletionRequest completionRequest = ChatCompletionRequest.builder()
// .prompt("Human:" + chat.prompt +"\nAI:")
.model("gpt-3.5-turbo")
.stream(true)
.messages(message).build();
ObjectMapper mapper = new ObjectMapper();
mapper.configure(DeserializationFeature.FAIL_ON_UNKNOWN_PROPERTIES, false);
mapper.setSerializationInclusion(JsonInclude.Include.NON_NULL);
mapper.setPropertyNamingStrategy(PropertyNamingStrategy.SNAKE_CASE);
var json = mapper.writeValueAsString(completionRequest);
log.info("json:{}", json);
HttpRequest request = HttpRequest.newBuilder()
.header("Authorization", "Bearer " + this.token)
.header( "Content-Type", "application/json")
.POST(HttpRequest.BodyPublishers.ofString(json))
.uri(URI.create("https://api.openai.com/v1/chat/completions"))
.build();
client.sendAsync(request, HttpResponse.BodyHandlers.ofLines())
.thenApply(HttpResponse::body).get()
.forEach(System.out::println);
}
@Builder
@Data
public static class ChatCompletionRequest{
private String model;
private Boolean stream;
private List<ChatMessage> messages;
}
@AllArgsConstructor
@Data
public static class ChatMessage{
private String role;
private String content;
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((1334, 1845), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1334, 1788), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1334, 1762), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1334, 1735), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1334, 1700), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1334, 1662), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1334, 1625), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1334, 1506), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((1334, 1463), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2122, 2153), 'java.net.http.HttpClient.newBuilder'), ((2202, 2718), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2202, 2693), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2202, 2663), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2202, 2637), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2202, 2610), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2202, 2575), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2202, 2537), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2202, 2500), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2202, 2467), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2202, 2382), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2202, 2339), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3135, 3449), 'java.net.http.HttpRequest.newBuilder'), ((3135, 3424), 'java.net.http.HttpRequest.newBuilder'), ((3135, 3350), 'java.net.http.HttpRequest.newBuilder'), ((3135, 3285), 'java.net.http.HttpRequest.newBuilder'), ((3135, 3224), 'java.net.http.HttpRequest.newBuilder'), ((3308, 3349), 'java.net.http.HttpRequest.BodyPublishers.ofString'), ((3486, 3521), 'java.net.http.HttpResponse.BodyHandlers.ofLines'), ((3950, 4466), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3950, 4441), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3950, 4411), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3950, 4385), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3950, 4358), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3950, 4323), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3950, 4285), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3950, 4248), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3950, 4215), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3950, 4130), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3950, 4087), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((5712, 5743), 'java.net.http.HttpClient.newBuilder'), ((6060, 6327), 'java.net.http.HttpRequest.newBuilder'), ((6060, 6302), 'java.net.http.HttpRequest.newBuilder'), ((6060, 6233), 'java.net.http.HttpRequest.newBuilder'), ((6060, 6210), 'java.net.http.HttpRequest.newBuilder'), ((6060, 6149), 'java.net.http.HttpRequest.newBuilder'), ((6379, 6415), 'java.net.http.HttpResponse.BodyHandlers.ofString'), ((6622, 6653), 'java.net.http.HttpClient.newBuilder'), ((7453, 7772), 'java.net.http.HttpRequest.newBuilder'), ((7453, 7747), 'java.net.http.HttpRequest.newBuilder'), ((7453, 7668), 'java.net.http.HttpRequest.newBuilder'), ((7453, 7603), 'java.net.http.HttpRequest.newBuilder'), ((7453, 7542), 'java.net.http.HttpRequest.newBuilder'), ((7626, 7667), 'java.net.http.HttpRequest.BodyPublishers.ofString'), ((7809, 7844), 'java.net.http.HttpResponse.BodyHandlers.ofLines')]
|
package br.com.alura.ecomart.chatbot.infra.openai;
import br.com.alura.ecomart.chatbot.domain.DadosCalculoFrete;
import br.com.alura.ecomart.chatbot.domain.service.CalculadorDeFrete;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.theokanning.openai.completion.chat.ChatFunction;
import com.theokanning.openai.completion.chat.ChatFunctionCall;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.messages.Message;
import com.theokanning.openai.messages.MessageRequest;
import com.theokanning.openai.runs.Run;
import com.theokanning.openai.runs.RunCreateRequest;
import com.theokanning.openai.runs.SubmitToolOutputRequestItem;
import com.theokanning.openai.runs.SubmitToolOutputsRequest;
import com.theokanning.openai.service.FunctionExecutor;
import com.theokanning.openai.service.OpenAiService;
import com.theokanning.openai.threads.ThreadRequest;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.stereotype.Component;
import java.time.Duration;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.Comparator;
import java.util.List;
import java.util.stream.Collectors;
@Component
public class OpenAIClient {
private final String apiKey;
private final String assistantId;
private String threadId;
private final OpenAiService service;
private final CalculadorDeFrete calculadorDeFrete;
public OpenAIClient(@Value("${app.openai.api.key}") String apiKey, @Value("${app.openai.assistant.id}") String assistantId, CalculadorDeFrete calculadorDeFrete) {
this.apiKey = apiKey;
this.service = new OpenAiService(apiKey, Duration.ofSeconds(60));
this.assistantId = assistantId;
this.calculadorDeFrete = calculadorDeFrete;
}
public String enviarRequisicaoChatCompletion(DadosRequisicaoChatCompletion dados) {
var messageRequest = MessageRequest
.builder()
.role(ChatMessageRole.USER.value())
.content(dados.promptUsuario())
.build();
if (this.threadId == null) {
var threadRequest = ThreadRequest
.builder()
.messages(Arrays.asList(messageRequest))
.build();
var thread = service.createThread(threadRequest);
this.threadId = thread.getId();
} else {
service.createMessage(this.threadId, messageRequest);
}
var runRequest = RunCreateRequest
.builder()
.assistantId(assistantId)
.build();
var run = service.createRun(threadId, runRequest);
var concluido = false;
var precisaChamarFuncao = false;
try {
while (!concluido && !precisaChamarFuncao) {
Thread.sleep(1000 * 10);
run = service.retrieveRun(threadId, run.getId());
concluido = run.getStatus().equalsIgnoreCase("completed");
precisaChamarFuncao = run.getRequiredAction() != null;
}
} catch (InterruptedException e) {
throw new RuntimeException(e);
}
if (precisaChamarFuncao) {
var precoDoFrete = chamarFuncao(run);
var submitRequest = SubmitToolOutputsRequest
.builder()
.toolOutputs(Arrays.asList(
new SubmitToolOutputRequestItem(
run
.getRequiredAction()
.getSubmitToolOutputs()
.getToolCalls()
.get(0)
.getId(),
precoDoFrete)
))
.build();
service.submitToolOutputs(threadId, run.getId(), submitRequest);
try {
while (!concluido) {
Thread.sleep(1000 * 10);
run = service.retrieveRun(threadId, run.getId());
concluido = run.getStatus().equalsIgnoreCase("completed");
}
} catch (InterruptedException e) {
throw new RuntimeException(e);
}
}
var mensagens = service.listMessages(threadId);
return mensagens
.getData()
.stream()
.sorted(Comparator.comparingInt(Message::getCreatedAt).reversed())
.findFirst().get().getContent().get(0).getText().getValue()
.replaceAll("\\\u3010.*?\\\u3011", "");
}
private String chamarFuncao(Run run) {
try {
var funcao = run.getRequiredAction().getSubmitToolOutputs().getToolCalls().get(0).getFunction();
var funcaoCalcularFrete = ChatFunction.builder()
.name("calcularFrete")
.executor(DadosCalculoFrete.class, d -> calculadorDeFrete.calcular(d))
.build();
var executorDeFuncoes = new FunctionExecutor(Arrays.asList(funcaoCalcularFrete));
var functionCall = new ChatFunctionCall(funcao.getName(), new ObjectMapper().readTree(funcao.getArguments()));
return executorDeFuncoes.execute(functionCall).toString();
} catch (Exception e) {
throw new RuntimeException(e);
}
}
public List<String> carregarHistoricoDeMensagens() {
var mensagens = new ArrayList<String>();
if (this.threadId != null) {
mensagens.addAll(
service
.listMessages(this.threadId)
.getData()
.stream()
.sorted(Comparator.comparingInt(Message::getCreatedAt))
.map(m -> m.getContent().get(0).getText().getValue())
.collect(Collectors.toList())
);
}
return mensagens;
}
public void apagarThread() {
if (this.threadId != null) {
service.deleteThread(this.threadId);
this.threadId = null;
}
}
}
|
[
"com.theokanning.openai.completion.chat.ChatFunction.builder",
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value"
] |
[((1972, 2000), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((4533, 4590), 'java.util.Comparator.comparingInt'), ((4935, 5120), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((4935, 5091), 'com.theokanning.openai.completion.chat.ChatFunction.builder'), ((4935, 5000), 'com.theokanning.openai.completion.chat.ChatFunction.builder')]
|
package learn.scraibe.controllers;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatCompletionResult;
import com.theokanning.openai.completion.chat.ChatMessage;
import com.theokanning.openai.completion.chat.ChatMessageRole;
import com.theokanning.openai.service.OpenAiService;
import learn.scraibe.models.Note;
import org.springframework.beans.factory.annotation.Value;
import org.springframework.http.HttpStatus;
import org.springframework.http.ResponseEntity;
import org.springframework.web.bind.annotation.PostMapping;
import org.springframework.web.bind.annotation.RequestBody;
import org.springframework.web.bind.annotation.RequestMapping;
import org.springframework.web.bind.annotation.RestController;
import java.time.Duration;
import java.util.ArrayList;
import java.util.List;
@RestController
@RequestMapping("/generate-completion")
public class OpenAIController {
@Value("${openai.api.key}")
private String openaiApiKey;
@PostMapping
public ResponseEntity<Object> generateCompletion(@RequestBody Note note) {
if(note.getContent() == null || note.getContent().isBlank()){
return new ResponseEntity<>("Cannot have blank notes", HttpStatus.BAD_REQUEST);
}
//create service that will route to OpenAI endpoint, provide key and timeout value incase openai takes a long time
OpenAiService service = new OpenAiService(openaiApiKey, Duration.ofSeconds(60));
//set up messages and Roles
List<ChatMessage> messages = new ArrayList<>();
ChatMessage userMessage = new ChatMessage(ChatMessageRole.USER.value(), "organize with bullet points, only respond with bullet points "+ note.getContent());
ChatMessage systemMessage = new ChatMessage(ChatMessageRole.ASSISTANT.value(), "you are a helpful assistant");
messages.add(userMessage);
messages.add((systemMessage));
// configure chatCompletionRequest object that will be sent over via the api
ChatCompletionRequest chatCompletionRequest = ChatCompletionRequest
.builder()
.model("gpt-3.5-turbo-0613")
.messages(messages)
.build();
//use service to make the request to OpenAI and then get the specific message to send back to the frontend.
ChatMessage responseMessage = service.createChatCompletion(chatCompletionRequest).getChoices().get(0).getMessage();
note.setContent(responseMessage.getContent());
return new ResponseEntity<>(note, HttpStatus.OK);
//TODO make a conditional statement based on the success of a response message,
//one previous error occurred because the request timed out(openai took too long to send back a request)
// but extending the duration seemed to solved the issue, just wondering what other issues to anticipate.
}
}
|
[
"com.theokanning.openai.completion.chat.ChatMessageRole.USER.value",
"com.theokanning.openai.completion.chat.ChatMessageRole.ASSISTANT.value"
] |
[((1638, 1666), 'com.theokanning.openai.completion.chat.ChatMessageRole.USER.value'), ((1805, 1838), 'com.theokanning.openai.completion.chat.ChatMessageRole.ASSISTANT.value')]
|
package com.theokanning.openai.service;
import com.theokanning.openai.moderation.Moderation;
import com.theokanning.openai.moderation.ModerationRequest;
import org.junit.jupiter.api.Test;
import static org.junit.jupiter.api.Assertions.assertTrue;
public class ModerationTest {
String token = System.getenv("OPENAI_TOKEN");
com.theokanning.openai.service.OpenAiService service = new OpenAiService(token);
@Test
void createModeration() {
ModerationRequest moderationRequest = ModerationRequest.builder()
.input("I want to kill them")
.model("text-moderation-latest")
.build();
Moderation moderationScore = service.createModeration(moderationRequest).getResults().get(0);
assertTrue(moderationScore.isFlagged());
}
}
|
[
"com.theokanning.openai.moderation.ModerationRequest.builder"
] |
[((504, 651), 'com.theokanning.openai.moderation.ModerationRequest.builder'), ((504, 626), 'com.theokanning.openai.moderation.ModerationRequest.builder'), ((504, 577), 'com.theokanning.openai.moderation.ModerationRequest.builder')]
|
package com.theokanning.openai.service;
import com.theokanning.openai.moderation.Moderation;
import com.theokanning.openai.moderation.ModerationRequest;
import org.junit.jupiter.api.Test;
import static org.junit.jupiter.api.Assertions.assertTrue;
public class ModerationTest {
String token = System.getenv("OPENAI_TOKEN");
com.theokanning.openai.service.OpenAiService service = new OpenAiService(token);
@Test
void createModeration() {
ModerationRequest moderationRequest = ModerationRequest.builder()
.input("I want to kill them")
.model("text-moderation-latest")
.build();
Moderation moderationScore = service.createModeration(moderationRequest).getResults().get(0);
assertTrue(moderationScore.isFlagged());
}
}
|
[
"com.theokanning.openai.moderation.ModerationRequest.builder"
] |
[((504, 651), 'com.theokanning.openai.moderation.ModerationRequest.builder'), ((504, 626), 'com.theokanning.openai.moderation.ModerationRequest.builder'), ((504, 577), 'com.theokanning.openai.moderation.ModerationRequest.builder')]
|
package com.theokanning.openai.service;
import com.fasterxml.jackson.annotation.JsonInclude;
import com.fasterxml.jackson.core.JsonProcessingException;
import com.fasterxml.jackson.core.type.TypeReference;
import com.fasterxml.jackson.databind.DeserializationFeature;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.fasterxml.jackson.databind.PropertyNamingStrategy;
import com.theokanning.openai.ListSearchParameters;
import com.theokanning.openai.OpenAiResponse;
import com.theokanning.openai.assistants.Assistant;
import com.theokanning.openai.assistants.AssistantFunction;
import com.theokanning.openai.assistants.AssistantRequest;
import com.theokanning.openai.assistants.AssistantToolsEnum;
import com.theokanning.openai.assistants.Tool;
import com.theokanning.openai.completion.chat.ChatCompletionRequest;
import com.theokanning.openai.completion.chat.ChatFunction;
import com.theokanning.openai.completion.chat.ChatFunctionCall;
import com.theokanning.openai.messages.Message;
import com.theokanning.openai.messages.MessageRequest;
import com.theokanning.openai.runs.RequiredAction;
import com.theokanning.openai.runs.Run;
import com.theokanning.openai.runs.RunCreateRequest;
import com.theokanning.openai.runs.RunStep;
import com.theokanning.openai.runs.SubmitToolOutputRequestItem;
import com.theokanning.openai.runs.SubmitToolOutputs;
import com.theokanning.openai.runs.SubmitToolOutputsRequest;
import com.theokanning.openai.runs.ToolCall;
import com.theokanning.openai.threads.Thread;
import com.theokanning.openai.threads.ThreadRequest;
import com.theokanning.openai.utils.TikTokensUtil;
import org.junit.jupiter.api.Test;
import java.time.Duration;
import java.util.ArrayList;
import java.util.List;
import java.util.Map;
import java.util.Objects;
import static org.junit.jupiter.api.Assertions.assertEquals;
import static org.junit.jupiter.api.Assertions.assertNotNull;
class AssistantFunctionTest {
String token = System.getenv("OPENAI_TOKEN");
OpenAiService service = new OpenAiService(token, Duration.ofMinutes(1));
@Test
void createRetrieveRun() throws JsonProcessingException {
ObjectMapper mapper = new ObjectMapper();
mapper.configure(DeserializationFeature.FAIL_ON_UNKNOWN_PROPERTIES, false);
mapper.setSerializationInclusion(JsonInclude.Include.NON_NULL);
mapper.setPropertyNamingStrategy(PropertyNamingStrategy.SNAKE_CASE);
mapper.addMixIn(ChatFunction.class, ChatFunctionMixIn.class);
mapper.addMixIn(ChatCompletionRequest.class, ChatCompletionRequestMixIn.class);
mapper.addMixIn(ChatFunctionCall.class, ChatFunctionCallMixIn.class);
String funcDef = "{\n" +
" \"type\": \"object\",\n" +
" \"properties\": {\n" +
" \"location\": {\n" +
" \"type\": \"string\",\n" +
" \"description\": \"The city and state, e.g. San Francisco, CA\"\n" +
" },\n" +
" \"unit\": {\n" +
" \"type\": \"string\",\n" +
" \"enum\": [\"celsius\", \"fahrenheit\"]\n" +
" }\n" +
" },\n" +
" \"required\": [\"location\"]\n" +
"}";
Map<String, Object> funcParameters = mapper.readValue(funcDef, new TypeReference<Map<String, Object>>() {});
AssistantFunction function = AssistantFunction.builder()
.name("weather_reporter")
.description("Get the current weather of a location")
.parameters(funcParameters)
.build();
List<Tool> toolList = new ArrayList<>();
Tool funcTool = new Tool(AssistantToolsEnum.FUNCTION, function);
toolList.add(funcTool);
AssistantRequest assistantRequest = AssistantRequest.builder()
.model(TikTokensUtil.ModelEnum.GPT_4_1106_preview.getName())
.name("MATH_TUTOR")
.instructions("You are a personal Math Tutor.")
.tools(toolList)
.build();
Assistant assistant = service.createAssistant(assistantRequest);
ThreadRequest threadRequest = ThreadRequest.builder()
.build();
Thread thread = service.createThread(threadRequest);
MessageRequest messageRequest = MessageRequest.builder()
.content("What's the weather of Xiamen?")
.build();
Message message = service.createMessage(thread.getId(), messageRequest);
RunCreateRequest runCreateRequest = RunCreateRequest.builder()
.assistantId(assistant.getId())
.build();
Run run = service.createRun(thread.getId(), runCreateRequest);
assertNotNull(run);
Run retrievedRun = service.retrieveRun(thread.getId(), run.getId());
while (!(retrievedRun.getStatus().equals("completed"))
&& !(retrievedRun.getStatus().equals("failed"))
&& !(retrievedRun.getStatus().equals("requires_action"))){
retrievedRun = service.retrieveRun(thread.getId(), run.getId());
}
if (retrievedRun.getStatus().equals("requires_action")) {
RequiredAction requiredAction = retrievedRun.getRequiredAction();
System.out.println("requiredAction");
System.out.println(mapper.writeValueAsString(requiredAction));
List<ToolCall> toolCalls = requiredAction.getSubmitToolOutputs().getToolCalls();
ToolCall toolCall = toolCalls.get(0);
String toolCallId = toolCall.getId();
SubmitToolOutputRequestItem toolOutputRequestItem = SubmitToolOutputRequestItem.builder()
.toolCallId(toolCallId)
.output("sunny")
.build();
List<SubmitToolOutputRequestItem> toolOutputRequestItems = new ArrayList<>();
toolOutputRequestItems.add(toolOutputRequestItem);
SubmitToolOutputsRequest submitToolOutputsRequest = SubmitToolOutputsRequest.builder()
.toolOutputs(toolOutputRequestItems)
.build();
retrievedRun = service.submitToolOutputs(retrievedRun.getThreadId(), retrievedRun.getId(), submitToolOutputsRequest);
while (!(retrievedRun.getStatus().equals("completed"))
&& !(retrievedRun.getStatus().equals("failed"))
&& !(retrievedRun.getStatus().equals("requires_action"))){
retrievedRun = service.retrieveRun(thread.getId(), run.getId());
}
OpenAiResponse<Message> response = service.listMessages(thread.getId());
List<Message> messages = response.getData();
System.out.println(mapper.writeValueAsString(messages));
}
}
}
|
[
"com.theokanning.openai.utils.TikTokensUtil.ModelEnum.GPT_4_1106_preview.getName",
"com.theokanning.openai.assistants.AssistantRequest.builder",
"com.theokanning.openai.messages.MessageRequest.builder",
"com.theokanning.openai.assistants.AssistantFunction.builder",
"com.theokanning.openai.runs.SubmitToolOutputsRequest.builder",
"com.theokanning.openai.runs.SubmitToolOutputRequestItem.builder",
"com.theokanning.openai.threads.ThreadRequest.builder",
"com.theokanning.openai.runs.RunCreateRequest.builder"
] |
[((3437, 3645), 'com.theokanning.openai.assistants.AssistantFunction.builder'), ((3437, 3620), 'com.theokanning.openai.assistants.AssistantFunction.builder'), ((3437, 3576), 'com.theokanning.openai.assistants.AssistantFunction.builder'), ((3437, 3506), 'com.theokanning.openai.assistants.AssistantFunction.builder'), ((3864, 4125), 'com.theokanning.openai.assistants.AssistantRequest.builder'), ((3864, 4100), 'com.theokanning.openai.assistants.AssistantRequest.builder'), ((3864, 4067), 'com.theokanning.openai.assistants.AssistantRequest.builder'), ((3864, 4003), 'com.theokanning.openai.assistants.AssistantRequest.builder'), ((3864, 3967), 'com.theokanning.openai.assistants.AssistantRequest.builder'), ((3914, 3966), 'com.theokanning.openai.utils.TikTokensUtil.ModelEnum.GPT_4_1106_preview.getName'), ((4239, 4287), 'com.theokanning.openai.threads.ThreadRequest.builder'), ((4391, 4498), 'com.theokanning.openai.messages.MessageRequest.builder'), ((4391, 4473), 'com.theokanning.openai.messages.MessageRequest.builder'), ((4627, 4726), 'com.theokanning.openai.runs.RunCreateRequest.builder'), ((4627, 4701), 'com.theokanning.openai.runs.RunCreateRequest.builder'), ((5724, 5871), 'com.theokanning.openai.runs.SubmitToolOutputRequestItem.builder'), ((5724, 5842), 'com.theokanning.openai.runs.SubmitToolOutputRequestItem.builder'), ((5724, 5805), 'com.theokanning.openai.runs.SubmitToolOutputRequestItem.builder'), ((6090, 6210), 'com.theokanning.openai.runs.SubmitToolOutputsRequest.builder'), ((6090, 6181), 'com.theokanning.openai.runs.SubmitToolOutputsRequest.builder')]
|
package com.theokanning.openai.service;
import com.theokanning.openai.audio.CreateSpeechRequest;
import com.theokanning.openai.audio.CreateTranscriptionRequest;
import com.theokanning.openai.audio.CreateTranslationRequest;
import com.theokanning.openai.audio.TranscriptionResult;
import com.theokanning.openai.audio.TranslationResult;
import org.junit.jupiter.api.Test;
import java.io.IOException;
import java.time.Duration;
import okhttp3.MediaType;
import okhttp3.ResponseBody;
import static org.junit.jupiter.api.Assertions.*;
public class AudioTest {
static String englishAudioFilePath = "src/test/resources/hello-world.mp3";
static String koreanAudioFilePath = "src/test/resources/korean-hello.mp3";
String token = System.getenv("OPENAI_TOKEN");
OpenAiService service = new OpenAiService(token, Duration.ofSeconds(30));
@Test
void createTranscription() {
CreateTranscriptionRequest createTranscriptionRequest = CreateTranscriptionRequest.builder()
.model("whisper-1")
.build();
String text = service.createTranscription(createTranscriptionRequest, englishAudioFilePath).getText();
assertEquals("Hello World.", text);
}
@Test
void createTranscriptionVerbose() {
CreateTranscriptionRequest createTranscriptionRequest = CreateTranscriptionRequest.builder()
.model("whisper-1")
.responseFormat("verbose_json")
.build();
TranscriptionResult result = service.createTranscription(createTranscriptionRequest, englishAudioFilePath);
assertEquals("Hello World.", result.getText());
assertEquals("transcribe", result.getTask());
assertEquals("english", result.getLanguage());
assertTrue(result.getDuration() > 0);
assertEquals(1, result.getSegments().size());
}
@Test
void createTranslation() {
CreateTranslationRequest createTranslationRequest = CreateTranslationRequest.builder()
.model("whisper-1")
.build();
String text = service.createTranslation(createTranslationRequest, koreanAudioFilePath).getText();
assertEquals("Hello, my name is Yoona. I am a Korean native speaker.", text);
}
@Test
void createTranslationVerbose() {
CreateTranslationRequest createTranslationRequest = CreateTranslationRequest.builder()
.model("whisper-1")
.responseFormat("verbose_json")
.build();
TranslationResult result = service.createTranslation(createTranslationRequest, koreanAudioFilePath);
assertEquals("Hello, my name is Yoona. I am a Korean native speaker.", result.getText());
assertEquals("translate", result.getTask());
assertEquals("english", result.getLanguage());
assertTrue(result.getDuration() > 0);
assertEquals(1, result.getSegments().size());
}
@Test
void createSpeech() throws IOException {
CreateSpeechRequest createSpeechRequest = CreateSpeechRequest.builder()
.model("tts-1")
.input("Hello World.")
.voice("alloy")
.build();
final ResponseBody speech = service.createSpeech(createSpeechRequest);
assertNotNull(speech);
assertEquals(MediaType.get("audio/mpeg"), speech.contentType());
assertTrue(speech.bytes().length > 0);
}
}
|
[
"com.theokanning.openai.audio.CreateTranslationRequest.builder",
"com.theokanning.openai.audio.CreateSpeechRequest.builder",
"com.theokanning.openai.audio.CreateTranscriptionRequest.builder"
] |
[((958, 1055), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((958, 1030), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((1334, 1479), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((1334, 1454), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((1334, 1406), 'com.theokanning.openai.audio.CreateTranscriptionRequest.builder'), ((1971, 2066), 'com.theokanning.openai.audio.CreateTranslationRequest.builder'), ((1971, 2041), 'com.theokanning.openai.audio.CreateTranslationRequest.builder'), ((2376, 2519), 'com.theokanning.openai.audio.CreateTranslationRequest.builder'), ((2376, 2494), 'com.theokanning.openai.audio.CreateTranslationRequest.builder'), ((2376, 2446), 'com.theokanning.openai.audio.CreateTranslationRequest.builder'), ((3049, 3206), 'com.theokanning.openai.audio.CreateSpeechRequest.builder'), ((3049, 3181), 'com.theokanning.openai.audio.CreateSpeechRequest.builder'), ((3049, 3149), 'com.theokanning.openai.audio.CreateSpeechRequest.builder'), ((3049, 3110), 'com.theokanning.openai.audio.CreateSpeechRequest.builder')]
|
package com.odde.doughnut.services.ai;
import static com.odde.doughnut.services.ai.builder.OpenAIChatRequestBuilder.askClarificationQuestion;
import static com.odde.doughnut.services.ai.tools.AiToolFactory.COMPLETE_NOTE_DETAILS;
import com.odde.doughnut.controllers.dto.*;
import com.odde.doughnut.entities.Note;
import com.odde.doughnut.services.ai.builder.OpenAIChatRequestBuilder;
import com.odde.doughnut.services.ai.tools.AiTool;
import com.odde.doughnut.services.openAiApis.OpenAiApiHandler;
import com.theokanning.openai.assistants.*;
import com.theokanning.openai.messages.MessageRequest;
import com.theokanning.openai.runs.RequiredAction;
import com.theokanning.openai.runs.Run;
import com.theokanning.openai.runs.ToolCall;
import com.theokanning.openai.threads.Thread;
import com.theokanning.openai.threads.ThreadRequest;
import java.util.stream.Stream;
public record ContentCompletionService(OpenAiApiHandler openAiApiHandler) {
public AiCompletionResponse getAiCompletion(
AiCompletionParams aiCompletionParams, Note note, String assistantId) {
String threadId = createThread(aiCompletionParams, note);
Run run = openAiApiHandler.createRun(threadId, assistantId);
return getThreadResponse(threadId, run);
}
public AiCompletionResponse answerAiCompletionClarifyingQuestion(
AiCompletionAnswerClarifyingQuestionParams answerClarifyingQuestionParams) {
String threadId = answerClarifyingQuestionParams.getThreadId();
Run retrievedRun = openAiApiHandler.submitToolOutputs(answerClarifyingQuestionParams);
return getThreadResponse(threadId, retrievedRun);
}
private String createThread(AiCompletionParams aiCompletionParams, Note note) {
ThreadRequest threadRequest = ThreadRequest.builder().build();
Thread thread = openAiApiHandler.createThread(threadRequest);
MessageRequest messageRequest =
MessageRequest.builder()
.content(
note.getNoteDescription()
+ "------------\n"
+ aiCompletionParams.getCompletionPrompt())
.build();
openAiApiHandler.createMessage(thread.getId(), messageRequest);
return thread.getId();
}
private AiCompletionResponse getThreadResponse(String threadId, Run currentRun) {
Run run = openAiApiHandler.retrieveUntilCompletedOrRequiresAction(threadId, currentRun);
AiCompletionResponse completionResponse = new AiCompletionResponse();
completionResponse.setThreadId(threadId);
completionResponse.setRunId(currentRun.getId());
if (run.getStatus().equals("requires_action")) {
RequiredAction requiredAction = run.getRequiredAction();
int size = requiredAction.getSubmitToolOutputs().getToolCalls().size();
if (size != 1) {
throw new RuntimeException("Unexpected number of tool calls: " + size);
}
ToolCall toolCall = requiredAction.getSubmitToolOutputs().getToolCalls().get(0);
AiCompletionRequiredAction actionRequired =
getTools()
.flatMap(t -> t.tryConsume(toolCall))
.findFirst()
.orElseThrow(
() ->
new RuntimeException(
"Unknown function name: " + toolCall.getFunction().getName()));
actionRequired.setToolCallId(toolCall.getId());
completionResponse.setRequiredAction(actionRequired);
} else {
String message =
openAiApiHandler
.getThreadLastMessage(threadId)
.getContent()
.getFirst()
.getText()
.getValue();
completionResponse.setLastMessage(message);
}
return completionResponse;
}
public Assistant createNoteCompletionAssistant(String modelName) {
AssistantRequest assistantRequest =
AssistantRequest.builder()
.model(modelName)
.name("Note details completion")
.instructions(OpenAIChatRequestBuilder.systemInstruction)
.tools(getTools().map(AiTool::getTool).toList())
.build();
return openAiApiHandler.createAssistant(assistantRequest);
}
private static Stream<AiTool> getTools() {
return Stream.of(
AiTool.build(
COMPLETE_NOTE_DETAILS,
"Text completion for the details of the note of focus",
NoteDetailsCompletion.class,
(noteDetailsCompletion) -> {
AiCompletionRequiredAction result = new AiCompletionRequiredAction();
result.setContentToAppend(noteDetailsCompletion.completion);
return result;
}),
AiTool.build(
askClarificationQuestion,
"Ask question to get more context",
ClarifyingQuestion.class,
(clarifyingQuestion) -> {
AiCompletionRequiredAction result = new AiCompletionRequiredAction();
result.setClarifyingQuestion(clarifyingQuestion);
return result;
}));
}
}
|
[
"com.theokanning.openai.threads.ThreadRequest.builder",
"com.theokanning.openai.messages.MessageRequest.builder"
] |
[((1732, 1763), 'com.theokanning.openai.threads.ThreadRequest.builder'), ((1875, 2087), 'com.theokanning.openai.messages.MessageRequest.builder'), ((1875, 2066), 'com.theokanning.openai.messages.MessageRequest.builder')]
|
package cn.shu.wechat.utils;
import cn.shu.wechat.configuration.OpenAIConfiguration;
import cn.shu.wechat.entity.Message;
import com.fasterxml.jackson.annotation.JsonInclude;
import com.fasterxml.jackson.databind.DeserializationFeature;
import com.fasterxml.jackson.databind.ObjectMapper;
import com.fasterxml.jackson.databind.PropertyNamingStrategy;
import com.theokanning.openai.OpenAiApi;
import com.theokanning.openai.OpenAiService;
import com.theokanning.openai.completion.CompletionRequest;
import com.theokanning.openai.completion.CompletionResult;
import okhttp3.*;
import retrofit2.Retrofit;
import retrofit2.adapter.rxjava2.RxJava2CallAdapterFactory;
import retrofit2.converter.jackson.JacksonConverterFactory;
import java.io.IOException;
import java.util.List;
import java.util.concurrent.TimeUnit;
import java.util.stream.Collectors;
import java.util.stream.Stream;
import static java.time.Duration.ofSeconds;
public class OpenAPIUtil {
private static final String BASE_URL = "https://api.openai.com/";
public static List<Message> chat(String q) {
ObjectMapper mapper = new ObjectMapper();
mapper.configure(DeserializationFeature.FAIL_ON_UNKNOWN_PROPERTIES, false);
mapper.setSerializationInclusion(JsonInclude.Include.NON_NULL);
mapper.setPropertyNamingStrategy(PropertyNamingStrategy.SNAKE_CASE);
OkHttpClient client = new OkHttpClient.Builder()
.addInterceptor(new Interceptor() {
@Override
public Response intercept(Chain chain) throws IOException {
Request request = chain.request()
.newBuilder()
.header("Authorization", "Bearer " + OpenAIConfiguration.getInstance().getOpenaiKey())
.build();
return chain.proceed(request);
}
})
.sslSocketFactory(TestSSLSocketClient.getSSLSocketFactory(), TestSSLSocketClient.getX509TrustManager())
.hostnameVerifier(TestSSLSocketClient.getHostnameVerifier())
.connectionPool(new ConnectionPool(5, 1, TimeUnit.SECONDS))
.readTimeout(ofSeconds(OpenAIConfiguration.getInstance().getExpire()).toMillis(), TimeUnit.MILLISECONDS)
.build();
Retrofit retrofit = new Retrofit.Builder()
.baseUrl(BASE_URL)
.client(client)
.addConverterFactory(JacksonConverterFactory.create(mapper))
.addCallAdapterFactory(RxJava2CallAdapterFactory.create())
.build();
OpenAiService service = new OpenAiService(retrofit.create(OpenAiApi.class));
CompletionRequest completionRequest = CompletionRequest.builder()
.prompt(q)
.maxTokens(1024)
.model("text-davinci-003")
.echo(true)
.build();
CompletionResult completion = service.createCompletion(completionRequest);
Stream<Message> messageStream = completion.getChoices().stream()
.map(e -> {
return Message.builder().content(e.getText().substring(e.getText().indexOf("\n\n") + 2)).build();
});
return messageStream.collect(Collectors.toList());
}
}
|
[
"com.theokanning.openai.completion.CompletionRequest.builder"
] |
[((1749, 1797), 'cn.shu.wechat.configuration.OpenAIConfiguration.getInstance'), ((2249, 2294), 'cn.shu.wechat.configuration.OpenAIConfiguration.getInstance'), ((2788, 2971), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2788, 2946), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2788, 2918), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2788, 2875), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((2788, 2842), 'com.theokanning.openai.completion.CompletionRequest.builder'), ((3184, 3273), 'cn.shu.wechat.entity.Message.builder'), ((3184, 3265), 'cn.shu.wechat.entity.Message.builder')]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.