comicocr / modules /translators /trans_sakura.py
fasdfsa's picture
init
f6f8d06
# 同步更新自manga-image-translator
import logging
import re
import time
from typing import List, Dict, Union, Callable
import time
import os
import json
import openai
from .base import BaseTranslator, register_translator
OPENAPI_V1_API = int(openai.__version__.split('.')[0]) >= 1
class InvalidNumTranslations(Exception):
pass
class SakuraDict():
"""
Sakura字典类,用于加载和管理Sakura字典。
属性:
--------
logger : logging.Logger
日志记录器对象
dict_str : str
字典内容字符串
version : str
Sakura字典版本号
path : str
字典文件路径
方法:
--------
__init__(self, path: str, logger: logging.Logger, version: str = "0.9") -> None:
初始化Sakura字典对象。
load_dict(self, dic_path: str) -> None:
根据字典类型加载字典。
get_dict_str(self) -> str:
获取字典内容字符串。
save_dict_to_file(self, dic_path: str, dict_type: str = "sakura") -> None:
将字典内容保存到文件。
"""
def __init__(self, path: str, logger: logging.Logger, version: str = "0.9") -> None:
"""
初始化Sakura字典对象。
参数:
--------
path : str
字典文件路径
logger : logging.Logger
日志记录器对象
version : str, optional
Sakura字典版本号,默认为"0.9"
"""
self.logger = logger
self.dict_str = ""
self.version = version
self.path = path
if not path:
return # 如果路径为空,直接返回,不加载字典
if not os.path.exists(path):
if self.version != "0.9":
self.logger.info(f"字典文件不存在: {path}\n 如果您不需要字典功能,请忽略此警告。")
return
if self.version == "1.0":
try:
self.load_dict(path)
except Exception as e:
self.logger.warning(f"载入字典失败: {e}")
elif self.version == "0.9":
pass
else:
self.logger.info("您当前选择了Sakura 0.9版本,暂不支持术语表")
def load_dict(self, dic_path: str) -> None:
"""
根据字典类型加载字典。
参数:
--------
dic_path : str
字典文件路径
"""
if self.version == "0.9" or not dic_path:
return
dic_type = self._detect_type(dic_path)
if dic_type == "galtransl":
self._load_galtransl_dic(dic_path)
elif dic_type == "sakura":
self._load_sakura_dict(dic_path)
elif dic_type == "json":
self._load_json_dict(dic_path)
else:
self.logger.warning(f"未知的字典类型: {dic_path}")
self.logger.debug(f"字典内容(转换后): {self.dict_str[:100]}")
def _load_galtransl_dic(self, dic_path: str) -> None:
"""
加载Galtransl格式的字典。
参数:
--------
dic_path : str
字典文件路径
"""
if self.version == "0.9":
return
with open(dic_path, encoding="utf8") as f:
dic_lines = f.readlines()
if not dic_lines:
return
dic_name = os.path.basename(dic_path)
gpt_dict = []
for line in dic_lines:
if line.startswith(("\n", "\\\\", "//")):
continue
line = line.replace(" ", "\t")
sp = line.rstrip("\r\n").split("\t")
if len(sp) < 2:
continue
src, dst, *info = sp
gpt_dict.append(
{"src": src, "dst": dst, "info": info[0] if info else None})
gpt_dict_text_list = [
f"{gpt['src']}->{gpt['dst']}{' #' + gpt['info'] if gpt['info'] else ''}" for gpt in gpt_dict]
self.dict_str = "\n".join(gpt_dict_text_list)
self.logger.info(f"载入 Galtransl 字典: {dic_name} {len(gpt_dict)}普通词条")
def _load_sakura_dict(self, dic_path: str) -> None:
"""
加载Sakura格式的字典。
参数:
--------
dic_path : str
字典文件路径
"""
if self.version == "0.9":
return
with open(dic_path, encoding="utf8") as f:
dic_lines = f.readlines()
if not dic_lines:
return
dic_name = os.path.basename(dic_path)
gpt_dict_text_list = []
for line in dic_lines:
if line.startswith(("\n", "\\\\", "//")):
continue
sp = line.rstrip("\r\n").split("->")
if len(sp) < 2:
continue
src, dst_info = sp
dst_info_sp = dst_info.split("#")
dst = dst_info_sp[0].strip()
info = dst_info_sp[1].strip() if len(dst_info_sp) > 1 else None
gpt_dict_text_list.append(
f"{src}->{dst}{' #' + info if info else ''}")
self.dict_str = "\n".join(gpt_dict_text_list)
self.logger.info(
f"载入标准Sakura字典: {dic_name} {len(gpt_dict_text_list)}普通词条")
def _load_json_dict(self, dic_path: str) -> None:
"""
加载JSON格式的字典。
参数:
--------
dic_path : str
字典文件路径
"""
if self.version == "0.9":
return
with open(dic_path, encoding="utf8") as f:
dic_json = json.load(f)
if not dic_json:
return
dic_name = os.path.basename(dic_path)
gpt_dict_text_list = []
for item in dic_json:
if not item:
continue
src = item.get("src", "")
dst = item.get("dst", "")
info = item.get("info", "")
gpt_dict_text_list.append(
f"{src}->{dst}{' #' + info if info else ''}")
self.dict_str = "\n".join(gpt_dict_text_list)
self.logger.info(f"载入JSON字典: {dic_name} {len(gpt_dict_text_list)}条记录")
def _detect_type(self, dic_path: str) -> str:
"""
检测字典文件的类型。
参数:
--------
dic_path : str
字典文件路径
返回:
--------
str
字典类型,可能的值有"galtransl"、"sakura"、"json"和"unknown"
"""
if self.version == "0.9":
return "unknown"
with open(dic_path, encoding="utf8") as f:
dic_lines = f.readlines()
self.logger.debug(f"检测字典类型: {dic_path}")
if not dic_lines:
return "unknown"
if dic_path.endswith(".json"):
return "json"
for line in dic_lines:
if line.startswith(("\n", "\\\\", "//")):
continue
if "\t" in line or " " in line:
return "galtransl"
elif "->" in line:
return "sakura"
return "unknown"
def get_dict_str(self) -> str:
"""
获取字典内容字符串。
返回:
--------
str
字典内容字符串
"""
if self.version == "0.9" or not self.path:
return ""
if not self.dict_str:
try:
self.load_dict(self.path)
except Exception as e:
self.logger.warning(f"载入字典失败: {e}")
return self.dict_str
def get_dict_str_within_text(self, text: str, force_apply_dict: bool = False) -> str:
"""
获取字典内容字符串,仅保留字典中出现的词条。
参数:
--------
text : str
待翻译文本
返回:
--------
str
字典内容字符串
"""
if force_apply_dict:
return self.get_dict_str()
if self.version == "0.9" or not self.path:
return ""
if not self.dict_str:
try:
self.load_dict(self.path)
except Exception as e:
self.logger.warning(f"载入字典失败: {e}")
return ""
# 初始化一个空列表用于存储匹配的字典行
matched_dict_lines = []
# 遍历字典中的每一行
for line in self.dict_str.splitlines():
if '->' in line:
src = line.split('->')[0]
# 检查 src 是否在输入文本中
# self.logger.debug(f"检查字典原文{src}是否在文本{text}中")
if src in text:
# self.logger.debug(f"匹配到字典行: {line}")
matched_dict_lines.append(line)
# 将匹配的字典行拼接成一个字符串并返回
return '\n'.join(matched_dict_lines)
def dict_to_json(self) -> str:
"""
将字典内容转换为JSON格式。
返回:
--------
str
字典内容的JSON格式字符串
"""
if self.version == "0.9" or not self.path:
return ""
if not self.dict_str:
try:
self.load_dict(self.path)
except Exception as e:
self.logger.warning(f"载入字典失败: {e}")
dict_json = []
for line in self.dict_str.split("\n"):
if not line:
continue
sp = line.split("->")
if len(sp) < 2:
continue
src, dst_info = sp
dst_info_sp = dst_info.split("#")
dst = dst_info_sp[0].strip()
info = dst_info_sp[1].strip() if len(dst_info_sp) > 1 else None
dict_json.append({"src": src, "dst": dst, "info": info})
return json.dumps(dict_json, ensure_ascii=False, indent=4)
def save_dict_to_file(self, dic_path: str, dict_type: str = "sakura") -> None:
"""
将字典内容保存到文件。
参数:
--------
dic_path : str
字典文件保存路径
dict_type : str, optional
字典类型,可选值有"sakura"、"galtransl"和"json",默认为"sakura"
"""
if self.version == "0.9" or not self.path:
return
if dict_type == "sakura":
with open(dic_path, "w", encoding="utf8") as f:
f.write(self.dict_str)
elif dict_type == "galtransl":
with open(dic_path, "w", encoding="utf8") as f:
f.write(self.dict_str.replace(
"->", " ").replace(" #", " "))
elif dict_type == "json":
json_data = self.dict_to_json()
with open(dic_path, "w", encoding="utf8") as f:
json.dump(json_data, f, ensure_ascii=False, indent=4)
else:
self.logger.warning(f"未知的字典类型: {dict_type}")
@register_translator('Sakura')
class SakuraTranslator(BaseTranslator):
concate_text = False
cht_require_convert = True
params: Dict = {
'low vram mode': {
'value': True,
'description': 'check it if you\'re running it locally on a single device and encountered a crash due to vram OOM',
'type': 'checkbox',
},
'api baseurl': 'http://127.0.0.1:8080/v1',
'dict path': '',
'version': {
'type': 'selector',
'options': [
'0.9',
'1.0',
'galtransl-v1'
],
'value': '0.9'
},
'retry attempts': 3,
'timeout': 999,
'max tokens': 1024,
'repeat detect threshold': 20,
'force apply dict': {
'value': False,
'description': 'Force apply the dictionary regardless of whether the terms appear in the original text \n DO NOT CHECK THIS IF YOU ARE NOT SURE WHAT IT MEANS',
'type': 'checkbox',
},
'do enlarge small kana': {
'value': False,
'description': 'Enlarge small kana to normal size',
'type': 'checkbox',
}
}
_CHAT_SYSTEM_TEMPLATE_009 = (
'你是一个轻小说翻译模型,可以流畅通顺地以日本轻小说的风格将日文翻译成简体中文,并联系上下文正确使用人称代词,不擅自添加原文中没有的代词。'
)
_CHAT_SYSTEM_TEMPLATE_100 = (
'你是一个轻小说翻译模型,可以流畅通顺地以日本轻小说的风格将日文翻译成简体中文,并联系上下文正确使用人称代词,不擅自添加原文中没有的代词。'
)
_CHAT_SYSTEM_TEMPLATE_GALTRANSL_V1 = (
'你是一个视觉小说翻译模型,可以通顺地使用给定的术语表以指定的风格将日文翻译成简体中文,并联系上下文正确使用人称代词,注意不要混淆使役态和被动态的主语和宾语,不要擅自添加原文中没有的代词,也不要擅自增加或减少换行。'
)
@property
def timeout(self) -> int:
return self.params['timeout']
@property
def retry_attempts(self) -> int:
return self.params['retry attempts']
@property
def repeat_detect_threshold(self) -> int:
return self.params['repeat detect threshold']
@property
def max_tokens(self) -> int:
return self.params['max tokens']
@property
def api_base_raw(self) -> str:
return self.params['api baseurl']
@property
def api_base(self) -> str:
url = self.api_base_raw
if url.endswith('/'):
url = url[:-1]
if not url.endswith('/v1'):
url += '/v1'
return url
@property
def sakura_version(self) -> str:
return self.params['version']['value']
@property
def dict_path(self) -> str:
return self.params['dict path']
@property
def force_apply_dict(self) -> bool:
return self.params['force apply dict']['value']
@property
def do_enlarge_small_kana(self) -> bool:
return self.params['do enlarge small kana']['value']
def _setup_translator(self):
self.lang_map['简体中文'] = 'Simplified Chinese'
self.lang_map['日本語'] = 'Japanese'
self.temperature = 0.1
self.top_p = 0.3
self.frequency_penalty = 0.05
self._current_style = "precise"
self._emoji_pattern = re.compile(r'[\U00010000-\U0010ffff]')
self._heart_pattern = re.compile(r'❤')
sakura_version = self.sakura_version if self.sakura_version!= 'galtransl-v1' else '1.0'
self.sakura_dict = SakuraDict(
self.dict_path, self.logger, sakura_version)
self.logger.info(f'当前选择的Sakura版本: {self.sakura_version}')
def updateParam(self, param_key: str, param_content):
super().updateParam(param_key, param_content)
if param_key == 'dict path' or param_key == 'version':
self.set_dict_path(self.params['dict path'])
def set_dict_path(self, path: str):
self.params['dict path'] = path
self.sakura_dict = SakuraDict(path, self.logger, self.sakura_version)
self.logger.debug(f'更新Sakura字典路径为: {path}')
@staticmethod
def enlarge_small_kana(text, ignore=''):
"""将小写平假名或片假名转换为普通大小
参数
----------
text : str
全角平假名或片假名字符串。
ignore : str, 可选
转换时要忽略的字符。
返回
------
str
平假名或片假名字符串,小写假名已转换为大写
示例
--------
>>> print(enlarge_small_kana('さくらきょうこ'))
さくらきようこ
>>> print(enlarge_small_kana('キュゥべえ'))
キユウべえ
"""
SMALL_KANA = list('ぁぃぅぇぉゃゅょっァィゥェォヵヶャュョッ')
SMALL_KANA_NORMALIZED = list('あいうえおやゆよつアイウエオカケヤユヨツ')
SMALL_KANA2BIG_KANA = dict(
zip(map(ord, SMALL_KANA), SMALL_KANA_NORMALIZED))
def _exclude_ignorechar(ignore, conv_map):
for character in map(ord, ignore):
del conv_map[character]
return conv_map
def _convert(text, conv_map):
return text.translate(conv_map)
def _translate(text, ignore, conv_map):
if ignore:
_conv_map = _exclude_ignorechar(ignore, conv_map.copy())
return _convert(text, _conv_map)
return _convert(text, conv_map)
return _translate(text, ignore, SMALL_KANA2BIG_KANA)
def detect_and_calculate_repeats(self, s: str, threshold: int = 20, remove_all=True) -> Union[bool, str, int, str, int]:
"""
检测文本中是否存在重复模式,并计算重复次数。
返回值: (是否重复, 去除重复后的文本, 重复次数, 重复模式, 实际阈值)
"""
# 初始化标记重复模式的变量
repeated = False
longest_pattern = '' # 存储最长的重复模式
longest_count = 0 # 存储最长模式的重复次数
counts = [] # 存储所有找到的重复次数
# 遍历所有可能的模式长度,从1到字符串长度的一半
for pattern_length in range(1, len(s) // 2 + 1):
# 构建正则表达式模式,匹配指定长度的重复模式
pattern = re.compile(r'(.{%d})\1+' % pattern_length)
# 查找所有匹配的重复模式
for match in re.finditer(pattern, s):
current_pattern = match.group(1) # 当前找到的重复模式
current_count = len(match.group(0)) // pattern_length # 计算重复次数
counts.append(current_count) # 将当前模式的重复次数添加到 counts 列表
# 如果当前模式的重复次数达到或超过阈值
if current_count >= threshold:
self.logger.warning(f"检测到重复模式: {current_pattern},重复次数: {current_count}")
repeated = True # 标记检测到重复模式
# 如果当前模式的重复次数大于最长的重复次数
if current_count > longest_count:
longest_count = current_count # 更新最长的重复次数
longest_pattern = current_pattern # 更新最长的重复模式
# 如果需要移除所有重复模式
if remove_all:
s = s[:match.start()] + s[match.end():] # 从字符串中移除重复模式
break # 跳出当前循环,检查下一个模式长度
if repeated:
break # 如果已经检测到重复模式,跳出外层循环
# 计算实际阈值,取默认阈值和所有找到的重复次数的最大众数中的最大值
actual_threshold = max(threshold, max(counts, default=0))
# 返回检测结果,包括是否重复、去除重复后的文本、重复次数、重复模式和实际阈值
return repeated, s, longest_count, longest_pattern, actual_threshold
def _format_prompt_log(self, prompt: str) -> str:
gpt_dict_raw_text = self.sakura_dict.get_dict_str_within_text(prompt, self.force_apply_dict)
prompt_009 = '\n'.join([
'System:',
self._CHAT_SYSTEM_TEMPLATE_009,
'User:',
'将下面的日文文本翻译成中文:',
prompt,
])
prompt_100 = '\n'.join([
'System:',
self._CHAT_SYSTEM_TEMPLATE_100,
'User:',
"根据以下术语表(可以为空):",
gpt_dict_raw_text,
"将下面的日文文本根据对应关系和备注翻译成中文:",
prompt,
])
prompt_galtransl_v1 = '\n'.join([
'System:',
self._CHAT_SYSTEM_TEMPLATE_GALTRANSL_V1,
'User:',
"根据以下术语表:",
gpt_dict_raw_text,
"将下面的日文文本根据上述术语表的对应关系和注释翻译成中文:",
prompt,
])
if self.sakura_version == '0.9':
return prompt_009
elif self.sakura_version == '1.0':
return prompt_100
else:
return prompt_galtransl_v1
def _split_text(self, text: str) -> List[str]:
"""
将字符串按换行符分割为列表。
"""
if isinstance(text, list):
return text
return text.split('\n')
def _preprocess_queries(self, queries: List[str]) -> List[str]:
"""
预处理查询文本,去除emoji,替换特殊字符,并添加「」标记。
"""
if self.do_enlarge_small_kana:
queries = [self.enlarge_small_kana(query) for query in queries]
queries = [self._emoji_pattern.sub('', query) for query in queries]
queries = [self._heart_pattern.sub('♥', query) for query in queries]
queries = [f'「{query}」' for query in queries]
self.logger.debug(f'预处理后的查询文本:{queries}')
return queries
def _check_translation_quality(self, queries: List[str], response: str) -> List[str]:
"""
检查翻译结果的质量,包括重复和行数对齐问题,如果存在问题则尝试重新翻译或返回原始文本。
"""
def _retry_translation(queries: List[str], check_func: Callable[[str], bool], error_message: str) -> str:
styles = ["precise", "normal", "aggressive", ]
for i in range(self.retry_attempts):
self._set_gpt_style(styles[i])
self.logger.warning(
f'{error_message} 尝试次数: {i + 1}。当前参数风格:{self._current_style}。')
response = self._handle_translation_request(queries)
if not check_func(response):
return response
return None
# 检查请求内容是否含有超过默认阈值的重复内容
if self.detect_and_calculate_repeats(''.join(queries), self.repeat_detect_threshold)[0]:
self.logger.warning(
f'请求内容本身含有超过默认阈值{self.repeat_detect_threshold}的重复内容。')
# 根据译文众数和默认阈值计算实际阈值
actual_threshold = max(max(self.detect_and_calculate_repeats(
query)[4] for query in queries), self.repeat_detect_threshold)
if self.detect_and_calculate_repeats(response, actual_threshold)[0]:
response = _retry_translation(queries, lambda r: self.detect_and_calculate_repeats(
r, actual_threshold)[0], f'检测到大量重复内容(当前阈值:{actual_threshold}),疑似模型退化,重新翻译。')
if response is None:
self.logger.warning(
f'疑似模型退化,尝试{self.retry_attempts}次仍未解决,进行单行翻译。')
return self._translate_single_lines(queries)
if not self.check_align(queries, response):
response = _retry_translation(queries, lambda r: not self.check_align(
queries, r), '因为检测到原文与译文行数不匹配,重新翻译。')
if response is None:
self.logger.warning(
f'原文与译文行数不匹配,尝试{self.retry_attempts}次仍未解决,进行单行翻译。')
return self._translate_single_lines(queries)
return self._split_text(response)
def _translate_single_lines(self, queries: List[str]) -> List[str]:
"""
逐行翻译查询文本。
"""
translations = []
for query in queries:
response = self._handle_translation_request(query)
if self.detect_and_calculate_repeats(response)[0]:
self.logger.warning(f"单行翻译结果存在重复内容: {response},返回原文。")
translations.append(query)
else:
translations.append(response)
return translations
def check_align(self, queries: List[str], response: str) -> bool:
"""
检查原始文本和翻译结果的行数是否对齐。
"""
translations = self._split_text(response)
is_aligned = len(queries) == len(translations)
if not is_aligned:
self.logger.warning(
f"行数不匹配 - 原文行数: {len(queries)},译文行数: {len(translations)}")
return is_aligned
def _delete_quotation_mark(self, texts: List[str]) -> List[str]:
"""
删除文本中的「」标记。
"""
new_texts = []
for text in texts:
text = text.strip('「」')
new_texts.append(text)
return new_texts
def _translate(self, src_list) -> List[str]:
self.logger.debug(
f'Temperature: {self.temperature}, TopP: {self.top_p}')
self.logger.debug(f'原文: {src_list}')
text_prompt = '\n'.join(src_list)
self.logger.debug('-- Sakura Prompt --\n' +
self._format_prompt_log(text_prompt) + '\n\n')
# 预处理查询文本
queries = self._preprocess_queries(src_list)
# 发送翻译请求
response = self._handle_translation_request(queries)
self.logger.debug('-- Sakura Response --\n' + response + '\n\n')
# 检查翻译结果是否存在重复或行数不匹配的问题
translations = self._check_translation_quality(queries, response)
return self._delete_quotation_mark(translations)
def _handle_translation_request(self, prompt):
ratelimit_attempt = 0
server_error_attempt = 0
timeout_attempt = 0
while True:
if OPENAPI_V1_API:
try:
response = self._request_translation(prompt)
break
except openai.RateLimitError:
ratelimit_attempt += 1
if ratelimit_attempt >= self.retry_attempts:
raise
self.logger.warning(
f'Sakura因被限速而进行重试。尝试次数: {ratelimit_attempt}')
time.sleep(2)
except openai.APIError as e:
server_error_attempt += 1
if server_error_attempt >= self.retry_attempts:
self.logger.warning(e)
self.logger.warning('Sakura翻译失败。返回原始文本。')
return '\n'.join(prompt)
self.logger.warning(
f'Sakura因服务器错误而进行重试。 当前API baseurl为"{self.api_base}",尝试次数: {server_error_attempt}, 错误信息: {e}')
time.sleep(1)
except FileNotFoundError:
self.logger.warning(
'Sakura因文件不存在而进行重试。')
time.sleep(30)
except TimeoutError:
timeout_attempt += 1
if timeout_attempt >= self.retry_attempts:
raise Exception('Sakura超时。')
self.logger.warning(
f'Sakura因超时而进行重试。尝试次数: {timeout_attempt}')
else:
try:
response = self._request_translation(prompt)
break
except openai.error.RateLimitError:
ratelimit_attempt += 1
if ratelimit_attempt >= self.retry_attempts:
raise
self.logger.warning(
f'Sakura因被限速而进行重试。尝试次数: {ratelimit_attempt}')
time.sleep(2)
except openai.error.APIError as e:
server_error_attempt += 1
if server_error_attempt >= self.retry_attempts:
self.logger.warning(
e, 'Sakura翻译失败。返回原始文本。')
return '\n'.join(prompt)
self.logger.warning(
f'Sakura因服务器错误而进行重试,请检查Sakura是否已经启动,API baseurl是否正确,并关闭一切代理软件后重试。\n 当前API baseurl为"{self.api_base}",尝试次数: {server_error_attempt}, 错误信息: {e}')
time.sleep(1)
except openai.error.APIConnectionError as e:
server_error_attempt += 1
if server_error_attempt >= self.retry_attempts:
self.logger.warning(
e, 'Sakura翻译失败。返回原始文本。')
return '\n'.join(prompt)
self.logger.warning(
f'Sakura因服务器连接错误而进行重试,请检查Sakura是否已经启动,API baseurl是否正确,并关闭一切代理软件后重试。\n 当前API baseurl为"{self.api_base}",尝试次数: {server_error_attempt}, 错误信息: {e}')
time.sleep(1)
except FileNotFoundError:
self.logger.warning(
'Sakura因文件不存在而进行重试。')
time.sleep(30)
except TimeoutError:
timeout_attempt += 1
if timeout_attempt >= self.retry_attempts:
raise Exception('Sakura超时。')
self.logger.warning(
f'Sakura因超时而进行重试。尝试次数: {timeout_attempt}')
return response
def _request_translation(self, input_text_list):
if isinstance(input_text_list, list):
raw_text = "\n".join(input_text_list)
else:
raw_text = input_text_list
extra_query = {
'do_sample': False,
'num_beams': 1,
'repetition_penalty': 1.0,
}
gpt_dict_raw_text = self.sakura_dict.get_dict_str_within_text(raw_text, self.force_apply_dict)
if self.sakura_version == "0.9" or gpt_dict_raw_text == "":
messages = [
{
"role": "system",
"content": f"{self._CHAT_SYSTEM_TEMPLATE_009}"
},
{
"role": "user",
"content": f"将下面的日文文本翻译成中文:{raw_text}"
}
]
elif self.sakura_version == "1.0":
messages = [
{
"role": "system",
"content": f"{self._CHAT_SYSTEM_TEMPLATE_100}"
},
{
"role": "user",
"content": f"根据以下术语表(可以为空):\n{gpt_dict_raw_text}\n将下面的日文文本根据对应关系和备注翻译成中文:{raw_text}"
}
]
else:
messages = [
{
"role": "system",
"content": f"{self._CHAT_SYSTEM_TEMPLATE_GALTRANSL_V1}"
},
{
"role": "user",
"content": f"根据以下术语表:\n{gpt_dict_raw_text}\n将下面的日文文本根据上述术语表的对应关系和注释翻译成中文:{raw_text}"
}
]
if OPENAPI_V1_API:
client = openai.Client(
api_key="sk-114514",
base_url=self.api_base
)
response = client.chat.completions.create(
model="sukinishiro",
messages=messages,
temperature=self.temperature,
top_p=self.top_p,
max_tokens=self.max_tokens,
frequency_penalty=self.frequency_penalty,
seed=-1,
extra_query=extra_query,
)
else:
openai.api_base = self.api_base
openai.api_key = "sk-114514"
response = openai.ChatCompletion.create(
model="sukinishiro",
messages=messages,
temperature=self.temperature,
top_p=self.top_p,
max_tokens=self.max_tokens,
frequency_penalty=self.frequency_penalty,
seed=-1,
extra_query=extra_query,
)
for choice in response.choices:
if OPENAPI_V1_API:
return choice.message.content
else:
if 'text' in choice:
return choice.text
return response.choices[0].message.content
def _set_gpt_style(self, style_name: str):
"""
设置GPT的生成风格。
"""
if self._current_style == style_name:
return
self._current_style = style_name
if style_name == "precise":
temperature, top_p = 0.1, 0.3
frequency_penalty = 0.05
elif style_name == "normal":
temperature, top_p = 0.3, 0.3
frequency_penalty = 0.2
elif style_name == "aggressive":
temperature, top_p = 0.3, 0.3
frequency_penalty = 0.3
self.temperature = temperature
self.top_p = top_p
self.frequency_penalty = frequency_penalty