repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/test_cmdline_crawl_with_pipeline/test_spider/__init__.py
tests/test_cmdline_crawl_with_pipeline/test_spider/__init__.py
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/test_cmdline_crawl_with_pipeline/test_spider/spiders/exception.py
tests/test_cmdline_crawl_with_pipeline/test_spider/spiders/exception.py
import scrapy class ExceptionSpider(scrapy.Spider): name = "exception" custom_settings = { "ITEM_PIPELINES": {"test_spider.pipelines.TestSpiderExceptionPipeline": 300} } def parse(self, response): pass
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/test_cmdline_crawl_with_pipeline/test_spider/spiders/__init__.py
tests/test_cmdline_crawl_with_pipeline/test_spider/spiders/__init__.py
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/test_cmdline_crawl_with_pipeline/test_spider/spiders/normal.py
tests/test_cmdline_crawl_with_pipeline/test_spider/spiders/normal.py
import scrapy class NormalSpider(scrapy.Spider): name = "normal" custom_settings = { "ITEM_PIPELINES": {"test_spider.pipelines.TestSpiderPipeline": 300} } def parse(self, response): pass
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/keys/__init__.py
tests/keys/__init__.py
from datetime import datetime, timedelta, timezone from pathlib import Path from cryptography.hazmat.backends import default_backend from cryptography.hazmat.primitives.asymmetric import rsa from cryptography.hazmat.primitives.hashes import SHA256 from cryptography.hazmat.primitives.serialization import ( Encoding, NoEncryption, PrivateFormat, ) from cryptography.x509 import ( CertificateBuilder, DNSName, Name, NameAttribute, SubjectAlternativeName, random_serial_number, ) from cryptography.x509.oid import NameOID # https://cryptography.io/en/latest/x509/tutorial/#creating-a-self-signed-certificate def generate_keys(): folder = Path(__file__).parent key = rsa.generate_private_key( public_exponent=65537, key_size=2048, backend=default_backend(), ) (folder / "localhost.key").write_bytes( key.private_bytes( encoding=Encoding.PEM, format=PrivateFormat.TraditionalOpenSSL, encryption_algorithm=NoEncryption(), ), ) subject = issuer = Name( [ NameAttribute(NameOID.COUNTRY_NAME, "IE"), NameAttribute(NameOID.ORGANIZATION_NAME, "Scrapy"), NameAttribute(NameOID.COMMON_NAME, "localhost"), ] ) cert = ( CertificateBuilder() .subject_name(subject) .issuer_name(issuer) .public_key(key.public_key()) .serial_number(random_serial_number()) .not_valid_before(datetime.now(tz=timezone.utc)) .not_valid_after(datetime.now(tz=timezone.utc) + timedelta(days=10)) .add_extension( SubjectAlternativeName([DNSName("localhost")]), critical=False, ) .sign(key, SHA256(), default_backend()) ) (folder / "localhost.crt").write_bytes(cert.public_bytes(Encoding.PEM))
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/mockserver/http.py
tests/mockserver/http.py
from __future__ import annotations from pathlib import Path from twisted.web import resource from twisted.web.static import Data, File from twisted.web.util import Redirect from tests import tests_datadir from .http_base import BaseMockServer, main_factory from .http_resources import ( ArbitraryLengthPayloadResource, BrokenChunkedResource, BrokenDownloadResource, ChunkedResource, Compress, ContentLengthHeaderResource, Delay, Drop, DuplicateHeaderResource, Echo, EmptyContentTypeHeaderResource, Follow, ForeverTakingResource, HostHeaderResource, LargeChunkedFileResource, NoMetaRefreshRedirect, Partial, PayloadResource, Raw, RedirectTo, ResponseHeadersResource, SetCookie, Status, ) class Root(resource.Resource): def __init__(self): super().__init__() self.putChild(b"status", Status()) self.putChild(b"follow", Follow()) self.putChild(b"delay", Delay()) self.putChild(b"partial", Partial()) self.putChild(b"drop", Drop()) self.putChild(b"raw", Raw()) self.putChild(b"echo", Echo()) self.putChild(b"payload", PayloadResource()) self.putChild(b"alpayload", ArbitraryLengthPayloadResource()) self.putChild(b"static", File(str(Path(tests_datadir, "test_site/")))) self.putChild(b"redirect-to", RedirectTo()) self.putChild(b"text", Data(b"Works", "text/plain")) self.putChild( b"html", Data( b"<body><p class='one'>Works</p><p class='two'>World</p></body>", "text/html", ), ) self.putChild( b"enc-gb18030", Data(b"<p>gb18030 encoding</p>", "text/html; charset=gb18030"), ) self.putChild(b"redirect", Redirect(b"/redirected")) self.putChild( b"redirect-no-meta-refresh", NoMetaRefreshRedirect(b"/redirected") ) self.putChild(b"redirected", Data(b"Redirected here", "text/plain")) numbers = [str(x).encode("utf8") for x in range(2**18)] self.putChild(b"numbers", Data(b"".join(numbers), "text/plain")) self.putChild(b"wait", ForeverTakingResource()) self.putChild(b"hang-after-headers", ForeverTakingResource(write=True)) self.putChild(b"host", HostHeaderResource()) self.putChild(b"broken", BrokenDownloadResource()) self.putChild(b"chunked", ChunkedResource()) self.putChild(b"broken-chunked", BrokenChunkedResource()) self.putChild(b"contentlength", ContentLengthHeaderResource()) self.putChild(b"nocontenttype", EmptyContentTypeHeaderResource()) self.putChild(b"largechunkedfile", LargeChunkedFileResource()) self.putChild(b"compress", Compress()) self.putChild(b"duplicate-header", DuplicateHeaderResource()) self.putChild(b"response-headers", ResponseHeadersResource()) self.putChild(b"set-cookie", SetCookie()) def getChild(self, name, request): return self def render(self, request): return b"Scrapy mock HTTP server\n" class MockServer(BaseMockServer): module_name = "tests.mockserver.http" main = main_factory(Root) if __name__ == "__main__": main()
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/mockserver/http_base.py
tests/mockserver/http_base.py
"""Base classes and functions for HTTP mockservers.""" from __future__ import annotations import argparse import sys from abc import ABC, abstractmethod from subprocess import PIPE, Popen from typing import TYPE_CHECKING from urllib.parse import urlparse from twisted.web.server import Site from tests.utils import get_script_run_env from .utils import ssl_context_factory if TYPE_CHECKING: from collections.abc import Callable from twisted.web import resource class BaseMockServer(ABC): listen_http: bool = True listen_https: bool = True @property @abstractmethod def module_name(self) -> str: raise NotImplementedError def __init__(self) -> None: if not self.listen_http and not self.listen_https: raise ValueError("At least one of listen_http and listen_https must be set") self.proc: Popen | None = None self.host: str = "127.0.0.1" self.http_port: int | None = None self.https_port: int | None = None def __enter__(self): self.proc = Popen( [sys.executable, "-u", "-m", self.module_name, *self.get_additional_args()], stdout=PIPE, env=get_script_run_env(), ) if self.listen_http: http_address = self.proc.stdout.readline().strip().decode("ascii") http_parsed = urlparse(http_address) self.http_port = http_parsed.port if self.listen_https: https_address = self.proc.stdout.readline().strip().decode("ascii") https_parsed = urlparse(https_address) self.https_port = https_parsed.port return self def __exit__(self, exc_type, exc_value, traceback): if self.proc: self.proc.kill() self.proc.communicate() def get_additional_args(self) -> list[str]: return [] def port(self, is_secure: bool = False) -> int: if not is_secure and not self.listen_http: raise ValueError("This server doesn't provide HTTP") if is_secure and not self.listen_https: raise ValueError("This server doesn't provide HTTPS") port = self.https_port if is_secure else self.http_port assert port is not None return port def url(self, path: str, is_secure: bool = False) -> str: port = self.port(is_secure) scheme = "https" if is_secure else "http" return f"{scheme}://{self.host}:{port}{path}" def main_factory( resource_class: type[resource.Resource], *, listen_http: bool = True, listen_https: bool = True, ) -> Callable[[], None]: if not listen_http and not listen_https: raise ValueError("At least one of listen_http and listen_https must be set") def main() -> None: from twisted.internet import reactor root = resource_class() factory = Site(root) if listen_http: http_port = reactor.listenTCP(0, factory) if listen_https: parser = argparse.ArgumentParser() parser.add_argument("--keyfile", help="SSL key file") parser.add_argument("--certfile", help="SSL certificate file") parser.add_argument( "--cipher-string", default=None, help="SSL cipher string (optional)", ) args = parser.parse_args() context_factory_kw = {} if args.keyfile: context_factory_kw["keyfile"] = args.keyfile if args.certfile: context_factory_kw["certfile"] = args.certfile if args.cipher_string: context_factory_kw["cipher_string"] = args.cipher_string context_factory = ssl_context_factory(**context_factory_kw) https_port = reactor.listenSSL(0, factory, context_factory) def print_listening(): if listen_http: http_host = http_port.getHost() http_address = f"http://{http_host.host}:{http_host.port}" print(http_address) if listen_https: https_host = https_port.getHost() https_address = f"https://{https_host.host}:{https_host.port}" print(https_address) reactor.callWhenRunning(print_listening) reactor.run() return main
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/mockserver/ftp.py
tests/mockserver/ftp.py
from __future__ import annotations import sys from argparse import ArgumentParser from pathlib import Path from shutil import rmtree from subprocess import PIPE, Popen from tempfile import mkdtemp from pyftpdlib.authorizers import DummyAuthorizer from pyftpdlib.handlers import FTPHandler from pyftpdlib.servers import FTPServer from tests.utils import get_script_run_env class MockFTPServer: """Creates an FTP server on port 2121 with a default passwordless user (anonymous) and a temporary root path that you can read from the :attr:`path` attribute.""" def __enter__(self): self.path = Path(mkdtemp()) self.proc = Popen( [sys.executable, "-u", "-m", "tests.mockserver.ftp", "-d", str(self.path)], stderr=PIPE, env=get_script_run_env(), ) for line in self.proc.stderr: if b"starting FTP server" in line: break return self def __exit__(self, exc_type, exc_value, traceback): rmtree(str(self.path)) self.proc.kill() self.proc.communicate() def url(self, path): return "ftp://127.0.0.1:2121/" + path def main() -> None: parser = ArgumentParser() parser.add_argument("-d", "--directory") args = parser.parse_args() authorizer = DummyAuthorizer() full_permissions = "elradfmwMT" authorizer.add_anonymous(args.directory, perm=full_permissions) handler = FTPHandler handler.authorizer = authorizer address = ("127.0.0.1", 2121) server = FTPServer(address, handler) server.serve_forever() if __name__ == "__main__": main()
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/mockserver/simple_https.py
tests/mockserver/simple_https.py
# This is only used by tests.test_downloader_handlers_http_base.TestSimpleHttpsBase from __future__ import annotations from twisted.web import resource from twisted.web.static import Data from .http_base import BaseMockServer, main_factory class Root(resource.Resource): def __init__(self): resource.Resource.__init__(self) self.putChild(b"file", Data(b"0123456789", "text/plain")) def getChild(self, name, request): return self class SimpleMockServer(BaseMockServer): listen_http = False module_name = "tests.mockserver.simple_https" def __init__(self, keyfile: str, certfile: str, cipher_string: str | None): super().__init__() self.keyfile = keyfile self.certfile = certfile self.cipher_string = cipher_string or "" def get_additional_args(self) -> list[str]: args = [ "--keyfile", self.keyfile, "--certfile", self.certfile, ] if self.cipher_string is not None: args.extend(["--cipher-string", self.cipher_string]) return args main = main_factory(Root, listen_http=False) if __name__ == "__main__": main()
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/mockserver/dns.py
tests/mockserver/dns.py
from __future__ import annotations import sys from subprocess import PIPE, Popen from twisted.internet import defer from twisted.names import dns, error from twisted.names.server import DNSServerFactory from tests.utils import get_script_run_env class MockDNSResolver: """ Implements twisted.internet.interfaces.IResolver partially """ def _resolve(self, name): record = dns.Record_A(address=b"127.0.0.1") answer = dns.RRHeader(name=name, payload=record) return [answer], [], [] def query(self, query, timeout=None): if query.type == dns.A: return defer.succeed(self._resolve(query.name.name)) return defer.fail(error.DomainError()) def lookupAllRecords(self, name, timeout=None): return defer.succeed(self._resolve(name)) class MockDNSServer: def __enter__(self): self.proc = Popen( [sys.executable, "-u", "-m", "tests.mockserver.dns"], stdout=PIPE, env=get_script_run_env(), ) self.host = "127.0.0.1" self.port = int( self.proc.stdout.readline().strip().decode("ascii").split(":")[1] ) return self def __exit__(self, exc_type, exc_value, traceback): self.proc.kill() self.proc.communicate() def main() -> None: from twisted.internet import reactor clients = [MockDNSResolver()] factory = DNSServerFactory(clients=clients) protocol = dns.DNSDatagramProtocol(controller=factory) listener = reactor.listenUDP(0, protocol) def print_listening(): host = listener.getHost() print(f"{host.host}:{host.port}") reactor.callWhenRunning(print_listening) reactor.run() if __name__ == "__main__": main()
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/mockserver/utils.py
tests/mockserver/utils.py
from __future__ import annotations from pathlib import Path from OpenSSL import SSL from twisted.internet import ssl from scrapy.utils.python import to_bytes def ssl_context_factory( keyfile="keys/localhost.key", certfile="keys/localhost.crt", cipher_string=None ): factory = ssl.DefaultOpenSSLContextFactory( str(Path(__file__).parent.parent / keyfile), str(Path(__file__).parent.parent / certfile), ) if cipher_string: ctx = factory.getContext() # disabling TLS1.3 because it unconditionally enables some strong ciphers ctx.set_options(SSL.OP_CIPHER_SERVER_PREFERENCE | SSL.OP_NO_TLSv1_3) ctx.set_cipher_list(to_bytes(cipher_string)) return factory
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/mockserver/__init__.py
tests/mockserver/__init__.py
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/mockserver/proxy_echo.py
tests/mockserver/proxy_echo.py
# This is only used by tests.test_downloader_handlers_http_base.TestHttpProxyBase from __future__ import annotations from .http_base import BaseMockServer, main_factory from .http_resources import UriResource class ProxyEchoMockServer(BaseMockServer): module_name = "tests.mockserver.proxy_echo" main = main_factory(UriResource) if __name__ == "__main__": main()
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/tests/mockserver/http_resources.py
tests/mockserver/http_resources.py
from __future__ import annotations import gzip import json import random from urllib.parse import urlencode from twisted.internet.task import deferLater from twisted.web import resource, server from twisted.web.server import NOT_DONE_YET from twisted.web.util import Redirect, redirectTo from scrapy.utils.python import to_bytes, to_unicode def getarg(request, name, default=None, type_=None): if name in request.args: value = request.args[name][0] if type_ is not None: value = type_(value) return value return default def close_connection(request): # We have to force a disconnection for HTTP/1.1 clients. Otherwise # client keeps the connection open waiting for more data. request.channel.loseConnection() request.finish() # most of the following resources are copied from twisted.web.test.test_webclient class ForeverTakingResource(resource.Resource): """ L{ForeverTakingResource} is a resource which never finishes responding to requests. """ def __init__(self, write=False): resource.Resource.__init__(self) self._write = write def render(self, request): if self._write: request.write(b"some bytes") return server.NOT_DONE_YET class HostHeaderResource(resource.Resource): """ A testing resource which renders itself as the value of the host header from the request. """ def render(self, request): return request.requestHeaders.getRawHeaders(b"host")[0] class PayloadResource(resource.Resource): """ A testing resource which renders itself as the contents of the request body as long as the request body is 100 bytes long, otherwise which renders itself as C{"ERROR"}. """ def render(self, request): data = request.content.read() contentLength = request.requestHeaders.getRawHeaders(b"content-length")[0] if len(data) != 100 or int(contentLength) != 100: return b"ERROR" return data class LeafResource(resource.Resource): isLeaf = True def deferRequest(self, request, delay, f, *a, **kw): from twisted.internet import reactor def _cancelrequest(_): # silence CancelledError d.addErrback(lambda _: None) d.cancel() d = deferLater(reactor, delay, f, *a, **kw) request.notifyFinish().addErrback(_cancelrequest) return d class Follow(LeafResource): def render(self, request): total = getarg(request, b"total", 100, type_=int) show = getarg(request, b"show", 1, type_=int) order = getarg(request, b"order", b"desc") maxlatency = getarg(request, b"maxlatency", 0, type_=float) n = getarg(request, b"n", total, type_=int) if order == b"rand": nlist = [random.randint(1, total) for _ in range(show)] else: # order == "desc" nlist = range(n, max(n - show, 0), -1) lag = random.random() * maxlatency self.deferRequest(request, lag, self.renderRequest, request, nlist) return NOT_DONE_YET def renderRequest(self, request, nlist): s = """<html> <head></head> <body>""" args = request.args.copy() for nl in nlist: args[b"n"] = [to_bytes(str(nl))] argstr = urlencode(args, doseq=True) s += f"<a href='/follow?{argstr}'>follow {nl}</a><br>" s += """</body>""" request.write(to_bytes(s)) request.finish() class Delay(LeafResource): def render_GET(self, request): n = getarg(request, b"n", 1, type_=float) b = getarg(request, b"b", 1, type_=int) if b: # send headers now and delay body request.write("") self.deferRequest(request, n, self._delayedRender, request, n) return NOT_DONE_YET def _delayedRender(self, request, n): request.write(to_bytes(f"Response delayed for {n:.3f} seconds\n")) request.finish() class Status(LeafResource): def render_GET(self, request): n = getarg(request, b"n", 200, type_=int) request.setResponseCode(n) return b"" class Raw(LeafResource): def render_GET(self, request): request.startedWriting = 1 self.deferRequest(request, 0, self._delayedRender, request) return NOT_DONE_YET render_POST = render_GET def _delayedRender(self, request): raw = getarg(request, b"raw", b"HTTP 1.1 200 OK\n") request.startedWriting = 1 request.write(raw) request.channel.transport.loseConnection() request.finish() class Echo(LeafResource): def render_GET(self, request): output = { "headers": { to_unicode(k): [to_unicode(v) for v in vs] for k, vs in request.requestHeaders.getAllRawHeaders() }, "body": to_unicode(request.content.read()), } return to_bytes(json.dumps(output)) render_POST = render_GET class RedirectTo(LeafResource): def render(self, request): goto = getarg(request, b"goto", b"/") # we force the body content, otherwise Twisted redirectTo() # returns HTML with <meta http-equiv="refresh" redirectTo(goto, request) return b"redirecting..." class Partial(LeafResource): def render_GET(self, request): request.setHeader(b"Content-Length", b"1024") self.deferRequest(request, 0, self._delayedRender, request) return NOT_DONE_YET def _delayedRender(self, request): request.write(b"partial content\n") request.finish() class Drop(Partial): def _delayedRender(self, request): abort = getarg(request, b"abort", 0, type_=int) request.write(b"this connection will be dropped\n") tr = request.channel.transport try: if abort and hasattr(tr, "abortConnection"): tr.abortConnection() else: tr.loseConnection() finally: request.finish() class ArbitraryLengthPayloadResource(LeafResource): def render(self, request): return request.content.read() class NoMetaRefreshRedirect(Redirect): def render(self, request: server.Request) -> bytes: content = Redirect.render(self, request) return content.replace( b'http-equiv="refresh"', b'http-no-equiv="do-not-refresh-me"' ) class ContentLengthHeaderResource(resource.Resource): """ A testing resource which renders itself as the value of the Content-Length header from the request. """ def render(self, request): return request.requestHeaders.getRawHeaders(b"content-length")[0] class ChunkedResource(resource.Resource): def render(self, request): from twisted.internet import reactor def response(): request.write(b"chunked ") request.write(b"content\n") request.finish() reactor.callLater(0, response) return server.NOT_DONE_YET class BrokenChunkedResource(resource.Resource): def render(self, request): from twisted.internet import reactor def response(): request.write(b"chunked ") request.write(b"content\n") # Disable terminating chunk on finish. request.chunked = False close_connection(request) reactor.callLater(0, response) return server.NOT_DONE_YET class BrokenDownloadResource(resource.Resource): def render(self, request): from twisted.internet import reactor def response(): request.setHeader(b"Content-Length", b"20") request.write(b"partial") close_connection(request) reactor.callLater(0, response) return server.NOT_DONE_YET class EmptyContentTypeHeaderResource(resource.Resource): """ A testing resource which renders itself as the value of request body without content-type header in response. """ def render(self, request): request.setHeader("content-type", "") return request.content.read() class LargeChunkedFileResource(resource.Resource): def render(self, request): from twisted.internet import reactor def response(): for i in range(1024): request.write(b"x" * 1024) request.finish() reactor.callLater(0, response) return server.NOT_DONE_YET class DuplicateHeaderResource(resource.Resource): def render(self, request): request.responseHeaders.setRawHeaders(b"Set-Cookie", [b"a=b", b"c=d"]) return b"" class UriResource(resource.Resource): """Return the full uri that was requested""" def getChild(self, path, request): return self def render(self, request): # Note: this is an ugly hack for CONNECT request timeout test. # Returning some data here fail SSL/TLS handshake # ToDo: implement proper HTTPS proxy tests, not faking them. if request.method != b"CONNECT": return request.uri return b"" class ResponseHeadersResource(resource.Resource): """Return a response with headers set from the JSON request body""" def render(self, request): body = json.loads(request.content.read().decode()) for header_name, header_value in body.items(): request.responseHeaders.addRawHeader(header_name, header_value) return json.dumps(body).encode("utf-8") class Compress(resource.Resource): """Compress the data sent in the request url params and set Content-Encoding header""" def render(self, request): data = request.args.get(b"data")[0] accept_encoding_header = request.getHeader(b"accept-encoding") # include common encoding schemes here if accept_encoding_header == b"gzip": request.setHeader(b"Content-Encoding", b"gzip") return gzip.compress(data) # just set this to trigger a test failure if no valid accept-encoding header was set request.setResponseCode(500) return b"Did not receive a valid accept-encoding header" class SetCookie(resource.Resource): """Return a response with a Set-Cookie header for each request url parameter""" def render(self, request): for cookie_name, cookie_values in request.args.items(): for cookie_value in cookie_values: cookie = (cookie_name.decode() + "=" + cookie_value.decode()).encode() request.setHeader(b"Set-Cookie", cookie) return b""
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/docs/conftest.py
docs/conftest.py
from doctest import ELLIPSIS, NORMALIZE_WHITESPACE from pathlib import Path from sybil import Sybil from sybil.parsers.doctest import DocTestParser from sybil.parsers.skip import skip try: # >2.0.1 from sybil.parsers.codeblock import PythonCodeBlockParser except ImportError: from sybil.parsers.codeblock import CodeBlockParser as PythonCodeBlockParser from scrapy.http.response.html import HtmlResponse def load_response(url: str, filename: str) -> HtmlResponse: input_path = Path(__file__).parent / "_tests" / filename return HtmlResponse(url, body=input_path.read_bytes()) def setup(namespace): namespace["load_response"] = load_response pytest_collect_file = Sybil( parsers=[ DocTestParser(optionflags=ELLIPSIS | NORMALIZE_WHITESPACE), PythonCodeBlockParser(future_imports=["print_function"]), skip, ], pattern="*.rst", setup=setup, ).pytest()
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/docs/conf.py
docs/conf.py
# Configuration file for the Sphinx documentation builder. # # For the full list of built-in configuration values, see the documentation: # https://www.sphinx-doc.org/en/master/usage/configuration.html import os import sys from collections.abc import Sequence from pathlib import Path # If your extensions are in another directory, add it here. If the directory # is relative to the documentation root, use Path.absolute to make it absolute. sys.path.append(str(Path(__file__).parent / "_ext")) sys.path.insert(0, str(Path(__file__).parent.parent)) # -- Project information ----------------------------------------------------- # https://www.sphinx-doc.org/en/master/usage/configuration.html#project-information project = "Scrapy" project_copyright = "Scrapy developers" author = "Scrapy developers" # -- General configuration --------------------------------------------------- # https://www.sphinx-doc.org/en/master/usage/configuration.html#general-configuration extensions = [ "notfound.extension", "scrapydocs", "sphinx.ext.autodoc", "scrapyfixautodoc", # Must be after "sphinx.ext.autodoc" "sphinx.ext.coverage", "sphinx.ext.intersphinx", "sphinx.ext.viewcode", "sphinx_rtd_dark_mode", ] templates_path = ["_templates"] exclude_patterns = ["build", "Thumbs.db", ".DS_Store"] # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. try: import scrapy version = ".".join(map(str, scrapy.version_info[:2])) release = scrapy.__version__ except ImportError: version = "" release = "" suppress_warnings = ["epub.unknown_project_files"] # -- Options for HTML output ------------------------------------------------- # https://www.sphinx-doc.org/en/master/usage/configuration.html#options-for-html-output html_theme = "sphinx_rtd_theme" html_static_path = ["_static"] html_last_updated_fmt = "%b %d, %Y" html_css_files = [ "custom.css", ] html_context = { "display_github": True, "github_user": "scrapy", "github_repo": "scrapy", "github_version": "master", "conf_py_path": "/docs/", } # Set canonical URL from the Read the Docs Domain html_baseurl = os.environ.get("READTHEDOCS_CANONICAL_URL", "") # -- Options for LaTeX output ------------------------------------------------ # https://www.sphinx-doc.org/en/master/usage/configuration.html#options-for-latex-output # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, author, document class [howto/manual]). latex_documents = [ ("index", "Scrapy.tex", "Scrapy Documentation", "Scrapy developers", "manual"), ] # -- Options for the linkcheck builder --------------------------------------- # https://www.sphinx-doc.org/en/master/usage/configuration.html#options-for-the-linkcheck-builder linkcheck_ignore = [ r"http://localhost:\d+", "http://hg.scrapy.org", r"https://github.com/scrapy/scrapy/commit/\w+", r"https://github.com/scrapy/scrapy/issues/\d+", ] linkcheck_anchors_ignore_for_url = ["https://github.com/pyca/cryptography/issues/2692"] # -- Options for the Coverage extension -------------------------------------- # https://www.sphinx-doc.org/en/master/usage/extensions/coverage.html#configuration coverage_ignore_pyobjects = [ # Contract’s add_pre_hook and add_post_hook are not documented because # they should be transparent to contract developers, for whom pre_hook and # post_hook should be the actual concern. r"\bContract\.add_(pre|post)_hook$", # ContractsManager is an internal class, developers are not expected to # interact with it directly in any way. r"\bContractsManager\b$", # For default contracts we only want to document their general purpose in # their __init__ method, the methods they reimplement to achieve that purpose # should be irrelevant to developers using those contracts. r"\w+Contract\.(adjust_request_args|(pre|post)_process)$", # Methods of downloader middlewares are not documented, only the classes # themselves, since downloader middlewares are controlled through Scrapy # settings. r"^scrapy\.downloadermiddlewares\.\w*?\.(\w*?Middleware|DownloaderStats)\.", # Base classes of downloader middlewares are implementation details that # are not meant for users. r"^scrapy\.downloadermiddlewares\.\w*?\.Base\w*?Middleware", # The interface methods of duplicate request filtering classes are already # covered in the interface documentation part of the DUPEFILTER_CLASS # setting documentation. r"^scrapy\.dupefilters\.[A-Z]\w*?\.(from_crawler|request_seen|open|close|log)$", # Private exception used by the command-line interface implementation. r"^scrapy\.exceptions\.UsageError", # Methods of BaseItemExporter subclasses are only documented in # BaseItemExporter. r"^scrapy\.exporters\.(?!BaseItemExporter\b)\w*?\.", # Extension behavior is only modified through settings. Methods of # extension classes, as well as helper functions, are implementation # details that are not documented. r"^scrapy\.extensions\.[a-z]\w*?\.[A-Z]\w*?\.", # methods r"^scrapy\.extensions\.[a-z]\w*?\.[a-z]", # helper functions # Never documented before, and deprecated now. r"^scrapy\.linkextractors\.FilteringLinkExtractor$", # Implementation detail of LxmlLinkExtractor r"^scrapy\.linkextractors\.lxmlhtml\.LxmlParserLinkExtractor", ] # -- Options for the InterSphinx extension ----------------------------------- # https://www.sphinx-doc.org/en/master/usage/extensions/intersphinx.html#configuration intersphinx_mapping = { "attrs": ("https://www.attrs.org/en/stable/", None), "coverage": ("https://coverage.readthedocs.io/en/latest", None), "cryptography": ("https://cryptography.io/en/latest/", None), "cssselect": ("https://cssselect.readthedocs.io/en/latest", None), "itemloaders": ("https://itemloaders.readthedocs.io/en/latest/", None), "parsel": ("https://parsel.readthedocs.io/en/latest/", None), "pytest": ("https://docs.pytest.org/en/latest", None), "python": ("https://docs.python.org/3", None), "sphinx": ("https://www.sphinx-doc.org/en/master", None), "tox": ("https://tox.wiki/en/latest/", None), "twisted": ("https://docs.twisted.org/en/stable/", None), "twistedapi": ("https://docs.twisted.org/en/stable/api/", None), "w3lib": ("https://w3lib.readthedocs.io/en/latest", None), } intersphinx_disabled_reftypes: Sequence[str] = [] # -- Other options ------------------------------------------------------------ default_dark_mode = False
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/docs/utils/linkfix.py
docs/utils/linkfix.py
#!/usr/bin/python """ Linkfix - a companion to sphinx's linkcheck builder. Uses the linkcheck's output file to fix links in docs. Originally created for this issue: https://github.com/scrapy/scrapy/issues/606 Author: dufferzafar """ import re import sys from pathlib import Path def main(): # Used for remembering the file (and its contents) # so we don't have to open the same file again. _filename = None _contents = None # A regex that matches standard linkcheck output lines line_re = re.compile(r"(.*)\:\d+\:\s\[(.*)\]\s(?:(.*)\sto\s(.*)|(.*))") # Read lines from the linkcheck output file try: with Path("build/linkcheck/output.txt").open(encoding="utf-8") as out: output_lines = out.readlines() except OSError: print("linkcheck output not found; please run linkcheck first.") sys.exit(1) # For every line, fix the respective file for line in output_lines: match = re.match(line_re, line) if match: newfilename = match.group(1) errortype = match.group(2) # Broken links can't be fixed and # I am not sure what do with the local ones. if errortype.lower() in ["broken", "local"]: print("Not Fixed: " + line) else: # If this is a new file if newfilename != _filename: # Update the previous file if _filename: Path(_filename).write_text(_contents, encoding="utf-8") _filename = newfilename # Read the new file to memory _contents = Path(_filename).read_text(encoding="utf-8") _contents = _contents.replace(match.group(3), match.group(4)) else: # We don't understand what the current line means! print("Not Understood: " + line) if __name__ == "__main__": main()
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/docs/_ext/scrapyfixautodoc.py
docs/_ext/scrapyfixautodoc.py
""" Must be included after 'sphinx.ext.autodoc'. Fixes unwanted 'alias of' behavior. https://github.com/sphinx-doc/sphinx/issues/4422 """ # pylint: disable=import-error from sphinx.application import Sphinx def maybe_skip_member(app: Sphinx, what, name: str, obj, skip: bool, options) -> bool: if not skip: # autodocs was generating a text "alias of" for the following members return name in {"default_item_class", "default_selector_class"} return skip def setup(app: Sphinx) -> None: app.connect("autodoc-skip-member", maybe_skip_member)
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/docs/_ext/scrapydocs.py
docs/_ext/scrapydocs.py
# pylint: disable=import-error from collections.abc import Sequence from operator import itemgetter from typing import Any, TypedDict from docutils import nodes from docutils.nodes import Element, General, Node, document from docutils.parsers.rst import Directive from sphinx.application import Sphinx from sphinx.util.nodes import make_refnode class SettingData(TypedDict): docname: str setting_name: str refid: str class SettingslistNode(General, Element): pass class SettingsListDirective(Directive): def run(self) -> Sequence[Node]: return [SettingslistNode()] def is_setting_index(node: Node) -> bool: if node.tagname == "index" and node["entries"]: # type: ignore[index,attr-defined] # index entries for setting directives look like: # [('pair', 'SETTING_NAME; setting', 'std:setting-SETTING_NAME', '')] entry_type, info, _ = node["entries"][0][:3] # type: ignore[index] return entry_type == "pair" and info.endswith("; setting") return False def get_setting_name_and_refid(node: Node) -> tuple[str, str]: """Extract setting name from directive index node""" _, info, refid = node["entries"][0][:3] # type: ignore[index] return info.replace("; setting", ""), refid def collect_scrapy_settings_refs(app: Sphinx, doctree: document) -> None: env = app.builder.env if not hasattr(env, "scrapy_all_settings"): emptyList: list[SettingData] = [] env.scrapy_all_settings = emptyList # type: ignore[attr-defined] for node in doctree.findall(is_setting_index): setting_name, refid = get_setting_name_and_refid(node) env.scrapy_all_settings.append( # type: ignore[attr-defined] SettingData( docname=env.docname, setting_name=setting_name, refid=refid, ) ) def make_setting_element( setting_data: SettingData, app: Sphinx, fromdocname: str ) -> Any: refnode = make_refnode( app.builder, fromdocname, todocname=setting_data["docname"], targetid=setting_data["refid"], child=nodes.Text(setting_data["setting_name"]), ) p = nodes.paragraph() p += refnode item = nodes.list_item() item += p return item def replace_settingslist_nodes( app: Sphinx, doctree: document, fromdocname: str ) -> None: env = app.builder.env for node in doctree.findall(SettingslistNode): settings_list = nodes.bullet_list() settings_list.extend( [ make_setting_element(d, app, fromdocname) for d in sorted(env.scrapy_all_settings, key=itemgetter("setting_name")) # type: ignore[attr-defined] if fromdocname != d["docname"] ] ) node.replace_self(settings_list) def source_role( name, rawtext, text: str, lineno, inliner, options=None, content=None ) -> tuple[list[Any], list[Any]]: ref = "https://github.com/scrapy/scrapy/blob/master/" + text node = nodes.reference(rawtext, text, refuri=ref, **options) return [node], [] def issue_role( name, rawtext, text: str, lineno, inliner, options=None, content=None ) -> tuple[list[Any], list[Any]]: ref = "https://github.com/scrapy/scrapy/issues/" + text node = nodes.reference(rawtext, "issue " + text, refuri=ref) return [node], [] def commit_role( name, rawtext, text: str, lineno, inliner, options=None, content=None ) -> tuple[list[Any], list[Any]]: ref = "https://github.com/scrapy/scrapy/commit/" + text node = nodes.reference(rawtext, "commit " + text, refuri=ref) return [node], [] def rev_role( name, rawtext, text: str, lineno, inliner, options=None, content=None ) -> tuple[list[Any], list[Any]]: ref = "http://hg.scrapy.org/scrapy/changeset/" + text node = nodes.reference(rawtext, "r" + text, refuri=ref) return [node], [] def setup(app: Sphinx) -> None: app.add_crossref_type( directivename="setting", rolename="setting", indextemplate="pair: %s; setting", ) app.add_crossref_type( directivename="signal", rolename="signal", indextemplate="pair: %s; signal", ) app.add_crossref_type( directivename="command", rolename="command", indextemplate="pair: %s; command", ) app.add_crossref_type( directivename="reqmeta", rolename="reqmeta", indextemplate="pair: %s; reqmeta", ) app.add_role("source", source_role) app.add_role("commit", commit_role) app.add_role("issue", issue_role) app.add_role("rev", rev_role) app.add_node(SettingslistNode) app.add_directive("settingslist", SettingsListDirective) app.connect("doctree-read", collect_scrapy_settings_refs) app.connect("doctree-resolved", replace_settingslist_nodes)
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/extras/qps-bench-server.py
extras/qps-bench-server.py
#!/usr/bin/env python from collections import deque from time import time from twisted.internet import reactor # noqa: TID253 from twisted.web.resource import Resource from twisted.web.server import NOT_DONE_YET, Site class Root(Resource): def __init__(self): Resource.__init__(self) self.concurrent = 0 self.tail = deque(maxlen=100) self._reset_stats() def _reset_stats(self): self.tail.clear() self.start = self.lastmark = self.lasttime = time() def getChild(self, request, name): return self def render(self, request): now = time() delta = now - self.lasttime # reset stats on high iter-request times caused by client restarts if delta > 3: # seconds self._reset_stats() return "" self.tail.appendleft(delta) self.lasttime = now self.concurrent += 1 if now - self.lastmark >= 3: self.lastmark = now qps = len(self.tail) / sum(self.tail) print( f"samplesize={len(self.tail)} concurrent={self.concurrent} qps={qps:0.2f}" ) if "latency" in request.args: latency = float(request.args["latency"][0]) reactor.callLater(latency, self._finish, request) return NOT_DONE_YET self.concurrent -= 1 return "" def _finish(self, request): self.concurrent -= 1 if not request.finished and not request._disconnected: request.finish() root = Root() factory = Site(root) reactor.listenTCP(8880, factory) reactor.run()
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
scrapy/scrapy
https://github.com/scrapy/scrapy/blob/d1bd8eb49f7aba9289e4ff692006cead8bcd9080/extras/qpsclient.py
extras/qpsclient.py
""" A spider that generate light requests to measure QPS throughput usage: scrapy runspider qpsclient.py --loglevel=INFO --set RANDOMIZE_DOWNLOAD_DELAY=0 --set CONCURRENT_REQUESTS=50 -a qps=10 -a latency=0.3 """ from scrapy.http import Request from scrapy.spiders import Spider class QPSSpider(Spider): name = "qps" benchurl = "http://localhost:8880/" # Max concurrency is limited by global CONCURRENT_REQUESTS setting max_concurrent_requests = 8 # Requests per second goal qps = None # same as: 1 / download_delay download_delay = None # time in seconds to delay server responses latency = None # number of slots to create slots = 1 def __init__(self, *a, **kw): super().__init__(*a, **kw) if self.qps is not None: self.qps = float(self.qps) self.download_delay = 1 / self.qps elif self.download_delay is not None: self.download_delay = float(self.download_delay) async def start(self): for item_or_request in self.start_requests(): yield item_or_request def start_requests(self): url = self.benchurl if self.latency is not None: url += f"?latency={self.latency}" slots = int(self.slots) if slots > 1: urls = [url.replace("localhost", f"127.0.0.{x + 1}") for x in range(slots)] else: urls = [url] idx = 0 while True: url = urls[idx % len(urls)] yield Request(url, dont_filter=True) idx += 1 def parse(self, response): pass
python
BSD-3-Clause
d1bd8eb49f7aba9289e4ff692006cead8bcd9080
2026-01-04T14:38:41.023839Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/download_deps.py
download_deps.py
#!/usr/bin/env python3 # PEP 723 metadata # /// script # requires-python = ">=3.10" # dependencies = [ # "nltk", # "huggingface-hub" # ] # /// import argparse import os import urllib.request from typing import Union import nltk from huggingface_hub import snapshot_download def get_urls(use_china_mirrors=False) -> list[Union[str, list[str]]]: if use_china_mirrors: return [ "http://mirrors.tuna.tsinghua.edu.cn/ubuntu/pool/main/o/openssl/libssl1.1_1.1.1f-1ubuntu2_amd64.deb", "http://mirrors.tuna.tsinghua.edu.cn/ubuntu-ports/pool/main/o/openssl/libssl1.1_1.1.1f-1ubuntu2_arm64.deb", "https://repo.huaweicloud.com/repository/maven/org/apache/tika/tika-server-standard/3.2.3/tika-server-standard-3.2.3.jar", "https://repo.huaweicloud.com/repository/maven/org/apache/tika/tika-server-standard/3.2.3/tika-server-standard-3.2.3.jar.md5", "https://openaipublic.blob.core.windows.net/encodings/cl100k_base.tiktoken", ["https://registry.npmmirror.com/-/binary/chrome-for-testing/121.0.6167.85/linux64/chrome-linux64.zip", "chrome-linux64-121-0-6167-85"], ["https://registry.npmmirror.com/-/binary/chrome-for-testing/121.0.6167.85/linux64/chromedriver-linux64.zip", "chromedriver-linux64-121-0-6167-85"], "https://github.com/astral-sh/uv/releases/download/0.9.16/uv-x86_64-unknown-linux-gnu.tar.gz", ] else: return [ "http://archive.ubuntu.com/ubuntu/pool/main/o/openssl/libssl1.1_1.1.1f-1ubuntu2_amd64.deb", "http://ports.ubuntu.com/pool/main/o/openssl/libssl1.1_1.1.1f-1ubuntu2_arm64.deb", "https://repo1.maven.org/maven2/org/apache/tika/tika-server-standard/3.2.3/tika-server-standard-3.2.3.jar", "https://repo1.maven.org/maven2/org/apache/tika/tika-server-standard/3.2.3/tika-server-standard-3.2.3.jar.md5", "https://openaipublic.blob.core.windows.net/encodings/cl100k_base.tiktoken", ["https://storage.googleapis.com/chrome-for-testing-public/121.0.6167.85/linux64/chrome-linux64.zip", "chrome-linux64-121-0-6167-85"], ["https://storage.googleapis.com/chrome-for-testing-public/121.0.6167.85/linux64/chromedriver-linux64.zip", "chromedriver-linux64-121-0-6167-85"], "https://github.com/astral-sh/uv/releases/download/0.9.16/uv-x86_64-unknown-linux-gnu.tar.gz", ] repos = [ "InfiniFlow/text_concat_xgb_v1.0", "InfiniFlow/deepdoc", ] def download_model(repository_id): local_directory = os.path.abspath(os.path.join("huggingface.co", repository_id)) os.makedirs(local_directory, exist_ok=True) snapshot_download(repo_id=repository_id, local_dir=local_directory) if __name__ == "__main__": parser = argparse.ArgumentParser(description="Download dependencies with optional China mirror support") parser.add_argument("--china-mirrors", action="store_true", help="Use China-accessible mirrors for downloads") args = parser.parse_args() urls = get_urls(args.china_mirrors) for url in urls: download_url = url[0] if isinstance(url, list) else url filename = url[1] if isinstance(url, list) else url.split("/")[-1] print(f"Downloading {filename} from {download_url}...") if not os.path.exists(filename): urllib.request.urlretrieve(download_url, filename) local_dir = os.path.abspath("nltk_data") for data in ["wordnet", "punkt", "punkt_tab"]: print(f"Downloading nltk {data}...") nltk.download(data, download_dir=local_dir) for repo_id in repos: print(f"Downloading huggingface repo {repo_id}...") download_model(repo_id)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/run_tests.py
run_tests.py
#!/usr/bin/env python3 # # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import sys import os import argparse import subprocess from pathlib import Path from typing import List class Colors: """ANSI color codes for terminal output""" RED = '\033[0;31m' GREEN = '\033[0;32m' YELLOW = '\033[1;33m' BLUE = '\033[0;34m' NC = '\033[0m' # No Color class TestRunner: """RAGFlow Unit Test Runner""" def __init__(self): self.project_root = Path(__file__).parent.resolve() self.ut_dir = Path(self.project_root / 'test' / 'unit_test') # Default options self.coverage = False self.parallel = False self.verbose = False self.markers = "" # Python interpreter path self.python = sys.executable @staticmethod def print_info(message: str) -> None: """Print informational message""" print(f"{Colors.BLUE}[INFO]{Colors.NC} {message}") @staticmethod def print_error(message: str) -> None: """Print error message""" print(f"{Colors.RED}[ERROR]{Colors.NC} {message}") @staticmethod def show_usage() -> None: """Display usage information""" usage = """ RAGFlow Unit Test Runner Usage: python run_tests.py [OPTIONS] OPTIONS: -h, --help Show this help message -c, --coverage Run tests with coverage report -p, --parallel Run tests in parallel (requires pytest-xdist) -v, --verbose Verbose output -t, --test FILE Run specific test file or directory -m, --markers MARKERS Run tests with specific markers (e.g., "unit", "integration") EXAMPLES: # Run all tests python run_tests.py # Run with coverage python run_tests.py --coverage # Run in parallel python run_tests.py --parallel # Run specific test file python run_tests.py --test services/test_dialog_service.py # Run only unit tests python run_tests.py --markers "unit" # Run tests with coverage and parallel execution python run_tests.py --coverage --parallel """ print(usage) def build_pytest_command(self) -> List[str]: """Build the pytest command arguments""" cmd = ["pytest", str(self.ut_dir)] # Add test path # Add markers if self.markers: cmd.extend(["-m", self.markers]) # Add verbose flag if self.verbose: cmd.extend(["-vv"]) else: cmd.append("-v") # Add coverage if self.coverage: # Relative path from test directory to source code source_path = str(self.project_root / "common") cmd.extend([ "--cov", source_path, "--cov-report", "html", "--cov-report", "term" ]) # Add parallel execution if self.parallel: # Try to get number of CPU cores try: import multiprocessing cpu_count = multiprocessing.cpu_count() cmd.extend(["-n", str(cpu_count)]) except ImportError: # Fallback to auto if multiprocessing not available cmd.extend(["-n", "auto"]) # Add default options from pyproject.toml if it exists pyproject_path = self.project_root / "pyproject.toml" if pyproject_path.exists(): cmd.extend(["--config-file", str(pyproject_path)]) return cmd def run_tests(self) -> bool: """Execute the pytest command""" # Change to test directory os.chdir(self.project_root) # Build command cmd = self.build_pytest_command() # Print test configuration self.print_info("Running RAGFlow Unit Tests") self.print_info("=" * 40) self.print_info(f"Test Directory: {self.ut_dir}") self.print_info(f"Coverage: {self.coverage}") self.print_info(f"Parallel: {self.parallel}") self.print_info(f"Verbose: {self.verbose}") if self.markers: self.print_info(f"Markers: {self.markers}") print(f"\n{Colors.BLUE}[EXECUTING]{Colors.NC} {' '.join(cmd)}\n") # Run pytest try: result = subprocess.run(cmd, check=False) if result.returncode == 0: print(f"\n{Colors.GREEN}[SUCCESS]{Colors.NC} All tests passed!") if self.coverage: coverage_dir = self.ut_dir / "htmlcov" if coverage_dir.exists(): index_file = coverage_dir / "index.html" print(f"\n{Colors.BLUE}[INFO]{Colors.NC} Coverage report generated:") print(f" {index_file}") print("\nOpen with:") print(f" - Windows: start {index_file}") print(f" - macOS: open {index_file}") print(f" - Linux: xdg-open {index_file}") return True else: print(f"\n{Colors.RED}[FAILURE]{Colors.NC} Some tests failed!") return False except KeyboardInterrupt: print(f"\n{Colors.YELLOW}[INTERRUPTED]{Colors.NC} Test execution interrupted by user") return False except Exception as e: self.print_error(f"Failed to execute tests: {e}") return False def parse_arguments(self) -> bool: """Parse command line arguments""" parser = argparse.ArgumentParser( description="RAGFlow Unit Test Runner", formatter_class=argparse.RawDescriptionHelpFormatter, epilog=""" Examples: python run_tests.py # Run all tests python run_tests.py --coverage # Run with coverage python run_tests.py --parallel # Run in parallel python run_tests.py --test services/test_dialog_service.py # Run specific test python run_tests.py --markers "unit" # Run only unit tests """ ) parser.add_argument( "-c", "--coverage", action="store_true", help="Run tests with coverage report" ) parser.add_argument( "-p", "--parallel", action="store_true", help="Run tests in parallel (requires pytest-xdist)" ) parser.add_argument( "-v", "--verbose", action="store_true", help="Verbose output" ) parser.add_argument( "-t", "--test", type=str, default="", help="Run specific test file or directory" ) parser.add_argument( "-m", "--markers", type=str, default="", help="Run tests with specific markers (e.g., 'unit', 'integration')" ) try: args = parser.parse_args() # Set options self.coverage = args.coverage self.parallel = args.parallel self.verbose = args.verbose self.markers = args.markers return True except SystemExit: # argparse already printed help, just exit return False except Exception as e: self.print_error(f"Error parsing arguments: {e}") return False def run(self) -> int: """Main execution method""" # Parse command line arguments if not self.parse_arguments(): return 1 # Run tests success = self.run_tests() return 0 if success else 1 def main(): """Entry point""" runner = TestRunner() return runner.run() if __name__ == "__main__": sys.exit(main())
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/check_comment_ascii.py
check_comment_ascii.py
#!/usr/bin/env python3 """ Check whether given python files contain non-ASCII comments. How to check the whole git repo: ``` $ git ls-files -z -- '*.py' | xargs -0 python3 check_comment_ascii.py ``` """ import sys import tokenize import ast import pathlib import re ASCII = re.compile(r"^[\n -~]*\Z") # Printable ASCII + newline def check(src: str, name: str) -> int: """ docstring line 1 docstring line 2 """ ok = 1 # A common comment begins with `#` with tokenize.open(src) as fp: for tk in tokenize.generate_tokens(fp.readline): if tk.type == tokenize.COMMENT and not ASCII.fullmatch(tk.string): print(f"{name}:{tk.start[0]}: non-ASCII comment: {tk.string}") ok = 0 # A docstring begins and ends with `'''` for node in ast.walk(ast.parse(pathlib.Path(src).read_text(), filename=name)): if isinstance(node, (ast.FunctionDef, ast.ClassDef, ast.Module)): if (doc := ast.get_docstring(node)) and not ASCII.fullmatch(doc): print(f"{name}:{node.lineno}: non-ASCII docstring: {doc}") ok = 0 return ok if __name__ == "__main__": status = 0 for file in sys.argv[1:]: if not check(file, file): status = 1 sys.exit(status)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/agentic_reasoning/prompts.py
agentic_reasoning/prompts.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # BEGIN_SEARCH_QUERY = "<|begin_search_query|>" END_SEARCH_QUERY = "<|end_search_query|>" BEGIN_SEARCH_RESULT = "<|begin_search_result|>" END_SEARCH_RESULT = "<|end_search_result|>" MAX_SEARCH_LIMIT = 6 REASON_PROMPT = f"""You are an advanced reasoning agent. Your goal is to answer the user's question by breaking it down into a series of verifiable steps. You have access to a powerful search tool to find information. **Your Task:** 1. Analyze the user's question. 2. If you need information, issue a search query to find a specific fact. 3. Review the search results. 4. Repeat the search process until you have all the facts needed to answer the question. 5. Once you have gathered sufficient information, synthesize the facts and provide the final answer directly. **Tool Usage:** - To search, you MUST write your query between the special tokens: {BEGIN_SEARCH_QUERY}your query{END_SEARCH_QUERY}. - The system will provide results between {BEGIN_SEARCH_RESULT}search results{END_SEARCH_RESULT}. - You have a maximum of {MAX_SEARCH_LIMIT} search attempts. --- **Example 1: Multi-hop Question** **Question:** "Are both the directors of Jaws and Casino Royale from the same country?" **Your Thought Process & Actions:** First, I need to identify the director of Jaws. {BEGIN_SEARCH_QUERY}who is the director of Jaws?{END_SEARCH_QUERY} [System returns search results] {BEGIN_SEARCH_RESULT} Jaws is a 1975 American thriller film directed by Steven Spielberg. {END_SEARCH_RESULT} Okay, the director of Jaws is Steven Spielberg. Now I need to find out his nationality. {BEGIN_SEARCH_QUERY}where is Steven Spielberg from?{END_SEARCH_QUERY} [System returns search results] {BEGIN_SEARCH_RESULT} Steven Allan Spielberg is an American filmmaker. Born in Cincinnati, Ohio... {END_SEARCH_RESULT} So, Steven Spielberg is from the USA. Next, I need to find the director of Casino Royale. {BEGIN_SEARCH_QUERY}who is the director of Casino Royale 2006?{END_SEARCH_QUERY} [System returns search results] {BEGIN_SEARCH_RESULT} Casino Royale is a 2006 spy film directed by Martin Campbell. {END_SEARCH_RESULT} The director of Casino Royale is Martin Campbell. Now I need his nationality. {BEGIN_SEARCH_QUERY}where is Martin Campbell from?{END_SEARCH_QUERY} [System returns search results] {BEGIN_SEARCH_RESULT} Martin Campbell (born 24 October 1943) is a New Zealand film and television director. {END_SEARCH_RESULT} I have all the information. Steven Spielberg is from the USA, and Martin Campbell is from New Zealand. They are not from the same country. Final Answer: No, the directors of Jaws and Casino Royale are not from the same country. Steven Spielberg is from the USA, and Martin Campbell is from New Zealand. --- **Example 2: Simple Fact Retrieval** **Question:** "When was the founder of craigslist born?" **Your Thought Process & Actions:** First, I need to know who founded craigslist. {BEGIN_SEARCH_QUERY}who founded craigslist?{END_SEARCH_QUERY} [System returns search results] {BEGIN_SEARCH_RESULT} Craigslist was founded in 1995 by Craig Newmark. {END_SEARCH_RESULT} The founder is Craig Newmark. Now I need his birth date. {BEGIN_SEARCH_QUERY}when was Craig Newmark born?{END_SEARCH_QUERY} [System returns search results] {BEGIN_SEARCH_RESULT} Craig Newmark was born on December 6, 1952. {END_SEARCH_RESULT} I have found the answer. Final Answer: The founder of craigslist, Craig Newmark, was born on December 6, 1952. --- **Important Rules:** - **One Fact at a Time:** Decompose the problem and issue one search query at a time to find a single, specific piece of information. - **Be Precise:** Formulate clear and precise search queries. If a search fails, rephrase it. - **Synthesize at the End:** Do not provide the final answer until you have completed all necessary searches. - **Language Consistency:** Your search queries should be in the same language as the user's question. Now, begin your work. Please answer the following question by thinking step-by-step. """ RELEVANT_EXTRACTION_PROMPT = """You are a highly efficient information extraction module. Your sole purpose is to extract the single most relevant piece of information from the provided `Searched Web Pages` that directly answers the `Current Search Query`. **Your Task:** 1. Read the `Current Search Query` to understand what specific information is needed. 2. Scan the `Searched Web Pages` to find the answer to that query. 3. Extract only the essential, factual information that answers the query. Be concise. **Context (For Your Information Only):** The `Previous Reasoning Steps` are provided to give you context on the overall goal, but your primary focus MUST be on answering the `Current Search Query`. Do not use information from the previous steps in your output. **Output Format:** Your response must follow one of two formats precisely. 1. **If a direct and relevant answer is found:** - Start your response immediately with `Final Information`. - Provide only the extracted fact(s). Do not add any extra conversational text. *Example:* `Current Search Query`: Where is Martin Campbell from? `Searched Web Pages`: [Long article snippet about Martin Campbell's career, which includes the sentence "Martin Campbell (born 24 October 1943) is a New Zealand film and television director..."] *Your Output:* Final Information Martin Campbell is a New Zealand film and television director. 2. **If no relevant answer that directly addresses the query is found in the web pages:** - Start your response immediately with `Final Information`. - Write the exact phrase: `No helpful information found.` --- **BEGIN TASK** **Inputs:** - **Previous Reasoning Steps:** {prev_reasoning} - **Current Search Query:** {search_query} - **Searched Web Pages:** {document} """
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/agentic_reasoning/__init__.py
agentic_reasoning/__init__.py
from .deep_research import DeepResearcher as DeepResearcher
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/agentic_reasoning/deep_research.py
agentic_reasoning/deep_research.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import re from functools import partial from agentic_reasoning.prompts import BEGIN_SEARCH_QUERY, BEGIN_SEARCH_RESULT, END_SEARCH_RESULT, MAX_SEARCH_LIMIT, \ END_SEARCH_QUERY, REASON_PROMPT, RELEVANT_EXTRACTION_PROMPT from api.db.services.llm_service import LLMBundle from rag.nlp import extract_between from rag.prompts import kb_prompt from rag.utils.tavily_conn import Tavily class DeepResearcher: def __init__(self, chat_mdl: LLMBundle, prompt_config: dict, kb_retrieve: partial = None, kg_retrieve: partial = None ): self.chat_mdl = chat_mdl self.prompt_config = prompt_config self._kb_retrieve = kb_retrieve self._kg_retrieve = kg_retrieve def _remove_tags(text: str, start_tag: str, end_tag: str) -> str: """General Tag Removal Method""" pattern = re.escape(start_tag) + r"(.*?)" + re.escape(end_tag) return re.sub(pattern, "", text) @staticmethod def _remove_query_tags(text: str) -> str: """Remove Query Tags""" return DeepResearcher._remove_tags(text, BEGIN_SEARCH_QUERY, END_SEARCH_QUERY) @staticmethod def _remove_result_tags(text: str) -> str: """Remove Result Tags""" return DeepResearcher._remove_tags(text, BEGIN_SEARCH_RESULT, END_SEARCH_RESULT) async def _generate_reasoning(self, msg_history): """Generate reasoning steps""" query_think = "" if msg_history[-1]["role"] != "user": msg_history.append({"role": "user", "content": "Continues reasoning with the new information.\n"}) else: msg_history[-1]["content"] += "\n\nContinues reasoning with the new information.\n" async for ans in self.chat_mdl.async_chat_streamly(REASON_PROMPT, msg_history, {"temperature": 0.7}): ans = re.sub(r"^.*</think>", "", ans, flags=re.DOTALL) if not ans: continue query_think = ans yield query_think query_think = "" yield query_think def _extract_search_queries(self, query_think, question, step_index): """Extract search queries from thinking""" queries = extract_between(query_think, BEGIN_SEARCH_QUERY, END_SEARCH_QUERY) if not queries and step_index == 0: # If this is the first step and no queries are found, use the original question as the query queries = [question] return queries def _truncate_previous_reasoning(self, all_reasoning_steps): """Truncate previous reasoning steps to maintain a reasonable length""" truncated_prev_reasoning = "" for i, step in enumerate(all_reasoning_steps): truncated_prev_reasoning += f"Step {i + 1}: {step}\n\n" prev_steps = truncated_prev_reasoning.split('\n\n') if len(prev_steps) <= 5: truncated_prev_reasoning = '\n\n'.join(prev_steps) else: truncated_prev_reasoning = '' for i, step in enumerate(prev_steps): if i == 0 or i >= len(prev_steps) - 4 or BEGIN_SEARCH_QUERY in step or BEGIN_SEARCH_RESULT in step: truncated_prev_reasoning += step + '\n\n' else: if truncated_prev_reasoning[-len('\n\n...\n\n'):] != '\n\n...\n\n': truncated_prev_reasoning += '...\n\n' return truncated_prev_reasoning.strip('\n') def _retrieve_information(self, search_query): """Retrieve information from different sources""" # 1. Knowledge base retrieval kbinfos = [] try: kbinfos = self._kb_retrieve(question=search_query) if self._kb_retrieve else {"chunks": [], "doc_aggs": []} except Exception as e: logging.error(f"Knowledge base retrieval error: {e}") # 2. Web retrieval (if Tavily API is configured) try: if self.prompt_config.get("tavily_api_key"): tav = Tavily(self.prompt_config["tavily_api_key"]) tav_res = tav.retrieve_chunks(search_query) kbinfos["chunks"].extend(tav_res["chunks"]) kbinfos["doc_aggs"].extend(tav_res["doc_aggs"]) except Exception as e: logging.error(f"Web retrieval error: {e}") # 3. Knowledge graph retrieval (if configured) try: if self.prompt_config.get("use_kg") and self._kg_retrieve: ck = self._kg_retrieve(question=search_query) if ck["content_with_weight"]: kbinfos["chunks"].insert(0, ck) except Exception as e: logging.error(f"Knowledge graph retrieval error: {e}") return kbinfos def _update_chunk_info(self, chunk_info, kbinfos): """Update chunk information for citations""" if not chunk_info["chunks"]: # If this is the first retrieval, use the retrieval results directly for k in chunk_info.keys(): chunk_info[k] = kbinfos[k] else: # Merge newly retrieved information, avoiding duplicates cids = [c["chunk_id"] for c in chunk_info["chunks"]] for c in kbinfos["chunks"]: if c["chunk_id"] not in cids: chunk_info["chunks"].append(c) dids = [d["doc_id"] for d in chunk_info["doc_aggs"]] for d in kbinfos["doc_aggs"]: if d["doc_id"] not in dids: chunk_info["doc_aggs"].append(d) async def _extract_relevant_info(self, truncated_prev_reasoning, search_query, kbinfos): """Extract and summarize relevant information""" summary_think = "" async for ans in self.chat_mdl.async_chat_streamly( RELEVANT_EXTRACTION_PROMPT.format( prev_reasoning=truncated_prev_reasoning, search_query=search_query, document="\n".join(kb_prompt(kbinfos, 4096)) ), [{"role": "user", "content": f'Now you should analyze each web page and find helpful information based on the current search query "{search_query}" and previous reasoning steps.'}], {"temperature": 0.7}): ans = re.sub(r"^.*</think>", "", ans, flags=re.DOTALL) if not ans: continue summary_think = ans yield summary_think summary_think = "" yield summary_think async def thinking(self, chunk_info: dict, question: str): executed_search_queries = [] msg_history = [{"role": "user", "content": f'Question:\"{question}\"\n'}] all_reasoning_steps = [] think = "<think>" for step_index in range(MAX_SEARCH_LIMIT + 1): # Check if the maximum search limit has been reached if step_index == MAX_SEARCH_LIMIT - 1: summary_think = f"\n{BEGIN_SEARCH_RESULT}\nThe maximum search limit is exceeded. You are not allowed to search.\n{END_SEARCH_RESULT}\n" yield {"answer": think + summary_think + "</think>", "reference": {}, "audio_binary": None} all_reasoning_steps.append(summary_think) msg_history.append({"role": "assistant", "content": summary_think}) break # Step 1: Generate reasoning query_think = "" async for ans in self._generate_reasoning(msg_history): query_think = ans yield {"answer": think + self._remove_query_tags(query_think) + "</think>", "reference": {}, "audio_binary": None} think += self._remove_query_tags(query_think) all_reasoning_steps.append(query_think) # Step 2: Extract search queries queries = self._extract_search_queries(query_think, question, step_index) if not queries and step_index > 0: # If not the first step and no queries, end the search process break # Process each search query for search_query in queries: logging.info(f"[THINK]Query: {step_index}. {search_query}") msg_history.append({"role": "assistant", "content": search_query}) think += f"\n\n> {step_index + 1}. {search_query}\n\n" yield {"answer": think + "</think>", "reference": {}, "audio_binary": None} # Check if the query has already been executed if search_query in executed_search_queries: summary_think = f"\n{BEGIN_SEARCH_RESULT}\nYou have searched this query. Please refer to previous results.\n{END_SEARCH_RESULT}\n" yield {"answer": think + summary_think + "</think>", "reference": {}, "audio_binary": None} all_reasoning_steps.append(summary_think) msg_history.append({"role": "user", "content": summary_think}) think += summary_think continue executed_search_queries.append(search_query) # Step 3: Truncate previous reasoning steps truncated_prev_reasoning = self._truncate_previous_reasoning(all_reasoning_steps) # Step 4: Retrieve information kbinfos = self._retrieve_information(search_query) # Step 5: Update chunk information self._update_chunk_info(chunk_info, kbinfos) # Step 6: Extract relevant information think += "\n\n" summary_think = "" async for ans in self._extract_relevant_info(truncated_prev_reasoning, search_query, kbinfos): summary_think = ans yield {"answer": think + self._remove_result_tags(summary_think) + "</think>", "reference": {}, "audio_binary": None} all_reasoning_steps.append(summary_think) msg_history.append( {"role": "user", "content": f"\n\n{BEGIN_SEARCH_RESULT}{summary_think}{END_SEARCH_RESULT}\n\n"}) think += self._remove_result_tags(summary_think) logging.info(f"[THINK]Summary: {step_index}. {summary_think}") yield think + "</think>"
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/plugin/common.py
plugin/common.py
PLUGIN_TYPE_LLM_TOOLS = "llm_tools"
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/plugin/llm_tool_plugin.py
plugin/llm_tool_plugin.py
from typing import Any, TypedDict import pluginlib from .common import PLUGIN_TYPE_LLM_TOOLS class LLMToolParameter(TypedDict): type: str description: str displayDescription: str required: bool class LLMToolMetadata(TypedDict): name: str displayName: str description: str displayDescription: str parameters: dict[str, LLMToolParameter] @pluginlib.Parent(PLUGIN_TYPE_LLM_TOOLS) class LLMToolPlugin: @classmethod @pluginlib.abstractmethod def get_metadata(cls) -> LLMToolMetadata: pass def invoke(self, **kwargs) -> str: raise NotImplementedError def llm_tool_metadata_to_openai_tool(llm_tool_metadata: LLMToolMetadata) -> dict[str, Any]: return { "type": "function", "function": { "name": llm_tool_metadata["name"], "description": llm_tool_metadata["description"], "parameters": { "type": "object", "properties": { k: { "type": p["type"], "description": p["description"] } for k, p in llm_tool_metadata["parameters"].items() }, "required": [k for k, p in llm_tool_metadata["parameters"].items() if p["required"]] } } }
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/plugin/__init__.py
plugin/__init__.py
from .plugin_manager import PluginManager GlobalPluginManager = PluginManager()
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/plugin/plugin_manager.py
plugin/plugin_manager.py
import logging import os from pathlib import Path import pluginlib from .common import PLUGIN_TYPE_LLM_TOOLS from .llm_tool_plugin import LLMToolPlugin class PluginManager: _llm_tool_plugins: dict[str, LLMToolPlugin] def __init__(self) -> None: self._llm_tool_plugins = {} def load_plugins(self) -> None: loader = pluginlib.PluginLoader( paths=[str(Path(os.path.dirname(__file__), "embedded_plugins"))] ) for type, plugins in loader.plugins.items(): for name, plugin in plugins.items(): logging.info(f"Loaded {type} plugin {name} version {plugin.version}") if type == PLUGIN_TYPE_LLM_TOOLS: metadata = plugin.get_metadata() self._llm_tool_plugins[metadata["name"]] = plugin def get_llm_tools(self) -> list[LLMToolPlugin]: return list(self._llm_tool_plugins.values()) def get_llm_tool_by_name(self, name: str) -> LLMToolPlugin | None: return self._llm_tool_plugins.get(name) def get_llm_tools_by_names(self, tool_names: list[str]) -> list[LLMToolPlugin]: results = [] for name in tool_names: plugin = self._llm_tool_plugins.get(name) if plugin is not None: results.append(plugin) return results
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/plugin/embedded_plugins/llm_tools/bad_calculator.py
plugin/embedded_plugins/llm_tools/bad_calculator.py
import logging from plugin.llm_tool_plugin import LLMToolMetadata, LLMToolPlugin class BadCalculatorPlugin(LLMToolPlugin): """ A sample LLM tool plugin, will add two numbers with 100. It only presents for demo purpose. Do not use it in production. """ _version_ = "1.0.0" @classmethod def get_metadata(cls) -> LLMToolMetadata: return { "name": "bad_calculator", "displayName": "$t:bad_calculator.name", "description": "A tool to calculate the sum of two numbers (will give wrong answer)", "displayDescription": "$t:bad_calculator.description", "parameters": { "a": { "type": "number", "description": "The first number", "displayDescription": "$t:bad_calculator.params.a", "required": True }, "b": { "type": "number", "description": "The second number", "displayDescription": "$t:bad_calculator.params.b", "required": True } } } def invoke(self, a: int, b: int) -> str: logging.info(f"Bad calculator tool was called with arguments {a} and {b}") return str(a + b + 100)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/raptor.py
rag/raptor.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import asyncio import logging import re import numpy as np import umap from sklearn.mixture import GaussianMixture from api.db.services.task_service import has_canceled from common.connection_utils import timeout from common.exceptions import TaskCanceledException from common.token_utils import truncate from graphrag.utils import ( chat_limiter, get_embed_cache, get_llm_cache, set_embed_cache, set_llm_cache, ) class RecursiveAbstractiveProcessing4TreeOrganizedRetrieval: def __init__( self, max_cluster, llm_model, embd_model, prompt, max_token=512, threshold=0.1, max_errors=3, ): self._max_cluster = max_cluster self._llm_model = llm_model self._embd_model = embd_model self._threshold = threshold self._prompt = prompt self._max_token = max_token self._max_errors = max(1, max_errors) self._error_count = 0 @timeout(60 * 20) async def _chat(self, system, history, gen_conf): cached = await asyncio.to_thread(get_llm_cache, self._llm_model.llm_name, system, history, gen_conf) if cached: return cached last_exc = None for attempt in range(3): try: response = await self._llm_model.async_chat(system, history, gen_conf) response = re.sub(r"^.*</think>", "", response, flags=re.DOTALL) if response.find("**ERROR**") >= 0: raise Exception(response) await asyncio.to_thread(set_llm_cache,self._llm_model.llm_name,system,response,history,gen_conf) return response except Exception as exc: last_exc = exc logging.warning("RAPTOR LLM call failed on attempt %d/3: %s", attempt + 1, exc) if attempt < 2: await asyncio.sleep(1 + attempt) raise last_exc if last_exc else Exception("LLM chat failed without exception") @timeout(20) async def _embedding_encode(self, txt): response = await asyncio.to_thread(get_embed_cache, self._embd_model.llm_name, txt) if response is not None: return response embds, _ = await asyncio.to_thread(self._embd_model.encode, [txt]) if len(embds) < 1 or len(embds[0]) < 1: raise Exception("Embedding error: ") embds = embds[0] await asyncio.to_thread(set_embed_cache, self._embd_model.llm_name, txt, embds) return embds def _get_optimal_clusters(self, embeddings: np.ndarray, random_state: int, task_id: str = ""): max_clusters = min(self._max_cluster, len(embeddings)) n_clusters = np.arange(1, max_clusters) bics = [] for n in n_clusters: if task_id: if has_canceled(task_id): logging.info(f"Task {task_id} cancelled during get optimal clusters.") raise TaskCanceledException(f"Task {task_id} was cancelled") gm = GaussianMixture(n_components=n, random_state=random_state) gm.fit(embeddings) bics.append(gm.bic(embeddings)) optimal_clusters = n_clusters[np.argmin(bics)] return optimal_clusters async def __call__(self, chunks, random_state, callback=None, task_id: str = ""): if len(chunks) <= 1: return [] chunks = [(s, a) for s, a in chunks if s and a is not None and len(a) > 0] layers = [(0, len(chunks))] start, end = 0, len(chunks) @timeout(60 * 20) async def summarize(ck_idx: list[int]): nonlocal chunks if task_id: if has_canceled(task_id): logging.info(f"Task {task_id} cancelled during RAPTOR summarization.") raise TaskCanceledException(f"Task {task_id} was cancelled") texts = [chunks[i][0] for i in ck_idx] len_per_chunk = int((self._llm_model.max_length - self._max_token) / len(texts)) cluster_content = "\n".join([truncate(t, max(1, len_per_chunk)) for t in texts]) try: async with chat_limiter: if task_id and has_canceled(task_id): logging.info(f"Task {task_id} cancelled before RAPTOR LLM call.") raise TaskCanceledException(f"Task {task_id} was cancelled") cnt = await self._chat( "You're a helpful assistant.", [ { "role": "user", "content": self._prompt.format(cluster_content=cluster_content), } ], {"max_tokens": max(self._max_token, 512)}, # fix issue: #10235 ) cnt = re.sub( "(······\n由于长度的原因,回答被截断了,要继续吗?|For the content length reason, it stopped, continue?)", "", cnt, ) logging.debug(f"SUM: {cnt}") if task_id and has_canceled(task_id): logging.info(f"Task {task_id} cancelled before RAPTOR embedding.") raise TaskCanceledException(f"Task {task_id} was cancelled") embds = await self._embedding_encode(cnt) chunks.append((cnt, embds)) except TaskCanceledException: raise except Exception as exc: self._error_count += 1 warn_msg = f"[RAPTOR] Skip cluster ({len(ck_idx)} chunks) due to error: {exc}" logging.warning(warn_msg) if callback: callback(msg=warn_msg) if self._error_count >= self._max_errors: raise RuntimeError(f"RAPTOR aborted after {self._error_count} errors. Last error: {exc}") from exc labels = [] while end - start > 1: if task_id: if has_canceled(task_id): logging.info(f"Task {task_id} cancelled during RAPTOR layer processing.") raise TaskCanceledException(f"Task {task_id} was cancelled") embeddings = [embd for _, embd in chunks[start:end]] if len(embeddings) == 2: await summarize([start, start + 1]) if callback: callback(msg="Cluster one layer: {} -> {}".format(end - start, len(chunks) - end)) labels.extend([0, 0]) layers.append((end, len(chunks))) start = end end = len(chunks) continue n_neighbors = int((len(embeddings) - 1) ** 0.8) reduced_embeddings = umap.UMAP( n_neighbors=max(2, n_neighbors), n_components=min(12, len(embeddings) - 2), metric="cosine", ).fit_transform(embeddings) n_clusters = self._get_optimal_clusters(reduced_embeddings, random_state, task_id=task_id) if n_clusters == 1: lbls = [0 for _ in range(len(reduced_embeddings))] else: gm = GaussianMixture(n_components=n_clusters, random_state=random_state) gm.fit(reduced_embeddings) probs = gm.predict_proba(reduced_embeddings) lbls = [np.where(prob > self._threshold)[0] for prob in probs] lbls = [lbl[0] if isinstance(lbl, np.ndarray) else lbl for lbl in lbls] tasks = [] for c in range(n_clusters): ck_idx = [i + start for i in range(len(lbls)) if lbls[i] == c] assert len(ck_idx) > 0 if task_id and has_canceled(task_id): logging.info(f"Task {task_id} cancelled before RAPTOR cluster processing.") raise TaskCanceledException(f"Task {task_id} was cancelled") tasks.append(asyncio.create_task(summarize(ck_idx))) try: await asyncio.gather(*tasks, return_exceptions=False) except Exception as e: logging.error(f"Error in RAPTOR cluster processing: {e}") for t in tasks: t.cancel() await asyncio.gather(*tasks, return_exceptions=True) raise assert len(chunks) - end == n_clusters, "{} vs. {}".format(len(chunks) - end, n_clusters) labels.extend(lbls) layers.append((end, len(chunks))) if callback: callback(msg="Cluster one layer: {} -> {}".format(end - start, len(chunks) - end)) start = end end = len(chunks) return chunks
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/settings.py
rag/settings.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/benchmark.py
rag/benchmark.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json import os import sys import time import argparse from collections import defaultdict from common import settings from common.constants import LLMType from api.db.services.llm_service import LLMBundle from api.db.services.knowledgebase_service import KnowledgebaseService from common.misc_utils import get_uuid from rag.nlp import tokenize, search from ranx import evaluate from ranx import Qrels, Run import pandas as pd from tqdm import tqdm global max_docs max_docs = sys.maxsize class Benchmark: def __init__(self, kb_id): self.kb_id = kb_id e, self.kb = KnowledgebaseService.get_by_id(kb_id) self.similarity_threshold = self.kb.similarity_threshold self.vector_similarity_weight = self.kb.vector_similarity_weight self.embd_mdl = LLMBundle(self.kb.tenant_id, LLMType.EMBEDDING, llm_name=self.kb.embd_id, lang=self.kb.language) self.tenant_id = '' self.index_name = '' self.initialized_index = False def _get_retrieval(self, qrels): # Need to wait for the ES and Infinity index to be ready time.sleep(20) run = defaultdict(dict) query_list = list(qrels.keys()) for query in query_list: ranks = settings.retriever.retrieval(query, self.embd_mdl, self.tenant_id, [self.kb.id], 1, 30, 0.0, self.vector_similarity_weight) if len(ranks["chunks"]) == 0: print(f"deleted query: {query}") del qrels[query] continue for c in ranks["chunks"]: c.pop("vector", None) run[query][c["chunk_id"]] = c["similarity"] return run def embedding(self, docs): texts = [d["content_with_weight"] for d in docs] embeddings, _ = self.embd_mdl.encode(texts) assert len(docs) == len(embeddings) vector_size = 0 for i, d in enumerate(docs): v = embeddings[i] vector_size = len(v) d["q_%d_vec" % len(v)] = v return docs, vector_size def init_index(self, vector_size: int): if self.initialized_index: return if settings.docStoreConn.index_exist(self.index_name, self.kb_id): settings.docStoreConn.delete_idx(self.index_name, self.kb_id) settings.docStoreConn.create_idx(self.index_name, self.kb_id, vector_size) self.initialized_index = True def ms_marco_index(self, file_path, index_name): qrels = defaultdict(dict) texts = defaultdict(dict) docs_count = 0 docs = [] filelist = sorted(os.listdir(file_path)) for fn in filelist: if docs_count >= max_docs: break if not fn.endswith(".parquet"): continue data = pd.read_parquet(os.path.join(file_path, fn)) for i in tqdm(range(len(data)), colour="green", desc="Tokenizing:" + fn): if docs_count >= max_docs: break query = data.iloc[i]['query'] for rel, text in zip(data.iloc[i]['passages']['is_selected'], data.iloc[i]['passages']['passage_text']): d = { "id": get_uuid(), "kb_id": self.kb.id, "docnm_kwd": "xxxxx", "doc_id": "ksksks" } tokenize(d, text, "english") docs.append(d) texts[d["id"]] = text qrels[query][d["id"]] = int(rel) if len(docs) >= 32: docs_count += len(docs) docs, vector_size = self.embedding(docs) self.init_index(vector_size) settings.docStoreConn.insert(docs, self.index_name, self.kb_id) docs = [] if docs: docs, vector_size = self.embedding(docs) self.init_index(vector_size) settings.docStoreConn.insert(docs, self.index_name, self.kb_id) return qrels, texts def trivia_qa_index(self, file_path, index_name): qrels = defaultdict(dict) texts = defaultdict(dict) docs_count = 0 docs = [] filelist = sorted(os.listdir(file_path)) for fn in filelist: if docs_count >= max_docs: break if not fn.endswith(".parquet"): continue data = pd.read_parquet(os.path.join(file_path, fn)) for i in tqdm(range(len(data)), colour="green", desc="Indexing:" + fn): if docs_count >= max_docs: break query = data.iloc[i]['question'] for rel, text in zip(data.iloc[i]["search_results"]['rank'], data.iloc[i]["search_results"]['search_context']): d = { "id": get_uuid(), "kb_id": self.kb.id, "docnm_kwd": "xxxxx", "doc_id": "ksksks" } tokenize(d, text, "english") docs.append(d) texts[d["id"]] = text qrels[query][d["id"]] = int(rel) if len(docs) >= 32: docs_count += len(docs) docs, vector_size = self.embedding(docs) self.init_index(vector_size) settings.docStoreConn.insert(docs,self.index_name) docs = [] docs, vector_size = self.embedding(docs) self.init_index(vector_size) settings.docStoreConn.insert(docs, self.index_name) return qrels, texts def miracl_index(self, file_path, corpus_path, index_name): corpus_total = {} for corpus_file in os.listdir(corpus_path): tmp_data = pd.read_json(os.path.join(corpus_path, corpus_file), lines=True) for index, i in tmp_data.iterrows(): corpus_total[i['docid']] = i['text'] topics_total = {} for topics_file in os.listdir(os.path.join(file_path, 'topics')): if 'test' in topics_file: continue tmp_data = pd.read_csv(os.path.join(file_path, 'topics', topics_file), sep='\t', names=['qid', 'query']) for index, i in tmp_data.iterrows(): topics_total[i['qid']] = i['query'] qrels = defaultdict(dict) texts = defaultdict(dict) docs_count = 0 docs = [] for qrels_file in os.listdir(os.path.join(file_path, 'qrels')): if 'test' in qrels_file: continue if docs_count >= max_docs: break tmp_data = pd.read_csv(os.path.join(file_path, 'qrels', qrels_file), sep='\t', names=['qid', 'Q0', 'docid', 'relevance']) for i in tqdm(range(len(tmp_data)), colour="green", desc="Indexing:" + qrels_file): if docs_count >= max_docs: break query = topics_total[tmp_data.iloc[i]['qid']] text = corpus_total[tmp_data.iloc[i]['docid']] rel = tmp_data.iloc[i]['relevance'] d = { "id": get_uuid(), "kb_id": self.kb.id, "docnm_kwd": "xxxxx", "doc_id": "ksksks" } tokenize(d, text, 'english') docs.append(d) texts[d["id"]] = text qrels[query][d["id"]] = int(rel) if len(docs) >= 32: docs_count += len(docs) docs, vector_size = self.embedding(docs) self.init_index(vector_size) settings.docStoreConn.insert(docs, self.index_name) docs = [] docs, vector_size = self.embedding(docs) self.init_index(vector_size) settings.docStoreConn.insert(docs, self.index_name) return qrels, texts def save_results(self, qrels, run, texts, dataset, file_path): keep_result = [] run_keys = list(run.keys()) for run_i in tqdm(range(len(run_keys)), desc="Calculating ndcg@10 for single query"): key = run_keys[run_i] keep_result.append({'query': key, 'qrel': qrels[key], 'run': run[key], 'ndcg@10': evaluate({key: qrels[key]}, {key: run[key]}, "ndcg@10")}) keep_result = sorted(keep_result, key=lambda kk: kk['ndcg@10']) with open(os.path.join(file_path, dataset + 'result.md'), 'w', encoding='utf-8') as f: f.write('## Score For Every Query\n') for keep_result_i in keep_result: f.write('### query: ' + keep_result_i['query'] + ' ndcg@10:' + str(keep_result_i['ndcg@10']) + '\n') scores = [[i[0], i[1]] for i in keep_result_i['run'].items()] scores = sorted(scores, key=lambda kk: kk[1]) for score in scores[:10]: f.write('- text: ' + str(texts[score[0]]) + '\t qrel: ' + str(score[1]) + '\n') json.dump(qrels, open(os.path.join(file_path, dataset + '.qrels.json'), "w+", encoding='utf-8'), indent=2) json.dump(run, open(os.path.join(file_path, dataset + '.run.json'), "w+", encoding='utf-8'), indent=2) print(os.path.join(file_path, dataset + '_result.md'), 'Saved!') def __call__(self, dataset, file_path, miracl_corpus=''): if dataset == "ms_marco_v1.1": self.tenant_id = "benchmark_ms_marco_v11" self.index_name = search.index_name(self.tenant_id) qrels, texts = self.ms_marco_index(file_path, "benchmark_ms_marco_v1.1") run = self._get_retrieval(qrels) print(dataset, evaluate(Qrels(qrels), Run(run), ["ndcg@10", "map@5", "mrr@10"])) self.save_results(qrels, run, texts, dataset, file_path) if dataset == "trivia_qa": self.tenant_id = "benchmark_trivia_qa" self.index_name = search.index_name(self.tenant_id) qrels, texts = self.trivia_qa_index(file_path, "benchmark_trivia_qa") run = self._get_retrieval(qrels) print(dataset, evaluate(Qrels(qrels), Run(run), ["ndcg@10", "map@5", "mrr@10"])) self.save_results(qrels, run, texts, dataset, file_path) if dataset == "miracl": for lang in ['ar', 'bn', 'de', 'en', 'es', 'fa', 'fi', 'fr', 'hi', 'id', 'ja', 'ko', 'ru', 'sw', 'te', 'th', 'yo', 'zh']: if not os.path.isdir(os.path.join(file_path, 'miracl-v1.0-' + lang)): print('Directory: ' + os.path.join(file_path, 'miracl-v1.0-' + lang) + ' not found!') continue if not os.path.isdir(os.path.join(file_path, 'miracl-v1.0-' + lang, 'qrels')): print('Directory: ' + os.path.join(file_path, 'miracl-v1.0-' + lang, 'qrels') + 'not found!') continue if not os.path.isdir(os.path.join(file_path, 'miracl-v1.0-' + lang, 'topics')): print('Directory: ' + os.path.join(file_path, 'miracl-v1.0-' + lang, 'topics') + 'not found!') continue if not os.path.isdir(os.path.join(miracl_corpus, 'miracl-corpus-v1.0-' + lang)): print('Directory: ' + os.path.join(miracl_corpus, 'miracl-corpus-v1.0-' + lang) + ' not found!') continue self.tenant_id = "benchmark_miracl_" + lang self.index_name = search.index_name(self.tenant_id) self.initialized_index = False qrels, texts = self.miracl_index(os.path.join(file_path, 'miracl-v1.0-' + lang), os.path.join(miracl_corpus, 'miracl-corpus-v1.0-' + lang), "benchmark_miracl_" + lang) run = self._get_retrieval(qrels) print(dataset, evaluate(Qrels(qrels), Run(run), ["ndcg@10", "map@5", "mrr@10"])) self.save_results(qrels, run, texts, dataset, file_path) if __name__ == '__main__': print('*****************RAGFlow Benchmark*****************') parser = argparse.ArgumentParser(usage="benchmark.py <max_docs> <kb_id> <dataset> <dataset_path> [<miracl_corpus_path>])", description='RAGFlow Benchmark') parser.add_argument('max_docs', metavar='max_docs', type=int, help='max docs to evaluate') parser.add_argument('kb_id', metavar='kb_id', help='dataset id') parser.add_argument('dataset', metavar='dataset', help='dataset name, shall be one of ms_marco_v1.1(https://huggingface.co/datasets/microsoft/ms_marco), trivia_qa(https://huggingface.co/datasets/mandarjoshi/trivia_qa>), miracl(https://huggingface.co/datasets/miracl/miracl') parser.add_argument('dataset_path', metavar='dataset_path', help='dataset path') parser.add_argument('miracl_corpus_path', metavar='miracl_corpus_path', nargs='?', default="", help='miracl corpus path. Only needed when dataset is miracl') args = parser.parse_args() max_docs = args.max_docs kb_id = args.kb_id ex = Benchmark(kb_id) dataset = args.dataset dataset_path = args.dataset_path if dataset == "ms_marco_v1.1" or dataset == "trivia_qa": ex(dataset, dataset_path) elif dataset == "miracl": if len(args) < 5: print('Please input the correct parameters!') exit(1) miracl_corpus_path = args[4] ex(dataset, dataset_path, miracl_corpus=args.miracl_corpus_path) else: print("Dataset: ", dataset, "not supported!")
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/__init__.py
rag/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # from beartype.claw import beartype_this_package # beartype_this_package()
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/table.py
rag/app/table.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import copy import csv import io import logging import re from io import BytesIO from xpinyin import Pinyin import numpy as np import pandas as pd from collections import Counter # from openpyxl import load_workbook, Workbook from dateutil.parser import parse as datetime_parse from api.db.services.knowledgebase_service import KnowledgebaseService from deepdoc.parser.figure_parser import vision_figure_parser_figure_xlsx_wrapper from deepdoc.parser.utils import get_text from rag.nlp import rag_tokenizer, tokenize, tokenize_table from deepdoc.parser import ExcelParser class Excel(ExcelParser): def __call__(self, fnm, binary=None, from_page=0, to_page=10000000000, callback=None, **kwargs): if not binary: wb = Excel._load_excel_to_workbook(fnm) else: wb = Excel._load_excel_to_workbook(BytesIO(binary)) total = 0 for sheet_name in wb.sheetnames: total += len(list(wb[sheet_name].rows)) res, fails, done = [], [], 0 rn = 0 flow_images = [] pending_cell_images = [] tables = [] for sheet_name in wb.sheetnames: ws = wb[sheet_name] images = Excel._extract_images_from_worksheet(ws, sheetname=sheet_name) if images: image_descriptions = vision_figure_parser_figure_xlsx_wrapper(images=images, callback=callback, **kwargs) if image_descriptions and len(image_descriptions) == len(images): for i, bf in enumerate(image_descriptions): images[i]["image_description"] = "\n".join(bf[0][1]) for img in images: if img["span_type"] == "single_cell" and img.get("image_description"): pending_cell_images.append(img) else: flow_images.append(img) try: rows = list(ws.rows) except Exception as e: logging.warning(f"Skip sheet '{sheet_name}' due to rows access error: {e}") continue if not rows: continue headers, header_rows = self._parse_headers(ws, rows) if not headers: continue data = [] for i, r in enumerate(rows[header_rows:]): rn += 1 if rn - 1 < from_page: continue if rn - 1 >= to_page: break row_data = self._extract_row_data(ws, r, header_rows + i, len(headers)) if row_data is None: fails.append(str(i)) continue if self._is_empty_row(row_data): continue data.append(row_data) done += 1 if len(data) == 0: continue df = pd.DataFrame(data, columns=headers) for img in pending_cell_images: excel_row = img["row_from"] - 1 excel_col = img["col_from"] - 1 df_row_idx = excel_row - header_rows if df_row_idx < 0 or df_row_idx >= len(df): flow_images.append(img) continue if excel_col < 0 or excel_col >= len(df.columns): flow_images.append(img) continue col_name = df.columns[excel_col] if not df.iloc[df_row_idx][col_name]: df.iat[df_row_idx, excel_col] = img["image_description"] res.append(df) for img in flow_images: tables.append( ( ( img["image"], # Image.Image [img["image_description"]] # description list (must be list) ), [ (0, 0, 0, 0, 0) # dummy position ] ) ) callback(0.3, ("Extract records: {}~{}".format(from_page + 1, min(to_page, from_page + rn)) + ( f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) return res, tables def _parse_headers(self, ws, rows): if len(rows) == 0: return [], 0 has_complex_structure = self._has_complex_header_structure(ws, rows) if has_complex_structure: return self._parse_multi_level_headers(ws, rows) else: return self._parse_simple_headers(rows) def _has_complex_header_structure(self, ws, rows): if len(rows) < 1: return False merged_ranges = list(ws.merged_cells.ranges) # 检查前两行是否涉及合并单元格 for rng in merged_ranges: if rng.min_row <= 2: # 只要合并区域涉及第1或第2行 return True return False def _row_looks_like_header(self, row): header_like_cells = 0 data_like_cells = 0 non_empty_cells = 0 for cell in row: if cell.value is not None: non_empty_cells += 1 val = str(cell.value).strip() if self._looks_like_header(val): header_like_cells += 1 elif self._looks_like_data(val): data_like_cells += 1 if non_empty_cells == 0: return False return header_like_cells >= data_like_cells def _parse_simple_headers(self, rows): if not rows: return [], 0 header_row = rows[0] headers = [] for cell in header_row: if cell.value is not None: header_value = str(cell.value).strip() if header_value: headers.append(header_value) else: pass final_headers = [] for i, cell in enumerate(header_row): if cell.value is not None: header_value = str(cell.value).strip() if header_value: final_headers.append(header_value) else: final_headers.append(f"Column_{i + 1}") else: final_headers.append(f"Column_{i + 1}") return final_headers, 1 def _parse_multi_level_headers(self, ws, rows): if len(rows) < 2: return [], 0 header_rows = self._detect_header_rows(rows) if header_rows == 1: return self._parse_simple_headers(rows) else: return self._build_hierarchical_headers(ws, rows, header_rows), header_rows def _detect_header_rows(self, rows): if len(rows) < 2: return 1 header_rows = 1 max_check_rows = min(5, len(rows)) for i in range(1, max_check_rows): row = rows[i] if self._row_looks_like_header(row): header_rows = i + 1 else: break return header_rows def _looks_like_header(self, value): if len(value) < 1: return False if any(ord(c) > 127 for c in value): return True if len([c for c in value if c.isalpha()]) >= 2: return True if any(c in value for c in ["(", ")", ":", ":", "(", ")", "_", "-"]): return True return False def _looks_like_data(self, value): if len(value) == 1 and value.upper() in ["Y", "N", "M", "X", "/", "-"]: return True if value.replace(".", "").replace("-", "").replace(",", "").isdigit(): return True if value.startswith("0x") and len(value) <= 10: return True return False def _build_hierarchical_headers(self, ws, rows, header_rows): headers = [] max_col = max(len(row) for row in rows[:header_rows]) if header_rows > 0 else 0 merged_ranges = list(ws.merged_cells.ranges) for col_idx in range(max_col): header_parts = [] for row_idx in range(header_rows): if col_idx < len(rows[row_idx]): cell_value = rows[row_idx][col_idx].value merged_value = self._get_merged_cell_value(ws, row_idx + 1, col_idx + 1, merged_ranges) if merged_value is not None: cell_value = merged_value if cell_value is not None: cell_value = str(cell_value).strip() if cell_value and cell_value not in header_parts and self._is_valid_header_part(cell_value): header_parts.append(cell_value) if header_parts: header = "-".join(header_parts) headers.append(header) else: headers.append(f"Column_{col_idx + 1}") final_headers = [h for h in headers if h and h != "-"] return final_headers def _is_valid_header_part(self, value): if len(value) == 1 and value.upper() in ["Y", "N", "M", "X"]: return False if value.replace(".", "").replace("-", "").replace(",", "").isdigit(): return False if value in ["/", "-", "+", "*", "="]: return False return True def _get_merged_cell_value(self, ws, row, col, merged_ranges): for merged_range in merged_ranges: if merged_range.min_row <= row <= merged_range.max_row and merged_range.min_col <= col <= merged_range.max_col: return ws.cell(merged_range.min_row, merged_range.min_col).value return None def _extract_row_data(self, ws, row, absolute_row_idx, expected_cols): row_data = [] merged_ranges = list(ws.merged_cells.ranges) actual_row_num = absolute_row_idx + 1 for col_idx in range(expected_cols): cell_value = None actual_col_num = col_idx + 1 try: cell_value = ws.cell(row=actual_row_num, column=actual_col_num).value except ValueError: if col_idx < len(row): cell_value = row[col_idx].value if cell_value is None: merged_value = self._get_merged_cell_value(ws, actual_row_num, actual_col_num, merged_ranges) if merged_value is not None: cell_value = merged_value else: cell_value = self._get_inherited_value(ws, actual_row_num, actual_col_num, merged_ranges) row_data.append(cell_value) return row_data def _get_inherited_value(self, ws, row, col, merged_ranges): for merged_range in merged_ranges: if merged_range.min_row <= row <= merged_range.max_row and merged_range.min_col <= col <= merged_range.max_col: return ws.cell(merged_range.min_row, merged_range.min_col).value return None def _is_empty_row(self, row_data): for val in row_data: if val is not None and str(val).strip() != "": return False return True def trans_datatime(s): try: return datetime_parse(s.strip()).strftime("%Y-%m-%d %H:%M:%S") except Exception as e: logging.warning(f"Failed to parse date from {s}, error: {e}") pass def trans_bool(s): if re.match(r"(true|yes|是|\*|✓|✔|☑|✅|√)$", str(s).strip(), flags=re.IGNORECASE): return "yes" if re.match(r"(false|no|否|⍻|×)$", str(s).strip(), flags=re.IGNORECASE): return "no" return None def column_data_type(arr): arr = list(arr) counts = {"int": 0, "float": 0, "text": 0, "datetime": 0, "bool": 0} trans = {t: f for f, t in [(int, "int"), (float, "float"), (trans_datatime, "datetime"), (trans_bool, "bool"), (str, "text")]} float_flag = False for a in arr: if a is None: continue if re.match(r"[+-]?[0-9]+$", str(a).replace("%%", "")) and not str(a).replace("%%", "").startswith("0"): counts["int"] += 1 if int(str(a)) > 2 ** 63 - 1: float_flag = True break elif re.match(r"[+-]?[0-9.]{,19}$", str(a).replace("%%", "")) and not str(a).replace("%%", "").startswith("0"): counts["float"] += 1 elif re.match(r"(true|yes|是|\*|✓|✔|☑|✅|√|false|no|否|⍻|×)$", str(a), flags=re.IGNORECASE): counts["bool"] += 1 elif trans_datatime(str(a)): counts["datetime"] += 1 else: counts["text"] += 1 if float_flag: ty = "float" else: counts = sorted(counts.items(), key=lambda x: x[1] * -1) ty = counts[0][0] for i in range(len(arr)): if arr[i] is None: continue try: arr[i] = trans[ty](str(arr[i])) except Exception as e: arr[i] = None logging.warning(f"Column {i}: {e}") # if ty == "text": # if len(arr) > 128 and uni / len(arr) < 0.1: # ty = "keyword" return arr, ty def chunk(filename, binary=None, from_page=0, to_page=10000000000, lang="Chinese", callback=None, **kwargs): """ Excel and csv(txt) format files are supported. For csv or txt file, the delimiter between columns is TAB. The first line must be column headers. Column headers must be meaningful terms inorder to make our NLP model understanding. It's good to enumerate some synonyms using slash '/' to separate, and even better to enumerate values using brackets like 'gender/sex(male, female)'. Here are some examples for headers: 1. supplier/vendor\tcolor(yellow, red, brown)\tgender/sex(male, female)\tsize(M,L,XL,XXL) 2. 姓名/名字\t电话/手机/微信\t最高学历(高中,职高,硕士,本科,博士,初中,中技,中专,专科,专升本,MPA,MBA,EMBA) Every row in table will be treated as a chunk. """ tbls = [] is_english = lang.lower() == "english" if re.search(r"\.xlsx?$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") excel_parser = Excel() dfs, tbls = excel_parser(filename, binary, from_page=from_page, to_page=to_page, callback=callback, **kwargs) elif re.search(r"\.txt$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") txt = get_text(filename, binary) lines = txt.split("\n") fails = [] headers = lines[0].split(kwargs.get("delimiter", "\t")) rows = [] for i, line in enumerate(lines[1:]): if i < from_page: continue if i >= to_page: break row = [field for field in line.split(kwargs.get("delimiter", "\t"))] if len(row) != len(headers): fails.append(str(i)) continue rows.append(row) callback(0.3, ("Extract records: {}~{}".format(from_page, min(len(lines), to_page)) + ( f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) dfs = [pd.DataFrame(np.array(rows), columns=headers)] elif re.search(r"\.csv$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") txt = get_text(filename, binary) delimiter = kwargs.get("delimiter", ",") reader = csv.reader(io.StringIO(txt), delimiter=delimiter) all_rows = list(reader) if not all_rows: raise ValueError("Empty CSV file") headers = all_rows[0] fails = [] rows = [] for i, row in enumerate(all_rows[1 + from_page: 1 + to_page]): if len(row) != len(headers): fails.append(str(i + from_page)) continue rows.append(row) callback( 0.3, (f"Extract records: {from_page}~{from_page + len(rows)}" + (f"{len(fails)} failure, line: {','.join(fails[:3])}..." if fails else "")) ) dfs = [pd.DataFrame(rows, columns=headers)] else: raise NotImplementedError("file type not supported yet(excel, text, csv supported)") res = [] PY = Pinyin() fieds_map = {"text": "_tks", "int": "_long", "keyword": "_kwd", "float": "_flt", "datetime": "_dt", "bool": "_kwd"} for df in dfs: for n in ["id", "_id", "index", "idx"]: if n in df.columns: del df[n] clmns = df.columns.values if len(clmns) != len(set(clmns)): col_counts = Counter(clmns) duplicates = [col for col, count in col_counts.items() if count > 1] if duplicates: raise ValueError(f"Duplicate column names detected: {duplicates}\nFrom: {clmns}") txts = list(copy.deepcopy(clmns)) py_clmns = [PY.get_pinyins(re.sub(r"(/.*|([^()]+?)|\([^()]+?\))", "", str(n)), "_")[0] for n in clmns] clmn_tys = [] for j in range(len(clmns)): cln, ty = column_data_type(df[clmns[j]]) clmn_tys.append(ty) df[clmns[j]] = cln if ty == "text": txts.extend([str(c) for c in cln if c]) clmns_map = [(py_clmns[i].lower() + fieds_map[clmn_tys[i]], str(clmns[i]).replace("_", " ")) for i in range(len(clmns))] eng = lang.lower() == "english" # is_english(txts) for ii, row in df.iterrows(): d = {"docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename))} row_txt = [] for j in range(len(clmns)): if row[clmns[j]] is None: continue if not str(row[clmns[j]]): continue if not isinstance(row[clmns[j]], pd.Series) and pd.isna(row[clmns[j]]): continue fld = clmns_map[j][0] d[fld] = row[clmns[j]] if clmn_tys[j] != "text" else rag_tokenizer.tokenize(row[clmns[j]]) row_txt.append("{}:{}".format(clmns[j], row[clmns[j]])) if not row_txt: continue tokenize(d, "; ".join(row_txt), eng) res.append(d) if tbls: doc = {"docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename))} res.extend(tokenize_table(tbls, doc, is_english)) KnowledgebaseService.update_parser_config(kwargs["kb_id"], {"field_map": {k: v for k, v in clmns_map}}) callback(0.35, "") return res if __name__ == "__main__": import sys def dummy(prog=None, msg=""): pass chunk(sys.argv[1], callback=dummy)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/one.py
rag/app/one.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging from io import BytesIO import re from deepdoc.parser.utils import get_text from rag.app import naive from rag.nlp import rag_tokenizer, tokenize from deepdoc.parser import PdfParser, ExcelParser, HtmlParser from deepdoc.parser.figure_parser import vision_figure_parser_docx_wrapper from rag.app.naive import by_plaintext, PARSERS from common.parser_config_utils import normalize_layout_recognizer class Pdf(PdfParser): def __call__(self, filename, binary=None, from_page=0, to_page=100000, zoomin=3, callback=None): from timeit import default_timer as timer start = timer() callback(msg="OCR started") self.__images__( filename if not binary else binary, zoomin, from_page, to_page, callback ) callback(msg="OCR finished ({:.2f}s)".format(timer() - start)) start = timer() self._layouts_rec(zoomin, drop=False) callback(0.63, "Layout analysis ({:.2f}s)".format(timer() - start)) logging.debug("layouts cost: {}s".format(timer() - start)) start = timer() self._table_transformer_job(zoomin) callback(0.65, "Table analysis ({:.2f}s)".format(timer() - start)) start = timer() self._text_merge() callback(0.67, "Text merged ({:.2f}s)".format(timer() - start)) tbls = self._extract_table_figure(True, zoomin, True, True) self._concat_downward() sections = [(b["text"], self.get_position(b, zoomin)) for i, b in enumerate(self.boxes)] return [(txt, "") for txt, _ in sorted(sections, key=lambda x: ( x[-1][0][0], x[-1][0][3], x[-1][0][1]))], tbls def chunk(filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, **kwargs): """ Supported file formats are docx, pdf, excel, txt. One file forms a chunk which maintains original text order. """ parser_config = kwargs.get( "parser_config", { "chunk_token_num": 512, "delimiter": "\n!?。;!?", "layout_recognize": "DeepDOC"}) eng = lang.lower() == "english" # is_english(cks) if re.search(r"\.docx$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") sections, tbls = naive.Docx()(filename, binary) tbls = vision_figure_parser_docx_wrapper(sections=sections, tbls=tbls, callback=callback, **kwargs) sections = [s for s, _ in sections if s] for (_, html), _ in tbls: sections.append(html) callback(0.8, "Finish parsing.") elif re.search(r"\.pdf$", filename, re.IGNORECASE): layout_recognizer, parser_model_name = normalize_layout_recognizer( parser_config.get("layout_recognize", "DeepDOC") ) if isinstance(layout_recognizer, bool): layout_recognizer = "DeepDOC" if layout_recognizer else "Plain Text" name = layout_recognizer.strip().lower() parser = PARSERS.get(name, by_plaintext) callback(0.1, "Start to parse.") sections, tbls, pdf_parser = parser( filename=filename, binary=binary, from_page=from_page, to_page=to_page, lang=lang, callback=callback, pdf_cls=Pdf, layout_recognizer=layout_recognizer, mineru_llm_name=parser_model_name, **kwargs ) if not sections and not tbls: return [] if name in ["tcadp", "docling", "mineru"]: parser_config["chunk_token_num"] = 0 callback(0.8, "Finish parsing.") for (img, rows), poss in tbls: if not rows: continue sections.append((rows if isinstance(rows, str) else rows[0], [(p[0] + 1 - from_page, p[1], p[2], p[3], p[4]) for p in poss])) sections = [s for s, _ in sections if s] elif re.search(r"\.xlsx?$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") excel_parser = ExcelParser() sections = excel_parser.html(binary, 1000000000) elif re.search(r"\.(txt|md|markdown|mdx)$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") txt = get_text(filename, binary) sections = txt.split("\n") sections = [s for s in sections if s] callback(0.8, "Finish parsing.") elif re.search(r"\.(htm|html)$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") sections = HtmlParser()(filename, binary) sections = [s for s in sections if s] callback(0.8, "Finish parsing.") elif re.search(r"\.doc$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") try: from tika import parser as tika_parser except Exception as e: callback(0.8, f"tika not available: {e}. Unsupported .doc parsing.") logging.warning(f"tika not available: {e}. Unsupported .doc parsing for {filename}.") return [] binary = BytesIO(binary) doc_parsed = tika_parser.from_buffer(binary) if doc_parsed.get('content', None) is not None: sections = doc_parsed['content'].split('\n') sections = [s for s in sections if s] callback(0.8, "Finish parsing.") else: raise NotImplementedError( "file type not supported yet(doc, docx, pdf, txt supported)") doc = { "docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename)) } doc["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(doc["title_tks"]) tokenize(doc, "\n".join(sections), eng) return [doc] if __name__ == "__main__": import sys def dummy(prog=None, msg=""): pass chunk(sys.argv[1], from_page=0, to_page=10, callback=dummy)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/audio.py
rag/app/audio.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import os import re import tempfile from common.constants import LLMType from api.db.services.llm_service import LLMBundle from rag.nlp import rag_tokenizer, tokenize def chunk(filename, binary, tenant_id, lang, callback=None, **kwargs): doc = {"docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename))} doc["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(doc["title_tks"]) # is it English is_english = lang.lower() == "english" # is_english(sections) try: _, ext = os.path.splitext(filename) if not ext: raise RuntimeError("No extension detected.") if ext not in [".da", ".wave", ".wav", ".mp3", ".wav", ".aac", ".flac", ".ogg", ".aiff", ".au", ".midi", ".wma", ".realaudio", ".vqf", ".oggvorbis", ".aac", ".ape"]: raise RuntimeError(f"Extension {ext} is not supported yet.") tmp_path = "" with tempfile.NamedTemporaryFile(suffix=ext, delete=False) as tmpf: tmpf.write(binary) tmpf.flush() tmp_path = os.path.abspath(tmpf.name) callback(0.1, "USE Sequence2Txt LLM to transcription the audio") seq2txt_mdl = LLMBundle(tenant_id, LLMType.SPEECH2TEXT, lang=lang) ans = seq2txt_mdl.transcription(tmp_path) callback(0.8, "Sequence2Txt LLM respond: %s ..." % ans[:32]) tokenize(doc, ans, is_english) return [doc] except Exception as e: callback(prog=-1, msg=str(e)) finally: if tmp_path and os.path.exists(tmp_path): try: os.unlink(tmp_path) except Exception as e: logging.exception(f"Failed to remove temporary file: {tmp_path}, exception: {e}") pass return []
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/naive.py
rag/app/naive.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import re import os from functools import reduce from io import BytesIO from timeit import default_timer as timer from docx import Document from docx.image.exceptions import InvalidImageStreamError, UnexpectedEndOfFileError, UnrecognizedImageError from docx.opc.pkgreader import _SerializedRelationships, _SerializedRelationship from docx.opc.oxml import parse_xml from markdown import markdown from PIL import Image from common.token_utils import num_tokens_from_string from common.constants import LLMType from api.db.services.llm_service import LLMBundle from rag.utils.file_utils import extract_embed_file, extract_links_from_pdf, extract_links_from_docx, extract_html from deepdoc.parser import DocxParser, ExcelParser, HtmlParser, JsonParser, MarkdownElementExtractor, MarkdownParser, \ PdfParser, TxtParser from deepdoc.parser.figure_parser import VisionFigureParser, vision_figure_parser_docx_wrapper, \ vision_figure_parser_pdf_wrapper from deepdoc.parser.pdf_parser import PlainParser, VisionParser from deepdoc.parser.docling_parser import DoclingParser from deepdoc.parser.tcadp_parser import TCADPParser from common.parser_config_utils import normalize_layout_recognizer from rag.nlp import concat_img, find_codec, naive_merge, naive_merge_with_images, naive_merge_docx, rag_tokenizer, \ tokenize_chunks, tokenize_chunks_with_images, tokenize_table, attach_media_context, append_context2table_image4pdf def by_deepdoc(filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, pdf_cls=None, **kwargs): callback = callback binary = binary pdf_parser = pdf_cls() if pdf_cls else Pdf() sections, tables = pdf_parser( filename if not binary else binary, from_page=from_page, to_page=to_page, callback=callback ) tables = vision_figure_parser_pdf_wrapper(tbls=tables, callback=callback, **kwargs) return sections, tables, pdf_parser def by_mineru( filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, pdf_cls=None, parse_method: str = "raw", mineru_llm_name: str | None = None, tenant_id: str | None = None, **kwargs, ): pdf_parser = None if tenant_id: if not mineru_llm_name: try: from api.db.services.tenant_llm_service import TenantLLMService env_name = TenantLLMService.ensure_mineru_from_env(tenant_id) candidates = TenantLLMService.query(tenant_id=tenant_id, llm_factory="MinerU", model_type=LLMType.OCR) if candidates: mineru_llm_name = candidates[0].llm_name elif env_name: mineru_llm_name = env_name except Exception as e: # best-effort fallback logging.warning(f"fallback to env mineru: {e}") if mineru_llm_name: try: ocr_model = LLMBundle(tenant_id=tenant_id, llm_type=LLMType.OCR, llm_name=mineru_llm_name, lang=lang) pdf_parser = ocr_model.mdl sections, tables = pdf_parser.parse_pdf( filepath=filename, binary=binary, callback=callback, parse_method=parse_method, lang=lang, **kwargs, ) return sections, tables, pdf_parser except Exception as e: logging.error(f"Failed to parse pdf via LLMBundle MinerU ({mineru_llm_name}): {e}") if callback: callback(-1, "MinerU not found.") return None, None, None def by_docling(filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, pdf_cls=None, **kwargs): pdf_parser = DoclingParser() parse_method = kwargs.get("parse_method", "raw") if not pdf_parser.check_installation(): callback(-1, "Docling not found.") return None, None, pdf_parser sections, tables = pdf_parser.parse_pdf( filepath=filename, binary=binary, callback=callback, output_dir=os.environ.get("MINERU_OUTPUT_DIR", ""), delete_output=bool(int(os.environ.get("MINERU_DELETE_OUTPUT", 1))), parse_method=parse_method ) return sections, tables, pdf_parser def by_tcadp(filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, pdf_cls=None, **kwargs): tcadp_parser = TCADPParser() if not tcadp_parser.check_installation(): callback(-1, "TCADP parser not available. Please check Tencent Cloud API configuration.") return None, None, tcadp_parser sections, tables = tcadp_parser.parse_pdf( filepath=filename, binary=binary, callback=callback, output_dir=os.environ.get("TCADP_OUTPUT_DIR", ""), file_type="PDF" ) return sections, tables, tcadp_parser def by_plaintext(filename, binary=None, from_page=0, to_page=100000, callback=None, **kwargs): layout_recognizer = (kwargs.get("layout_recognizer") or "").strip() if (not layout_recognizer) or (layout_recognizer == "Plain Text"): pdf_parser = PlainParser() else: tenant_id = kwargs.get("tenant_id") if not tenant_id: raise ValueError("tenant_id is required when using vision layout recognizer") vision_model = LLMBundle( tenant_id, LLMType.IMAGE2TEXT, llm_name=layout_recognizer, lang=kwargs.get("lang", "Chinese"), ) pdf_parser = VisionParser(vision_model=vision_model, **kwargs) sections, tables = pdf_parser( filename if not binary else binary, from_page=from_page, to_page=to_page, callback=callback ) return sections, tables, pdf_parser PARSERS = { "deepdoc": by_deepdoc, "mineru": by_mineru, "docling": by_docling, "tcadp": by_tcadp, "plaintext": by_plaintext, # default } class Docx(DocxParser): def __init__(self): pass def get_picture(self, document, paragraph): imgs = paragraph._element.xpath('.//pic:pic') if not imgs: return None res_img = None for img in imgs: embed = img.xpath('.//a:blip/@r:embed') if not embed: continue embed = embed[0] try: related_part = document.part.related_parts[embed] image_blob = related_part.image.blob except UnrecognizedImageError: logging.info("Unrecognized image format. Skipping image.") continue except UnexpectedEndOfFileError: logging.info("EOF was unexpectedly encountered while reading an image stream. Skipping image.") continue except InvalidImageStreamError: logging.info("The recognized image stream appears to be corrupted. Skipping image.") continue except UnicodeDecodeError: logging.info("The recognized image stream appears to be corrupted. Skipping image.") continue except Exception as e: logging.warning(f"The recognized image stream appears to be corrupted. Skipping image, exception: {e}") continue try: image = Image.open(BytesIO(image_blob)).convert('RGB') if res_img is None: res_img = image else: res_img = concat_img(res_img, image) except Exception as e: logging.warning(f"Fail to open or concat images, exception: {e}") continue return res_img def __clean(self, line): line = re.sub(r"\u3000", " ", line).strip() return line def __get_nearest_title(self, table_index, filename): """Get the hierarchical title structure before the table""" import re from docx.text.paragraph import Paragraph titles = [] blocks = [] # Get document name from filename parameter doc_name = re.sub(r"\.[a-zA-Z]+$", "", filename) if not doc_name: doc_name = "Untitled Document" # Collect all document blocks while maintaining document order try: # Iterate through all paragraphs and tables in document order for i, block in enumerate(self.doc._element.body): if block.tag.endswith('p'): # Paragraph p = Paragraph(block, self.doc) blocks.append(('p', i, p)) elif block.tag.endswith('tbl'): # Table blocks.append(('t', i, None)) # Table object will be retrieved later except Exception as e: logging.error(f"Error collecting blocks: {e}") return "" # Find the target table position target_table_pos = -1 table_count = 0 for i, (block_type, pos, _) in enumerate(blocks): if block_type == 't': if table_count == table_index: target_table_pos = pos break table_count += 1 if target_table_pos == -1: return "" # Target table not found # Find the nearest heading paragraph in reverse order nearest_title = None for i in range(len(blocks) - 1, -1, -1): block_type, pos, block = blocks[i] if pos >= target_table_pos: # Skip blocks after the table continue if block_type != 'p': continue if block.style and block.style.name and re.search(r"Heading\s*(\d+)", block.style.name, re.I): try: level_match = re.search(r"(\d+)", block.style.name) if level_match: level = int(level_match.group(1)) if level <= 7: # Support up to 7 heading levels title_text = block.text.strip() if title_text: # Avoid empty titles nearest_title = (level, title_text) break except Exception as e: logging.error(f"Error parsing heading level: {e}") if nearest_title: # Add current title titles.append(nearest_title) current_level = nearest_title[0] # Find all parent headings, allowing cross-level search while current_level > 1: found = False for i in range(len(blocks) - 1, -1, -1): block_type, pos, block = blocks[i] if pos >= target_table_pos: # Skip blocks after the table continue if block_type != 'p': continue if block.style and re.search(r"Heading\s*(\d+)", block.style.name, re.I): try: level_match = re.search(r"(\d+)", block.style.name) if level_match: level = int(level_match.group(1)) # Find any heading with a higher level if level < current_level: title_text = block.text.strip() if title_text: # Avoid empty titles titles.append((level, title_text)) current_level = level found = True break except Exception as e: logging.error(f"Error parsing parent heading: {e}") if not found: # Break if no parent heading is found break # Sort by level (ascending, from highest to lowest) titles.sort(key=lambda x: x[0]) # Organize titles (from highest to lowest) hierarchy = [doc_name] + [t[1] for t in titles] return " > ".join(hierarchy) return "" def __call__(self, filename, binary=None, from_page=0, to_page=100000): self.doc = Document( filename) if not binary else Document(BytesIO(binary)) pn = 0 lines = [] last_image = None for p in self.doc.paragraphs: if pn > to_page: break if from_page <= pn < to_page: if p.text.strip(): if p.style and p.style.name == 'Caption': former_image = None if lines and lines[-1][1] and lines[-1][2] != 'Caption': former_image = lines[-1][1].pop() elif last_image: former_image = last_image last_image = None lines.append((self.__clean(p.text), [former_image], p.style.name)) else: current_image = self.get_picture(self.doc, p) image_list = [current_image] if last_image: image_list.insert(0, last_image) last_image = None lines.append((self.__clean(p.text), image_list, p.style.name if p.style else "")) else: if current_image := self.get_picture(self.doc, p): if lines: lines[-1][1].append(current_image) else: last_image = current_image for run in p.runs: if 'lastRenderedPageBreak' in run._element.xml: pn += 1 continue if 'w:br' in run._element.xml and 'type="page"' in run._element.xml: pn += 1 new_line = [(line[0], reduce(concat_img, line[1]) if line[1] else None) for line in lines] tbls = [] for i, tb in enumerate(self.doc.tables): title = self.__get_nearest_title(i, filename) html = "<table>" if title: html += f"<caption>Table Location: {title}</caption>" for r in tb.rows: html += "<tr>" i = 0 try: while i < len(r.cells): span = 1 c = r.cells[i] for j in range(i + 1, len(r.cells)): if c.text == r.cells[j].text: span += 1 i = j else: break i += 1 html += f"<td>{c.text}</td>" if span == 1 else f"<td colspan='{span}'>{c.text}</td>" except Exception as e: logging.warning(f"Error parsing table, ignore: {e}") html += "</tr>" html += "</table>" tbls.append(((None, html), "")) return new_line, tbls def to_markdown(self, filename=None, binary=None, inline_images: bool = True): """ This function uses mammoth, licensed under the BSD 2-Clause License. """ import base64 import uuid import mammoth from markdownify import markdownify docx_file = BytesIO(binary) if binary else open(filename, "rb") def _convert_image_to_base64(image): try: with image.open() as image_file: image_bytes = image_file.read() encoded = base64.b64encode(image_bytes).decode("utf-8") base64_url = f"data:{image.content_type};base64,{encoded}" alt_name = "image" alt_name = f"img_{uuid.uuid4().hex[:8]}" return {"src": base64_url, "alt": alt_name} except Exception as e: logging.warning(f"Failed to convert image to base64: {e}") return {"src": "", "alt": "image"} try: if inline_images: result = mammoth.convert_to_html(docx_file, convert_image=mammoth.images.img_element(_convert_image_to_base64)) else: result = mammoth.convert_to_html(docx_file) html = result.value markdown_text = markdownify(html) return markdown_text finally: if not binary: docx_file.close() class Pdf(PdfParser): def __init__(self): super().__init__() def __call__(self, filename, binary=None, from_page=0, to_page=100000, zoomin=3, callback=None, separate_tables_figures=False): start = timer() first_start = start callback(msg="OCR started") self.__images__( filename if not binary else binary, zoomin, from_page, to_page, callback ) callback(msg="OCR finished ({:.2f}s)".format(timer() - start)) logging.info("OCR({}~{}): {:.2f}s".format(from_page, to_page, timer() - start)) start = timer() self._layouts_rec(zoomin) callback(0.63, "Layout analysis ({:.2f}s)".format(timer() - start)) start = timer() self._table_transformer_job(zoomin) callback(0.65, "Table analysis ({:.2f}s)".format(timer() - start)) start = timer() self._text_merge(zoomin=zoomin) callback(0.67, "Text merged ({:.2f}s)".format(timer() - start)) if separate_tables_figures: tbls, figures = self._extract_table_figure(True, zoomin, True, True, True) self._concat_downward() logging.info("layouts cost: {}s".format(timer() - first_start)) return [(b["text"], self._line_tag(b, zoomin)) for b in self.boxes], tbls, figures else: tbls = self._extract_table_figure(True, zoomin, True, True) self._naive_vertical_merge() self._concat_downward() # self._final_reading_order_merge() # self._filter_forpages() logging.info("layouts cost: {}s".format(timer() - first_start)) return [(b["text"], self._line_tag(b, zoomin)) for b in self.boxes], tbls class Markdown(MarkdownParser): def md_to_html(self, sections): if not sections: return [] if isinstance(sections, type("")): text = sections elif isinstance(sections[0], type("")): text = sections[0] else: return [] from bs4 import BeautifulSoup html_content = markdown(text) soup = BeautifulSoup(html_content, 'html.parser') return soup def get_hyperlink_urls(self, soup): if soup: return set([a.get('href') for a in soup.find_all('a') if a.get('href')]) return [] def extract_image_urls_with_lines(self, text): md_img_re = re.compile(r"!\[[^\]]*\]\(([^)\s]+)") html_img_re = re.compile(r'src=["\\\']([^"\\\'>\\s]+)', re.IGNORECASE) urls = [] seen = set() lines = text.splitlines() for idx, line in enumerate(lines): for url in md_img_re.findall(line): if (url, idx) not in seen: urls.append({"url": url, "line": idx}) seen.add((url, idx)) for url in html_img_re.findall(line): if (url, idx) not in seen: urls.append({"url": url, "line": idx}) seen.add((url, idx)) # cross-line try: from bs4 import BeautifulSoup soup = BeautifulSoup(text, 'html.parser') newline_offsets = [m.start() for m in re.finditer(r"\n", text)] + [len(text)] for img_tag in soup.find_all('img'): src = img_tag.get('src') if not src: continue tag_str = str(img_tag) pos = text.find(tag_str) if pos == -1: # fallback pos = max(text.find(src), 0) line_no = 0 for i, off in enumerate(newline_offsets): if pos <= off: line_no = i break if (src, line_no) not in seen: urls.append({"url": src, "line": line_no}) seen.add((src, line_no)) except Exception as e: logging.error("Failed to extract image urls: {}".format(e)) pass return urls def load_images_from_urls(self, urls, cache=None): import requests from pathlib import Path cache = cache or {} images = [] for url in urls: if url in cache: if cache[url]: images.append(cache[url]) continue img_obj = None try: if url.startswith(('http://', 'https://')): response = requests.get(url, stream=True, timeout=30) if response.status_code == 200 and response.headers.get('Content-Type', '').startswith('image/'): img_obj = Image.open(BytesIO(response.content)).convert('RGB') else: local_path = Path(url) if local_path.exists(): img_obj = Image.open(url).convert('RGB') else: logging.warning(f"Local image file not found: {url}") except Exception as e: logging.error(f"Failed to download/open image from {url}: {e}") cache[url] = img_obj if img_obj: images.append(img_obj) return images, cache def __call__(self, filename, binary=None, separate_tables=True, delimiter=None, return_section_images=False): if binary: encoding = find_codec(binary) txt = binary.decode(encoding, errors="ignore") else: with open(filename, "r") as f: txt = f.read() remainder, tables = self.extract_tables_and_remainder(f'{txt}\n', separate_tables=separate_tables) # To eliminate duplicate tables in chunking result, uncomment code below and set separate_tables to True in line 410. # extractor = MarkdownElementExtractor(remainder) extractor = MarkdownElementExtractor(txt) image_refs = self.extract_image_urls_with_lines(txt) element_sections = extractor.extract_elements(delimiter, include_meta=True) sections = [] section_images = [] image_cache = {} for element in element_sections: content = element["content"] start_line = element["start_line"] end_line = element["end_line"] urls_in_section = [ref["url"] for ref in image_refs if start_line <= ref["line"] <= end_line] imgs = [] if urls_in_section: imgs, image_cache = self.load_images_from_urls(urls_in_section, image_cache) combined_image = None if imgs: combined_image = reduce(concat_img, imgs) if len(imgs) > 1 else imgs[0] sections.append((content, "")) section_images.append(combined_image) tbls = [] for table in tables: tbls.append(((None, markdown(table, extensions=['markdown.extensions.tables'])), "")) if return_section_images: return sections, tbls, section_images return sections, tbls def load_from_xml_v2(baseURI, rels_item_xml): """ Return |_SerializedRelationships| instance loaded with the relationships contained in *rels_item_xml*. Returns an empty collection if *rels_item_xml* is |None|. """ srels = _SerializedRelationships() if rels_item_xml is not None: rels_elm = parse_xml(rels_item_xml) for rel_elm in rels_elm.Relationship_lst: if rel_elm.target_ref in ('../NULL', 'NULL'): continue srels._srels.append(_SerializedRelationship(baseURI, rel_elm)) return srels def chunk(filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, **kwargs): """ Supported file formats are docx, pdf, excel, txt. This method apply the naive ways to chunk files. Successive text will be sliced into pieces using 'delimiter'. Next, these successive pieces are merge into chunks whose token number is no more than 'Max token number'. """ urls = set() url_res = [] is_english = lang.lower() == "english" # is_english(cks) parser_config = kwargs.get( "parser_config", { "chunk_token_num": 512, "delimiter": "\n!?。;!?", "layout_recognize": "DeepDOC", "analyze_hyperlink": True}) child_deli = (parser_config.get("children_delimiter") or "").encode('utf-8').decode('unicode_escape').encode( 'latin1').decode('utf-8') cust_child_deli = re.findall(r"`([^`]+)`", child_deli) child_deli = "|".join(re.sub(r"`([^`]+)`", "", child_deli)) if cust_child_deli: cust_child_deli = sorted(set(cust_child_deli), key=lambda x: -len(x)) cust_child_deli = "|".join(re.escape(t) for t in cust_child_deli if t) child_deli += cust_child_deli is_markdown = False table_context_size = max(0, int(parser_config.get("table_context_size", 0) or 0)) image_context_size = max(0, int(parser_config.get("image_context_size", 0) or 0)) doc = { "docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename)) } doc["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(doc["title_tks"]) res = [] pdf_parser = None section_images = None is_root = kwargs.get("is_root", True) embed_res = [] if is_root: # Only extract embedded files at the root call embeds = [] if binary is not None: embeds = extract_embed_file(binary) else: raise Exception("Embedding extraction from file path is not supported.") # Recursively chunk each embedded file and collect results for embed_filename, embed_bytes in embeds: try: sub_res = chunk(embed_filename, binary=embed_bytes, lang=lang, callback=callback, is_root=False, **kwargs) or [] embed_res.extend(sub_res) except Exception as e: error_msg = f"Failed to chunk embed {embed_filename}: {e}" logging.error(error_msg) if callback: callback(0.05, error_msg) continue if re.search(r"\.docx$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") if parser_config.get("analyze_hyperlink", False) and is_root: urls = extract_links_from_docx(binary) for index, url in enumerate(urls): html_bytes, metadata = extract_html(url) if not html_bytes: continue try: sub_url_res = chunk(url, html_bytes, callback=callback, lang=lang, is_root=False, **kwargs) except Exception as e: logging.info(f"Failed to chunk url in registered file type {url}: {e}") sub_url_res = chunk(f"{index}.html", html_bytes, callback=callback, lang=lang, is_root=False, **kwargs) url_res.extend(sub_url_res) # fix "There is no item named 'word/NULL' in the archive", referring to https://github.com/python-openxml/python-docx/issues/1105#issuecomment-1298075246 _SerializedRelationships.load_from_xml = load_from_xml_v2 sections, tables = Docx()(filename, binary) tables = vision_figure_parser_docx_wrapper(sections=sections, tbls=tables, callback=callback, **kwargs) res = tokenize_table(tables, doc, is_english) callback(0.8, "Finish parsing.") st = timer() chunks, images = naive_merge_docx( sections, int(parser_config.get( "chunk_token_num", 128)), parser_config.get( "delimiter", "\n!?。;!?")) res.extend(tokenize_chunks_with_images(chunks, doc, is_english, images, child_delimiters_pattern=child_deli)) logging.info("naive_merge({}): {}".format(filename, timer() - st)) res.extend(embed_res) res.extend(url_res) if table_context_size or image_context_size: attach_media_context(res, table_context_size, image_context_size) return res elif re.search(r"\.pdf$", filename, re.IGNORECASE): layout_recognizer, parser_model_name = normalize_layout_recognizer( parser_config.get("layout_recognize", "DeepDOC") ) if parser_config.get("analyze_hyperlink", False) and is_root: urls = extract_links_from_pdf(binary) if isinstance(layout_recognizer, bool): layout_recognizer = "DeepDOC" if layout_recognizer else "Plain Text" name = layout_recognizer.strip().lower() parser = PARSERS.get(name, by_plaintext) callback(0.1, "Start to parse.") sections, tables, pdf_parser = parser( filename=filename, binary=binary, from_page=from_page, to_page=to_page, lang=lang, callback=callback, layout_recognizer=layout_recognizer, mineru_llm_name=parser_model_name, **kwargs ) if not sections and not tables: return [] if table_context_size or image_context_size: tables = append_context2table_image4pdf(sections, tables, image_context_size) if name in ["tcadp", "docling", "mineru"]: parser_config["chunk_token_num"] = 0 res = tokenize_table(tables, doc, is_english) callback(0.8, "Finish parsing.") elif re.search(r"\.(csv|xlsx?)$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") # Check if tcadp_parser is selected for spreadsheet files layout_recognizer = parser_config.get("layout_recognize", "DeepDOC") if layout_recognizer == "TCADP Parser": table_result_type = parser_config.get("table_result_type", "1") markdown_image_response_type = parser_config.get("markdown_image_response_type", "1") tcadp_parser = TCADPParser( table_result_type=table_result_type, markdown_image_response_type=markdown_image_response_type ) if not tcadp_parser.check_installation(): callback(-1, "TCADP parser not available. Please check Tencent Cloud API configuration.") return res # Determine file type based on extension file_type = "XLSX" if re.search(r"\.xlsx?$", filename, re.IGNORECASE) else "CSV" sections, tables = tcadp_parser.parse_pdf( filepath=filename, binary=binary, callback=callback, output_dir=os.environ.get("TCADP_OUTPUT_DIR", ""), file_type=file_type ) parser_config["chunk_token_num"] = 0 res = tokenize_table(tables, doc, is_english) callback(0.8, "Finish parsing.") else: # Default DeepDOC parser
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/tag.py
rag/app/tag.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json import re import csv from copy import deepcopy from deepdoc.parser.utils import get_text from rag.app.qa import Excel from rag.nlp import rag_tokenizer from common import settings def beAdoc(d, q, a, eng, row_num=-1): d["content_with_weight"] = q d["content_ltks"] = rag_tokenizer.tokenize(q) d["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(d["content_ltks"]) d["tag_kwd"] = [t.strip().replace(".", "_") for t in a.split(",") if t.strip()] if row_num >= 0: d["top_int"] = [row_num] return d def chunk(filename, binary=None, lang="Chinese", callback=None, **kwargs): """ Excel and csv(txt) format files are supported. If the file is in Excel format, there should be 2 column content and tags without header. And content column is ahead of tags column. And it's O.K if it has multiple sheets as long as the columns are rightly composed. If it's in csv format, it should be UTF-8 encoded. Use TAB as delimiter to separate content and tags. All the deformed lines will be ignored. Every pair will be treated as a chunk. """ eng = lang.lower() == "english" res = [] doc = { "docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename)) } if re.search(r"\.xlsx?$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") excel_parser = Excel() for ii, (q, a) in enumerate(excel_parser(filename, binary, callback)): res.append(beAdoc(deepcopy(doc), q, a, eng, ii)) return res elif re.search(r"\.(txt)$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") txt = get_text(filename, binary) lines = txt.split("\n") comma, tab = 0, 0 for line in lines: if len(line.split(",")) == 2: comma += 1 if len(line.split("\t")) == 2: tab += 1 delimiter = "\t" if tab >= comma else "," fails = [] content = "" i = 0 while i < len(lines): arr = lines[i].split(delimiter) if len(arr) != 2: content += "\n" + lines[i] elif len(arr) == 2: content += "\n" + arr[0] res.append(beAdoc(deepcopy(doc), content, arr[1], eng, i)) content = "" i += 1 if len(res) % 999 == 0: callback(len(res) * 0.6 / len(lines), ("Extract TAG: {}".format(len(res)) + ( f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) callback(0.6, ("Extract TAG: {}".format(len(res)) + ( f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) return res elif re.search(r"\.(csv)$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") txt = get_text(filename, binary) lines = txt.split("\n") fails = [] content = "" res = [] reader = csv.reader(lines) for i, row in enumerate(reader): row = [r.strip() for r in row if r.strip()] if len(row) != 2: content += "\n" + lines[i] elif len(row) == 2: content += "\n" + row[0] res.append(beAdoc(deepcopy(doc), content, row[1], eng, i)) content = "" if len(res) % 999 == 0: callback(len(res) * 0.6 / len(lines), ("Extract Tags: {}".format(len(res)) + ( f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) callback(0.6, ("Extract TAG : {}".format(len(res)) + ( f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) return res raise NotImplementedError( "Excel, csv(txt) format files are supported.") def label_question(question, kbs): from api.db.services.knowledgebase_service import KnowledgebaseService from graphrag.utils import get_tags_from_cache, set_tags_to_cache tags = None tag_kb_ids = [] for kb in kbs: if kb.parser_config.get("tag_kb_ids"): tag_kb_ids.extend(kb.parser_config["tag_kb_ids"]) if tag_kb_ids: all_tags = get_tags_from_cache(tag_kb_ids) if not all_tags: all_tags = settings.retriever.all_tags_in_portion(kb.tenant_id, tag_kb_ids) set_tags_to_cache(tags=all_tags, kb_ids=tag_kb_ids) else: all_tags = json.loads(all_tags) tag_kbs = KnowledgebaseService.get_by_ids(tag_kb_ids) if not tag_kbs: return tags tags = settings.retriever.tag_query(question, list(set([kb.tenant_id for kb in tag_kbs])), tag_kb_ids, all_tags, kb.parser_config.get("topn_tags", 3) ) return tags if __name__ == "__main__": import sys def dummy(prog=None, msg=""): pass chunk(sys.argv[1], from_page=0, to_page=10, callback=dummy)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/qa.py
rag/app/qa.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import re import csv from copy import deepcopy from io import BytesIO from timeit import default_timer as timer from openpyxl import load_workbook from deepdoc.parser.utils import get_text from rag.nlp import is_english, random_choices, qbullets_category, add_positions, has_qbullet, docx_question_level from rag.nlp import rag_tokenizer, tokenize_table, concat_img from deepdoc.parser import PdfParser, ExcelParser, DocxParser from docx import Document from PIL import Image from markdown import markdown from common.float_utils import get_float class Excel(ExcelParser): def __call__(self, fnm, binary=None, callback=None): if not binary: wb = load_workbook(fnm) else: wb = load_workbook(BytesIO(binary)) total = 0 for sheetname in wb.sheetnames: total += len(list(wb[sheetname].rows)) res, fails = [], [] for sheetname in wb.sheetnames: ws = wb[sheetname] rows = list(ws.rows) for i, r in enumerate(rows): q, a = "", "" for cell in r: if not cell.value: continue if not q: q = str(cell.value) elif not a: a = str(cell.value) else: break if q and a: res.append((q, a)) else: fails.append(str(i + 1)) if len(res) % 999 == 0: callback(len(res) * 0.6 / total, ("Extract pairs: {}".format(len(res)) + (f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) callback(0.6, ("Extract pairs: {}. ".format(len(res)) + ( f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) self.is_english = is_english( [rmPrefix(q) for q, _ in random_choices(res, k=30) if len(q) > 1]) return res class Pdf(PdfParser): def __call__(self, filename, binary=None, from_page=0, to_page=100000, zoomin=3, callback=None): start = timer() callback(msg="OCR started") self.__images__( filename if not binary else binary, zoomin, from_page, to_page, callback ) callback(msg="OCR finished ({:.2f}s)".format(timer() - start)) logging.debug("OCR({}~{}): {:.2f}s".format(from_page, to_page, timer() - start)) start = timer() self._layouts_rec(zoomin, drop=False) callback(0.63, "Layout analysis ({:.2f}s)".format(timer() - start)) start = timer() self._table_transformer_job(zoomin) callback(0.65, "Table analysis ({:.2f}s)".format(timer() - start)) start = timer() self._text_merge() callback(0.67, "Text merged ({:.2f}s)".format(timer() - start)) tbls = self._extract_table_figure(True, zoomin, True, True) # self._naive_vertical_merge() # self._concat_downward() # self._filter_forpages() logging.debug("layouts: {}".format(timer() - start)) sections = [b["text"] for b in self.boxes] bull_x0_list = [] q_bull, reg = qbullets_category(sections) if q_bull == -1: raise ValueError("Unable to recognize Q&A structure.") qai_list = [] last_q, last_a, last_tag = '', '', '' last_index = -1 last_box = {'text': ''} last_bull = None def sort_key(element): tbls_pn = element[1][0][0] tbls_top = element[1][0][3] return tbls_pn, tbls_top tbls.sort(key=sort_key) tbl_index = 0 last_pn, last_bottom = 0, 0 tbl_pn, tbl_left, tbl_right, tbl_top, tbl_bottom, tbl_tag, tbl_text = 1, 0, 0, 0, 0, '@@0\t0\t0\t0\t0##', '' for box in self.boxes: section, line_tag = box['text'], self._line_tag(box, zoomin) has_bull, index = has_qbullet(reg, box, last_box, last_index, last_bull, bull_x0_list) last_box, last_index, last_bull = box, index, has_bull line_pn = get_float(line_tag.lstrip('@@').split('\t')[0]) line_top = get_float(line_tag.rstrip('##').split('\t')[3]) tbl_pn, tbl_left, tbl_right, tbl_top, tbl_bottom, tbl_tag, tbl_text = self.get_tbls_info(tbls, tbl_index) if not has_bull: # No question bullet if not last_q: if tbl_pn < line_pn or (tbl_pn == line_pn and tbl_top <= line_top): # image passed tbl_index += 1 continue else: sum_tag = line_tag sum_section = section while ((tbl_pn == last_pn and tbl_top >= last_bottom) or (tbl_pn > last_pn)) \ and ((tbl_pn == line_pn and tbl_top <= line_top) or ( tbl_pn < line_pn)): # add image at the middle of current answer sum_tag = f'{tbl_tag}{sum_tag}' sum_section = f'{tbl_text}{sum_section}' tbl_index += 1 tbl_pn, tbl_left, tbl_right, tbl_top, tbl_bottom, tbl_tag, tbl_text = self.get_tbls_info(tbls, tbl_index) last_a = f'{last_a}{sum_section}' last_tag = f'{last_tag}{sum_tag}' else: if last_q: while ((tbl_pn == last_pn and tbl_top >= last_bottom) or (tbl_pn > last_pn)) \ and ((tbl_pn == line_pn and tbl_top <= line_top) or ( tbl_pn < line_pn)): # add image at the end of last answer last_tag = f'{last_tag}{tbl_tag}' last_a = f'{last_a}{tbl_text}' tbl_index += 1 tbl_pn, tbl_left, tbl_right, tbl_top, tbl_bottom, tbl_tag, tbl_text = self.get_tbls_info(tbls, tbl_index) image, poss = self.crop(last_tag, need_position=True) qai_list.append((last_q, last_a, image, poss)) last_q, last_a, last_tag = '', '', '' last_q = has_bull.group() _, end = has_bull.span() last_a = section[end:] last_tag = line_tag last_bottom = float(line_tag.rstrip('##').split('\t')[4]) last_pn = line_pn if last_q: qai_list.append((last_q, last_a, *self.crop(last_tag, need_position=True))) return qai_list, tbls def get_tbls_info(self, tbls, tbl_index): if tbl_index >= len(tbls): return 1, 0, 0, 0, 0, '@@0\t0\t0\t0\t0##', '' tbl_pn = tbls[tbl_index][1][0][0] + 1 tbl_left = tbls[tbl_index][1][0][1] tbl_right = tbls[tbl_index][1][0][2] tbl_top = tbls[tbl_index][1][0][3] tbl_bottom = tbls[tbl_index][1][0][4] tbl_tag = "@@{}\t{:.1f}\t{:.1f}\t{:.1f}\t{:.1f}##" \ .format(tbl_pn, tbl_left, tbl_right, tbl_top, tbl_bottom) _tbl_text = ''.join(tbls[tbl_index][0][1]) return tbl_pn, tbl_left, tbl_right, tbl_top, tbl_bottom, tbl_tag, _tbl_text class Docx(DocxParser): def __init__(self): pass def get_picture(self, document, paragraph): img = paragraph._element.xpath('.//pic:pic') if not img: return None img = img[0] embed = img.xpath('.//a:blip/@r:embed')[0] related_part = document.part.related_parts[embed] image = related_part.image image = Image.open(BytesIO(image.blob)).convert('RGB') return image def __call__(self, filename, binary=None, from_page=0, to_page=100000, callback=None): self.doc = Document( filename) if not binary else Document(BytesIO(binary)) pn = 0 last_answer, last_image = "", None question_stack, level_stack = [], [] qai_list = [] for p in self.doc.paragraphs: if pn > to_page: break question_level, p_text = 0, '' if from_page <= pn < to_page and p.text.strip(): question_level, p_text = docx_question_level(p) if not question_level or question_level > 6: # not a question last_answer = f'{last_answer}\n{p_text}' current_image = self.get_picture(self.doc, p) last_image = concat_img(last_image, current_image) else: # is a question if last_answer or last_image: sum_question = '\n'.join(question_stack) if sum_question: qai_list.append((sum_question, last_answer, last_image)) last_answer, last_image = '', None i = question_level while question_stack and i <= level_stack[-1]: question_stack.pop() level_stack.pop() question_stack.append(p_text) level_stack.append(question_level) for run in p.runs: if 'lastRenderedPageBreak' in run._element.xml: pn += 1 continue if 'w:br' in run._element.xml and 'type="page"' in run._element.xml: pn += 1 if last_answer: sum_question = '\n'.join(question_stack) if sum_question: qai_list.append((sum_question, last_answer, last_image)) tbls = [] for tb in self.doc.tables: html = "<table>" for r in tb.rows: html += "<tr>" i = 0 while i < len(r.cells): span = 1 c = r.cells[i] for j in range(i + 1, len(r.cells)): if c.text == r.cells[j].text: span += 1 i = j i += 1 html += f"<td>{c.text}</td>" if span == 1 else f"<td colspan='{span}'>{c.text}</td>" html += "</tr>" html += "</table>" tbls.append(((None, html), "")) return qai_list, tbls def rmPrefix(txt): return re.sub( r"^(问题|答案|回答|user|assistant|Q|A|Question|Answer|问|答)[\t:: ]+", "", txt.strip(), flags=re.IGNORECASE) def beAdocPdf(d, q, a, eng, image, poss): qprefix = "Question: " if eng else "问题:" aprefix = "Answer: " if eng else "回答:" d["content_with_weight"] = "\t".join( [qprefix + rmPrefix(q), aprefix + rmPrefix(a)]) d["content_ltks"] = rag_tokenizer.tokenize(q) d["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(d["content_ltks"]) if image: d["image"] = image d["doc_type_kwd"] = "image" add_positions(d, poss) return d def beAdocDocx(d, q, a, eng, image, row_num=-1): qprefix = "Question: " if eng else "问题:" aprefix = "Answer: " if eng else "回答:" d["content_with_weight"] = "\t".join( [qprefix + rmPrefix(q), aprefix + rmPrefix(a)]) d["content_ltks"] = rag_tokenizer.tokenize(q) d["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(d["content_ltks"]) if image: d["image"] = image d["doc_type_kwd"] = "image" if row_num >= 0: d["top_int"] = [row_num] return d def beAdoc(d, q, a, eng, row_num=-1): qprefix = "Question: " if eng else "问题:" aprefix = "Answer: " if eng else "回答:" d["content_with_weight"] = "\t".join( [qprefix + rmPrefix(q), aprefix + rmPrefix(a)]) d["content_ltks"] = rag_tokenizer.tokenize(q) d["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(d["content_ltks"]) if row_num >= 0: d["top_int"] = [row_num] return d def mdQuestionLevel(s): match = re.match(r'#*', s) return (len(match.group(0)), s.lstrip('#').lstrip()) if match else (0, s) def chunk(filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, **kwargs): """ Excel and csv(txt) format files are supported. If the file is in Excel format, there should be 2 column question and answer without header. And question column is ahead of answer column. And it's O.K if it has multiple sheets as long as the columns are rightly composed. If it's in csv format, it should be UTF-8 encoded. Use TAB as delimiter to separate question and answer. All the deformed lines will be ignored. Every pair of Q&A will be treated as a chunk. """ eng = lang.lower() == "english" res = [] doc = { "docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename)) } if re.search(r"\.xlsx?$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") excel_parser = Excel() for ii, (q, a) in enumerate(excel_parser(filename, binary, callback)): res.append(beAdoc(deepcopy(doc), q, a, eng, ii)) return res elif re.search(r"\.(txt)$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") txt = get_text(filename, binary) lines = txt.split("\n") comma, tab = 0, 0 for line in lines: if len(line.split(",")) == 2: comma += 1 if len(line.split("\t")) == 2: tab += 1 delimiter = "\t" if tab >= comma else "," fails = [] question, answer = "", "" i = 0 while i < len(lines): arr = lines[i].split(delimiter) if len(arr) != 2: if question: answer += "\n" + lines[i] else: fails.append(str(i + 1)) elif len(arr) == 2: if question and answer: res.append(beAdoc(deepcopy(doc), question, answer, eng, i)) question, answer = arr i += 1 if len(res) % 999 == 0: callback(len(res) * 0.6 / len(lines), ("Extract Q&A: {}".format(len(res)) + ( f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) if question: res.append(beAdoc(deepcopy(doc), question, answer, eng, len(lines))) callback(0.6, ("Extract Q&A: {}".format(len(res)) + ( f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) return res elif re.search(r"\.(csv)$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") txt = get_text(filename, binary) lines = txt.split("\n") delimiter = "\t" if any("\t" in line for line in lines) else "," fails = [] question, answer = "", "" res = [] reader = csv.reader(lines, delimiter=delimiter) for i, row in enumerate(reader): if len(row) != 2: if question: answer += "\n" + lines[i] else: fails.append(str(i + 1)) elif len(row) == 2: if question and answer: res.append(beAdoc(deepcopy(doc), question, answer, eng, i)) question, answer = row if len(res) % 999 == 0: callback(len(res) * 0.6 / len(lines), ("Extract Q&A: {}".format(len(res)) + ( f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) if question: res.append(beAdoc(deepcopy(doc), question, answer, eng, len(list(reader)))) callback(0.6, ("Extract Q&A: {}".format(len(res)) + ( f"{len(fails)} failure, line: %s..." % (",".join(fails[:3])) if fails else ""))) return res elif re.search(r"\.pdf$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") pdf_parser = Pdf() qai_list, tbls = pdf_parser(filename if not binary else binary, from_page=from_page, to_page=to_page, callback=callback) for q, a, image, poss in qai_list: res.append(beAdocPdf(deepcopy(doc), q, a, eng, image, poss)) return res elif re.search(r"\.(md|markdown|mdx)$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") txt = get_text(filename, binary) lines = txt.split("\n") _last_question, last_answer = "", "" question_stack, level_stack = [], [] code_block = False for index, line in enumerate(lines): if line.strip().startswith('```'): code_block = not code_block question_level, question = 0, '' if not code_block: question_level, question = mdQuestionLevel(line) if not question_level or question_level > 6: # not a question last_answer = f'{last_answer}\n{line}' else: # is a question if last_answer.strip(): sum_question = '\n'.join(question_stack) if sum_question: res.append(beAdoc(deepcopy(doc), sum_question, markdown(last_answer, extensions=['markdown.extensions.tables']), eng, index)) last_answer = '' i = question_level while question_stack and i <= level_stack[-1]: question_stack.pop() level_stack.pop() question_stack.append(question) level_stack.append(question_level) if last_answer.strip(): sum_question = '\n'.join(question_stack) if sum_question: res.append(beAdoc(deepcopy(doc), sum_question, markdown(last_answer, extensions=['markdown.extensions.tables']), eng, index)) return res elif re.search(r"\.docx$", filename, re.IGNORECASE): docx_parser = Docx() qai_list, tbls = docx_parser(filename, binary, from_page=0, to_page=10000, callback=callback) res = tokenize_table(tbls, doc, eng) for i, (q, a, image) in enumerate(qai_list): res.append(beAdocDocx(deepcopy(doc), q, a, eng, image, i)) return res raise NotImplementedError( "Excel, csv(txt), pdf, markdown and docx format files are supported.") if __name__ == "__main__": import sys def dummy(prog=None, msg=""): pass chunk(sys.argv[1], from_page=0, to_page=10, callback=dummy)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/book.py
rag/app/book.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import re from io import BytesIO from deepdoc.parser.utils import get_text from rag.app import naive from rag.app.naive import by_plaintext, PARSERS from common.parser_config_utils import normalize_layout_recognizer from rag.nlp import bullets_category, is_english, remove_contents_table, \ hierarchical_merge, make_colon_as_title, naive_merge, random_choices, tokenize_table, \ tokenize_chunks, attach_media_context from rag.nlp import rag_tokenizer from deepdoc.parser import PdfParser, HtmlParser from deepdoc.parser.figure_parser import vision_figure_parser_docx_wrapper from PIL import Image class Pdf(PdfParser): def __call__(self, filename, binary=None, from_page=0, to_page=100000, zoomin=3, callback=None): from timeit import default_timer as timer start = timer() callback(msg="OCR started") self.__images__( filename if not binary else binary, zoomin, from_page, to_page, callback) callback(msg="OCR finished ({:.2f}s)".format(timer() - start)) start = timer() self._layouts_rec(zoomin) callback(0.67, "Layout analysis ({:.2f}s)".format(timer() - start)) logging.debug("layouts: {}".format(timer() - start)) start = timer() self._table_transformer_job(zoomin) callback(0.68, "Table analysis ({:.2f}s)".format(timer() - start)) start = timer() self._text_merge() tbls = self._extract_table_figure(True, zoomin, True, True) self._naive_vertical_merge() self._filter_forpages() self._merge_with_same_bullet() callback(0.8, "Text extraction ({:.2f}s)".format(timer() - start)) return [(b["text"] + self._line_tag(b, zoomin), b.get("layoutno", "")) for b in self.boxes], tbls def chunk(filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, **kwargs): """ Supported file formats are docx, pdf, txt. Since a book is long and not all the parts are useful, if it's a PDF, please set up the page ranges for every book in order eliminate negative effects and save elapsed computing time. """ parser_config = kwargs.get( "parser_config", { "chunk_token_num": 512, "delimiter": "\n!?。;!?", "layout_recognize": "DeepDOC"}) doc = { "docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename)) } doc["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(doc["title_tks"]) pdf_parser = None sections, tbls = [], [] if re.search(r"\.docx$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") doc_parser = naive.Docx() # TODO: table of contents need to be removed sections, tbls = doc_parser( filename, binary=binary, from_page=from_page, to_page=to_page) remove_contents_table(sections, eng=is_english( random_choices([t for t, _ in sections], k=200))) tbls = vision_figure_parser_docx_wrapper(sections=sections, tbls=tbls, callback=callback, **kwargs) # tbls = [((None, lns), None) for lns in tbls] sections = [(item[0], item[1] if item[1] is not None else "") for item in sections if not isinstance(item[1], Image.Image)] callback(0.8, "Finish parsing.") elif re.search(r"\.pdf$", filename, re.IGNORECASE): layout_recognizer, parser_model_name = normalize_layout_recognizer( parser_config.get("layout_recognize", "DeepDOC") ) if isinstance(layout_recognizer, bool): layout_recognizer = "DeepDOC" if layout_recognizer else "Plain Text" name = layout_recognizer.strip().lower() parser = PARSERS.get(name, by_plaintext) callback(0.1, "Start to parse.") sections, tables, pdf_parser = parser( filename=filename, binary=binary, from_page=from_page, to_page=to_page, lang=lang, callback=callback, pdf_cls=Pdf, layout_recognizer=layout_recognizer, mineru_llm_name=parser_model_name, **kwargs ) if not sections and not tables: return [] if name in ["tcadp", "docling", "mineru"]: parser_config["chunk_token_num"] = 0 callback(0.8, "Finish parsing.") elif re.search(r"\.txt$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") txt = get_text(filename, binary) sections = txt.split("\n") sections = [(line, "") for line in sections if line] remove_contents_table(sections, eng=is_english( random_choices([t for t, _ in sections], k=200))) callback(0.8, "Finish parsing.") elif re.search(r"\.(htm|html)$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") sections = HtmlParser()(filename, binary) sections = [(line, "") for line in sections if line] remove_contents_table(sections, eng=is_english( random_choices([t for t, _ in sections], k=200))) callback(0.8, "Finish parsing.") elif re.search(r"\.doc$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") try: from tika import parser as tika_parser except Exception as e: callback(0.8, f"tika not available: {e}. Unsupported .doc parsing.") logging.warning(f"tika not available: {e}. Unsupported .doc parsing for {filename}.") return [] binary = BytesIO(binary) doc_parsed = tika_parser.from_buffer(binary) if doc_parsed.get('content', None) is not None: sections = doc_parsed['content'].split('\n') sections = [(line, "") for line in sections if line] remove_contents_table(sections, eng=is_english( random_choices([t for t, _ in sections], k=200))) callback(0.8, "Finish parsing.") else: raise NotImplementedError( "file type not supported yet(doc, docx, pdf, txt supported)") make_colon_as_title(sections) bull = bullets_category( [t for t in random_choices([t for t, _ in sections], k=100)]) if bull >= 0: chunks = ["\n".join(ck) for ck in hierarchical_merge(bull, sections, 5)] else: sections = [s.split("@") for s, _ in sections] sections = [(pr[0], "@" + pr[1]) if len(pr) == 2 else (pr[0], '') for pr in sections] chunks = naive_merge( sections, parser_config.get("chunk_token_num", 256), parser_config.get("delimiter", "\n。;!?") ) # is it English # is_english(random_choices([t for t, _ in sections], k=218)) eng = lang.lower() == "english" res = tokenize_table(tbls, doc, eng) res.extend(tokenize_chunks(chunks, doc, eng, pdf_parser)) table_ctx = max(0, int(parser_config.get("table_context_size", 0) or 0)) image_ctx = max(0, int(parser_config.get("image_context_size", 0) or 0)) if table_ctx or image_ctx: attach_media_context(res, table_ctx, image_ctx) return res if __name__ == "__main__": import sys def dummy(prog=None, msg=""): pass chunk(sys.argv[1], from_page=1, to_page=10, callback=dummy)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/resume.py
rag/app/resume.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import base64 import datetime import json import re import pandas as pd import requests from api.db.services.knowledgebase_service import KnowledgebaseService from rag.nlp import rag_tokenizer from deepdoc.parser.resume import refactor from deepdoc.parser.resume import step_one, step_two from common.string_utils import remove_redundant_spaces forbidden_select_fields4resume = [ "name_pinyin_kwd", "edu_first_fea_kwd", "degree_kwd", "sch_rank_kwd", "edu_fea_kwd" ] def remote_call(filename, binary): q = { "header": { "uid": 1, "user": "kevinhu", "log_id": filename }, "request": { "p": { "request_id": "1", "encrypt_type": "base64", "filename": filename, "langtype": '', "fileori": base64.b64encode(binary).decode('utf-8') }, "c": "resume_parse_module", "m": "resume_parse" } } for _ in range(3): try: resume = requests.post( "http://127.0.0.1:61670/tog", data=json.dumps(q)) resume = resume.json()["response"]["results"] resume = refactor(resume) for k in ["education", "work", "project", "training", "skill", "certificate", "language"]: if not resume.get(k) and k in resume: del resume[k] resume = step_one.refactor(pd.DataFrame([{"resume_content": json.dumps(resume), "tob_resume_id": "x", "updated_at": datetime.datetime.now().strftime( "%Y-%m-%d %H:%M:%S")}])) resume = step_two.parse(resume) return resume except Exception: logging.exception("Resume parser has not been supported yet!") return {} def chunk(filename, binary=None, callback=None, **kwargs): """ The supported file formats are pdf, docx and txt. To maximize the effectiveness, parse the resume correctly, please concat us: https://github.com/infiniflow/ragflow """ if not re.search(r"\.(pdf|doc|docx|txt)$", filename, flags=re.IGNORECASE): raise NotImplementedError("file type not supported yet(pdf supported)") if not binary: with open(filename, "rb") as f: binary = f.read() callback(0.2, "Resume parsing is going on...") resume = remote_call(filename, binary) if len(resume.keys()) < 7: callback(-1, "Resume is not successfully parsed.") raise Exception("Resume parser remote call fail!") callback(0.6, "Done parsing. Chunking...") logging.debug("chunking resume: " + json.dumps(resume, ensure_ascii=False, indent=2)) field_map = { "name_kwd": "姓名/名字", "name_pinyin_kwd": "姓名拼音/名字拼音", "gender_kwd": "性别(男,女)", "age_int": "年龄/岁/年纪", "phone_kwd": "电话/手机/微信", "email_tks": "email/e-mail/邮箱", "position_name_tks": "职位/职能/岗位/职责", "expect_city_names_tks": "期望城市", "work_exp_flt": "工作年限/工作年份/N年经验/毕业了多少年", "corporation_name_tks": "最近就职(上班)的公司/上一家公司", "first_school_name_tks": "第一学历毕业学校", "first_degree_kwd": "第一学历(高中,职高,硕士,本科,博士,初中,中技,中专,专科,专升本,MPA,MBA,EMBA)", "highest_degree_kwd": "最高学历(高中,职高,硕士,本科,博士,初中,中技,中专,专科,专升本,MPA,MBA,EMBA)", "first_major_tks": "第一学历专业", "edu_first_fea_kwd": "第一学历标签(211,留学,双一流,985,海外知名,重点大学,中专,专升本,专科,本科,大专)", "degree_kwd": "过往学历(高中,职高,硕士,本科,博士,初中,中技,中专,专科,专升本,MPA,MBA,EMBA)", "major_tks": "学过的专业/过往专业", "school_name_tks": "学校/毕业院校", "sch_rank_kwd": "学校标签(顶尖学校,精英学校,优质学校,一般学校)", "edu_fea_kwd": "教育标签(211,留学,双一流,985,海外知名,重点大学,中专,专升本,专科,本科,大专)", "corp_nm_tks": "就职过的公司/之前的公司/上过班的公司", "edu_end_int": "毕业年份", "industry_name_tks": "所在行业", "birth_dt": "生日/出生年份", "expect_position_name_tks": "期望职位/期望职能/期望岗位", } titles = [] for n in ["name_kwd", "gender_kwd", "position_name_tks", "age_int"]: v = resume.get(n, "") if isinstance(v, list): v = v[0] if n.find("tks") > 0: v = remove_redundant_spaces(v) titles.append(str(v)) doc = { "docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize("-".join(titles) + "-简历") } doc["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(doc["title_tks"]) pairs = [] for n, m in field_map.items(): if not resume.get(n): continue v = resume[n] if isinstance(v, list): v = " ".join(v) if n.find("tks") > 0: v = remove_redundant_spaces(v) pairs.append((m, str(v))) doc["content_with_weight"] = "\n".join( ["{}: {}".format(re.sub(r"([^()]+)", "", k), v) for k, v in pairs]) doc["content_ltks"] = rag_tokenizer.tokenize(doc["content_with_weight"]) doc["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(doc["content_ltks"]) for n, _ in field_map.items(): if n not in resume: continue if isinstance(resume[n], list) and ( len(resume[n]) == 1 or n not in forbidden_select_fields4resume): resume[n] = resume[n][0] if n.find("_tks") > 0: resume[n] = rag_tokenizer.fine_grained_tokenize(resume[n]) doc[n] = resume[n] logging.debug("chunked resume to " + str(doc)) KnowledgebaseService.update_parser_config( kwargs["kb_id"], {"field_map": field_map}) return [doc] if __name__ == "__main__": import sys def dummy(a, b): pass chunk(sys.argv[1], callback=dummy)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/presentation.py
rag/app/presentation.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import copy import re from collections import defaultdict from io import BytesIO from PIL import Image from PyPDF2 import PdfReader as pdf2_read from deepdoc.parser import PdfParser, PptParser, PlainParser from rag.app.naive import by_plaintext, PARSERS from common.parser_config_utils import normalize_layout_recognizer from rag.nlp import rag_tokenizer from rag.nlp import tokenize, is_english class Ppt(PptParser): def __call__(self, fnm, from_page, to_page, callback=None): txts = super().__call__(fnm, from_page, to_page) callback(0.5, "Text extraction finished.") import aspose.slides as slides import aspose.pydrawing as drawing imgs = [] with slides.Presentation(BytesIO(fnm)) as presentation: for i, slide in enumerate(presentation.slides[from_page: to_page]): try: with BytesIO() as buffered: slide.get_thumbnail( 0.1, 0.1).save( buffered, drawing.imaging.ImageFormat.jpeg) buffered.seek(0) imgs.append(Image.open(buffered).copy()) except RuntimeError as e: raise RuntimeError( f'ppt parse error at page {i + 1}, original error: {str(e)}') from e assert len(imgs) == len( txts), "Slides text and image do not match: {} vs. {}".format( len(imgs), len(txts)) callback(0.9, "Image extraction finished") self.is_english = is_english(txts) return [(txts[i], imgs[i]) for i in range(len(txts))] class Pdf(PdfParser): def __init__(self): super().__init__() def __call__(self, filename, binary=None, from_page=0, to_page=100000, zoomin=3, callback=None, **kwargs): # 1. OCR callback(msg="OCR started") self.__images__(filename if not binary else binary, zoomin, from_page, to_page, callback) # 2. Layout Analysis callback(msg="Layout Analysis") self._layouts_rec(zoomin) # 3. Table Analysis callback(msg="Table Analysis") self._table_transformer_job(zoomin) # 4. Text Merge self._text_merge() # 5. Extract Tables (Force HTML) tbls = self._extract_table_figure(True, zoomin, True, True) # 6. Re-assemble Page Content page_items = defaultdict(list) # (A) Add text for b in self.boxes: # b["page_number"] is relative page number,must + from_page global_page_num = b["page_number"] + from_page if not (from_page < global_page_num <= to_page + from_page): continue page_items[global_page_num].append({ "top": b["top"], "x0": b["x0"], "text": b["text"], "type": "text" }) # (B) Add table and figure for (img, content), positions in tbls: if not positions: continue if isinstance(content, list): final_text = "\n".join(content) elif isinstance(content, str): final_text = content else: final_text = str(content) try: pn_index = positions[0][0] if isinstance(pn_index, list): pn_index = pn_index[0] # pn_index in tbls is absolute page number current_page_num = int(pn_index) + 1 except Exception as e: print(f"Error parsing position: {e}") continue if not (from_page < current_page_num <= to_page + from_page): continue top = positions[0][3] left = positions[0][1] page_items[current_page_num].append({ "top": top, "x0": left, "text": final_text, "type": "table_or_figure" }) # 7. Generate result res = [] for i in range(len(self.page_images)): current_pn = from_page + i + 1 items = page_items.get(current_pn, []) # Sort by vertical position items.sort(key=lambda x: (x["top"], x["x0"])) full_page_text = "\n\n".join([item["text"] for item in items]) if not full_page_text.strip(): full_page_text = f"[No text or data found in Page {current_pn}]" page_img = self.page_images[i] res.append((full_page_text, page_img)) callback(0.9, "Parsing finished") return res, [] class PlainPdf(PlainParser): def __call__(self, filename, binary=None, from_page=0, to_page=100000, callback=None, **kwargs): self.pdf = pdf2_read(filename if not binary else BytesIO(binary)) page_txt = [] for page in self.pdf.pages[from_page: to_page]: page_txt.append(page.extract_text()) callback(0.9, "Parsing finished") return [(txt, None) for txt in page_txt], [] def chunk(filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, parser_config=None, **kwargs): """ The supported file formats are pdf, pptx. Every page will be treated as a chunk. And the thumbnail of every page will be stored. PPT file will be parsed by using this method automatically, setting-up for every PPT file is not necessary. """ if parser_config is None: parser_config = {} eng = lang.lower() == "english" doc = { "docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize( re.sub(r"\.[a-zA-Z]+$", "", filename)) } doc["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(doc["title_tks"]) res = [] if re.search(r"\.pptx?$", filename, re.IGNORECASE): ppt_parser = Ppt() for pn, (txt, img) in enumerate(ppt_parser( filename if not binary else binary, from_page, 1000000, callback)): d = copy.deepcopy(doc) pn += from_page d["image"] = img d["doc_type_kwd"] = "image" d["page_num_int"] = [pn + 1] d["top_int"] = [0] d["position_int"] = [(pn + 1, 0, img.size[0], 0, img.size[1])] tokenize(d, txt, eng) res.append(d) return res elif re.search(r"\.pdf$", filename, re.IGNORECASE): layout_recognizer, parser_model_name = normalize_layout_recognizer( parser_config.get("layout_recognize", "DeepDOC") ) if isinstance(layout_recognizer, bool): layout_recognizer = "DeepDOC" if layout_recognizer else "Plain Text" name = layout_recognizer.strip().lower() parser = PARSERS.get(name, by_plaintext) callback(0.1, "Start to parse.") sections, _, _ = parser( filename=filename, binary=binary, from_page=from_page, to_page=to_page, lang=lang, callback=callback, pdf_cls=Pdf, layout_recognizer=layout_recognizer, mineru_llm_name=parser_model_name, **kwargs ) if not sections: return [] if name in ["tcadp", "docling", "mineru"]: parser_config["chunk_token_num"] = 0 callback(0.8, "Finish parsing.") for pn, (txt, img) in enumerate(sections): d = copy.deepcopy(doc) pn += from_page if not isinstance(img, Image.Image): img = None d["image"] = img d["page_num_int"] = [pn + 1] d["top_int"] = [0] d["position_int"] = [(pn + 1, 0, img.size[0] if img else 0, 0, img.size[1] if img else 0)] tokenize(d, txt, eng) res.append(d) return res raise NotImplementedError( "file type not supported yet(pptx, pdf supported)") if __name__ == "__main__": import sys def dummy(a, b): pass chunk(sys.argv[1], callback=dummy)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/picture.py
rag/app/picture.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import asyncio import io import re import numpy as np from PIL import Image from api.db.services.llm_service import LLMBundle from common.constants import LLMType from common.string_utils import clean_markdown_block from deepdoc.vision import OCR from rag.nlp import attach_media_context, rag_tokenizer, tokenize ocr = OCR() # Gemini supported MIME types VIDEO_EXTS = [".mp4", ".mov", ".avi", ".flv", ".mpeg", ".mpg", ".webm", ".wmv", ".3gp", ".3gpp", ".mkv"] def chunk(filename, binary, tenant_id, lang, callback=None, **kwargs): doc = { "docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename)), } eng = lang.lower() == "english" parser_config = kwargs.get("parser_config", {}) or {} image_ctx = max(0, int(parser_config.get("image_context_size", 0) or 0)) if any(filename.lower().endswith(ext) for ext in VIDEO_EXTS): try: doc.update( { "doc_type_kwd": "video", } ) cv_mdl = LLMBundle(tenant_id, llm_type=LLMType.IMAGE2TEXT, lang=lang) ans = asyncio.run( cv_mdl.async_chat(system="", history=[], gen_conf={}, video_bytes=binary, filename=filename)) callback(0.8, "CV LLM respond: %s ..." % ans[:32]) ans += "\n" + ans tokenize(doc, ans, eng) return [doc] except Exception as e: callback(prog=-1, msg=str(e)) else: img = Image.open(io.BytesIO(binary)).convert("RGB") doc.update( { "image": img, "doc_type_kwd": "image", } ) bxs = ocr(np.array(img)) txt = "\n".join([t[0] for _, t in bxs if t[0]]) callback(0.4, "Finish OCR: (%s ...)" % txt[:12]) if (eng and len(txt.split()) > 32) or len(txt) > 32: tokenize(doc, txt, eng) callback(0.8, "OCR results is too long to use CV LLM.") return attach_media_context([doc], 0, image_ctx) try: callback(0.4, "Use CV LLM to describe the picture.") cv_mdl = LLMBundle(tenant_id, LLMType.IMAGE2TEXT, lang=lang) img_binary = io.BytesIO() img.save(img_binary, format="JPEG") img_binary.seek(0) ans = cv_mdl.describe(img_binary.read()) callback(0.8, "CV LLM respond: %s ..." % ans[:32]) txt += "\n" + ans tokenize(doc, txt, eng) return attach_media_context([doc], 0, image_ctx) except Exception as e: callback(prog=-1, msg=str(e)) return [] def vision_llm_chunk(binary, vision_model, prompt=None, callback=None): """ A simple wrapper to process image to markdown texts via VLM. Returns: Simple markdown texts generated by VLM. """ callback = callback or (lambda prog, msg: None) img = binary txt = "" try: with io.BytesIO() as img_binary: try: img.save(img_binary, format="JPEG") except Exception: img_binary.seek(0) img_binary.truncate() img.save(img_binary, format="PNG") img_binary.seek(0) ans = clean_markdown_block(vision_model.describe_with_prompt(img_binary.read(), prompt)) txt += "\n" + ans return txt except Exception as e: callback(-1, str(e)) return ""
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/__init__.py
rag/app/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/email.py
rag/app/email.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging from email import policy from email.parser import BytesParser from rag.app.naive import chunk as naive_chunk import re from rag.nlp import rag_tokenizer, naive_merge, tokenize_chunks from deepdoc.parser import HtmlParser, TxtParser from timeit import default_timer as timer import io def chunk( filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, **kwargs, ): """ Only eml is supported """ eng = lang.lower() == "english" # is_english(cks) parser_config = kwargs.get( "parser_config", {"chunk_token_num": 512, "delimiter": "\n!?。;!?", "layout_recognize": "DeepDOC"}, ) doc = { "docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename)), } doc["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(doc["title_tks"]) main_res = [] attachment_res = [] if binary: with io.BytesIO(binary) as buffer: msg = BytesParser(policy=policy.default).parse(buffer) else: with open(filename, "rb") as buffer: msg = BytesParser(policy=policy.default).parse(buffer) text_txt, html_txt = [], [] # get the email header info for header, value in msg.items(): text_txt.append(f"{header}: {value}") # get the email main info def _add_content(msg, content_type): def _decode_payload(payload, charset, target_list): try: target_list.append(payload.decode(charset)) except (UnicodeDecodeError, LookupError): for enc in ["utf-8", "gb2312", "gbk", "gb18030", "latin1"]: try: target_list.append(payload.decode(enc)) break except UnicodeDecodeError: continue else: target_list.append(payload.decode("utf-8", errors="ignore")) if content_type == "text/plain": payload = msg.get_payload(decode=True) charset = msg.get_content_charset() or "utf-8" _decode_payload(payload, charset, text_txt) elif content_type == "text/html": payload = msg.get_payload(decode=True) charset = msg.get_content_charset() or "utf-8" _decode_payload(payload, charset, html_txt) elif "multipart" in content_type: if msg.is_multipart(): for part in msg.iter_parts(): _add_content(part, part.get_content_type()) _add_content(msg, msg.get_content_type()) sections = TxtParser.parser_txt("\n".join(text_txt)) + [ (line, "") for line in HtmlParser.parser_txt("\n".join(html_txt), chunk_token_num=parser_config["chunk_token_num"]) if line ] st = timer() chunks = naive_merge( sections, int(parser_config.get("chunk_token_num", 128)), parser_config.get("delimiter", "\n!?。;!?"), ) main_res.extend(tokenize_chunks(chunks, doc, eng, None)) logging.debug("naive_merge({}): {}".format(filename, timer() - st)) # get the attachment info for part in msg.iter_attachments(): content_disposition = part.get("Content-Disposition") if content_disposition: dispositions = content_disposition.strip().split(";") if dispositions[0].lower() == "attachment": filename = part.get_filename() payload = part.get_payload(decode=True) try: attachment_res.extend( naive_chunk(filename, payload, callback=callback, **kwargs) ) except Exception: pass return main_res + attachment_res if __name__ == "__main__": import sys def dummy(prog=None, msg=""): pass chunk(sys.argv[1], callback=dummy)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/paper.py
rag/app/paper.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import copy import re from deepdoc.parser.figure_parser import vision_figure_parser_pdf_wrapper from common.constants import ParserType from rag.nlp import rag_tokenizer, tokenize, tokenize_table, add_positions, bullets_category, title_frequency, \ tokenize_chunks, attach_media_context from deepdoc.parser import PdfParser import numpy as np from rag.app.naive import by_plaintext, PARSERS from common.parser_config_utils import normalize_layout_recognizer class Pdf(PdfParser): def __init__(self): self.model_speciess = ParserType.PAPER.value super().__init__() def __call__(self, filename, binary=None, from_page=0, to_page=100000, zoomin=3, callback=None): from timeit import default_timer as timer start = timer() callback(msg="OCR started") self.__images__( filename if not binary else binary, zoomin, from_page, to_page, callback ) callback(msg="OCR finished ({:.2f}s)".format(timer() - start)) start = timer() self._layouts_rec(zoomin) callback(0.63, "Layout analysis ({:.2f}s)".format(timer() - start)) logging.debug(f"layouts cost: {timer() - start}s") start = timer() self._table_transformer_job(zoomin) callback(0.68, "Table analysis ({:.2f}s)".format(timer() - start)) start = timer() self._text_merge() tbls = self._extract_table_figure(True, zoomin, True, True) column_width = np.median([b["x1"] - b["x0"] for b in self.boxes]) self._concat_downward() self._filter_forpages() callback(0.75, "Text merged ({:.2f}s)".format(timer() - start)) # clean mess if column_width < self.page_images[0].size[0] / zoomin / 2: logging.debug("two_column................... {} {}".format(column_width, self.page_images[0].size[0] / zoomin / 2)) self.boxes = self.sort_X_by_page(self.boxes, column_width / 2) for b in self.boxes: b["text"] = re.sub(r"([\t  ]|\u3000){2,}", " ", b["text"].strip()) def _begin(txt): return re.match( "[0-9. 一、i]*(introduction|abstract|摘要|引言|keywords|key words|关键词|background|背景|目录|前言|contents)", txt.lower().strip()) if from_page > 0: return { "title": "", "authors": "", "abstract": "", "sections": [(b["text"] + self._line_tag(b, zoomin), b.get("layoutno", "")) for b in self.boxes if re.match(r"(text|title)", b.get("layoutno", "text"))], "tables": tbls } # get title and authors title = "" authors = [] i = 0 while i < min(32, len(self.boxes) - 1): b = self.boxes[i] i += 1 if b.get("layoutno", "").find("title") >= 0: title = b["text"] if _begin(title): title = "" break for j in range(3): if _begin(self.boxes[i + j]["text"]): break authors.append(self.boxes[i + j]["text"]) break break # get abstract abstr = "" i = 0 while i + 1 < min(32, len(self.boxes)): b = self.boxes[i] i += 1 txt = b["text"].lower().strip() if re.match("(abstract|摘要)", txt): if len(txt.split()) > 32 or len(txt) > 64: abstr = txt + self._line_tag(b, zoomin) break txt = self.boxes[i]["text"].lower().strip() if len(txt.split()) > 32 or len(txt) > 64: abstr = txt + self._line_tag(self.boxes[i], zoomin) i += 1 break if not abstr: i = 0 callback( 0.8, "Page {}~{}: Text merging finished".format( from_page, min( to_page, self.total_page))) for b in self.boxes: logging.debug("{} {}".format(b["text"], b.get("layoutno"))) logging.debug("{}".format(tbls)) return { "title": title, "authors": " ".join(authors), "abstract": abstr, "sections": [(b["text"] + self._line_tag(b, zoomin), b.get("layoutno", "")) for b in self.boxes[i:] if re.match(r"(text|title)", b.get("layoutno", "text"))], "tables": tbls } def chunk(filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, **kwargs): """ Only pdf is supported. The abstract of the paper will be sliced as an entire chunk, and will not be sliced partly. """ parser_config = kwargs.get( "parser_config", { "chunk_token_num": 512, "delimiter": "\n!?。;!?", "layout_recognize": "DeepDOC"}) if re.search(r"\.pdf$", filename, re.IGNORECASE): layout_recognizer, parser_model_name = normalize_layout_recognizer( parser_config.get("layout_recognize", "DeepDOC") ) if isinstance(layout_recognizer, bool): layout_recognizer = "DeepDOC" if layout_recognizer else "Plain Text" name = layout_recognizer.strip().lower() pdf_parser = PARSERS.get(name, by_plaintext) callback(0.1, "Start to parse.") if name == "deepdoc": pdf_parser = Pdf() paper = pdf_parser(filename if not binary else binary, from_page=from_page, to_page=to_page, callback=callback) else: kwargs.pop("parse_method", None) kwargs.pop("mineru_llm_name", None) sections, tables, pdf_parser = pdf_parser( filename=filename, binary=binary, from_page=from_page, to_page=to_page, lang=lang, callback=callback, pdf_cls=Pdf, layout_recognizer=layout_recognizer, mineru_llm_name=parser_model_name, parse_method="paper", **kwargs ) paper = { "title": filename, "authors": " ", "abstract": "", "sections": sections, "tables": tables } tbls = paper["tables"] tbls = vision_figure_parser_pdf_wrapper(tbls=tbls, callback=callback, **kwargs) paper["tables"] = tbls else: raise NotImplementedError("file type not supported yet(pdf supported)") doc = {"docnm_kwd": filename, "authors_tks": rag_tokenizer.tokenize(paper["authors"]), "title_tks": rag_tokenizer.tokenize(paper["title"] if paper["title"] else filename)} doc["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(doc["title_tks"]) doc["authors_sm_tks"] = rag_tokenizer.fine_grained_tokenize(doc["authors_tks"]) # is it English eng = lang.lower() == "english" # pdf_parser.is_english logging.debug("It's English.....{}".format(eng)) res = tokenize_table(paper["tables"], doc, eng) if paper["abstract"]: d = copy.deepcopy(doc) txt = pdf_parser.remove_tag(paper["abstract"]) d["important_kwd"] = ["abstract", "总结", "概括", "summary", "summarize"] d["important_tks"] = " ".join(d["important_kwd"]) d["image"], poss = pdf_parser.crop( paper["abstract"], need_position=True) add_positions(d, poss) tokenize(d, txt, eng) res.append(d) sorted_sections = paper["sections"] # set pivot using the most frequent type of title, # then merge between 2 pivot bull = bullets_category([txt for txt, _ in sorted_sections]) most_level, levels = title_frequency(bull, sorted_sections) assert len(sorted_sections) == len(levels) sec_ids = [] sid = 0 for i, lvl in enumerate(levels): if lvl <= most_level and i > 0 and lvl != levels[i - 1]: sid += 1 sec_ids.append(sid) logging.debug("{} {} {} {}".format(lvl, sorted_sections[i][0], most_level, sid)) chunks = [] last_sid = -2 for (txt, _), sec_id in zip(sorted_sections, sec_ids): if sec_id == last_sid: if chunks: chunks[-1] += "\n" + txt continue chunks.append(txt) last_sid = sec_id res.extend(tokenize_chunks(chunks, doc, eng, pdf_parser)) table_ctx = max(0, int(parser_config.get("table_context_size", 0) or 0)) image_ctx = max(0, int(parser_config.get("image_context_size", 0) or 0)) if table_ctx or image_ctx: attach_media_context(res, table_ctx, image_ctx) return res """ readed = [0] * len(paper["lines"]) # find colon firstly i = 0 while i + 1 < len(paper["lines"]): txt = pdf_parser.remove_tag(paper["lines"][i][0]) j = i if txt.strip("\n").strip()[-1] not in "::": i += 1 continue i += 1 while i < len(paper["lines"]) and not paper["lines"][i][0]: i += 1 if i >= len(paper["lines"]): break proj = [paper["lines"][i][0].strip()] i += 1 while i < len(paper["lines"]) and paper["lines"][i][0].strip()[0] == proj[-1][0]: proj.append(paper["lines"][i]) i += 1 for k in range(j, i): readed[k] = True txt = txt[::-1] if eng: r = re.search(r"(.*?) ([\\.;?!]|$)", txt) txt = r.group(1)[::-1] if r else txt[::-1] else: r = re.search(r"(.*?) ([。?;!]|$)", txt) txt = r.group(1)[::-1] if r else txt[::-1] for p in proj: d = copy.deepcopy(doc) txt += "\n" + pdf_parser.remove_tag(p) d["image"], poss = pdf_parser.crop(p, need_position=True) add_positions(d, poss) tokenize(d, txt, eng) res.append(d) i = 0 chunk = [] tk_cnt = 0 def add_chunk(): nonlocal chunk, res, doc, pdf_parser, tk_cnt d = copy.deepcopy(doc) ck = "\n".join(chunk) tokenize(d, pdf_parser.remove_tag(ck), pdf_parser.is_english) d["image"], poss = pdf_parser.crop(ck, need_position=True) add_positions(d, poss) res.append(d) chunk = [] tk_cnt = 0 while i < len(paper["lines"]): if tk_cnt > 128: add_chunk() if readed[i]: i += 1 continue readed[i] = True txt, layouts = paper["lines"][i] txt_ = pdf_parser.remove_tag(txt) i += 1 cnt = num_tokens_from_string(txt_) if any([ layouts.find("title") >= 0 and chunk, cnt + tk_cnt > 128 and tk_cnt > 32, ]): add_chunk() chunk = [txt] tk_cnt = cnt else: chunk.append(txt) tk_cnt += cnt if chunk: add_chunk() for i, d in enumerate(res): print(d) # d["image"].save(f"./logs/{i}.jpg") return res """ if __name__ == "__main__": import sys def dummy(prog=None, msg=""): pass chunk(sys.argv[1], callback=dummy)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/manual.py
rag/app/manual.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import copy import re from common.constants import ParserType from io import BytesIO from rag.nlp import rag_tokenizer, tokenize, tokenize_table, bullets_category, title_frequency, tokenize_chunks, \ docx_question_level, attach_media_context from common.token_utils import num_tokens_from_string from deepdoc.parser import PdfParser, DocxParser from deepdoc.parser.figure_parser import vision_figure_parser_pdf_wrapper, vision_figure_parser_docx_wrapper from docx import Document from PIL import Image from rag.app.naive import by_plaintext, PARSERS from common.parser_config_utils import normalize_layout_recognizer class Pdf(PdfParser): def __init__(self): self.model_speciess = ParserType.MANUAL.value super().__init__() def __call__(self, filename, binary=None, from_page=0, to_page=100000, zoomin=3, callback=None): from timeit import default_timer as timer start = timer() callback(msg="OCR started") self.__images__( filename if not binary else binary, zoomin, from_page, to_page, callback ) callback(msg="OCR finished ({:.2f}s)".format(timer() - start)) logging.debug("OCR: {}".format(timer() - start)) start = timer() self._layouts_rec(zoomin) callback(0.65, "Layout analysis ({:.2f}s)".format(timer() - start)) logging.debug("layouts: {}".format(timer() - start)) start = timer() self._table_transformer_job(zoomin) callback(0.67, "Table analysis ({:.2f}s)".format(timer() - start)) start = timer() self._text_merge() tbls = self._extract_table_figure(True, zoomin, True, True) self._concat_downward() self._filter_forpages() callback(0.68, "Text merged ({:.2f}s)".format(timer() - start)) # clean mess for b in self.boxes: b["text"] = re.sub(r"([\t  ]|\u3000){2,}", " ", b["text"].strip()) return [(b["text"], b.get("layoutno", ""), self.get_position(b, zoomin)) for i, b in enumerate(self.boxes)], tbls class Docx(DocxParser): def __init__(self): pass def get_picture(self, document, paragraph): img = paragraph._element.xpath('.//pic:pic') if not img: return None try: img = img[0] embed = img.xpath('.//a:blip/@r:embed')[0] related_part = document.part.related_parts[embed] image = related_part.image if image is not None: image = Image.open(BytesIO(image.blob)) return image elif related_part.blob is not None: image = Image.open(BytesIO(related_part.blob)) return image else: return None except Exception: return None def concat_img(self, img1, img2): if img1 and not img2: return img1 if not img1 and img2: return img2 if not img1 and not img2: return None width1, height1 = img1.size width2, height2 = img2.size new_width = max(width1, width2) new_height = height1 + height2 new_image = Image.new('RGB', (new_width, new_height)) new_image.paste(img1, (0, 0)) new_image.paste(img2, (0, height1)) return new_image def __call__(self, filename, binary=None, from_page=0, to_page=100000, callback=None): self.doc = Document( filename) if not binary else Document(BytesIO(binary)) pn = 0 last_answer, last_image = "", None question_stack, level_stack = [], [] ti_list = [] for p in self.doc.paragraphs: if pn > to_page: break question_level, p_text = 0, '' if from_page <= pn < to_page and p.text.strip(): question_level, p_text = docx_question_level(p) if not question_level or question_level > 6: # not a question last_answer = f'{last_answer}\n{p_text}' current_image = self.get_picture(self.doc, p) last_image = self.concat_img(last_image, current_image) else: # is a question if last_answer or last_image: sum_question = '\n'.join(question_stack) if sum_question: ti_list.append((f'{sum_question}\n{last_answer}', last_image)) last_answer, last_image = '', None i = question_level while question_stack and i <= level_stack[-1]: question_stack.pop() level_stack.pop() question_stack.append(p_text) level_stack.append(question_level) for run in p.runs: if 'lastRenderedPageBreak' in run._element.xml: pn += 1 continue if 'w:br' in run._element.xml and 'type="page"' in run._element.xml: pn += 1 if last_answer: sum_question = '\n'.join(question_stack) if sum_question: ti_list.append((f'{sum_question}\n{last_answer}', last_image)) tbls = [] for tb in self.doc.tables: html = "<table>" for r in tb.rows: html += "<tr>" i = 0 while i < len(r.cells): span = 1 c = r.cells[i] for j in range(i + 1, len(r.cells)): if c.text == r.cells[j].text: span += 1 i = j else: break i += 1 html += f"<td>{c.text}</td>" if span == 1 else f"<td colspan='{span}'>{c.text}</td>" html += "</tr>" html += "</table>" tbls.append(((None, html), "")) return ti_list, tbls def chunk(filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, **kwargs): """ Only pdf is supported. """ parser_config = kwargs.get( "parser_config", { "chunk_token_num": 512, "delimiter": "\n!?。;!?", "layout_recognize": "DeepDOC"}) pdf_parser = None doc = { "docnm_kwd": filename } doc["title_tks"] = rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", doc["docnm_kwd"])) doc["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(doc["title_tks"]) # is it English eng = lang.lower() == "english" # pdf_parser.is_english if re.search(r"\.pdf$", filename, re.IGNORECASE): layout_recognizer, parser_model_name = normalize_layout_recognizer( parser_config.get("layout_recognize", "DeepDOC") ) if isinstance(layout_recognizer, bool): layout_recognizer = "DeepDOC" if layout_recognizer else "Plain Text" name = layout_recognizer.strip().lower() pdf_parser = PARSERS.get(name, by_plaintext) callback(0.1, "Start to parse.") kwargs.pop("parse_method", None) kwargs.pop("mineru_llm_name", None) sections, tbls, pdf_parser = pdf_parser( filename=filename, binary=binary, from_page=from_page, to_page=to_page, lang=lang, callback=callback, pdf_cls=Pdf, layout_recognizer=layout_recognizer, mineru_llm_name=parser_model_name, parse_method="manual", **kwargs ) def _normalize_section(section): # pad section to length 3: (txt, sec_id, poss) if len(section) == 1: section = (section[0], "", []) elif len(section) == 2: section = (section[0], "", section[1]) elif len(section) != 3: raise ValueError(f"Unexpected section length: {len(section)} (value={section!r})") txt, layoutno, poss = section if isinstance(poss, str): poss = pdf_parser.extract_positions(poss) if poss: first = poss[0] # tuple: ([pn], x1, x2, y1, y2) pn = first[0] if isinstance(pn, list) and pn: pn = pn[0] # [pn] -> pn poss[0] = (pn, *first[1:]) return (txt, layoutno, poss) sections = [_normalize_section(sec) for sec in sections] if not sections and not tbls: return [] if name in ["tcadp", "docling", "mineru"]: parser_config["chunk_token_num"] = 0 callback(0.8, "Finish parsing.") if len(sections) > 0 and len(pdf_parser.outlines) / len(sections) > 0.03: max_lvl = max([lvl for _, lvl in pdf_parser.outlines]) most_level = max(0, max_lvl - 1) levels = [] for txt, _, _ in sections: for t, lvl in pdf_parser.outlines: tks = set([t[i] + t[i + 1] for i in range(len(t) - 1)]) tks_ = set([txt[i] + txt[i + 1] for i in range(min(len(t), len(txt) - 1))]) if len(set(tks & tks_)) / max([len(tks), len(tks_), 1]) > 0.8: levels.append(lvl) break else: levels.append(max_lvl + 1) else: bull = bullets_category([txt for txt, _, _ in sections]) most_level, levels = title_frequency( bull, [(txt, lvl) for txt, lvl, _ in sections]) assert len(sections) == len(levels) sec_ids = [] sid = 0 for i, lvl in enumerate(levels): if lvl <= most_level and i > 0 and lvl != levels[i - 1]: sid += 1 sec_ids.append(sid) sections = [(txt, sec_ids[i], poss) for i, (txt, _, poss) in enumerate(sections)] for (img, rows), poss in tbls: if not rows: continue sections.append((rows if isinstance(rows, str) else rows[0], -1, [(p[0] + 1 - from_page, p[1], p[2], p[3], p[4]) for p in poss])) def tag(pn, left, right, top, bottom): if pn + left + right + top + bottom == 0: return "" return "@@{}\t{:.1f}\t{:.1f}\t{:.1f}\t{:.1f}##" \ .format(pn, left, right, top, bottom) chunks = [] last_sid = -2 tk_cnt = 0 for txt, sec_id, poss in sorted(sections, key=lambda x: ( x[-1][0][0], x[-1][0][3], x[-1][0][1])): poss = "\t".join([tag(*pos) for pos in poss]) if tk_cnt < 32 or (tk_cnt < 1024 and (sec_id == last_sid or sec_id == -1)): if chunks: chunks[-1] += "\n" + txt + poss tk_cnt += num_tokens_from_string(txt) continue chunks.append(txt + poss) tk_cnt = num_tokens_from_string(txt) if sec_id > -1: last_sid = sec_id tbls = vision_figure_parser_pdf_wrapper(tbls=tbls, callback=callback, **kwargs) res = tokenize_table(tbls, doc, eng) res.extend(tokenize_chunks(chunks, doc, eng, pdf_parser)) table_ctx = max(0, int(parser_config.get("table_context_size", 0) or 0)) image_ctx = max(0, int(parser_config.get("image_context_size", 0) or 0)) if table_ctx or image_ctx: attach_media_context(res, table_ctx, image_ctx) return res elif re.search(r"\.docx?$", filename, re.IGNORECASE): docx_parser = Docx() ti_list, tbls = docx_parser(filename, binary, from_page=0, to_page=10000, callback=callback) tbls = vision_figure_parser_docx_wrapper(sections=ti_list, tbls=tbls, callback=callback, **kwargs) res = tokenize_table(tbls, doc, eng) for text, image in ti_list: d = copy.deepcopy(doc) if image: d['image'] = image d["doc_type_kwd"] = "image" tokenize(d, text, eng) res.append(d) table_ctx = max(0, int(parser_config.get("table_context_size", 0) or 0)) image_ctx = max(0, int(parser_config.get("image_context_size", 0) or 0)) if table_ctx or image_ctx: attach_media_context(res, table_ctx, image_ctx) return res else: raise NotImplementedError("file type not supported yet(pdf and docx supported)") if __name__ == "__main__": import sys def dummy(prog=None, msg=""): pass chunk(sys.argv[1], callback=dummy)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/app/laws.py
rag/app/laws.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import re from io import BytesIO from docx import Document from common.constants import ParserType from deepdoc.parser.utils import get_text from rag.nlp import bullets_category, remove_contents_table, \ make_colon_as_title, tokenize_chunks, docx_question_level, tree_merge from rag.nlp import rag_tokenizer, Node from deepdoc.parser import PdfParser, DocxParser, HtmlParser from rag.app.naive import by_plaintext, PARSERS from common.parser_config_utils import normalize_layout_recognizer class Docx(DocxParser): def __init__(self): pass def __clean(self, line): line = re.sub(r"\u3000", " ", line).strip() return line def old_call(self, filename, binary=None, from_page=0, to_page=100000): self.doc = Document( filename) if not binary else Document(BytesIO(binary)) pn = 0 lines = [] for p in self.doc.paragraphs: if pn > to_page: break if from_page <= pn < to_page and p.text.strip(): lines.append(self.__clean(p.text)) for run in p.runs: if 'lastRenderedPageBreak' in run._element.xml: pn += 1 continue if 'w:br' in run._element.xml and 'type="page"' in run._element.xml: pn += 1 return [line for line in lines if line] def __call__(self, filename, binary=None, from_page=0, to_page=100000): self.doc = Document( filename) if not binary else Document(BytesIO(binary)) pn = 0 lines = [] level_set = set() bull = bullets_category([p.text for p in self.doc.paragraphs]) for p in self.doc.paragraphs: if pn > to_page: break question_level, p_text = docx_question_level(p, bull) if not p_text.strip("\n"): continue lines.append((question_level, p_text)) level_set.add(question_level) for run in p.runs: if 'lastRenderedPageBreak' in run._element.xml: pn += 1 continue if 'w:br' in run._element.xml and 'type="page"' in run._element.xml: pn += 1 sorted_levels = sorted(level_set) h2_level = sorted_levels[1] if len(sorted_levels) > 1 else 1 h2_level = sorted_levels[-2] if h2_level == sorted_levels[-1] and len(sorted_levels) > 2 else h2_level root = Node(level=0, depth=h2_level, texts=[]) root.build_tree(lines) return [element for element in root.get_tree() if element] def __str__(self) -> str: return f''' question:{self.question}, answer:{self.answer}, level:{self.level}, childs:{self.childs} ''' class Pdf(PdfParser): def __init__(self): self.model_speciess = ParserType.LAWS.value super().__init__() def __call__(self, filename, binary=None, from_page=0, to_page=100000, zoomin=3, callback=None): from timeit import default_timer as timer start = timer() callback(msg="OCR started") self.__images__( filename if not binary else binary, zoomin, from_page, to_page, callback ) callback(msg="OCR finished ({:.2f}s)".format(timer() - start)) start = timer() self._layouts_rec(zoomin) callback(0.67, "Layout analysis ({:.2f}s)".format(timer() - start)) logging.debug("layouts: {}".format((timer() - start))) self._naive_vertical_merge() callback(0.8, "Text extraction ({:.2f}s)".format(timer() - start)) return [(b["text"], self._line_tag(b, zoomin)) for b in self.boxes], None def chunk(filename, binary=None, from_page=0, to_page=100000, lang="Chinese", callback=None, **kwargs): """ Supported file formats are docx, pdf, txt. """ parser_config = kwargs.get( "parser_config", { "chunk_token_num": 512, "delimiter": "\n!?。;!?", "layout_recognize": "DeepDOC"}) doc = { "docnm_kwd": filename, "title_tks": rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", filename)) } doc["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(doc["title_tks"]) pdf_parser = None sections = [] # is it English eng = lang.lower() == "english" # is_english(sections) if re.search(r"\.docx$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") chunks = Docx()(filename, binary) callback(0.7, "Finish parsing.") return tokenize_chunks(chunks, doc, eng, None) elif re.search(r"\.pdf$", filename, re.IGNORECASE): layout_recognizer, parser_model_name = normalize_layout_recognizer( parser_config.get("layout_recognize", "DeepDOC") ) if isinstance(layout_recognizer, bool): layout_recognizer = "DeepDOC" if layout_recognizer else "Plain Text" name = layout_recognizer.strip().lower() parser = PARSERS.get(name, by_plaintext) callback(0.1, "Start to parse.") raw_sections, tables, pdf_parser = parser( filename=filename, binary=binary, from_page=from_page, to_page=to_page, lang=lang, callback=callback, pdf_cls=Pdf, layout_recognizer=layout_recognizer, mineru_llm_name=parser_model_name, **kwargs ) if not raw_sections and not tables: return [] if name in ["tcadp", "docling", "mineru"]: parser_config["chunk_token_num"] = 0 for txt, poss in raw_sections: sections.append(txt + poss) callback(0.8, "Finish parsing.") elif re.search(r"\.(txt|md|markdown|mdx)$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") txt = get_text(filename, binary) sections = txt.split("\n") sections = [s for s in sections if s] callback(0.8, "Finish parsing.") elif re.search(r"\.(htm|html)$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") sections = HtmlParser()(filename, binary) sections = [s for s in sections if s] callback(0.8, "Finish parsing.") elif re.search(r"\.doc$", filename, re.IGNORECASE): callback(0.1, "Start to parse.") try: from tika import parser as tika_parser except Exception as e: callback(0.8, f"tika not available: {e}. Unsupported .doc parsing.") logging.warning(f"tika not available: {e}. Unsupported .doc parsing for {filename}.") return [] binary = BytesIO(binary) doc_parsed = tika_parser.from_buffer(binary) if doc_parsed.get('content', None) is not None: sections = doc_parsed['content'].split('\n') sections = [s for s in sections if s] callback(0.8, "Finish parsing.") else: callback(0.8, f"tika.parser got empty content from {filename}.") logging.warning(f"tika.parser got empty content from {filename}.") return [] else: raise NotImplementedError( "file type not supported yet(doc, docx, pdf, txt supported)") # Remove 'Contents' part remove_contents_table(sections, eng) make_colon_as_title(sections) bull = bullets_category(sections) res = tree_merge(bull, sections, 2) if not res: callback(0.99, "No chunk parsed out.") return tokenize_chunks(res, doc, eng, pdf_parser) # chunks = hierarchical_merge(bull, sections, 5) # return tokenize_chunks(["\n".join(ck)for ck in chunks], doc, eng, pdf_parser) if __name__ == "__main__": import sys def dummy(prog=None, msg=""): pass chunk(sys.argv[1], callback=dummy)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/file.py
rag/flow/file.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from api.db.services.document_service import DocumentService from rag.flow.base import ProcessBase, ProcessParamBase class FileParam(ProcessParamBase): def __init__(self): super().__init__() def check(self): pass def get_input_form(self) -> dict[str, dict]: return {} class File(ProcessBase): component_name = "File" async def _invoke(self, **kwargs): if self._canvas._doc_id: e, doc = DocumentService.get_by_id(self._canvas._doc_id) if not e: self.set_output("_ERROR", f"Document({self._canvas._doc_id}) not found!") return #b, n = File2DocumentService.get_storage_address(doc_id=self._canvas._doc_id) #self.set_output("blob", STORAGE_IMPL.get(b, n)) self.set_output("name", doc.name) else: file = kwargs.get("file") self.set_output("name", file["name"]) self.set_output("file", file) #self.set_output("blob", FileService.get_blob(file["created_by"], file["id"])) self.callback(1, "File fetched.")
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/pipeline.py
rag/flow/pipeline.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import asyncio import datetime import json import logging import random from timeit import default_timer as timer from agent.canvas import Graph from api.db.services.document_service import DocumentService from api.db.services.task_service import has_canceled, TaskService, CANVAS_DEBUG_DOC_ID from rag.utils.redis_conn import REDIS_CONN class Pipeline(Graph): def __init__(self, dsl: str|dict, tenant_id=None, doc_id=None, task_id=None, flow_id=None): if isinstance(dsl, dict): dsl = json.dumps(dsl, ensure_ascii=False) super().__init__(dsl, tenant_id, task_id) if doc_id == CANVAS_DEBUG_DOC_ID: doc_id = None self._doc_id = doc_id self._flow_id = flow_id self._kb_id = None if self._doc_id: self._kb_id = DocumentService.get_knowledgebase_id(doc_id) if not self._kb_id: self._doc_id = None def callback(self, component_name: str, progress: float | int | None = None, message: str = "") -> None: from common.exceptions import TaskCanceledException log_key = f"{self._flow_id}-{self.task_id}-logs" timestamp = timer() if has_canceled(self.task_id): progress = -1 message += "[CANCEL]" try: bin = REDIS_CONN.get(log_key) obj = json.loads(bin.encode("utf-8")) if obj: if obj[-1]["component_id"] == component_name: obj[-1]["trace"].append( { "progress": progress, "message": message, "datetime": datetime.datetime.now().strftime("%H:%M:%S"), "timestamp": timestamp, "elapsed_time": timestamp - obj[-1]["trace"][-1]["timestamp"], } ) else: obj.append( { "component_id": component_name, "trace": [{"progress": progress, "message": message, "datetime": datetime.datetime.now().strftime("%H:%M:%S"), "timestamp": timestamp, "elapsed_time": 0}], } ) else: obj = [ { "component_id": component_name, "trace": [{"progress": progress, "message": message, "datetime": datetime.datetime.now().strftime("%H:%M:%S"), "timestamp": timestamp, "elapsed_time": 0}], } ] if component_name != "END" and self._doc_id and self.task_id: percentage = 1.0 / len(self.components.items()) finished = 0.0 for o in obj: for t in o["trace"]: if t["progress"] < 0: finished = -1 break if finished < 0: break finished += o["trace"][-1]["progress"] * percentage msg = "" if len(obj[-1]["trace"]) == 1: msg += f"\n-------------------------------------\n[{self.get_component_name(o['component_id'])}]:\n" t = obj[-1]["trace"][-1] msg += "%s: %s\n" % (t["datetime"], t["message"]) TaskService.update_progress(self.task_id, {"progress": finished, "progress_msg": msg}) elif component_name == "END" and not self._doc_id: obj[-1]["trace"][-1]["dsl"] = json.loads(str(self)) REDIS_CONN.set_obj(log_key, obj, 60 * 30) except Exception as e: logging.exception(e) if has_canceled(self.task_id): raise TaskCanceledException(message) def fetch_logs(self): log_key = f"{self._flow_id}-{self.task_id}-logs" try: bin = REDIS_CONN.get(log_key) if bin: return json.loads(bin.encode("utf-8")) except Exception as e: logging.exception(e) return [] async def run(self, **kwargs): log_key = f"{self._flow_id}-{self.task_id}-logs" try: REDIS_CONN.set_obj(log_key, [], 60 * 10) except Exception as e: logging.exception(e) self.error = "" if not self.path: self.path.append("File") cpn_obj = self.get_component_obj(self.path[0]) await cpn_obj.invoke(**kwargs) if cpn_obj.error(): self.error = "[ERROR]" + cpn_obj.error() self.callback(cpn_obj.component_name, -1, self.error) if self._doc_id: TaskService.update_progress(self.task_id, { "progress": random.randint(0, 5) / 100.0, "progress_msg": "Start the pipeline...", "begin_at": datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S")}) idx = len(self.path) - 1 cpn_obj = self.get_component_obj(self.path[idx]) idx += 1 self.path.extend(cpn_obj.get_downstream()) while idx < len(self.path) and not self.error: last_cpn = self.get_component_obj(self.path[idx - 1]) cpn_obj = self.get_component_obj(self.path[idx]) async def invoke(): nonlocal last_cpn, cpn_obj await cpn_obj.invoke(**last_cpn.output()) #if inspect.iscoroutinefunction(cpn_obj.invoke): # await cpn_obj.invoke(**last_cpn.output()) #else: # cpn_obj.invoke(**last_cpn.output()) tasks = [] tasks.append(asyncio.create_task(invoke())) await asyncio.gather(*tasks) if cpn_obj.error(): self.error = "[ERROR]" + cpn_obj.error() self.callback(cpn_obj._id, -1, self.error) break idx += 1 self.path.extend(cpn_obj.get_downstream()) self.callback("END", 1 if not self.error else -1, json.dumps(self.get_component_obj(self.path[-1]).output(), ensure_ascii=False)) if not self.error: return self.get_component_obj(self.path[-1]).output() TaskService.update_progress(self.task_id, { "progress": -1, "progress_msg": f"[ERROR]: {self.error}"}) return {}
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/__init__.py
rag/flow/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import importlib import inspect import pkgutil from pathlib import Path from types import ModuleType from typing import Dict, Type __all_classes: Dict[str, Type] = {} _pkg_dir = Path(__file__).resolve().parent _pkg_name = __name__ def _should_skip_module(mod_name: str) -> bool: leaf = mod_name.rsplit(".", 1)[-1] return leaf in {"__init__"} or leaf.startswith("__") or leaf.startswith("_") or leaf.startswith("base") def _import_submodules() -> None: for modinfo in pkgutil.walk_packages([str(_pkg_dir)], prefix=_pkg_name + "."): # noqa: F821 mod_name = modinfo.name if _should_skip_module(mod_name): # noqa: F821 continue try: module = importlib.import_module(mod_name) _extract_classes_from_module(module) # noqa: F821 except ImportError as e: print(f"Warning: Failed to import module {mod_name}: {e}") def _extract_classes_from_module(module: ModuleType) -> None: for name, obj in inspect.getmembers(module): if inspect.isclass(obj) and obj.__module__ == module.__name__ and not name.startswith("_"): __all_classes[name] = obj globals()[name] = obj _import_submodules() __all__ = list(__all_classes.keys()) + ["__all_classes"] del _pkg_dir, _pkg_name, _import_submodules, _extract_classes_from_module
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/base.py
rag/flow/base.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import asyncio import logging import os import time from functools import partial from typing import Any from agent.component.base import ComponentBase, ComponentParamBase from common.connection_utils import timeout class ProcessParamBase(ComponentParamBase): def __init__(self): super().__init__() self.timeout = 100000000 self.persist_logs = True class ProcessBase(ComponentBase): def __init__(self, pipeline, id, param: ProcessParamBase): super().__init__(pipeline, id, param) if hasattr(self._canvas, "callback"): self.callback = partial(self._canvas.callback, id) else: self.callback = partial(lambda *args, **kwargs: None, id) async def invoke(self, **kwargs) -> dict[str, Any]: self.set_output("_created_time", time.perf_counter()) for k, v in kwargs.items(): self.set_output(k, v) try: await asyncio.wait_for( self._invoke(**kwargs), timeout=self._param.timeout ) self.callback(1, "Done") except Exception as e: if self.get_exception_default_value(): self.set_exception_default_value() else: self.set_output("_ERROR", str(e)) logging.exception(e) self.callback(-1, str(e)) self.set_output("_elapsed_time", time.perf_counter() - self.output("_created_time")) return self.output() @timeout(int(os.environ.get("COMPONENT_EXEC_TIMEOUT", 10 * 60))) async def _invoke(self, **kwargs): raise NotImplementedError()
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/extractor/extractor.py
rag/flow/extractor/extractor.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import logging import random from copy import deepcopy import xxhash from agent.component.llm import LLMParam, LLM from rag.flow.base import ProcessBase, ProcessParamBase from rag.prompts.generator import run_toc_from_text class ExtractorParam(ProcessParamBase, LLMParam): def __init__(self): super().__init__() self.field_name = "" def check(self): super().check() self.check_empty(self.field_name, "Result Destination") class Extractor(ProcessBase, LLM): component_name = "Extractor" async def _build_TOC(self, docs): self.callback(0.2,message="Start to generate table of content ...") docs = sorted(docs, key=lambda d:( d.get("page_num_int", 0)[0] if isinstance(d.get("page_num_int", 0), list) else d.get("page_num_int", 0), d.get("top_int", 0)[0] if isinstance(d.get("top_int", 0), list) else d.get("top_int", 0) )) toc = await run_toc_from_text([d["text"] for d in docs], self.chat_mdl) logging.info("------------ T O C -------------\n"+json.dumps(toc, ensure_ascii=False, indent=' ')) ii = 0 while ii < len(toc): try: idx = int(toc[ii]["chunk_id"]) del toc[ii]["chunk_id"] toc[ii]["ids"] = [docs[idx]["id"]] if ii == len(toc) -1: break for jj in range(idx+1, int(toc[ii+1]["chunk_id"])+1): toc[ii]["ids"].append(docs[jj]["id"]) except Exception as e: logging.exception(e) ii += 1 if toc: d = deepcopy(docs[-1]) d["doc_id"] = self._canvas._doc_id d["content_with_weight"] = json.dumps(toc, ensure_ascii=False) d["toc_kwd"] = "toc" d["available_int"] = 0 d["page_num_int"] = [100000000] d["id"] = xxhash.xxh64((d["content_with_weight"] + str(d["doc_id"])).encode("utf-8", "surrogatepass")).hexdigest() return d return None async def _invoke(self, **kwargs): self.set_output("output_format", "chunks") self.callback(random.randint(1, 5) / 100.0, "Start to generate.") inputs = self.get_input_elements() chunks = [] chunks_key = "" args = {} for k, v in inputs.items(): args[k] = v["value"] if isinstance(args[k], list): chunks = deepcopy(args[k]) chunks_key = k if chunks: if self._param.field_name == "toc": for ck in chunks: ck["doc_id"] = self._canvas._doc_id ck["id"] = xxhash.xxh64((ck["text"] + str(ck["doc_id"])).encode("utf-8")).hexdigest() toc =await self._build_TOC(chunks) chunks.append(toc) self.set_output("chunks", chunks) return prog = 0 for i, ck in enumerate(chunks): args[chunks_key] = ck["text"] msg, sys_prompt = self._sys_prompt_and_msg([], args) msg.insert(0, {"role": "system", "content": sys_prompt}) ck[self._param.field_name] = await self._generate_async(msg) prog += 1./len(chunks) if i % (len(chunks)//100+1) == 1: self.callback(prog, f"{i+1} / {len(chunks)}") self.set_output("chunks", chunks) else: msg, sys_prompt = self._sys_prompt_and_msg([], args) msg.insert(0, {"role": "system", "content": sys_prompt}) self.set_output("chunks", [{self._param.field_name: await self._generate_async(msg)}])
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/extractor/schema.py
rag/flow/extractor/schema.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import Any, Literal from pydantic import BaseModel, ConfigDict, Field class ExtractorFromUpstream(BaseModel): created_time: float | None = Field(default=None, alias="_created_time") elapsed_time: float | None = Field(default=None, alias="_elapsed_time") name: str file: dict | None = Field(default=None) chunks: list[dict[str, Any]] | None = Field(default=None) output_format: Literal["json", "markdown", "text", "html", "chunks"] | None = Field(default=None) json_result: list[dict[str, Any]] | None = Field(default=None, alias="json") markdown_result: str | None = Field(default=None, alias="markdown") text_result: str | None = Field(default=None, alias="text") html_result: str | None = Field(default=None, alias="html") model_config = ConfigDict(populate_by_name=True, extra="forbid") # def to_dict(self, *, exclude_none: bool = True) -> dict: # return self.model_dump(by_alias=True, exclude_none=exclude_none)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/extractor/__init__.py
rag/flow/extractor/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/hierarchical_merger/schema.py
rag/flow/hierarchical_merger/schema.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import Any, Literal from pydantic import BaseModel, ConfigDict, Field class HierarchicalMergerFromUpstream(BaseModel): created_time: float | None = Field(default=None, alias="_created_time") elapsed_time: float | None = Field(default=None, alias="_elapsed_time") name: str file: dict | None = Field(default=None) chunks: list[dict[str, Any]] | None = Field(default=None) output_format: Literal["json", "markdown", "text", "html", "chunks"] | None = Field(default=None) json_result: list[dict[str, Any]] | None = Field(default=None, alias="json") markdown_result: str | None = Field(default=None, alias="markdown") text_result: str | None = Field(default=None, alias="text") html_result: str | None = Field(default=None, alias="html") model_config = ConfigDict(populate_by_name=True, extra="forbid") # def to_dict(self, *, exclude_none: bool = True) -> dict: # return self.model_dump(by_alias=True, exclude_none=exclude_none)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/hierarchical_merger/__init__.py
rag/flow/hierarchical_merger/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/hierarchical_merger/hierarchical_merger.py
rag/flow/hierarchical_merger/hierarchical_merger.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import asyncio import logging import random import re from copy import deepcopy from functools import partial from common.misc_utils import get_uuid from rag.utils.base64_image import id2image, image2id from deepdoc.parser.pdf_parser import RAGFlowPdfParser from rag.flow.base import ProcessBase, ProcessParamBase from rag.flow.hierarchical_merger.schema import HierarchicalMergerFromUpstream from rag.nlp import concat_img from common import settings class HierarchicalMergerParam(ProcessParamBase): def __init__(self): super().__init__() self.levels = [] self.hierarchy = None def check(self): self.check_empty(self.levels, "Hierarchical setups.") self.check_empty(self.hierarchy, "Hierarchy number.") def get_input_form(self) -> dict[str, dict]: return {} class HierarchicalMerger(ProcessBase): component_name = "HierarchicalMerger" async def _invoke(self, **kwargs): try: from_upstream = HierarchicalMergerFromUpstream.model_validate(kwargs) except Exception as e: self.set_output("_ERROR", f"Input error: {str(e)}") return self.set_output("output_format", "chunks") self.callback(random.randint(1, 5) / 100.0, "Start to merge hierarchically.") if from_upstream.output_format in ["markdown", "text", "html"]: if from_upstream.output_format == "markdown": payload = from_upstream.markdown_result elif from_upstream.output_format == "text": payload = from_upstream.text_result else: # == "html" payload = from_upstream.html_result if not payload: payload = "" lines = [ln for ln in payload.split("\n") if ln] else: arr = from_upstream.chunks if from_upstream.output_format == "chunks" else from_upstream.json_result lines = [o.get("text", "") for o in arr] sections, section_images = [], [] for o in arr or []: sections.append((o.get("text", ""), o.get("position_tag", ""))) section_images.append(o.get("img_id")) matches = [] for txt in lines: good = False for lvl, regs in enumerate(self._param.levels): for reg in regs: if re.search(reg, txt): matches.append(lvl) good = True break if good: break if not good: matches.append(len(self._param.levels)) assert len(matches) == len(lines), f"{len(matches)} vs. {len(lines)}" root = { "level": -1, "index": -1, "texts": [], "children": [] } for i, m in enumerate(matches): if m == 0: root["children"].append({ "level": m, "index": i, "texts": [], "children": [] }) elif m == len(self._param.levels): def dfs(b): if not b["children"]: b["texts"].append(i) else: dfs(b["children"][-1]) dfs(root) else: def dfs(b): nonlocal m, i if not b["children"] or m == b["level"] + 1: b["children"].append({ "level": m, "index": i, "texts": [], "children": [] }) return dfs(b["children"][-1]) dfs(root) all_pathes = [] def dfs(n, path, depth): nonlocal all_pathes if not n["children"] and path: all_pathes.append(path) for nn in n["children"]: if depth < self._param.hierarchy: _path = deepcopy(path) else: _path = path _path.extend([nn["index"], *nn["texts"]]) dfs(nn, _path, depth+1) if depth == self._param.hierarchy: all_pathes.append(_path) for i in range(len(lines)): print(i, lines[i]) dfs(root, [], 0) if root["texts"]: all_pathes.insert(0, root["texts"]) if from_upstream.output_format in ["markdown", "text", "html"]: cks = [] for path in all_pathes: txt = "" for i in path: txt += lines[i] + "\n" cks.append(txt) self.set_output("chunks", [{"text": c} for c in cks if c]) else: cks = [] images = [] for path in all_pathes: txt = "" img = None for i in path: txt += lines[i] + "\n" concat_img(img, id2image(section_images[i], partial(settings.STORAGE_IMPL.get, tenant_id=self._canvas._tenant_id))) cks.append(txt) images.append(img) cks = [ { "text": RAGFlowPdfParser.remove_tag(c), "image": img, "positions": RAGFlowPdfParser.extract_positions(c), } for c, img in zip(cks, images) ] tasks = [] for d in cks: tasks.append(asyncio.create_task(image2id(d, partial(settings.STORAGE_IMPL.put, tenant_id=self._canvas._tenant_id), get_uuid()))) try: await asyncio.gather(*tasks, return_exceptions=False) except Exception as e: logging.error(f"Error in image2id: {e}") for t in tasks: t.cancel() await asyncio.gather(*tasks, return_exceptions=True) raise self.set_output("chunks", cks) self.callback(1, "Done.")
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/tests/client.py
rag/flow/tests/client.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import argparse import asyncio import json import os import time from concurrent.futures import ThreadPoolExecutor from common import settings from rag.flow.pipeline import Pipeline def print_logs(pipeline: Pipeline): last_logs = "[]" while True: time.sleep(5) logs = pipeline.fetch_logs() logs_str = json.dumps(logs, ensure_ascii=False) if logs_str != last_logs: print(logs_str) last_logs = logs_str if __name__ == "__main__": parser = argparse.ArgumentParser() dsl_default_path = os.path.join( os.path.dirname(os.path.realpath(__file__)), "dsl_examples", "general_pdf_all.json", ) parser.add_argument("-s", "--dsl", default=dsl_default_path, help="input dsl", action="store", required=False) parser.add_argument("-d", "--doc_id", default=False, help="Document ID", action="store", required=True) parser.add_argument("-t", "--tenant_id", default=False, help="Tenant ID", action="store", required=True) args = parser.parse_args() settings.init_settings() pipeline = Pipeline(open(args.dsl, "r").read(), tenant_id=args.tenant_id, doc_id=args.doc_id, task_id="xxxx", flow_id="xxx") pipeline.reset() exe = ThreadPoolExecutor(max_workers=5) thr = exe.submit(print_logs, pipeline) # queue_dataflow(dsl=open(args.dsl, "r").read(), tenant_id=args.tenant_id, doc_id=args.doc_id, task_id="xxxx", flow_id="xxx", priority=0) asyncio.run(pipeline.run()) thr.result()
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/tokenizer/schema.py
rag/flow/tokenizer/schema.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import Any, Literal from pydantic import BaseModel, ConfigDict, Field, model_validator class TokenizerFromUpstream(BaseModel): created_time: float | None = Field(default=None, alias="_created_time") elapsed_time: float | None = Field(default=None, alias="_elapsed_time") name: str = "" file: dict | None = Field(default=None) output_format: Literal["json", "markdown", "text", "html", "chunks"] | None = Field(default=None) chunks: list[dict[str, Any]] | None = Field(default=None) json_result: list[dict[str, Any]] | None = Field(default=None, alias="json") markdown_result: str | None = Field(default=None, alias="markdown") text_result: str | None = Field(default=None, alias="text") html_result: str | None = Field(default=None, alias="html") model_config = ConfigDict(populate_by_name=True, extra="forbid") @model_validator(mode="after") def _check_payloads(self) -> "TokenizerFromUpstream": if self.chunks: return self if self.output_format in {"markdown", "text", "html"}: if self.output_format == "markdown" and not self.markdown_result: raise ValueError("output_format=markdown requires a markdown payload (field: 'markdown' or 'markdown_result').") if self.output_format == "text" and not self.text_result: raise ValueError("output_format=text requires a text payload (field: 'text' or 'text_result').") if self.output_format == "html" and not self.html_result: raise ValueError("output_format=text requires a html payload (field: 'html' or 'html_result').") else: if not self.json_result and not self.chunks: raise ValueError("When no chunks are provided and output_format is not markdown/text, a JSON list payload is required (field: 'json' or 'json_result').") return self
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/tokenizer/__init__.py
rag/flow/tokenizer/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/tokenizer/tokenizer.py
rag/flow/tokenizer/tokenizer.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import asyncio import logging import random import re import numpy as np from common.constants import LLMType from api.db.services.knowledgebase_service import KnowledgebaseService from api.db.services.llm_service import LLMBundle from api.db.services.user_service import TenantService from common.connection_utils import timeout from rag.flow.base import ProcessBase, ProcessParamBase from rag.flow.tokenizer.schema import TokenizerFromUpstream from rag.nlp import rag_tokenizer from common import settings from rag.svr.task_executor import embed_limiter from common.token_utils import truncate class TokenizerParam(ProcessParamBase): def __init__(self): super().__init__() self.search_method = ["full_text", "embedding"] self.filename_embd_weight = 0.1 self.fields = ["text"] def check(self): for v in self.search_method: self.check_valid_value(v.lower(), "Chunk method abnormal.", ["full_text", "embedding"]) def get_input_form(self) -> dict[str, dict]: return {} class Tokenizer(ProcessBase): component_name = "Tokenizer" async def _embedding(self, name, chunks): parts = sum(["full_text" in self._param.search_method, "embedding" in self._param.search_method]) token_count = 0 if self._canvas._kb_id: e, kb = KnowledgebaseService.get_by_id(self._canvas._kb_id) embedding_id = kb.embd_id else: e, ten = TenantService.get_by_id(self._canvas._tenant_id) embedding_id = ten.embd_id embedding_model = LLMBundle(self._canvas._tenant_id, LLMType.EMBEDDING, llm_name=embedding_id) texts = [] for c in chunks: txt = "" if isinstance(self._param.fields, str): self._param.fields=[self._param.fields] for f in self._param.fields: f = c.get(f) if isinstance(f, str): txt += f elif isinstance(f, list): txt += "\n".join(f) texts.append(re.sub(r"</?(table|td|caption|tr|th)( [^<>]{0,12})?>", " ", txt)) vts, c = embedding_model.encode([name]) token_count += c tts = np.concatenate([vts[0] for _ in range(len(texts))], axis=0) @timeout(60) def batch_encode(txts): nonlocal embedding_model return embedding_model.encode([truncate(c, embedding_model.max_length - 10) for c in txts]) cnts_ = np.array([]) for i in range(0, len(texts), settings.EMBEDDING_BATCH_SIZE): async with embed_limiter: vts, c = await asyncio.to_thread(batch_encode,texts[i : i + settings.EMBEDDING_BATCH_SIZE],) if len(cnts_) == 0: cnts_ = vts else: cnts_ = np.concatenate((cnts_, vts), axis=0) token_count += c if i % 33 == 32: self.callback(i * 1.0 / len(texts) / parts / settings.EMBEDDING_BATCH_SIZE + 0.5 * (parts - 1)) cnts = cnts_ title_w = float(self._param.filename_embd_weight) vects = (title_w * tts + (1 - title_w) * cnts) if len(tts) == len(cnts) else cnts assert len(vects) == len(chunks) for i, ck in enumerate(chunks): v = vects[i].tolist() ck["q_%d_vec" % len(v)] = v return chunks, token_count async def _invoke(self, **kwargs): try: chunks = kwargs.get("chunks") kwargs["chunks"] = [c for c in chunks if c is not None] from_upstream = TokenizerFromUpstream.model_validate(kwargs) except Exception as e: self.set_output("_ERROR", f"Input error: {str(e)}") return self.set_output("output_format", "chunks") parts = sum(["full_text" in self._param.search_method, "embedding" in self._param.search_method]) if "full_text" in self._param.search_method: self.callback(random.randint(1, 5) / 100.0, "Start to tokenize.") if from_upstream.chunks: chunks = from_upstream.chunks for i, ck in enumerate(chunks): ck["title_tks"] = rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", from_upstream.name)) ck["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(ck["title_tks"]) if ck.get("questions"): ck["question_kwd"] = ck["questions"].split("\n") ck["question_tks"] = rag_tokenizer.tokenize(str(ck["questions"])) if ck.get("keywords"): ck["important_kwd"] = ck["keywords"].split(",") ck["important_tks"] = rag_tokenizer.tokenize(str(ck["keywords"])) if ck.get("summary"): ck["content_ltks"] = rag_tokenizer.tokenize(str(ck["summary"])) ck["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(ck["content_ltks"]) elif ck.get("text"): ck["content_ltks"] = rag_tokenizer.tokenize(ck["text"]) ck["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(ck["content_ltks"]) if i % 100 == 99: self.callback(i * 1.0 / len(chunks) / parts) elif from_upstream.output_format in ["markdown", "text", "html"]: if from_upstream.output_format == "markdown": payload = from_upstream.markdown_result elif from_upstream.output_format == "text": payload = from_upstream.text_result else: payload = from_upstream.html_result if not payload: return "" ck = {"text": payload} if "full_text" in self._param.search_method: ck["title_tks"] = rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", from_upstream.name)) ck["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(ck["title_tks"]) ck["content_ltks"] = rag_tokenizer.tokenize(payload) ck["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(ck["content_ltks"]) chunks = [ck] else: chunks = from_upstream.json_result for i, ck in enumerate(chunks): ck["title_tks"] = rag_tokenizer.tokenize(re.sub(r"\.[a-zA-Z]+$", "", from_upstream.name)) ck["title_sm_tks"] = rag_tokenizer.fine_grained_tokenize(ck["title_tks"]) if not ck.get("text"): continue ck["content_ltks"] = rag_tokenizer.tokenize(ck["text"]) ck["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(ck["content_ltks"]) if i % 100 == 99: self.callback(i * 1.0 / len(chunks) / parts) self.callback(1.0 / parts, "Finish tokenizing.") if "embedding" in self._param.search_method: self.callback(random.randint(1, 5) / 100.0 + 0.5 * (parts - 1), "Start embedding inference.") if from_upstream.name.strip() == "": logging.warning("Tokenizer: empty name provided from upstream, embedding may be not accurate.") chunks, token_count = await self._embedding(from_upstream.name, chunks) self.set_output("embedding_token_consumption", token_count) self.callback(1.0, "Finish embedding.") self.set_output("chunks", chunks)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/splitter/splitter.py
rag/flow/splitter/splitter.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import asyncio import logging import random import re from copy import deepcopy from functools import partial from common.misc_utils import get_uuid from rag.utils.base64_image import id2image, image2id from deepdoc.parser.pdf_parser import RAGFlowPdfParser from rag.flow.base import ProcessBase, ProcessParamBase from rag.flow.splitter.schema import SplitterFromUpstream from rag.nlp import attach_media_context, naive_merge, naive_merge_with_images from common import settings class SplitterParam(ProcessParamBase): def __init__(self): super().__init__() self.chunk_token_size = 512 self.delimiters = ["\n"] self.overlapped_percent = 0 self.children_delimiters = [] self.table_context_size = 0 self.image_context_size = 0 def check(self): self.check_empty(self.delimiters, "Delimiters.") self.check_positive_integer(self.chunk_token_size, "Chunk token size.") self.check_decimal_float(self.overlapped_percent, "Overlapped percentage: [0, 1)") self.check_nonnegative_number(self.table_context_size, "Table context size.") self.check_nonnegative_number(self.image_context_size, "Image context size.") def get_input_form(self) -> dict[str, dict]: return {} class Splitter(ProcessBase): component_name = "Splitter" async def _invoke(self, **kwargs): try: from_upstream = SplitterFromUpstream.model_validate(kwargs) except Exception as e: self.set_output("_ERROR", f"Input error: {str(e)}") return deli = "" for d in self._param.delimiters: if len(d) > 1: deli += f"`{d}`" else: deli += d custom_pattern = "|".join(re.escape(t) for t in sorted(set(self._param.children_delimiters), key=len, reverse=True)) self.set_output("output_format", "chunks") self.callback(random.randint(1, 5) / 100.0, "Start to split into chunks.") if from_upstream.output_format in ["markdown", "text", "html"]: if from_upstream.output_format == "markdown": payload = from_upstream.markdown_result elif from_upstream.output_format == "text": payload = from_upstream.text_result else: # == "html" payload = from_upstream.html_result if not payload: payload = "" cks = naive_merge( payload, self._param.chunk_token_size, deli, self._param.overlapped_percent, ) if custom_pattern: docs = [] for c in cks: if not c.strip(): continue split_sec = re.split(r"(%s)" % custom_pattern, c, flags=re.DOTALL) if split_sec: for j in range(0, len(split_sec), 2): docs.append({ "text": split_sec[j], "mom": c }) else: docs.append({"text": c}) self.set_output("chunks", docs) else: self.set_output("chunks", [{"text": c.strip()} for c in cks if c.strip()]) self.callback(1, "Done.") return # json json_result = from_upstream.json_result or [] if self._param.table_context_size or self._param.image_context_size: for ck in json_result: if "image" not in ck and ck.get("img_id") and not (isinstance(ck.get("text"), str) and ck.get("text").strip()): ck["image"] = True attach_media_context(json_result, self._param.table_context_size, self._param.image_context_size) for ck in json_result: if ck.get("image") is True: del ck["image"] sections, section_images = [], [] for o in json_result: sections.append((o.get("text", ""), o.get("position_tag", ""))) section_images.append(id2image(o.get("img_id"), partial(settings.STORAGE_IMPL.get, tenant_id=self._canvas._tenant_id))) chunks, images = naive_merge_with_images( sections, section_images, self._param.chunk_token_size, deli, self._param.overlapped_percent, ) cks = [ { "text": RAGFlowPdfParser.remove_tag(c), "image": img, "positions": [[pos[0][-1], *pos[1:]] for pos in RAGFlowPdfParser.extract_positions(c)] } for c, img in zip(chunks, images) if c.strip() ] tasks = [] for d in cks: tasks.append(asyncio.create_task(image2id(d, partial(settings.STORAGE_IMPL.put, tenant_id=self._canvas._tenant_id), get_uuid()))) try: await asyncio.gather(*tasks, return_exceptions=False) except Exception as e: logging.error(f"error when splitting: {e}") for t in tasks: t.cancel() await asyncio.gather(*tasks, return_exceptions=True) raise if custom_pattern: docs = [] for c in cks: split_sec = re.split(r"(%s)" % custom_pattern, c["text"], flags=re.DOTALL) if split_sec: c["mom"] = c["text"] for j in range(0, len(split_sec), 2): cc = deepcopy(c) cc["text"] = split_sec[j] docs.append(cc) else: docs.append(c) self.set_output("chunks", docs) else: self.set_output("chunks", cks) self.callback(1, "Done.")
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/splitter/schema.py
rag/flow/splitter/schema.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import Any, Literal from pydantic import BaseModel, ConfigDict, Field class SplitterFromUpstream(BaseModel): created_time: float | None = Field(default=None, alias="_created_time") elapsed_time: float | None = Field(default=None, alias="_elapsed_time") name: str file: dict | None = Field(default=None) chunks: list[dict[str, Any]] | None = Field(default=None) output_format: Literal["json", "markdown", "text", "html", "chunks"] | None = Field(default=None) json_result: list[dict[str, Any]] | None = Field(default=None, alias="json") markdown_result: str | None = Field(default=None, alias="markdown") text_result: str | None = Field(default=None, alias="text") html_result: str | None = Field(default=None, alias="html") model_config = ConfigDict(populate_by_name=True, extra="forbid") # def to_dict(self, *, exclude_none: bool = True) -> dict: # return self.model_dump(by_alias=True, exclude_none=exclude_none)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/splitter/__init__.py
rag/flow/splitter/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/parser/parser.py
rag/flow/parser/parser.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import asyncio import io import json import os import random import re from functools import partial from litellm import logging import numpy as np from PIL import Image from api.db.services.file2document_service import File2DocumentService from api.db.services.file_service import FileService from api.db.services.llm_service import LLMBundle from common import settings from common.constants import LLMType from common.misc_utils import get_uuid from deepdoc.parser import ExcelParser from deepdoc.parser.pdf_parser import PlainParser, RAGFlowPdfParser, VisionParser from deepdoc.parser.tcadp_parser import TCADPParser from rag.app.naive import Docx from rag.flow.base import ProcessBase, ProcessParamBase from rag.flow.parser.schema import ParserFromUpstream from rag.llm.cv_model import Base as VLM from rag.utils.base64_image import image2id class ParserParam(ProcessParamBase): def __init__(self): super().__init__() self.allowed_output_format = { "pdf": [ "json", "markdown", ], "spreadsheet": [ "json", "markdown", "html", ], "word": [ "json", "markdown", ], "slides": [ "json", ], "image": [ "text", ], "email": [ "text", "json", ], "text&markdown": [ "text", "json", ], "audio": [ "json", ], "video": [], } self.setups = { "pdf": { "parse_method": "deepdoc", # deepdoc/plain_text/tcadp_parser/vlm "lang": "Chinese", "suffix": [ "pdf", ], "output_format": "json", }, "spreadsheet": { "parse_method": "deepdoc", # deepdoc/tcadp_parser "output_format": "html", "suffix": [ "xls", "xlsx", "csv", ], }, "word": { "suffix": [ "doc", "docx", ], "output_format": "json", }, "text&markdown": { "suffix": ["md", "markdown", "mdx", "txt"], "output_format": "json", }, "slides": { "parse_method": "deepdoc", # deepdoc/tcadp_parser "suffix": [ "pptx", "ppt", ], "output_format": "json", }, "image": { "parse_method": "ocr", "llm_id": "", "lang": "Chinese", "system_prompt": "", "suffix": ["jpg", "jpeg", "png", "gif"], "output_format": "text", }, "email": { "suffix": [ "eml", "msg", ], "fields": ["from", "to", "cc", "bcc", "date", "subject", "body", "attachments", "metadata"], "output_format": "json", }, "audio": { "suffix": [ "da", "wave", "wav", "mp3", "aac", "flac", "ogg", "aiff", "au", "midi", "wma", "realaudio", "vqf", "oggvorbis", "ape", ], "output_format": "text", }, "video": { "suffix": [ "mp4", "avi", "mkv", ], "output_format": "text", }, } def check(self): pdf_config = self.setups.get("pdf", {}) if pdf_config: pdf_parse_method = pdf_config.get("parse_method", "") self.check_empty(pdf_parse_method, "Parse method abnormal.") if pdf_parse_method.lower() not in ["deepdoc", "plain_text", "mineru", "tcadp parser"]: self.check_empty(pdf_config.get("lang", ""), "PDF VLM language") pdf_output_format = pdf_config.get("output_format", "") self.check_valid_value(pdf_output_format, "PDF output format abnormal.", self.allowed_output_format["pdf"]) spreadsheet_config = self.setups.get("spreadsheet", "") if spreadsheet_config: spreadsheet_output_format = spreadsheet_config.get("output_format", "") self.check_valid_value(spreadsheet_output_format, "Spreadsheet output format abnormal.", self.allowed_output_format["spreadsheet"]) doc_config = self.setups.get("word", "") if doc_config: doc_output_format = doc_config.get("output_format", "") self.check_valid_value(doc_output_format, "Word processer document output format abnormal.", self.allowed_output_format["word"]) slides_config = self.setups.get("slides", "") if slides_config: slides_output_format = slides_config.get("output_format", "") self.check_valid_value(slides_output_format, "Slides output format abnormal.", self.allowed_output_format["slides"]) image_config = self.setups.get("image", "") if image_config: image_parse_method = image_config.get("parse_method", "") if image_parse_method not in ["ocr"]: self.check_empty(image_config.get("lang", ""), "Image VLM language") text_config = self.setups.get("text&markdown", "") if text_config: text_output_format = text_config.get("output_format", "") self.check_valid_value(text_output_format, "Text output format abnormal.", self.allowed_output_format["text&markdown"]) audio_config = self.setups.get("audio", "") if audio_config: self.check_empty(audio_config.get("llm_id"), "Audio VLM") video_config = self.setups.get("video", "") if video_config: self.check_empty(video_config.get("llm_id"), "Video VLM") email_config = self.setups.get("email", "") if email_config: email_output_format = email_config.get("output_format", "") self.check_valid_value(email_output_format, "Email output format abnormal.", self.allowed_output_format["email"]) def get_input_form(self) -> dict[str, dict]: return {} class Parser(ProcessBase): component_name = "Parser" def _pdf(self, name, blob): self.callback(random.randint(1, 5) / 100.0, "Start to work on a PDF.") conf = self._param.setups["pdf"] self.set_output("output_format", conf["output_format"]) raw_parse_method = conf.get("parse_method", "") parser_model_name = None parse_method = raw_parse_method parse_method = parse_method or "" if isinstance(raw_parse_method, str): lowered = raw_parse_method.lower() if lowered.endswith("@mineru"): parser_model_name = raw_parse_method.rsplit("@", 1)[0] parse_method = "MinerU" if parse_method.lower() == "deepdoc": bboxes = RAGFlowPdfParser().parse_into_bboxes(blob, callback=self.callback) elif parse_method.lower() == "plain_text": lines, _ = PlainParser()(blob) bboxes = [{"text": t} for t, _ in lines] elif parse_method.lower() == "mineru": def resolve_mineru_llm_name(): configured = parser_model_name or conf.get("mineru_llm_name") if configured: return configured tenant_id = self._canvas._tenant_id if not tenant_id: return None from api.db.services.tenant_llm_service import TenantLLMService env_name = TenantLLMService.ensure_mineru_from_env(tenant_id) candidates = TenantLLMService.query(tenant_id=tenant_id, llm_factory="MinerU", model_type=LLMType.OCR.value) if candidates: return candidates[0].llm_name return env_name parser_model_name = resolve_mineru_llm_name() if not parser_model_name: raise RuntimeError("MinerU model not configured. Please add MinerU in Model Providers or set MINERU_* env.") tenant_id = self._canvas._tenant_id ocr_model = LLMBundle(tenant_id, LLMType.OCR, llm_name=parser_model_name, lang=conf.get("lang", "Chinese")) pdf_parser = ocr_model.mdl lines, _ = pdf_parser.parse_pdf( filepath=name, binary=blob, callback=self.callback, parse_method=conf.get("mineru_parse_method", "raw"), lang=conf.get("lang", "Chinese"), ) bboxes = [] for t, poss in lines: box = { "image": pdf_parser.crop(poss, 1), "positions": [[pos[0][-1], *pos[1:]] for pos in pdf_parser.extract_positions(poss)], "text": t, } bboxes.append(box) elif parse_method.lower() == "tcadp parser": # ADP is a document parsing tool using Tencent Cloud API table_result_type = conf.get("table_result_type", "1") markdown_image_response_type = conf.get("markdown_image_response_type", "1") tcadp_parser = TCADPParser( table_result_type=table_result_type, markdown_image_response_type=markdown_image_response_type, ) sections, _ = tcadp_parser.parse_pdf( filepath=name, binary=blob, callback=self.callback, file_type="PDF", file_start_page=1, file_end_page=1000, ) bboxes = [] for section, position_tag in sections: if position_tag: # Extract position information from TCADP's position tag # Format: @@{page_number}\t{x0}\t{x1}\t{top}\t{bottom}## import re match = re.match(r"@@([0-9-]+)\t([0-9.]+)\t([0-9.]+)\t([0-9.]+)\t([0-9.]+)##", position_tag) if match: pn, x0, x1, top, bott = match.groups() bboxes.append( { "page_number": int(pn.split("-")[0]), # Take the first page number "x0": float(x0), "x1": float(x1), "top": float(top), "bottom": float(bott), "text": section, } ) else: # If no position info, add as text without position bboxes.append({"text": section}) else: bboxes.append({"text": section}) else: vision_model = LLMBundle(self._canvas._tenant_id, LLMType.IMAGE2TEXT, llm_name=conf.get("parse_method"), lang=self._param.setups["pdf"].get("lang")) lines, _ = VisionParser(vision_model=vision_model)(blob, callback=self.callback) bboxes = [] for t, poss in lines: for pn, x0, x1, top, bott in RAGFlowPdfParser.extract_positions(poss): bboxes.append( { "page_number": int(pn[0]), "x0": float(x0), "x1": float(x1), "top": float(top), "bottom": float(bott), "text": t, } ) for b in bboxes: text_val = b.get("text", "") has_text = isinstance(text_val, str) and text_val.strip() layout = b.get("layout_type") if layout == "figure" or (b.get("image") and not has_text): b["doc_type_kwd"] = "image" elif layout == "table": b["doc_type_kwd"] = "table" if conf.get("output_format") == "json": self.set_output("json", bboxes) if conf.get("output_format") == "markdown": mkdn = "" for b in bboxes: if b.get("layout_type", "") == "title": mkdn += "\n## " if b.get("layout_type", "") == "figure": mkdn += "\n![Image]({})".format(VLM.image2base64(b["image"])) continue mkdn += b.get("text", "") + "\n" self.set_output("markdown", mkdn) def _spreadsheet(self, name, blob): self.callback(random.randint(1, 5) / 100.0, "Start to work on a Spreadsheet.") conf = self._param.setups["spreadsheet"] self.set_output("output_format", conf["output_format"]) parse_method = conf.get("parse_method", "deepdoc") # Handle TCADP parser if parse_method.lower() == "tcadp parser": table_result_type = conf.get("table_result_type", "1") markdown_image_response_type = conf.get("markdown_image_response_type", "1") tcadp_parser = TCADPParser( table_result_type=table_result_type, markdown_image_response_type=markdown_image_response_type, ) if not tcadp_parser.check_installation(): raise RuntimeError("TCADP parser not available. Please check Tencent Cloud API configuration.") # Determine file type based on extension if re.search(r"\.xlsx?$", name, re.IGNORECASE): file_type = "XLSX" else: file_type = "CSV" self.callback(0.2, f"Using TCADP parser for {file_type} file.") sections, tables = tcadp_parser.parse_pdf( filepath=name, binary=blob, callback=self.callback, file_type=file_type, file_start_page=1, file_end_page=1000, ) # Process TCADP parser output based on configured output_format output_format = conf.get("output_format", "html") if output_format == "html": # For HTML output, combine sections and tables into HTML html_content = "" for section, position_tag in sections: if section: html_content += section + "\n" for table in tables: if table: html_content += table + "\n" self.set_output("html", html_content) elif output_format == "json": # For JSON output, create a list of text items result = [] # Add sections as text for section, position_tag in sections: if section: result.append({"text": section}) # Add tables as text for table in tables: if table: result.append({"text": table, "doc_type_kwd": "table"}) self.set_output("json", result) elif output_format == "markdown": # For markdown output, combine into markdown md_content = "" for section, position_tag in sections: if section: md_content += section + "\n\n" for table in tables: if table: md_content += table + "\n\n" self.set_output("markdown", md_content) else: # Default DeepDOC parser spreadsheet_parser = ExcelParser() if conf.get("output_format") == "html": htmls = spreadsheet_parser.html(blob, 1000000000) self.set_output("html", htmls[0]) elif conf.get("output_format") == "json": self.set_output("json", [{"text": txt} for txt in spreadsheet_parser(blob) if txt]) elif conf.get("output_format") == "markdown": self.set_output("markdown", spreadsheet_parser.markdown(blob)) def _word(self, name, blob): self.callback(random.randint(1, 5) / 100.0, "Start to work on a Word Processor Document") conf = self._param.setups["word"] self.set_output("output_format", conf["output_format"]) docx_parser = Docx() if conf.get("output_format") == "json": sections, tbls = docx_parser(name, binary=blob) sections = [{"text": section[0], "image": section[1]} for section in sections if section] sections.extend([{"text": tb, "image": None, "doc_type_kwd": "table"} for ((_, tb), _) in tbls]) self.set_output("json", sections) elif conf.get("output_format") == "markdown": markdown_text = docx_parser.to_markdown(name, binary=blob) self.set_output("markdown", markdown_text) def _slides(self, name, blob): self.callback(random.randint(1, 5) / 100.0, "Start to work on a PowerPoint Document") conf = self._param.setups["slides"] self.set_output("output_format", conf["output_format"]) parse_method = conf.get("parse_method", "deepdoc") # Handle TCADP parser if parse_method.lower() == "tcadp parser": table_result_type = conf.get("table_result_type", "1") markdown_image_response_type = conf.get("markdown_image_response_type", "1") tcadp_parser = TCADPParser( table_result_type=table_result_type, markdown_image_response_type=markdown_image_response_type, ) if not tcadp_parser.check_installation(): raise RuntimeError("TCADP parser not available. Please check Tencent Cloud API configuration.") # Determine file type based on extension if re.search(r"\.pptx?$", name, re.IGNORECASE): file_type = "PPTX" else: file_type = "PPT" self.callback(0.2, f"Using TCADP parser for {file_type} file.") sections, tables = tcadp_parser.parse_pdf( filepath=name, binary=blob, callback=self.callback, file_type=file_type, file_start_page=1, file_end_page=1000, ) # Process TCADP parser output - PPT only supports json format output_format = conf.get("output_format", "json") if output_format == "json": # For JSON output, create a list of text items result = [] # Add sections as text for section, position_tag in sections: if section: result.append({"text": section}) # Add tables as text for table in tables: if table: result.append({"text": table, "doc_type_kwd": "table"}) self.set_output("json", result) else: # Default DeepDOC parser (supports .pptx format) from deepdoc.parser.ppt_parser import RAGFlowPptParser as ppt_parser ppt_parser = ppt_parser() txts = ppt_parser(blob, 0, 100000, None) sections = [{"text": section} for section in txts if section.strip()] # json assert conf.get("output_format") == "json", "have to be json for ppt" if conf.get("output_format") == "json": self.set_output("json", sections) def _markdown(self, name, blob): from functools import reduce from rag.app.naive import Markdown as naive_markdown_parser from rag.nlp import concat_img self.callback(random.randint(1, 5) / 100.0, "Start to work on a markdown.") conf = self._param.setups["text&markdown"] self.set_output("output_format", conf["output_format"]) markdown_parser = naive_markdown_parser() sections, tables, section_images = markdown_parser( name, blob, separate_tables=False, delimiter=conf.get("delimiter"), return_section_images=True, ) if conf.get("output_format") == "json": json_results = [] for idx, (section_text, _) in enumerate(sections): json_result = { "text": section_text, } images = [] if section_images and len(section_images) > idx and section_images[idx] is not None: images.append(section_images[idx]) if images: # If multiple images found, combine them using concat_img combined_image = reduce(concat_img, images) if len(images) > 1 else images[0] json_result["image"] = combined_image json_results.append(json_result) self.set_output("json", json_results) else: self.set_output("text", "\n".join([section_text for section_text, _ in sections])) def _image(self, name, blob): from deepdoc.vision import OCR self.callback(random.randint(1, 5) / 100.0, "Start to work on an image.") conf = self._param.setups["image"] self.set_output("output_format", conf["output_format"]) img = Image.open(io.BytesIO(blob)).convert("RGB") if conf["parse_method"] == "ocr": # use ocr, recognize chars only ocr = OCR() bxs = ocr(np.array(img)) # return boxes and recognize result txt = "\n".join([t[0] for _, t in bxs if t[0]]) else: lang = conf["lang"] # use VLM to describe the picture cv_model = LLMBundle(self._canvas.get_tenant_id(), LLMType.IMAGE2TEXT, llm_name=conf["parse_method"], lang=lang) img_binary = io.BytesIO() img.save(img_binary, format="JPEG") img_binary.seek(0) system_prompt = conf.get("system_prompt") if system_prompt: txt = cv_model.describe_with_prompt(img_binary.read(), system_prompt) else: txt = cv_model.describe(img_binary.read()) self.set_output("text", txt) def _audio(self, name, blob): import os import tempfile self.callback(random.randint(1, 5) / 100.0, "Start to work on an audio.") conf = self._param.setups["audio"] self.set_output("output_format", conf["output_format"]) _, ext = os.path.splitext(name) with tempfile.NamedTemporaryFile(suffix=ext) as tmpf: tmpf.write(blob) tmpf.flush() tmp_path = os.path.abspath(tmpf.name) seq2txt_mdl = LLMBundle(self._canvas.get_tenant_id(), LLMType.SPEECH2TEXT, llm_name=conf["llm_id"]) txt = seq2txt_mdl.transcription(tmp_path) self.set_output("text", txt) def _video(self, name, blob): self.callback(random.randint(1, 5) / 100.0, "Start to work on an video.") conf = self._param.setups["video"] self.set_output("output_format", conf["output_format"]) cv_mdl = LLMBundle(self._canvas.get_tenant_id(), LLMType.IMAGE2TEXT, llm_name=conf["llm_id"]) txt = asyncio.run(cv_mdl.async_chat(system="", history=[], gen_conf={}, video_bytes=blob, filename=name)) self.set_output("text", txt) def _email(self, name, blob): self.callback(random.randint(1, 5) / 100.0, "Start to work on an email.") email_content = {} conf = self._param.setups["email"] self.set_output("output_format", conf["output_format"]) target_fields = conf["fields"] _, ext = os.path.splitext(name) if ext == ".eml": # handle eml file from email import policy from email.parser import BytesParser msg = BytesParser(policy=policy.default).parse(io.BytesIO(blob)) email_content["metadata"] = {} # handle header info for header, value in msg.items(): # get fields like from, to, cc, bcc, date, subject if header.lower() in target_fields: email_content[header.lower()] = value # get metadata elif header.lower() not in ["from", "to", "cc", "bcc", "date", "subject"]: email_content["metadata"][header.lower()] = value # get body if "body" in target_fields: body_text, body_html = [], [] def _add_content(m, content_type): def _decode_payload(payload, charset, target_list): try: target_list.append(payload.decode(charset)) except (UnicodeDecodeError, LookupError): for enc in ["utf-8", "gb2312", "gbk", "gb18030", "latin1"]: try: target_list.append(payload.decode(enc)) break except UnicodeDecodeError: continue else: target_list.append(payload.decode("utf-8", errors="ignore")) if content_type == "text/plain": payload = msg.get_payload(decode=True) charset = msg.get_content_charset() or "utf-8" _decode_payload(payload, charset, body_text) elif content_type == "text/html": payload = msg.get_payload(decode=True) charset = msg.get_content_charset() or "utf-8" _decode_payload(payload, charset, body_html) elif "multipart" in content_type: if m.is_multipart(): for part in m.iter_parts(): _add_content(part, part.get_content_type()) _add_content(msg, msg.get_content_type()) email_content["text"] = "\n".join(body_text) email_content["text_html"] = "\n".join(body_html) # get attachment if "attachments" in target_fields: attachments = [] for part in msg.iter_attachments(): content_disposition = part.get("Content-Disposition") if content_disposition: dispositions = content_disposition.strip().split(";") if dispositions[0].lower() == "attachment": filename = part.get_filename() payload = part.get_payload(decode=True).decode(part.get_content_charset()) attachments.append( { "filename": filename, "payload": payload, } ) email_content["attachments"] = attachments else: # handle msg file import extract_msg print("handle a msg file.") msg = extract_msg.Message(blob) # handle header info basic_content = { "from": msg.sender, "to": msg.to, "cc": msg.cc, "bcc": msg.bcc, "date": msg.date, "subject": msg.subject, } email_content.update({k: v for k, v in basic_content.items() if k in target_fields}) # get metadata email_content["metadata"] = { "message_id": msg.messageId, "in_reply_to": msg.inReplyTo, } # get body if "body" in target_fields: email_content["text"] = msg.body[0] if isinstance(msg.body, list) and msg.body else msg.body if not email_content["text"] and msg.htmlBody: email_content["text"] = msg.htmlBody[0] if isinstance(msg.htmlBody, list) and msg.htmlBody else msg.htmlBody # get attachments if "attachments" in target_fields: attachments = [] for t in msg.attachments: attachments.append( { "filename": t.name, "payload": t.data.decode("utf-8"), } ) email_content["attachments"] = attachments if conf["output_format"] == "json": self.set_output("json", [email_content]) else: content_txt = "" for k, v in email_content.items(): if isinstance(v, str): # basic info content_txt += f"{k}:{v}" + "\n" elif isinstance(v, dict): # metadata content_txt += f"{k}:{json.dumps(v)}" + "\n" elif isinstance(v, list): # attachments or others for fb in v: if isinstance(fb, dict): # attachments content_txt += f"{fb['filename']}:{fb['payload']}" + "\n" else: # str, usually plain text content_txt += fb self.set_output("text", content_txt) async def _invoke(self, **kwargs): function_map = { "pdf": self._pdf, "text&markdown": self._markdown, "spreadsheet": self._spreadsheet, "slides": self._slides, "word": self._word, "image": self._image, "audio": self._audio, "video": self._video, "email": self._email, } try: from_upstream = ParserFromUpstream.model_validate(kwargs) except Exception as e: self.set_output("_ERROR", f"Input error: {str(e)}") return name = from_upstream.name if self._canvas._doc_id: b, n = File2DocumentService.get_storage_address(doc_id=self._canvas._doc_id) blob = settings.STORAGE_IMPL.get(b, n) else: blob = FileService.get_blob(from_upstream.file["created_by"], from_upstream.file["id"]) done = False for p_type, conf in self._param.setups.items(): if from_upstream.name.split(".")[-1].lower() not in conf.get("suffix", []): continue await asyncio.to_thread(function_map[p_type], name, blob) done = True break if not done: raise Exception("No suitable for file extension: `.%s`" % from_upstream.name.split(".")[-1].lower()) outs = self.output() tasks = [] for d in outs.get("json", []): tasks.append(asyncio.create_task(image2id(d,partial(settings.STORAGE_IMPL.put, tenant_id=self._canvas._tenant_id),get_uuid()))) try: await asyncio.gather(*tasks, return_exceptions=False)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/parser/schema.py
rag/flow/parser/schema.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from pydantic import BaseModel, ConfigDict, Field class ParserFromUpstream(BaseModel): created_time: float | None = Field(default=None, alias="_created_time") elapsed_time: float | None = Field(default=None, alias="_elapsed_time") name: str file: dict | None = Field(default=None) model_config = ConfigDict(populate_by_name=True, extra="forbid")
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/flow/parser/__init__.py
rag/flow/parser/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/svr/cache_file_svr.py
rag/svr/cache_file_svr.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import time import traceback from api.db.db_models import close_connection from api.db.services.task_service import TaskService from rag.utils.redis_conn import REDIS_CONN from common import settings def collect(): doc_locations = TaskService.get_ongoing_doc_name() logging.debug(doc_locations) if len(doc_locations) == 0: time.sleep(1) return None return doc_locations def main(): locations = collect() if not locations: return logging.info(f"TASKS: {len(locations)}") for kb_id, loc in locations: try: if REDIS_CONN.is_alive(): try: key = "{}/{}".format(kb_id, loc) if REDIS_CONN.exist(key): continue file_bin = settings.STORAGE_IMPL.get(kb_id, loc) REDIS_CONN.transaction(key, file_bin, 12 * 60) logging.info("CACHE: {}".format(loc)) except Exception as e: logging.error(f"Error to get data from REDIS: {e}") traceback.print_stack() except Exception as e: logging.error(f"Error to check REDIS connection: {e}") traceback.print_stack() if __name__ == "__main__": while True: main() close_connection() time.sleep(1)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/svr/task_executor.py
rag/svr/task_executor.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import asyncio import socket import concurrent # from beartype import BeartypeConf # from beartype.claw import beartype_all # <-- you didn't sign up for this # beartype_all(conf=BeartypeConf(violation_type=UserWarning)) # <-- emit warnings from all code import random import sys import threading import time from api.db import PIPELINE_SPECIAL_PROGRESS_FREEZE_TASK_TYPES from api.db.services.knowledgebase_service import KnowledgebaseService from api.db.services.pipeline_operation_log_service import PipelineOperationLogService from api.db.joint_services.memory_message_service import handle_save_to_memory_task from common.connection_utils import timeout from common.metadata_utils import update_metadata_to, metadata_schema from rag.utils.base64_image import image2id from rag.utils.raptor_utils import should_skip_raptor, get_skip_reason from common.log_utils import init_root_logger from common.config_utils import show_configs from graphrag.general.index import run_graphrag_for_kb from graphrag.utils import get_llm_cache, set_llm_cache, get_tags_from_cache, set_tags_to_cache from rag.prompts.generator import keyword_extraction, question_proposal, content_tagging, run_toc_from_text, \ gen_metadata import logging import os from datetime import datetime import json import xxhash import copy import re from functools import partial from multiprocessing.context import TimeoutError from timeit import default_timer as timer import signal import exceptiongroup import faulthandler import numpy as np from peewee import DoesNotExist from common.constants import LLMType, ParserType, PipelineTaskType from api.db.services.document_service import DocumentService from api.db.services.llm_service import LLMBundle from api.db.services.task_service import TaskService, has_canceled, CANVAS_DEBUG_DOC_ID, GRAPH_RAPTOR_FAKE_DOC_ID from api.db.services.file2document_service import File2DocumentService from common.versions import get_ragflow_version from api.db.db_models import close_connection from rag.app import laws, paper, presentation, manual, qa, table, book, resume, picture, naive, one, audio, \ email, tag from rag.nlp import search, rag_tokenizer, add_positions from rag.raptor import RecursiveAbstractiveProcessing4TreeOrganizedRetrieval as Raptor from common.token_utils import num_tokens_from_string, truncate from rag.utils.redis_conn import REDIS_CONN, RedisDistributedLock from graphrag.utils import chat_limiter from common.signal_utils import start_tracemalloc_and_snapshot, stop_tracemalloc from common.exceptions import TaskCanceledException from common import settings from common.constants import PAGERANK_FLD, TAG_FLD, SVR_CONSUMER_GROUP_NAME BATCH_SIZE = 64 FACTORY = { "general": naive, ParserType.NAIVE.value: naive, ParserType.PAPER.value: paper, ParserType.BOOK.value: book, ParserType.PRESENTATION.value: presentation, ParserType.MANUAL.value: manual, ParserType.LAWS.value: laws, ParserType.QA.value: qa, ParserType.TABLE.value: table, ParserType.RESUME.value: resume, ParserType.PICTURE.value: picture, ParserType.ONE.value: one, ParserType.AUDIO.value: audio, ParserType.EMAIL.value: email, ParserType.KG.value: naive, ParserType.TAG.value: tag } TASK_TYPE_TO_PIPELINE_TASK_TYPE = { "dataflow": PipelineTaskType.PARSE, "raptor": PipelineTaskType.RAPTOR, "graphrag": PipelineTaskType.GRAPH_RAG, "mindmap": PipelineTaskType.MINDMAP, "memory": PipelineTaskType.MEMORY, } UNACKED_ITERATOR = None CONSUMER_NO = "0" if len(sys.argv) < 2 else sys.argv[1] CONSUMER_NAME = "task_executor_" + CONSUMER_NO BOOT_AT = datetime.now().astimezone().isoformat(timespec="milliseconds") PENDING_TASKS = 0 LAG_TASKS = 0 DONE_TASKS = 0 FAILED_TASKS = 0 CURRENT_TASKS = {} MAX_CONCURRENT_TASKS = int(os.environ.get('MAX_CONCURRENT_TASKS', "5")) MAX_CONCURRENT_CHUNK_BUILDERS = int(os.environ.get('MAX_CONCURRENT_CHUNK_BUILDERS', "1")) MAX_CONCURRENT_MINIO = int(os.environ.get('MAX_CONCURRENT_MINIO', '10')) task_limiter = asyncio.Semaphore(MAX_CONCURRENT_TASKS) chunk_limiter = asyncio.Semaphore(MAX_CONCURRENT_CHUNK_BUILDERS) embed_limiter = asyncio.Semaphore(MAX_CONCURRENT_CHUNK_BUILDERS) minio_limiter = asyncio.Semaphore(MAX_CONCURRENT_MINIO) kg_limiter = asyncio.Semaphore(2) WORKER_HEARTBEAT_TIMEOUT = int(os.environ.get('WORKER_HEARTBEAT_TIMEOUT', '120')) stop_event = threading.Event() def signal_handler(sig, frame): logging.info("Received interrupt signal, shutting down...") stop_event.set() time.sleep(1) sys.exit(0) def set_progress(task_id, from_page=0, to_page=-1, prog=None, msg="Processing..."): try: if prog is not None and prog < 0: msg = "[ERROR]" + msg cancel = has_canceled(task_id) if cancel: msg += " [Canceled]" prog = -1 if to_page > 0: if msg: if from_page < to_page: msg = f"Page({from_page + 1}~{to_page + 1}): " + msg if msg: msg = datetime.now().strftime("%H:%M:%S") + " " + msg d = {"progress_msg": msg} if prog is not None: d["progress"] = prog TaskService.update_progress(task_id, d) close_connection() if cancel: raise TaskCanceledException(msg) logging.info(f"set_progress({task_id}), progress: {prog}, progress_msg: {msg}") except DoesNotExist: logging.warning(f"set_progress({task_id}) got exception DoesNotExist") except Exception as e: logging.exception(f"set_progress({task_id}), progress: {prog}, progress_msg: {msg}, got exception: {e}") async def collect(): global CONSUMER_NAME, DONE_TASKS, FAILED_TASKS global UNACKED_ITERATOR svr_queue_names = settings.get_svr_queue_names() redis_msg = None try: if not UNACKED_ITERATOR: UNACKED_ITERATOR = REDIS_CONN.get_unacked_iterator(svr_queue_names, SVR_CONSUMER_GROUP_NAME, CONSUMER_NAME) try: redis_msg = next(UNACKED_ITERATOR) except StopIteration: for svr_queue_name in svr_queue_names: redis_msg = REDIS_CONN.queue_consumer(svr_queue_name, SVR_CONSUMER_GROUP_NAME, CONSUMER_NAME) if redis_msg: break except Exception as e: logging.exception(f"collect got exception: {e}") return None, None if not redis_msg: return None, None msg = redis_msg.get_message() if not msg: logging.error(f"collect got empty message of {redis_msg.get_msg_id()}") redis_msg.ack() return None, None canceled = False if msg.get("doc_id", "") in [GRAPH_RAPTOR_FAKE_DOC_ID, CANVAS_DEBUG_DOC_ID]: task = msg if task["task_type"] in PIPELINE_SPECIAL_PROGRESS_FREEZE_TASK_TYPES: task = TaskService.get_task(msg["id"], msg["doc_ids"]) if task: task["doc_id"] = msg["doc_id"] task["doc_ids"] = msg.get("doc_ids", []) or [] elif msg.get("task_type") == PipelineTaskType.MEMORY.lower(): _, task_obj = TaskService.get_by_id(msg["id"]) task = task_obj.to_dict() else: task = TaskService.get_task(msg["id"]) if task: canceled = has_canceled(task["id"]) if not task or canceled: state = "is unknown" if not task else "has been cancelled" FAILED_TASKS += 1 logging.warning(f"collect task {msg['id']} {state}") redis_msg.ack() return None, None task_type = msg.get("task_type", "") task["task_type"] = task_type if task_type[:8] == "dataflow": task["tenant_id"] = msg["tenant_id"] task["dataflow_id"] = msg["dataflow_id"] task["kb_id"] = msg.get("kb_id", "") if task_type[:6] == "memory": task["memory_id"] = msg["memory_id"] task["source_id"] = msg["source_id"] task["message_dict"] = msg["message_dict"] return redis_msg, task async def get_storage_binary(bucket, name): return await asyncio.to_thread(settings.STORAGE_IMPL.get, bucket, name) @timeout(60 * 80, 1) async def build_chunks(task, progress_callback): if task["size"] > settings.DOC_MAXIMUM_SIZE: set_progress(task["id"], prog=-1, msg="File size exceeds( <= %dMb )" % (int(settings.DOC_MAXIMUM_SIZE / 1024 / 1024))) return [] chunker = FACTORY[task["parser_id"].lower()] try: st = timer() bucket, name = File2DocumentService.get_storage_address(doc_id=task["doc_id"]) binary = await get_storage_binary(bucket, name) logging.info("From minio({}) {}/{}".format(timer() - st, task["location"], task["name"])) except TimeoutError: progress_callback(-1, "Internal server error: Fetch file from minio timeout. Could you try it again.") logging.exception( "Minio {}/{} got timeout: Fetch file from minio timeout.".format(task["location"], task["name"])) raise except Exception as e: if re.search("(No such file|not found)", str(e)): progress_callback(-1, "Can not find file <%s> from minio. Could you try it again?" % task["name"]) else: progress_callback(-1, "Get file from minio: %s" % str(e).replace("'", "")) logging.exception("Chunking {}/{} got exception".format(task["location"], task["name"])) raise try: async with chunk_limiter: cks = await asyncio.to_thread( chunker.chunk, task["name"], binary=binary, from_page=task["from_page"], to_page=task["to_page"], lang=task["language"], callback=progress_callback, kb_id=task["kb_id"], parser_config=task["parser_config"], tenant_id=task["tenant_id"], ) logging.info("Chunking({}) {}/{} done".format(timer() - st, task["location"], task["name"])) except TaskCanceledException: raise except Exception as e: progress_callback(-1, "Internal server error while chunking: %s" % str(e).replace("'", "")) logging.exception("Chunking {}/{} got exception".format(task["location"], task["name"])) raise docs = [] doc = { "doc_id": task["doc_id"], "kb_id": str(task["kb_id"]) } if task["pagerank"]: doc[PAGERANK_FLD] = int(task["pagerank"]) st = timer() @timeout(60) async def upload_to_minio(document, chunk): try: d = copy.deepcopy(document) d.update(chunk) d["id"] = xxhash.xxh64( (chunk["content_with_weight"] + str(d["doc_id"])).encode("utf-8", "surrogatepass")).hexdigest() d["create_time"] = str(datetime.now()).replace("T", " ")[:19] d["create_timestamp_flt"] = datetime.now().timestamp() if not d.get("image"): _ = d.pop("image", None) d["img_id"] = "" docs.append(d) return await image2id(d, partial(settings.STORAGE_IMPL.put, tenant_id=task["tenant_id"]), d["id"], task["kb_id"]) docs.append(d) except Exception: logging.exception( "Saving image of chunk {}/{}/{} got exception".format(task["location"], task["name"], d["id"])) raise tasks = [] for ck in cks: tasks.append(asyncio.create_task(upload_to_minio(doc, ck))) try: await asyncio.gather(*tasks, return_exceptions=False) except Exception as e: logging.error(f"MINIO PUT({task['name']}) got exception: {e}") for t in tasks: t.cancel() await asyncio.gather(*tasks, return_exceptions=True) raise el = timer() - st logging.info("MINIO PUT({}) cost {:.3f} s".format(task["name"], el)) if task["parser_config"].get("auto_keywords", 0): st = timer() progress_callback(msg="Start to generate keywords for every chunk ...") chat_mdl = LLMBundle(task["tenant_id"], LLMType.CHAT, llm_name=task["llm_id"], lang=task["language"]) async def doc_keyword_extraction(chat_mdl, d, topn): cached = get_llm_cache(chat_mdl.llm_name, d["content_with_weight"], "keywords", {"topn": topn}) if not cached: if has_canceled(task["id"]): progress_callback(-1, msg="Task has been canceled.") return async with chat_limiter: cached = await keyword_extraction(chat_mdl, d["content_with_weight"], topn) set_llm_cache(chat_mdl.llm_name, d["content_with_weight"], cached, "keywords", {"topn": topn}) if cached: d["important_kwd"] = cached.split(",") d["important_tks"] = rag_tokenizer.tokenize(" ".join(d["important_kwd"])) return tasks = [] for d in docs: tasks.append( asyncio.create_task(doc_keyword_extraction(chat_mdl, d, task["parser_config"]["auto_keywords"]))) try: await asyncio.gather(*tasks, return_exceptions=False) except Exception as e: logging.error("Error in doc_keyword_extraction: {}".format(e)) for t in tasks: t.cancel() await asyncio.gather(*tasks, return_exceptions=True) raise progress_callback(msg="Keywords generation {} chunks completed in {:.2f}s".format(len(docs), timer() - st)) if task["parser_config"].get("auto_questions", 0): st = timer() progress_callback(msg="Start to generate questions for every chunk ...") chat_mdl = LLMBundle(task["tenant_id"], LLMType.CHAT, llm_name=task["llm_id"], lang=task["language"]) async def doc_question_proposal(chat_mdl, d, topn): cached = get_llm_cache(chat_mdl.llm_name, d["content_with_weight"], "question", {"topn": topn}) if not cached: if has_canceled(task["id"]): progress_callback(-1, msg="Task has been canceled.") return async with chat_limiter: cached = await question_proposal(chat_mdl, d["content_with_weight"], topn) set_llm_cache(chat_mdl.llm_name, d["content_with_weight"], cached, "question", {"topn": topn}) if cached: d["question_kwd"] = cached.split("\n") d["question_tks"] = rag_tokenizer.tokenize("\n".join(d["question_kwd"])) tasks = [] for d in docs: tasks.append( asyncio.create_task(doc_question_proposal(chat_mdl, d, task["parser_config"]["auto_questions"]))) try: await asyncio.gather(*tasks, return_exceptions=False) except Exception as e: logging.error("Error in doc_question_proposal", exc_info=e) for t in tasks: t.cancel() await asyncio.gather(*tasks, return_exceptions=True) raise progress_callback(msg="Question generation {} chunks completed in {:.2f}s".format(len(docs), timer() - st)) if task["parser_config"].get("enable_metadata", False) and task["parser_config"].get("metadata"): st = timer() progress_callback(msg="Start to generate meta-data for every chunk ...") chat_mdl = LLMBundle(task["tenant_id"], LLMType.CHAT, llm_name=task["llm_id"], lang=task["language"]) async def gen_metadata_task(chat_mdl, d): cached = get_llm_cache(chat_mdl.llm_name, d["content_with_weight"], "metadata", task["parser_config"]["metadata"]) if not cached: if has_canceled(task["id"]): progress_callback(-1, msg="Task has been canceled.") return async with chat_limiter: cached = await gen_metadata(chat_mdl, metadata_schema(task["parser_config"]["metadata"]), d["content_with_weight"]) set_llm_cache(chat_mdl.llm_name, d["content_with_weight"], cached, "metadata", task["parser_config"]["metadata"]) if cached: d["metadata_obj"] = cached tasks = [] for d in docs: tasks.append(asyncio.create_task(gen_metadata_task(chat_mdl, d))) try: await asyncio.gather(*tasks, return_exceptions=False) except Exception as e: logging.error("Error in doc_question_proposal", exc_info=e) for t in tasks: t.cancel() await asyncio.gather(*tasks, return_exceptions=True) raise metadata = {} for doc in docs: metadata = update_metadata_to(metadata, doc["metadata_obj"]) del doc["metadata_obj"] if metadata: e, doc = DocumentService.get_by_id(task["doc_id"]) if e: if isinstance(doc.meta_fields, str): doc.meta_fields = json.loads(doc.meta_fields) metadata = update_metadata_to(metadata, doc.meta_fields) DocumentService.update_by_id(task["doc_id"], {"meta_fields": metadata}) progress_callback(msg="Question generation {} chunks completed in {:.2f}s".format(len(docs), timer() - st)) if task["kb_parser_config"].get("tag_kb_ids", []): progress_callback(msg="Start to tag for every chunk ...") kb_ids = task["kb_parser_config"]["tag_kb_ids"] tenant_id = task["tenant_id"] topn_tags = task["kb_parser_config"].get("topn_tags", 3) S = 1000 st = timer() examples = [] all_tags = get_tags_from_cache(kb_ids) if not all_tags: all_tags = settings.retriever.all_tags_in_portion(tenant_id, kb_ids, S) set_tags_to_cache(kb_ids, all_tags) else: all_tags = json.loads(all_tags) chat_mdl = LLMBundle(task["tenant_id"], LLMType.CHAT, llm_name=task["llm_id"], lang=task["language"]) docs_to_tag = [] for d in docs: task_canceled = has_canceled(task["id"]) if task_canceled: progress_callback(-1, msg="Task has been canceled.") return None if settings.retriever.tag_content(tenant_id, kb_ids, d, all_tags, topn_tags=topn_tags, S=S) and len( d[TAG_FLD]) > 0: examples.append({"content": d["content_with_weight"], TAG_FLD: d[TAG_FLD]}) else: docs_to_tag.append(d) async def doc_content_tagging(chat_mdl, d, topn_tags): cached = get_llm_cache(chat_mdl.llm_name, d["content_with_weight"], all_tags, {"topn": topn_tags}) if not cached: if has_canceled(task["id"]): progress_callback(-1, msg="Task has been canceled.") return picked_examples = random.choices(examples, k=2) if len(examples) > 2 else examples if not picked_examples: picked_examples.append({"content": "This is an example", TAG_FLD: {'example': 1}}) async with chat_limiter: cached = await content_tagging( chat_mdl, d["content_with_weight"], all_tags, picked_examples, topn_tags, ) if cached: cached = json.dumps(cached) if cached: set_llm_cache(chat_mdl.llm_name, d["content_with_weight"], cached, all_tags, {"topn": topn_tags}) d[TAG_FLD] = json.loads(cached) tasks = [] for d in docs_to_tag: tasks.append(asyncio.create_task(doc_content_tagging(chat_mdl, d, topn_tags))) try: await asyncio.gather(*tasks, return_exceptions=False) except Exception as e: logging.error("Error tagging docs: {}".format(e)) for t in tasks: t.cancel() await asyncio.gather(*tasks, return_exceptions=True) raise progress_callback(msg="Tagging {} chunks completed in {:.2f}s".format(len(docs), timer() - st)) return docs def build_TOC(task, docs, progress_callback): progress_callback(msg="Start to generate table of content ...") chat_mdl = LLMBundle(task["tenant_id"], LLMType.CHAT, llm_name=task["llm_id"], lang=task["language"]) docs = sorted(docs, key=lambda d: ( d.get("page_num_int", 0)[0] if isinstance(d.get("page_num_int", 0), list) else d.get("page_num_int", 0), d.get("top_int", 0)[0] if isinstance(d.get("top_int", 0), list) else d.get("top_int", 0) )) toc: list[dict] = asyncio.run( run_toc_from_text([d["content_with_weight"] for d in docs], chat_mdl, progress_callback)) logging.info("------------ T O C -------------\n" + json.dumps(toc, ensure_ascii=False, indent=' ')) ii = 0 while ii < len(toc): try: idx = int(toc[ii]["chunk_id"]) del toc[ii]["chunk_id"] toc[ii]["ids"] = [docs[idx]["id"]] if ii == len(toc) - 1: break for jj in range(idx + 1, int(toc[ii + 1]["chunk_id"]) + 1): toc[ii]["ids"].append(docs[jj]["id"]) except Exception as e: logging.exception(e) ii += 1 if toc: d = copy.deepcopy(docs[-1]) d["content_with_weight"] = json.dumps(toc, ensure_ascii=False) d["toc_kwd"] = "toc" d["available_int"] = 0 d["page_num_int"] = [100000000] d["id"] = xxhash.xxh64( (d["content_with_weight"] + str(d["doc_id"])).encode("utf-8", "surrogatepass")).hexdigest() return d return None def init_kb(row, vector_size: int): idxnm = search.index_name(row["tenant_id"]) return settings.docStoreConn.create_idx(idxnm, row.get("kb_id", ""), vector_size) async def embedding(docs, mdl, parser_config=None, callback=None): if parser_config is None: parser_config = {} tts, cnts = [], [] for d in docs: tts.append(d.get("docnm_kwd", "Title")) c = "\n".join(d.get("question_kwd", [])) if not c: c = d["content_with_weight"] c = re.sub(r"</?(table|td|caption|tr|th)( [^<>]{0,12})?>", " ", c) if not c: c = "None" cnts.append(c) tk_count = 0 if len(tts) == len(cnts): vts, c = await asyncio.to_thread(mdl.encode, tts[0:1]) tts = np.tile(vts[0], (len(cnts), 1)) tk_count += c @timeout(60) def batch_encode(txts): nonlocal mdl return mdl.encode([truncate(c, mdl.max_length - 10) for c in txts]) cnts_ = np.array([]) for i in range(0, len(cnts), settings.EMBEDDING_BATCH_SIZE): async with embed_limiter: vts, c = await asyncio.to_thread(batch_encode, cnts[i: i + settings.EMBEDDING_BATCH_SIZE]) if len(cnts_) == 0: cnts_ = vts else: cnts_ = np.concatenate((cnts_, vts), axis=0) tk_count += c callback(prog=0.7 + 0.2 * (i + 1) / len(cnts), msg="") cnts = cnts_ filename_embd_weight = parser_config.get("filename_embd_weight", 0.1) # due to the db support none value if not filename_embd_weight: filename_embd_weight = 0.1 title_w = float(filename_embd_weight) if tts.ndim == 2 and cnts.ndim == 2 and tts.shape == cnts.shape: vects = title_w * tts + (1 - title_w) * cnts else: vects = cnts assert len(vects) == len(docs) vector_size = 0 for i, d in enumerate(docs): v = vects[i].tolist() vector_size = len(v) d["q_%d_vec" % len(v)] = v return tk_count, vector_size async def run_dataflow(task: dict): from api.db.services.canvas_service import UserCanvasService from rag.flow.pipeline import Pipeline task_start_ts = timer() dataflow_id = task["dataflow_id"] doc_id = task["doc_id"] task_id = task["id"] task_dataset_id = task["kb_id"] if task["task_type"] == "dataflow": e, cvs = UserCanvasService.get_by_id(dataflow_id) assert e, "User pipeline not found." dsl = cvs.dsl else: e, pipeline_log = PipelineOperationLogService.get_by_id(dataflow_id) assert e, "Pipeline log not found." dsl = pipeline_log.dsl dataflow_id = pipeline_log.pipeline_id pipeline = Pipeline(dsl, tenant_id=task["tenant_id"], doc_id=doc_id, task_id=task_id, flow_id=dataflow_id) chunks = await pipeline.run(file=task["file"]) if task.get("file") else await pipeline.run() if doc_id == CANVAS_DEBUG_DOC_ID: return if not chunks: PipelineOperationLogService.create(document_id=doc_id, pipeline_id=dataflow_id, task_type=PipelineTaskType.PARSE, dsl=str(pipeline)) return embedding_token_consumption = chunks.get("embedding_token_consumption", 0) if chunks.get("chunks"): chunks = copy.deepcopy(chunks["chunks"]) elif chunks.get("json"): chunks = copy.deepcopy(chunks["json"]) elif chunks.get("markdown"): chunks = [{"text": [chunks["markdown"]]}] elif chunks.get("text"): chunks = [{"text": [chunks["text"]]}] elif chunks.get("html"): chunks = [{"text": [chunks["html"]]}] keys = [k for o in chunks for k in list(o.keys())] if not any([re.match(r"q_[0-9]+_vec", k) for k in keys]): try: set_progress(task_id, prog=0.82, msg="\n-------------------------------------\nStart to embedding...") e, kb = KnowledgebaseService.get_by_id(task["kb_id"]) embedding_id = kb.embd_id embedding_model = LLMBundle(task["tenant_id"], LLMType.EMBEDDING, llm_name=embedding_id) @timeout(60) def batch_encode(txts): nonlocal embedding_model return embedding_model.encode([truncate(c, embedding_model.max_length - 10) for c in txts]) vects = np.array([]) texts = [o.get("questions", o.get("summary", o["text"])) for o in chunks] delta = 0.20 / (len(texts) // settings.EMBEDDING_BATCH_SIZE + 1) prog = 0.8 for i in range(0, len(texts), settings.EMBEDDING_BATCH_SIZE): async with embed_limiter: vts, c = await asyncio.to_thread(batch_encode, texts[i: i + settings.EMBEDDING_BATCH_SIZE]) if len(vects) == 0: vects = vts else: vects = np.concatenate((vects, vts), axis=0) embedding_token_consumption += c prog += delta if i % (len(texts) // settings.EMBEDDING_BATCH_SIZE / 100 + 1) == 1: set_progress(task_id, prog=prog, msg=f"{i + 1} / {len(texts) // settings.EMBEDDING_BATCH_SIZE}") assert len(vects) == len(chunks) for i, ck in enumerate(chunks): v = vects[i].tolist() ck["q_%d_vec" % len(v)] = v except Exception as e: set_progress(task_id, prog=-1, msg=f"[ERROR]: {e}") PipelineOperationLogService.create(document_id=doc_id, pipeline_id=dataflow_id, task_type=PipelineTaskType.PARSE, dsl=str(pipeline)) return metadata = {} for ck in chunks: ck["doc_id"] = doc_id ck["kb_id"] = [str(task["kb_id"])] ck["docnm_kwd"] = task["name"] ck["create_time"] = str(datetime.now()).replace("T", " ")[:19] ck["create_timestamp_flt"] = datetime.now().timestamp() if not ck.get("id"): ck["id"] = xxhash.xxh64((ck["text"] + str(ck["doc_id"])).encode("utf-8")).hexdigest() if "questions" in ck: if "question_tks" not in ck: ck["question_kwd"] = ck["questions"].split("\n") ck["question_tks"] = rag_tokenizer.tokenize(str(ck["questions"])) del ck["questions"] if "keywords" in ck: if "important_tks" not in ck: ck["important_kwd"] = ck["keywords"].split(",") ck["important_tks"] = rag_tokenizer.tokenize(str(ck["keywords"])) del ck["keywords"] if "summary" in ck: if "content_ltks" not in ck: ck["content_ltks"] = rag_tokenizer.tokenize(str(ck["summary"])) ck["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(ck["content_ltks"]) del ck["summary"] if "metadata" in ck: metadata = update_metadata_to(metadata, ck["metadata"]) del ck["metadata"] if "content_with_weight" not in ck: ck["content_with_weight"] = ck["text"] del ck["text"] if "positions" in ck: add_positions(ck, ck["positions"]) del ck["positions"] if metadata: e, doc = DocumentService.get_by_id(doc_id) if e: if isinstance(doc.meta_fields, str): doc.meta_fields = json.loads(doc.meta_fields) metadata = update_metadata_to(metadata, doc.meta_fields) DocumentService.update_by_id(doc_id, {"meta_fields": metadata}) start_ts = timer() set_progress(task_id, prog=0.82, msg="[DOC Engine]:\nStart to index...") e = await insert_es(task_id, task["tenant_id"], task["kb_id"], chunks, partial(set_progress, task_id, 0, 100000000)) if not e: PipelineOperationLogService.create(document_id=doc_id, pipeline_id=dataflow_id, task_type=PipelineTaskType.PARSE, dsl=str(pipeline)) return time_cost = timer() - start_ts task_time_cost = timer() - task_start_ts set_progress(task_id, prog=1., msg="Indexing done ({:.2f}s). Task done ({:.2f}s)".format(time_cost, task_time_cost)) DocumentService.increment_chunk_num(doc_id, task_dataset_id, embedding_token_consumption, len(chunks), task_time_cost) logging.info("[Done], chunks({}), token({}), elapsed:{:.2f}".format(len(chunks), embedding_token_consumption, task_time_cost)) PipelineOperationLogService.create(document_id=doc_id, pipeline_id=dataflow_id, task_type=PipelineTaskType.PARSE, dsl=str(pipeline)) @timeout(3600) async def run_raptor_for_kb(row, kb_parser_config, chat_mdl, embd_mdl, vector_size, callback=None, doc_ids=[]): fake_doc_id = GRAPH_RAPTOR_FAKE_DOC_ID raptor_config = kb_parser_config.get("raptor", {}) vctr_nm = "q_%d_vec" % vector_size res = [] tk_count = 0 max_errors = int(os.environ.get("RAPTOR_MAX_ERRORS", 3)) async def generate(chunks, did): nonlocal tk_count, res raptor = Raptor( raptor_config.get("max_cluster", 64), chat_mdl, embd_mdl, raptor_config["prompt"], raptor_config["max_token"], raptor_config["threshold"], max_errors=max_errors, ) original_length = len(chunks) chunks = await raptor(chunks, kb_parser_config["raptor"]["random_seed"], callback, row["id"]) doc = { "doc_id": did, "kb_id": [str(row["kb_id"])], "docnm_kwd": row["name"], "title_tks": rag_tokenizer.tokenize(row["name"]), "raptor_kwd": "raptor" } if row["pagerank"]: doc[PAGERANK_FLD] = int(row["pagerank"]) for content, vctr in chunks[original_length:]: d = copy.deepcopy(doc) d["id"] = xxhash.xxh64((content + str(fake_doc_id)).encode("utf-8")).hexdigest()
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/svr/sync_data_source.py
rag/svr/sync_data_source.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # from beartype import BeartypeConf # from beartype.claw import beartype_all # <-- you didn't sign up for this # beartype_all(conf=BeartypeConf(violation_type=UserWarning)) # <-- emit warnings from all code import asyncio import copy import faulthandler import logging import os import signal import sys import threading import time import traceback from datetime import datetime, timezone from typing import Any from flask import json from api.db.services.connector_service import ConnectorService, SyncLogsService from api.db.services.knowledgebase_service import KnowledgebaseService from common import settings from common.config_utils import show_configs from common.data_source import ( BlobStorageConnector, NotionConnector, DiscordConnector, GoogleDriveConnector, MoodleConnector, JiraConnector, DropboxConnector, AirtableConnector, AsanaConnector, ImapConnector, ZendeskConnector, ) from common.constants import FileSource, TaskStatus from common.data_source.config import INDEX_BATCH_SIZE from common.data_source.models import ConnectorFailure from common.data_source.webdav_connector import WebDAVConnector from common.data_source.confluence_connector import ConfluenceConnector from common.data_source.gmail_connector import GmailConnector from common.data_source.box_connector import BoxConnector from common.data_source.github.connector import GithubConnector from common.data_source.gitlab_connector import GitlabConnector from common.data_source.bitbucket.connector import BitbucketConnector from common.data_source.interfaces import CheckpointOutputWrapper from common.log_utils import init_root_logger from common.signal_utils import start_tracemalloc_and_snapshot, stop_tracemalloc from common.versions import get_ragflow_version from box_sdk_gen import BoxOAuth, OAuthConfig, AccessToken MAX_CONCURRENT_TASKS = int(os.environ.get("MAX_CONCURRENT_TASKS", "5")) task_limiter = asyncio.Semaphore(MAX_CONCURRENT_TASKS) class SyncBase: SOURCE_NAME: str = None def __init__(self, conf: dict) -> None: self.conf = conf async def __call__(self, task: dict): SyncLogsService.start(task["id"], task["connector_id"]) async with task_limiter: try: await asyncio.wait_for(self._run_task_logic(task), timeout=task["timeout_secs"]) except asyncio.TimeoutError: msg = f"Task timeout after {task['timeout_secs']} seconds" SyncLogsService.update_by_id(task["id"], {"status": TaskStatus.FAIL, "error_msg": msg}) return except Exception as ex: msg = "\n".join([ "".join(traceback.format_exception_only(None, ex)).strip(), "".join(traceback.format_exception(None, ex, ex.__traceback__)).strip(), ]) SyncLogsService.update_by_id(task["id"], { "status": TaskStatus.FAIL, "full_exception_trace": msg, "error_msg": str(ex) }) return SyncLogsService.schedule(task["connector_id"], task["kb_id"], task["poll_range_start"]) async def _run_task_logic(self, task: dict): document_batch_generator = await self._generate(task) doc_num = 0 failed_docs = 0 next_update = datetime(1970, 1, 1, tzinfo=timezone.utc) if task["poll_range_start"]: next_update = task["poll_range_start"] for document_batch in document_batch_generator: if not document_batch: continue min_update = min(doc.doc_updated_at for doc in document_batch) max_update = max(doc.doc_updated_at for doc in document_batch) next_update = max(next_update, max_update) docs = [] for doc in document_batch: d = { "id": doc.id, "connector_id": task["connector_id"], "source": self.SOURCE_NAME, "semantic_identifier": doc.semantic_identifier, "extension": doc.extension, "size_bytes": doc.size_bytes, "doc_updated_at": doc.doc_updated_at, "blob": doc.blob, } if doc.metadata: d["metadata"] = doc.metadata docs.append(d) try: e, kb = KnowledgebaseService.get_by_id(task["kb_id"]) err, dids = SyncLogsService.duplicate_and_parse( kb, docs, task["tenant_id"], f"{self.SOURCE_NAME}/{task['connector_id']}", task["auto_parse"] ) SyncLogsService.increase_docs( task["id"], min_update, max_update, len(docs), "\n".join(err), len(err) ) doc_num += len(docs) except Exception as batch_ex: msg = str(batch_ex) code = getattr(batch_ex, "args", [None])[0] if code == 1267 or "collation" in msg.lower(): logging.warning(f"Skipping {len(docs)} document(s) due to collation conflict") else: logging.error(f"Error processing batch: {msg}") failed_docs += len(docs) continue prefix = self._get_source_prefix() if failed_docs > 0: logging.info(f"{prefix}{doc_num} docs synchronized till {next_update} ({failed_docs} skipped)") else: logging.info(f"{prefix}{doc_num} docs synchronized till {next_update}") SyncLogsService.done(task["id"], task["connector_id"]) task["poll_range_start"] = next_update async def _generate(self, task: dict): raise NotImplementedError def _get_source_prefix(self): return "" class _BlobLikeBase(SyncBase): DEFAULT_BUCKET_TYPE: str = "s3" async def _generate(self, task: dict): bucket_type = self.conf.get("bucket_type", self.DEFAULT_BUCKET_TYPE) self.connector = BlobStorageConnector( bucket_type=bucket_type, bucket_name=self.conf["bucket_name"], prefix=self.conf.get("prefix", ""), ) self.connector.load_credentials(self.conf["credentials"]) document_batch_generator = ( self.connector.load_from_state() if task["reindex"] == "1" or not task["poll_range_start"] else self.connector.poll_source( task["poll_range_start"].timestamp(), datetime.now(timezone.utc).timestamp(), ) ) begin_info = ( "totally" if task["reindex"] == "1" or not task["poll_range_start"] else "from {}".format(task["poll_range_start"]) ) logging.info( "Connect to {}: {}(prefix/{}) {}".format( bucket_type, self.conf["bucket_name"], self.conf.get("prefix", ""), begin_info, ) ) return document_batch_generator class S3(_BlobLikeBase): SOURCE_NAME: str = FileSource.S3 DEFAULT_BUCKET_TYPE: str = "s3" class R2(_BlobLikeBase): SOURCE_NAME: str = FileSource.R2 DEFAULT_BUCKET_TYPE: str = "r2" class OCI_STORAGE(_BlobLikeBase): SOURCE_NAME: str = FileSource.OCI_STORAGE DEFAULT_BUCKET_TYPE: str = "oci_storage" class GOOGLE_CLOUD_STORAGE(_BlobLikeBase): SOURCE_NAME: str = FileSource.GOOGLE_CLOUD_STORAGE DEFAULT_BUCKET_TYPE: str = "google_cloud_storage" class Confluence(SyncBase): SOURCE_NAME: str = FileSource.CONFLUENCE async def _generate(self, task: dict): from common.data_source.config import DocumentSource from common.data_source.interfaces import StaticCredentialsProvider index_mode = (self.conf.get("index_mode") or "everything").lower() if index_mode not in {"everything", "space", "page"}: index_mode = "everything" space = "" page_id = "" index_recursively = False if index_mode == "space": space = (self.conf.get("space") or "").strip() if not space: raise ValueError("Space Key is required when indexing a specific Confluence space.") elif index_mode == "page": page_id = (self.conf.get("page_id") or "").strip() if not page_id: raise ValueError("Page ID is required when indexing a specific Confluence page.") index_recursively = bool(self.conf.get("index_recursively", False)) self.connector = ConfluenceConnector( wiki_base=self.conf["wiki_base"], is_cloud=self.conf.get("is_cloud", True), space=space, page_id=page_id, index_recursively=index_recursively, ) credentials_provider = StaticCredentialsProvider(tenant_id=task["tenant_id"], connector_name=DocumentSource.CONFLUENCE, credential_json=self.conf["credentials"]) self.connector.set_credentials_provider(credentials_provider) # Determine the time range for synchronization based on reindex or poll_range_start if task["reindex"] == "1" or not task["poll_range_start"]: start_time = 0.0 begin_info = "totally" else: start_time = task["poll_range_start"].timestamp() begin_info = f"from {task['poll_range_start']}" end_time = datetime.now(timezone.utc).timestamp() raw_batch_size = self.conf.get("sync_batch_size") or self.conf.get("batch_size") or INDEX_BATCH_SIZE try: batch_size = int(raw_batch_size) except (TypeError, ValueError): batch_size = INDEX_BATCH_SIZE if batch_size <= 0: batch_size = INDEX_BATCH_SIZE def document_batches(): checkpoint = self.connector.build_dummy_checkpoint() pending_docs = [] iterations = 0 iteration_limit = 100_000 while checkpoint.has_more: wrapper = CheckpointOutputWrapper() doc_generator = wrapper(self.connector.load_from_checkpoint(start_time, end_time, checkpoint)) for document, failure, next_checkpoint in doc_generator: if failure is not None: logging.warning("Confluence connector failure: %s", getattr(failure, "failure_message", failure)) continue if document is not None: pending_docs.append(document) if len(pending_docs) >= batch_size: yield pending_docs pending_docs = [] if next_checkpoint is not None: checkpoint = next_checkpoint iterations += 1 if iterations > iteration_limit: raise RuntimeError("Too many iterations while loading Confluence documents.") if pending_docs: yield pending_docs def wrapper(): for batch in document_batches(): yield batch logging.info("Connect to Confluence: {} {}".format(self.conf["wiki_base"], begin_info)) return wrapper() class Notion(SyncBase): SOURCE_NAME: str = FileSource.NOTION async def _generate(self, task: dict): self.connector = NotionConnector(root_page_id=self.conf["root_page_id"]) self.connector.load_credentials(self.conf["credentials"]) document_generator = ( self.connector.load_from_state() if task["reindex"] == "1" or not task["poll_range_start"] else self.connector.poll_source(task["poll_range_start"].timestamp(), datetime.now(timezone.utc).timestamp()) ) begin_info = "totally" if task["reindex"] == "1" or not task["poll_range_start"] else "from {}".format( task["poll_range_start"]) logging.info("Connect to Notion: root({}) {}".format(self.conf["root_page_id"], begin_info)) return document_generator class Discord(SyncBase): SOURCE_NAME: str = FileSource.DISCORD async def _generate(self, task: dict): server_ids: str | None = self.conf.get("server_ids", None) # "channel1,channel2" channel_names: str | None = self.conf.get("channel_names", None) self.connector = DiscordConnector( server_ids=server_ids.split(",") if server_ids else [], channel_names=channel_names.split(",") if channel_names else [], start_date=datetime(1970, 1, 1, tzinfo=timezone.utc).strftime("%Y-%m-%d"), batch_size=self.conf.get("batch_size", 1024), ) self.connector.load_credentials(self.conf["credentials"]) document_generator = ( self.connector.load_from_state() if task["reindex"] == "1" or not task["poll_range_start"] else self.connector.poll_source(task["poll_range_start"].timestamp(), datetime.now(timezone.utc).timestamp()) ) begin_info = "totally" if task["reindex"] == "1" or not task["poll_range_start"] else "from {}".format( task["poll_range_start"]) logging.info("Connect to Discord: servers({}), channel({}) {}".format(server_ids, channel_names, begin_info)) return document_generator class Gmail(SyncBase): SOURCE_NAME: str = FileSource.GMAIL async def _generate(self, task: dict): # Gmail sync reuses the generic LoadConnector/PollConnector interface # implemented by common.data_source.gmail_connector.GmailConnector. # # Config expectations (self.conf): # credentials: Gmail / Workspace OAuth JSON (with primary admin email) # batch_size: optional, defaults to INDEX_BATCH_SIZE batch_size = self.conf.get("batch_size", INDEX_BATCH_SIZE) self.connector = GmailConnector(batch_size=batch_size) credentials = self.conf.get("credentials") if not credentials: raise ValueError("Gmail connector is missing credentials.") new_credentials = self.connector.load_credentials(credentials) if new_credentials: # Persist rotated / refreshed credentials back to connector config try: updated_conf = copy.deepcopy(self.conf) updated_conf["credentials"] = new_credentials ConnectorService.update_by_id(task["connector_id"], {"config": updated_conf}) self.conf = updated_conf logging.info( "Persisted refreshed Gmail credentials for connector %s", task["connector_id"], ) except Exception: logging.exception( "Failed to persist refreshed Gmail credentials for connector %s", task["connector_id"], ) # Decide between full reindex and incremental polling by time range. if task["reindex"] == "1" or not task.get("poll_range_start"): start_time = None end_time = None begin_info = "totally" document_generator = self.connector.load_from_state() else: poll_start = task["poll_range_start"] # Defensive: if poll_start is somehow None, fall back to full load if poll_start is None: start_time = None end_time = None begin_info = "totally" document_generator = self.connector.load_from_state() else: start_time = poll_start.timestamp() end_time = datetime.now(timezone.utc).timestamp() begin_info = f"from {poll_start}" document_generator = self.connector.poll_source(start_time, end_time) try: admin_email = self.connector.primary_admin_email except RuntimeError: admin_email = "unknown" logging.info(f"Connect to Gmail as {admin_email} {begin_info}") return document_generator class Dropbox(SyncBase): SOURCE_NAME: str = FileSource.DROPBOX async def _generate(self, task: dict): self.connector = DropboxConnector(batch_size=self.conf.get("batch_size", INDEX_BATCH_SIZE)) self.connector.load_credentials(self.conf["credentials"]) if task["reindex"] == "1" or not task["poll_range_start"]: document_generator = self.connector.load_from_state() begin_info = "totally" else: poll_start = task["poll_range_start"] document_generator = self.connector.poll_source( poll_start.timestamp(), datetime.now(timezone.utc).timestamp() ) begin_info = f"from {poll_start}" logging.info(f"[Dropbox] Connect to Dropbox {begin_info}") return document_generator class GoogleDrive(SyncBase): SOURCE_NAME: str = FileSource.GOOGLE_DRIVE async def _generate(self, task: dict): connector_kwargs = { "include_shared_drives": self.conf.get("include_shared_drives", False), "include_my_drives": self.conf.get("include_my_drives", False), "include_files_shared_with_me": self.conf.get("include_files_shared_with_me", False), "shared_drive_urls": self.conf.get("shared_drive_urls"), "my_drive_emails": self.conf.get("my_drive_emails"), "shared_folder_urls": self.conf.get("shared_folder_urls"), "specific_user_emails": self.conf.get("specific_user_emails"), "batch_size": self.conf.get("batch_size", INDEX_BATCH_SIZE), } self.connector = GoogleDriveConnector(**connector_kwargs) self.connector.set_allow_images(self.conf.get("allow_images", False)) credentials = self.conf.get("credentials") if not credentials: raise ValueError("Google Drive connector is missing credentials.") new_credentials = self.connector.load_credentials(credentials) if new_credentials: self._persist_rotated_credentials(task["connector_id"], new_credentials) if task["reindex"] == "1" or not task["poll_range_start"]: start_time = 0.0 begin_info = "totally" else: start_time = task["poll_range_start"].timestamp() begin_info = f"from {task['poll_range_start']}" end_time = datetime.now(timezone.utc).timestamp() raw_batch_size = self.conf.get("sync_batch_size") or self.conf.get("batch_size") or INDEX_BATCH_SIZE try: batch_size = int(raw_batch_size) except (TypeError, ValueError): batch_size = INDEX_BATCH_SIZE if batch_size <= 0: batch_size = INDEX_BATCH_SIZE def document_batches(): checkpoint = self.connector.build_dummy_checkpoint() pending_docs = [] iterations = 0 iteration_limit = 100_000 while checkpoint.has_more: wrapper = CheckpointOutputWrapper() doc_generator = wrapper(self.connector.load_from_checkpoint(start_time, end_time, checkpoint)) for document, failure, next_checkpoint in doc_generator: if failure is not None: logging.warning("Google Drive connector failure: %s", getattr(failure, "failure_message", failure)) continue if document is not None: pending_docs.append(document) if len(pending_docs) >= batch_size: yield pending_docs pending_docs = [] if next_checkpoint is not None: checkpoint = next_checkpoint iterations += 1 if iterations > iteration_limit: raise RuntimeError("Too many iterations while loading Google Drive documents.") if pending_docs: yield pending_docs try: admin_email = self.connector.primary_admin_email except RuntimeError: admin_email = "unknown" logging.info(f"Connect to Google Drive as {admin_email} {begin_info}") return document_batches() def _persist_rotated_credentials(self, connector_id: str, credentials: dict[str, Any]) -> None: try: updated_conf = copy.deepcopy(self.conf) updated_conf["credentials"] = credentials ConnectorService.update_by_id(connector_id, {"config": updated_conf}) self.conf = updated_conf logging.info("Persisted refreshed Google Drive credentials for connector %s", connector_id) except Exception: logging.exception("Failed to persist refreshed Google Drive credentials for connector %s", connector_id) class Jira(SyncBase): SOURCE_NAME: str = FileSource.JIRA def _get_source_prefix(self): return "[Jira]" async def _generate(self, task: dict): connector_kwargs = { "jira_base_url": self.conf["base_url"], "project_key": self.conf.get("project_key"), "jql_query": self.conf.get("jql_query"), "batch_size": self.conf.get("batch_size", INDEX_BATCH_SIZE), "include_comments": self.conf.get("include_comments", True), "include_attachments": self.conf.get("include_attachments", False), "labels_to_skip": self._normalize_list(self.conf.get("labels_to_skip")), "comment_email_blacklist": self._normalize_list(self.conf.get("comment_email_blacklist")), "scoped_token": self.conf.get("scoped_token", False), "attachment_size_limit": self.conf.get("attachment_size_limit"), "timezone_offset": self.conf.get("timezone_offset"), } self.connector = JiraConnector(**connector_kwargs) credentials = self.conf.get("credentials") if not credentials: raise ValueError("Jira connector is missing credentials.") self.connector.load_credentials(credentials) self.connector.validate_connector_settings() if task["reindex"] == "1" or not task["poll_range_start"]: start_time = 0.0 begin_info = "totally" else: start_time = task["poll_range_start"].timestamp() begin_info = f"from {task['poll_range_start']}" end_time = datetime.now(timezone.utc).timestamp() raw_batch_size = self.conf.get("sync_batch_size") or self.conf.get("batch_size") or INDEX_BATCH_SIZE try: batch_size = int(raw_batch_size) except (TypeError, ValueError): batch_size = INDEX_BATCH_SIZE if batch_size <= 0: batch_size = INDEX_BATCH_SIZE def document_batches(): checkpoint = self.connector.build_dummy_checkpoint() pending_docs = [] iterations = 0 iteration_limit = 100_000 while checkpoint.has_more: wrapper = CheckpointOutputWrapper() generator = wrapper( self.connector.load_from_checkpoint( start_time, end_time, checkpoint, ) ) for document, failure, next_checkpoint in generator: if failure is not None: logging.warning( f"[Jira] Jira connector failure: {getattr(failure, 'failure_message', failure)}" ) continue if document is not None: pending_docs.append(document) if len(pending_docs) >= batch_size: yield pending_docs pending_docs = [] if next_checkpoint is not None: checkpoint = next_checkpoint iterations += 1 if iterations > iteration_limit: logging.error(f"[Jira] Task {task.get('id')} exceeded iteration limit ({iteration_limit}).") raise RuntimeError("Too many iterations while loading Jira documents.") if pending_docs: yield pending_docs logging.info(f"[Jira] Connect to Jira {connector_kwargs['jira_base_url']} {begin_info}") return document_batches() @staticmethod def _normalize_list(values: Any) -> list[str] | None: if values is None: return None if isinstance(values, str): values = [item.strip() for item in values.split(",")] return [str(value).strip() for value in values if value is not None and str(value).strip()] class SharePoint(SyncBase): SOURCE_NAME: str = FileSource.SHAREPOINT async def _generate(self, task: dict): pass class Slack(SyncBase): SOURCE_NAME: str = FileSource.SLACK async def _generate(self, task: dict): pass class Teams(SyncBase): SOURCE_NAME: str = FileSource.TEAMS async def _generate(self, task: dict): pass class WebDAV(SyncBase): SOURCE_NAME: str = FileSource.WEBDAV async def _generate(self, task: dict): self.connector = WebDAVConnector( base_url=self.conf["base_url"], remote_path=self.conf.get("remote_path", "/") ) self.connector.load_credentials(self.conf["credentials"]) logging.info(f"Task info: reindex={task['reindex']}, poll_range_start={task['poll_range_start']}") if task["reindex"] == "1" or not task["poll_range_start"]: logging.info("Using load_from_state (full sync)") document_batch_generator = self.connector.load_from_state() begin_info = "totally" else: start_ts = task["poll_range_start"].timestamp() end_ts = datetime.now(timezone.utc).timestamp() logging.info(f"Polling WebDAV from {task['poll_range_start']} (ts: {start_ts}) to now (ts: {end_ts})") document_batch_generator = self.connector.poll_source(start_ts, end_ts) begin_info = "from {}".format(task["poll_range_start"]) logging.info("Connect to WebDAV: {}(path: {}) {}".format( self.conf["base_url"], self.conf.get("remote_path", "/"), begin_info )) def wrapper(): for document_batch in document_batch_generator: yield document_batch return wrapper() class Moodle(SyncBase): SOURCE_NAME: str = FileSource.MOODLE async def _generate(self, task: dict): self.connector = MoodleConnector( moodle_url=self.conf["moodle_url"], batch_size=self.conf.get("batch_size", INDEX_BATCH_SIZE) ) self.connector.load_credentials(self.conf["credentials"]) # Determine the time range for synchronization based on reindex or poll_range_start poll_start = task.get("poll_range_start") if task["reindex"] == "1" or poll_start is None: document_generator = self.connector.load_from_state() begin_info = "totally" else: document_generator = self.connector.poll_source( poll_start.timestamp(), datetime.now(timezone.utc).timestamp(), ) begin_info = f"from {poll_start}" logging.info("Connect to Moodle: {} {}".format(self.conf["moodle_url"], begin_info)) return document_generator class BOX(SyncBase): SOURCE_NAME: str = FileSource.BOX async def _generate(self, task: dict): self.connector = BoxConnector( folder_id=self.conf.get("folder_id", "0"), ) credential = json.loads(self.conf['credentials']['box_tokens']) auth = BoxOAuth( OAuthConfig( client_id=credential['client_id'], client_secret=credential['client_secret'], ) ) token = AccessToken( access_token=credential['access_token'], refresh_token=credential['refresh_token'], ) auth.token_storage.store(token) self.connector.load_credentials(auth) poll_start = task["poll_range_start"] if task["reindex"] == "1" or poll_start is None: document_generator = self.connector.load_from_state() begin_info = "totally" else: document_generator = self.connector.poll_source( poll_start.timestamp(), datetime.now(timezone.utc).timestamp(), ) begin_info = f"from {poll_start}" logging.info("Connect to Box: folder_id({}) {}".format(self.conf["folder_id"], begin_info)) return document_generator class Airtable(SyncBase): SOURCE_NAME: str = FileSource.AIRTABLE async def _generate(self, task: dict): """ Sync files from Airtable attachments. """ self.connector = AirtableConnector( base_id=self.conf.get("base_id"), table_name_or_id=self.conf.get("table_name_or_id"), ) credentials = self.conf.get("credentials", {}) if "airtable_access_token" not in credentials: raise ValueError("Missing airtable_access_token in credentials") self.connector.load_credentials( {"airtable_access_token": credentials["airtable_access_token"]} ) poll_start = task.get("poll_range_start") if task.get("reindex") == "1" or poll_start is None: document_generator = self.connector.load_from_state() begin_info = "totally" else: document_generator = self.connector.poll_source( poll_start.timestamp(), datetime.now(timezone.utc).timestamp(), ) begin_info = f"from {poll_start}" logging.info( "Connect to Airtable: base_id(%s), table(%s) %s", self.conf.get("base_id"), self.conf.get("table_name_or_id"), begin_info, ) return document_generator class Asana(SyncBase): SOURCE_NAME: str = FileSource.ASANA async def _generate(self, task: dict): self.connector = AsanaConnector( self.conf.get("asana_workspace_id"), self.conf.get("asana_project_ids"), self.conf.get("asana_team_id"), ) credentials = self.conf.get("credentials", {}) if "asana_api_token_secret" not in credentials: raise ValueError("Missing asana_api_token_secret in credentials") self.connector.load_credentials( {"asana_api_token_secret": credentials["asana_api_token_secret"]} ) if task.get("reindex") == "1" or not task.get("poll_range_start"): document_generator = self.connector.load_from_state() begin_info = "totally" else: poll_start = task.get("poll_range_start") if poll_start is None: document_generator = self.connector.load_from_state() begin_info = "totally" else: document_generator = self.connector.poll_source( poll_start.timestamp(), datetime.now(timezone.utc).timestamp(), ) begin_info = f"from {poll_start}" logging.info(
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/svr/discord_svr.py
rag/svr/discord_svr.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import discord import requests import base64 import asyncio URL = '{YOUR_IP_ADDRESS:PORT}/v1/api/completion_aibotk' # Default: https://demo.ragflow.io/v1/api/completion_aibotk JSON_DATA = { "conversation_id": "xxxxxxxxxxxxxxxxxxxxxxxxxxx", # Get conversation id from /api/new_conversation "Authorization": "ragflow-xxxxxxxxxxxxxxxxxxxxxxxxxxxxx", # RAGFlow Assistant Chat Bot API Key "word": "" # User question, don't need to initialize } DISCORD_BOT_KEY = "xxxxxxxxxxxxxxxxxxxxxxxxxx" # Get DISCORD_BOT_KEY from Discord Application intents = discord.Intents.default() intents.message_content = True client = discord.Client(intents=intents) @client.event async def on_ready(): logging.info(f'We have logged in as {client.user}') @client.event async def on_message(message): if message.author == client.user: return if client.user.mentioned_in(message): if len(message.content.split('> ')) == 1: await message.channel.send("Hi~ How can I help you? ") else: JSON_DATA['word'] = message.content.split('> ')[1] response = requests.post(URL, json=JSON_DATA) response_data = response.json().get('data', []) image_bool = False for i in response_data: if i['type'] == 1: res = i['content'] if i['type'] == 3: image_bool = True image_data = base64.b64decode(i['url']) with open('tmp_image.png', 'wb') as file: file.write(image_data) image = discord.File('tmp_image.png') await message.channel.send(f"{message.author.mention}{res}") if image_bool: await message.channel.send(file=image) loop = asyncio.get_event_loop() try: loop.run_until_complete(client.start(DISCORD_BOT_KEY)) except KeyboardInterrupt: loop.run_until_complete(client.close()) finally: loop.close()
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/infinity_conn.py
rag/utils/infinity_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import re import json import copy from infinity.common import InfinityException, SortType from infinity.errors import ErrorCode from common.decorator import singleton import pandas as pd from common.constants import PAGERANK_FLD, TAG_FLD from common.doc_store.doc_store_base import MatchExpr, MatchTextExpr, MatchDenseExpr, FusionExpr, OrderByExpr from common.doc_store.infinity_conn_base import InfinityConnectionBase @singleton class InfinityConnection(InfinityConnectionBase): """ Dataframe and fields convert """ @staticmethod def field_keyword(field_name: str): # Treat "*_kwd" tag-like columns as keyword lists except knowledge_graph_kwd; source_id is also keyword-like. if field_name == "source_id" or ( field_name.endswith("_kwd") and field_name not in ["knowledge_graph_kwd", "docnm_kwd", "important_kwd", "question_kwd"]): return True return False def convert_select_fields(self, output_fields: list[str]) -> list[str]: for i, field in enumerate(output_fields): if field in ["docnm_kwd", "title_tks", "title_sm_tks"]: output_fields[i] = "docnm" elif field in ["important_kwd", "important_tks"]: output_fields[i] = "important_keywords" elif field in ["question_kwd", "question_tks"]: output_fields[i] = "questions" elif field in ["content_with_weight", "content_ltks", "content_sm_ltks"]: output_fields[i] = "content" elif field in ["authors_tks", "authors_sm_tks"]: output_fields[i] = "authors" return list(set(output_fields)) @staticmethod def convert_matching_field(field_weight_str: str) -> str: tokens = field_weight_str.split("^") field = tokens[0] if field == "docnm_kwd" or field == "title_tks": field = "docnm@ft_docnm_rag_coarse" elif field == "title_sm_tks": field = "docnm@ft_docnm_rag_fine" elif field == "important_kwd": field = "important_keywords@ft_important_keywords_rag_coarse" elif field == "important_tks": field = "important_keywords@ft_important_keywords_rag_fine" elif field == "question_kwd": field = "questions@ft_questions_rag_coarse" elif field == "question_tks": field = "questions@ft_questions_rag_fine" elif field == "content_with_weight" or field == "content_ltks": field = "content@ft_content_rag_coarse" elif field == "content_sm_ltks": field = "content@ft_content_rag_fine" elif field == "authors_tks": field = "authors@ft_authors_rag_coarse" elif field == "authors_sm_tks": field = "authors@ft_authors_rag_fine" tokens[0] = field return "^".join(tokens) """ CRUD operations """ def search( self, select_fields: list[str], highlight_fields: list[str], condition: dict, match_expressions: list[MatchExpr], order_by: OrderByExpr, offset: int, limit: int, index_names: str | list[str], knowledgebase_ids: list[str], agg_fields: list[str] | None = None, rank_feature: dict | None = None, ) -> tuple[pd.DataFrame, int]: """ BUG: Infinity returns empty for a highlight field if the query string doesn't use that field. """ if isinstance(index_names, str): index_names = index_names.split(",") assert isinstance(index_names, list) and len(index_names) > 0 inf_conn = self.connPool.get_conn() db_instance = inf_conn.get_database(self.dbName) df_list = list() table_list = list() output = select_fields.copy() output = self.convert_select_fields(output) if agg_fields is None: agg_fields = [] for essential_field in ["id"] + agg_fields: if essential_field not in output: output.append(essential_field) score_func = "" score_column = "" for matchExpr in match_expressions: if isinstance(matchExpr, MatchTextExpr): score_func = "score()" score_column = "SCORE" break if not score_func: for matchExpr in match_expressions: if isinstance(matchExpr, MatchDenseExpr): score_func = "similarity()" score_column = "SIMILARITY" break if match_expressions: if score_func not in output: output.append(score_func) if PAGERANK_FLD not in output: output.append(PAGERANK_FLD) output = [f for f in output if f != "_score"] if limit <= 0: # ElasticSearch default limit is 10000 limit = 10000 # Prepare expressions common to all tables filter_cond = None filter_fulltext = "" if condition: table_found = False for indexName in index_names: for kb_id in knowledgebase_ids: table_name = f"{indexName}_{kb_id}" try: filter_cond = self.equivalent_condition_to_str(condition, db_instance.get_table(table_name)) table_found = True break except Exception: pass if table_found: break if not table_found: self.logger.error( f"No valid tables found for indexNames {index_names} and knowledgebaseIds {knowledgebase_ids}") return pd.DataFrame(), 0 for matchExpr in match_expressions: if isinstance(matchExpr, MatchTextExpr): if filter_cond and "filter" not in matchExpr.extra_options: matchExpr.extra_options.update({"filter": filter_cond}) matchExpr.fields = [self.convert_matching_field(field) for field in matchExpr.fields] fields = ",".join(matchExpr.fields) filter_fulltext = f"filter_fulltext('{fields}', '{matchExpr.matching_text}')" if filter_cond: filter_fulltext = f"({filter_cond}) AND {filter_fulltext}" minimum_should_match = matchExpr.extra_options.get("minimum_should_match", 0.0) if isinstance(minimum_should_match, float): str_minimum_should_match = str(int(minimum_should_match * 100)) + "%" matchExpr.extra_options["minimum_should_match"] = str_minimum_should_match # Add rank_feature support if rank_feature and "rank_features" not in matchExpr.extra_options: # Convert rank_feature dict to Infinity's rank_features string format # Format: "field^feature_name^weight,field^feature_name^weight" rank_features_list = [] for feature_name, weight in rank_feature.items(): # Use TAG_FLD as the field containing rank features rank_features_list.append(f"{TAG_FLD}^{feature_name}^{weight}") if rank_features_list: matchExpr.extra_options["rank_features"] = ",".join(rank_features_list) for k, v in matchExpr.extra_options.items(): if not isinstance(v, str): matchExpr.extra_options[k] = str(v) self.logger.debug(f"INFINITY search MatchTextExpr: {json.dumps(matchExpr.__dict__)}") elif isinstance(matchExpr, MatchDenseExpr): if filter_fulltext and "filter" not in matchExpr.extra_options: matchExpr.extra_options.update({"filter": filter_fulltext}) for k, v in matchExpr.extra_options.items(): if not isinstance(v, str): matchExpr.extra_options[k] = str(v) similarity = matchExpr.extra_options.get("similarity") if similarity: matchExpr.extra_options["threshold"] = similarity del matchExpr.extra_options["similarity"] self.logger.debug(f"INFINITY search MatchDenseExpr: {json.dumps(matchExpr.__dict__)}") elif isinstance(matchExpr, FusionExpr): if matchExpr.method == "weighted_sum": # The default is "minmax" which gives a zero score for the last doc. matchExpr.fusion_params["normalize"] = "atan" self.logger.debug(f"INFINITY search FusionExpr: {json.dumps(matchExpr.__dict__)}") order_by_expr_list = list() if order_by.fields: for order_field in order_by.fields: if order_field[1] == 0: order_by_expr_list.append((order_field[0], SortType.Asc)) else: order_by_expr_list.append((order_field[0], SortType.Desc)) total_hits_count = 0 # Scatter search tables and gather the results for indexName in index_names: for knowledgebaseId in knowledgebase_ids: table_name = f"{indexName}_{knowledgebaseId}" try: table_instance = db_instance.get_table(table_name) except Exception: continue table_list.append(table_name) builder = table_instance.output(output) if len(match_expressions) > 0: for matchExpr in match_expressions: if isinstance(matchExpr, MatchTextExpr): fields = ",".join(matchExpr.fields) builder = builder.match_text( fields, matchExpr.matching_text, matchExpr.topn, matchExpr.extra_options.copy(), ) elif isinstance(matchExpr, MatchDenseExpr): builder = builder.match_dense( matchExpr.vector_column_name, matchExpr.embedding_data, matchExpr.embedding_data_type, matchExpr.distance_type, matchExpr.topn, matchExpr.extra_options.copy(), ) elif isinstance(matchExpr, FusionExpr): builder = builder.fusion(matchExpr.method, matchExpr.topn, matchExpr.fusion_params) else: if filter_cond and len(filter_cond) > 0: builder.filter(filter_cond) if order_by.fields: builder.sort(order_by_expr_list) builder.offset(offset).limit(limit) kb_res, extra_result = builder.option({"total_hits_count": True}).to_df() if extra_result: total_hits_count += int(extra_result["total_hits_count"]) self.logger.debug(f"INFINITY search table: {str(table_name)}, result: {str(kb_res)}") df_list.append(kb_res) self.connPool.release_conn(inf_conn) res = self.concat_dataframes(df_list, output) if match_expressions: res["_score"] = res[score_column] + res[PAGERANK_FLD] res = res.sort_values(by="_score", ascending=False).reset_index(drop=True) res = res.head(limit) self.logger.debug(f"INFINITY search final result: {str(res)}") return res, total_hits_count def get(self, chunk_id: str, index_name: str, knowledgebase_ids: list[str]) -> dict | None: inf_conn = self.connPool.get_conn() db_instance = inf_conn.get_database(self.dbName) df_list = list() assert isinstance(knowledgebase_ids, list) table_list = list() for knowledgebaseId in knowledgebase_ids: table_name = f"{index_name}_{knowledgebaseId}" table_list.append(table_name) try: table_instance = db_instance.get_table(table_name) except Exception: self.logger.warning( f"Table not found: {table_name}, this dataset isn't created in Infinity. Maybe it is created in other document engine.") continue kb_res, _ = table_instance.output(["*"]).filter(f"id = '{chunk_id}'").to_df() self.logger.debug(f"INFINITY get table: {str(table_list)}, result: {str(kb_res)}") df_list.append(kb_res) self.connPool.release_conn(inf_conn) res = self.concat_dataframes(df_list, ["id"]) fields = set(res.columns.tolist()) for field in ["docnm_kwd", "title_tks", "title_sm_tks", "important_kwd", "important_tks", "question_kwd", "question_tks", "content_with_weight", "content_ltks", "content_sm_ltks", "authors_tks", "authors_sm_tks"]: fields.add(field) res_fields = self.get_fields(res, list(fields)) return res_fields.get(chunk_id, None) def insert(self, documents: list[dict], index_name: str, knowledgebase_id: str = None) -> list[str]: inf_conn = self.connPool.get_conn() db_instance = inf_conn.get_database(self.dbName) table_name = f"{index_name}_{knowledgebase_id}" try: table_instance = db_instance.get_table(table_name) except InfinityException as e: # src/common/status.cppm, kTableNotExist = 3022 if e.error_code != ErrorCode.TABLE_NOT_EXIST: raise vector_size = 0 patt = re.compile(r"q_(?P<vector_size>\d+)_vec") for k in documents[0].keys(): m = patt.match(k) if m: vector_size = int(m.group("vector_size")) break if vector_size == 0: raise ValueError("Cannot infer vector size from documents") self.create_idx(index_name, knowledgebase_id, vector_size) table_instance = db_instance.get_table(table_name) # embedding fields can't have a default value.... embedding_clmns = [] clmns = table_instance.show_columns().rows() for n, ty, _, _ in clmns: r = re.search(r"Embedding\([a-z]+,([0-9]+)\)", ty) if not r: continue embedding_clmns.append((n, int(r.group(1)))) docs = copy.deepcopy(documents) for d in docs: assert "_id" not in d assert "id" in d for k, v in list(d.items()): if k == "docnm_kwd": d["docnm"] = v elif k == "title_kwd": if not d.get("docnm_kwd"): d["docnm"] = self.list2str(v) elif k == "title_sm_tks": if not d.get("docnm_kwd"): d["docnm"] = self.list2str(v) elif k == "important_kwd": d["important_keywords"] = self.list2str(v) elif k == "important_tks": if not d.get("important_kwd"): d["important_keywords"] = v elif k == "content_with_weight": d["content"] = v elif k == "content_ltks": if not d.get("content_with_weight"): d["content"] = v elif k == "content_sm_ltks": if not d.get("content_with_weight"): d["content"] = v elif k == "authors_tks": d["authors"] = v elif k == "authors_sm_tks": if not d.get("authors_tks"): d["authors"] = v elif k == "question_kwd": d["questions"] = self.list2str(v, "\n") elif k == "question_tks": if not d.get("question_kwd"): d["questions"] = self.list2str(v) elif self.field_keyword(k): if isinstance(v, list): d[k] = "###".join(v) else: d[k] = v elif re.search(r"_feas$", k): d[k] = json.dumps(v) elif k == "kb_id": if isinstance(d[k], list): d[k] = d[k][0] # since d[k] is a list, but we need a str elif k == "position_int": assert isinstance(v, list) arr = [num for row in v for num in row] d[k] = "_".join(f"{num:08x}" for num in arr) elif k in ["page_num_int", "top_int"]: assert isinstance(v, list) d[k] = "_".join(f"{num:08x}" for num in v) else: d[k] = v for k in ["docnm_kwd", "title_tks", "title_sm_tks", "important_kwd", "important_tks", "content_with_weight", "content_ltks", "content_sm_ltks", "authors_tks", "authors_sm_tks", "question_kwd", "question_tks"]: if k in d: del d[k] for n, vs in embedding_clmns: if n in d: continue d[n] = [0] * vs ids = ["'{}'".format(d["id"]) for d in docs] str_ids = ", ".join(ids) str_filter = f"id IN ({str_ids})" table_instance.delete(str_filter) # for doc in documents: # logger.info(f"insert position_int: {doc['position_int']}") # logger.info(f"InfinityConnection.insert {json.dumps(documents)}") table_instance.insert(docs) self.connPool.release_conn(inf_conn) self.logger.debug(f"INFINITY inserted into {table_name} {str_ids}.") return [] def update(self, condition: dict, new_value: dict, index_name: str, knowledgebase_id: str) -> bool: # if 'position_int' in newValue: # logger.info(f"update position_int: {newValue['position_int']}") inf_conn = self.connPool.get_conn() db_instance = inf_conn.get_database(self.dbName) table_name = f"{index_name}_{knowledgebase_id}" table_instance = db_instance.get_table(table_name) # if "exists" in condition: # del condition["exists"] clmns = {} if table_instance: for n, ty, de, _ in table_instance.show_columns().rows(): clmns[n] = (ty, de) filter = self.equivalent_condition_to_str(condition, table_instance) removeValue = {} for k, v in list(new_value.items()): if k == "docnm_kwd": new_value["docnm"] = self.list2str(v) elif k == "title_kwd": if not new_value.get("docnm_kwd"): new_value["docnm"] = self.list2str(v) elif k == "title_sm_tks": if not new_value.get("docnm_kwd"): new_value["docnm"] = v elif k == "important_kwd": new_value["important_keywords"] = self.list2str(v) elif k == "important_tks": if not new_value.get("important_kwd"): new_value["important_keywords"] = v elif k == "content_with_weight": new_value["content"] = v elif k == "content_ltks": if not new_value.get("content_with_weight"): new_value["content"] = v elif k == "content_sm_ltks": if not new_value.get("content_with_weight"): new_value["content"] = v elif k == "authors_tks": new_value["authors"] = v elif k == "authors_sm_tks": if not new_value.get("authors_tks"): new_value["authors"] = v elif k == "question_kwd": new_value["questions"] = "\n".join(v) elif k == "question_tks": if not new_value.get("question_kwd"): new_value["questions"] = self.list2str(v) elif self.field_keyword(k): if isinstance(v, list): new_value[k] = "###".join(v) else: new_value[k] = v elif re.search(r"_feas$", k): new_value[k] = json.dumps(v) elif k == "kb_id": if isinstance(new_value[k], list): new_value[k] = new_value[k][0] # since d[k] is a list, but we need a str elif k == "position_int": assert isinstance(v, list) arr = [num for row in v for num in row] new_value[k] = "_".join(f"{num:08x}" for num in arr) elif k in ["page_num_int", "top_int"]: assert isinstance(v, list) new_value[k] = "_".join(f"{num:08x}" for num in v) elif k == "remove": if isinstance(v, str): assert v in clmns, f"'{v}' should be in '{clmns}'." ty, de = clmns[v] if ty.lower().find("cha"): if not de: de = "" new_value[v] = de else: for kk, vv in v.items(): removeValue[kk] = vv del new_value[k] else: new_value[k] = v for k in ["docnm_kwd", "title_tks", "title_sm_tks", "important_kwd", "important_tks", "content_with_weight", "content_ltks", "content_sm_ltks", "authors_tks", "authors_sm_tks", "question_kwd", "question_tks"]: if k in new_value: del new_value[k] remove_opt = {} # "[k,new_value]": [id_to_update, ...] if removeValue: col_to_remove = list(removeValue.keys()) row_to_opt = table_instance.output(col_to_remove + ["id"]).filter(filter).to_df() self.logger.debug(f"INFINITY search table {str(table_name)}, filter {filter}, result: {str(row_to_opt[0])}") row_to_opt = self.get_fields(row_to_opt, col_to_remove) for id, old_v in row_to_opt.items(): for k, remove_v in removeValue.items(): if remove_v in old_v[k]: new_v = old_v[k].copy() new_v.remove(remove_v) kv_key = json.dumps([k, new_v]) if kv_key not in remove_opt: remove_opt[kv_key] = [id] else: remove_opt[kv_key].append(id) self.logger.debug(f"INFINITY update table {table_name}, filter {filter}, newValue {new_value}.") for update_kv, ids in remove_opt.items(): k, v = json.loads(update_kv) table_instance.update(filter + " AND id in ({0})".format(",".join([f"'{id}'" for id in ids])), {k: "###".join(v)}) table_instance.update(filter, new_value) self.connPool.release_conn(inf_conn) return True """ Helper functions for search result """ def get_fields(self, res: tuple[pd.DataFrame, int] | pd.DataFrame, fields: list[str]) -> dict[str, dict]: if isinstance(res, tuple): res = res[0] if not fields: return {} fields_all = fields.copy() fields_all.append("id") fields_all = set(fields_all) if "docnm" in res.columns: for field in ["docnm_kwd", "title_tks", "title_sm_tks"]: if field in fields_all: res[field] = res["docnm"] if "important_keywords" in res.columns: if "important_kwd" in fields_all: res["important_kwd"] = res["important_keywords"].apply(lambda v: v.split()) if "important_tks" in fields_all: res["important_tks"] = res["important_keywords"] if "questions" in res.columns: if "question_kwd" in fields_all: res["question_kwd"] = res["questions"].apply(lambda v: v.splitlines()) if "question_tks" in fields_all: res["question_tks"] = res["questions"] if "content" in res.columns: for field in ["content_with_weight", "content_ltks", "content_sm_ltks"]: if field in fields_all: res[field] = res["content"] if "authors" in res.columns: for field in ["authors_tks", "authors_sm_tks"]: if field in fields_all: res[field] = res["authors"] column_map = {col.lower(): col for col in res.columns} matched_columns = {column_map[col.lower()]: col for col in fields_all if col.lower() in column_map} none_columns = [col for col in fields_all if col.lower() not in column_map] res2 = res[matched_columns.keys()] res2 = res2.rename(columns=matched_columns) res2.drop_duplicates(subset=["id"], inplace=True) for column in list(res2.columns): k = column.lower() if self.field_keyword(k): res2[column] = res2[column].apply(lambda v: [kwd for kwd in v.split("###") if kwd]) elif re.search(r"_feas$", k): res2[column] = res2[column].apply(lambda v: json.loads(v) if v else {}) elif k == "position_int": def to_position_int(v): if v: arr = [int(hex_val, 16) for hex_val in v.split("_")] v = [arr[i: i + 5] for i in range(0, len(arr), 5)] else: v = [] return v res2[column] = res2[column].apply(to_position_int) elif k in ["page_num_int", "top_int"]: res2[column] = res2[column].apply(lambda v: [int(hex_val, 16) for hex_val in v.split("_")] if v else []) else: pass for column in ["docnm", "important_keywords", "questions", "content", "authors"]: if column in res2: del res2[column] for column in none_columns: res2[column] = None return res2.set_index("id").to_dict(orient="index")
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/file_utils.py
rag/utils/file_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import io import hashlib import zipfile import requests from requests.exceptions import Timeout, RequestException from io import BytesIO from typing import List, Union, Tuple, Optional, Dict import PyPDF2 from docx import Document import olefile def _is_zip(h: bytes) -> bool: return h.startswith(b"PK\x03\x04") or h.startswith(b"PK\x05\x06") or h.startswith(b"PK\x07\x08") def _is_pdf(h: bytes) -> bool: return h.startswith(b"%PDF-") def _is_ole(h: bytes) -> bool: return h.startswith(b"\xD0\xCF\x11\xE0\xA1\xB1\x1A\xE1") def _sha10(b: bytes) -> str: return hashlib.sha256(b).hexdigest()[:10] def _guess_ext(b: bytes) -> str: h = b[:8] if _is_zip(h): try: with zipfile.ZipFile(io.BytesIO(b), "r") as z: names = [n.lower() for n in z.namelist()] if any(n.startswith("word/") for n in names): return ".docx" if any(n.startswith("ppt/") for n in names): return ".pptx" if any(n.startswith("xl/") for n in names): return ".xlsx" except Exception: pass return ".zip" if _is_pdf(h): return ".pdf" if _is_ole(h): return ".doc" return ".bin" # Try to extract the real embedded payload from OLE's Ole10Native def _extract_ole10native_payload(data: bytes) -> bytes: try: pos = 0 if len(data) < 4: return data _ = int.from_bytes(data[pos:pos + 4], "little") pos += 4 # filename/src/tmp (NUL-terminated ANSI) for _ in range(3): z = data.index(b"\x00", pos) pos = z + 1 # skip unknown 4 bytes pos += 4 if pos + 4 > len(data): return data size = int.from_bytes(data[pos:pos + 4], "little") pos += 4 if pos + size <= len(data): return data[pos:pos + size] except Exception: pass return data def extract_embed_file(target: Union[bytes, bytearray]) -> List[Tuple[str, bytes]]: """ Only extract the 'first layer' of embedding, returning raw (filename, bytes). """ top = bytes(target) head = top[:8] out: List[Tuple[str, bytes]] = [] seen = set() def push(b: bytes, name_hint: str = ""): h10 = _sha10(b) if h10 in seen: return seen.add(h10) ext = _guess_ext(b) # If name_hint has an extension use its basename; else fallback to guessed ext if "." in name_hint: fname = name_hint.split("/")[-1] else: fname = f"{h10}{ext}" out.append((fname, b)) # OOXML/ZIP container (docx/xlsx/pptx) if _is_zip(head): try: with zipfile.ZipFile(io.BytesIO(top), "r") as z: embed_dirs = ( "word/embeddings/", "word/objects/", "word/activex/", "xl/embeddings/", "ppt/embeddings/" ) for name in z.namelist(): low = name.lower() if any(low.startswith(d) for d in embed_dirs): try: b = z.read(name) push(b, name) except Exception: pass except Exception: pass return out # OLE container (doc/ppt/xls) if _is_ole(head): try: with olefile.OleFileIO(io.BytesIO(top)) as ole: for entry in ole.listdir(): p = "/".join(entry) try: data = ole.openstream(entry).read() except Exception: continue if not data: continue if "Ole10Native" in p or "ole10native" in p.lower(): data = _extract_ole10native_payload(data) push(data, p) except Exception: pass return out return out def extract_links_from_docx(docx_bytes: bytes): """ Extract all hyperlinks from a Word (.docx) document binary stream. Args: docx_bytes (bytes): Raw bytes of a .docx file. Returns: set[str]: A set of unique hyperlink URLs. """ links = set() with BytesIO(docx_bytes) as bio: document = Document(bio) # Each relationship may represent a hyperlink, image, footer, etc. for rel in document.part.rels.values(): if rel.reltype == ( "http://schemas.openxmlformats.org/officeDocument/2006/relationships/hyperlink" ): links.add(rel.target_ref) return links def extract_links_from_pdf(pdf_bytes: bytes): """ Extract all clickable hyperlinks from a PDF binary stream. Args: pdf_bytes (bytes): Raw bytes of a PDF file. Returns: set[str]: A set of unique hyperlink URLs (unordered). """ links = set() with BytesIO(pdf_bytes) as bio: pdf = PyPDF2.PdfReader(bio) for page in pdf.pages: annots = page.get("/Annots") if not annots or isinstance(annots, PyPDF2.generic.IndirectObject): continue for annot in annots: obj = annot.get_object() a = obj.get("/A") if a and a.get("/URI"): links.add(a["/URI"]) return links _GLOBAL_SESSION: Optional[requests.Session] = None def _get_session(headers: Optional[Dict[str, str]] = None) -> requests.Session: """Get or create a global reusable session.""" global _GLOBAL_SESSION if _GLOBAL_SESSION is None: _GLOBAL_SESSION = requests.Session() _GLOBAL_SESSION.headers.update({ "User-Agent": ( "Mozilla/5.0 (X11; Linux x86_64) " "AppleWebKit/537.36 (KHTML, like Gecko) " "Chrome/121.0 Safari/537.36" ) }) if headers: _GLOBAL_SESSION.headers.update(headers) return _GLOBAL_SESSION def extract_html( url: str, timeout: float = 60.0, headers: Optional[Dict[str, str]] = None, max_retries: int = 2, ) -> Tuple[Optional[bytes], Dict[str, str]]: """ Extract the full HTML page as raw bytes from a given URL. Automatically reuses a persistent HTTP session and applies robust timeout & retry logic. Args: url (str): Target webpage URL. timeout (float): Request timeout in seconds (applies to connect + read). headers (dict, optional): Extra HTTP headers. max_retries (int): Number of retries on timeout or transient errors. Returns: tuple(bytes|None, dict): - html_bytes: Raw HTML content (or None if failed) - metadata: HTTP info (status_code, content_type, final_url, error if any) """ session = _get_session(headers=headers) metadata = {"final_url": url, "status_code": "", "content_type": "", "error": ""} for attempt in range(1, max_retries + 1): try: resp = session.get(url, timeout=timeout) resp.raise_for_status() html_bytes = resp.content metadata.update({ "final_url": resp.url, "status_code": str(resp.status_code), "content_type": resp.headers.get("Content-Type", ""), }) return html_bytes, metadata except Timeout: metadata["error"] = f"Timeout after {timeout}s (attempt {attempt}/{max_retries})" if attempt >= max_retries: continue except RequestException as e: metadata["error"] = f"Request failed: {e}" continue return None, metadata
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/tavily_conn.py
rag/utils/tavily_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging from tavily import TavilyClient from common.misc_utils import get_uuid from rag.nlp import rag_tokenizer class Tavily: def __init__(self, api_key: str): self.tavily_client = TavilyClient(api_key=api_key) def search(self, query): try: response = self.tavily_client.search( query=query, search_depth="advanced", max_results=6 ) return [{"url": res["url"], "title": res["title"], "content": res["content"], "score": res["score"]} for res in response["results"]] except Exception as e: logging.exception(e) return [] def retrieve_chunks(self, question): chunks = [] aggs = [] logging.info("[Tavily]Q: " + question) for r in self.search(question): id = get_uuid() chunks.append({ "chunk_id": id, "content_ltks": rag_tokenizer.tokenize(r["content"]), "content_with_weight": r["content"], "doc_id": id, "docnm_kwd": r["title"], "kb_id": [], "important_kwd": [], "image_id": "", "similarity": r["score"], "vector_similarity": 1., "term_similarity": 0, "vector": [], "positions": [], "url": r["url"] }) aggs.append({ "doc_name": r["title"], "doc_id": id, "count": 1, "url": r["url"] }) logging.info("[Tavily]R: " + r["content"][:128] + "...") return {"chunks": chunks, "doc_aggs": aggs}
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/storage_factory.py
rag/utils/storage_factory.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/raptor_utils.py
rag/utils/raptor_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Utility functions for Raptor processing decisions. """ import logging from typing import Optional # File extensions for structured data types EXCEL_EXTENSIONS = {".xls", ".xlsx", ".xlsm", ".xlsb"} CSV_EXTENSIONS = {".csv", ".tsv"} STRUCTURED_EXTENSIONS = EXCEL_EXTENSIONS | CSV_EXTENSIONS def is_structured_file_type(file_type: Optional[str]) -> bool: """ Check if a file type is structured data (Excel, CSV, etc.) Args: file_type: File extension (e.g., ".xlsx", ".csv") Returns: True if file is structured data type """ if not file_type: return False # Normalize to lowercase and ensure leading dot file_type = file_type.lower() if not file_type.startswith("."): file_type = f".{file_type}" return file_type in STRUCTURED_EXTENSIONS def is_tabular_pdf(parser_id: str = "", parser_config: Optional[dict] = None) -> bool: """ Check if a PDF is being parsed as tabular data. Args: parser_id: Parser ID (e.g., "table", "naive") parser_config: Parser configuration dict Returns: True if PDF is being parsed as tabular data """ parser_config = parser_config or {} # If using table parser, it's tabular if parser_id and parser_id.lower() == "table": return True # Check if html4excel is enabled (Excel-like table parsing) if parser_config.get("html4excel", False): return True return False def should_skip_raptor( file_type: Optional[str] = None, parser_id: str = "", parser_config: Optional[dict] = None, raptor_config: Optional[dict] = None ) -> bool: """ Determine if Raptor should be skipped for a given document. This function implements the logic to automatically disable Raptor for: 1. Excel files (.xls, .xlsx, .csv, etc.) 2. PDFs with tabular data (using table parser or html4excel) Args: file_type: File extension (e.g., ".xlsx", ".pdf") parser_id: Parser ID being used parser_config: Parser configuration dict raptor_config: Raptor configuration dict (can override with auto_disable_for_structured_data) Returns: True if Raptor should be skipped, False otherwise """ parser_config = parser_config or {} raptor_config = raptor_config or {} # Check if auto-disable is explicitly disabled in config if raptor_config.get("auto_disable_for_structured_data", True) is False: logging.info("Raptor auto-disable is turned off via configuration") return False # Check for Excel/CSV files if is_structured_file_type(file_type): logging.info(f"Skipping Raptor for structured file type: {file_type}") return True # Check for tabular PDFs if file_type and file_type.lower() in [".pdf", "pdf"]: if is_tabular_pdf(parser_id, parser_config): logging.info(f"Skipping Raptor for tabular PDF (parser_id={parser_id})") return True return False def get_skip_reason( file_type: Optional[str] = None, parser_id: str = "", parser_config: Optional[dict] = None ) -> str: """ Get a human-readable reason why Raptor was skipped. Args: file_type: File extension parser_id: Parser ID being used parser_config: Parser configuration dict Returns: Reason string, or empty string if Raptor should not be skipped """ parser_config = parser_config or {} if is_structured_file_type(file_type): return f"Structured data file ({file_type}) - Raptor auto-disabled" if file_type and file_type.lower() in [".pdf", "pdf"]: if is_tabular_pdf(parser_id, parser_config): return f"Tabular PDF (parser={parser_id}) - Raptor auto-disabled" return ""
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/oss_conn.py
rag/utils/oss_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import logging import boto3 from botocore.exceptions import ClientError from botocore.config import Config import time from io import BytesIO from common.decorator import singleton from common import settings @singleton class RAGFlowOSS: def __init__(self): self.conn = None self.oss_config = settings.OSS self.access_key = self.oss_config.get('access_key', None) self.secret_key = self.oss_config.get('secret_key', None) self.endpoint_url = self.oss_config.get('endpoint_url', None) self.region = self.oss_config.get('region', None) self.bucket = self.oss_config.get('bucket', None) self.prefix_path = self.oss_config.get('prefix_path', None) self.__open__() @staticmethod def use_default_bucket(method): def wrapper(self, bucket, *args, **kwargs): # If there is a default bucket, use the default bucket actual_bucket = self.bucket if self.bucket else bucket return method(self, actual_bucket, *args, **kwargs) return wrapper @staticmethod def use_prefix_path(method): def wrapper(self, bucket, fnm, *args, **kwargs): # If the prefix path is set, use the prefix path fnm = f"{self.prefix_path}/{fnm}" if self.prefix_path else fnm return method(self, bucket, fnm, *args, **kwargs) return wrapper def __open__(self): try: if self.conn: self.__close__() except Exception: pass try: # Reference:https://help.aliyun.com/zh/oss/developer-reference/use-amazon-s3-sdks-to-access-oss self.conn = boto3.client( 's3', region_name=self.region, aws_access_key_id=self.access_key, aws_secret_access_key=self.secret_key, endpoint_url=self.endpoint_url, config=Config(s3={"addressing_style": "virtual"}, signature_version='v4') ) except Exception: logging.exception(f"Fail to connect at region {self.region}") def __close__(self): del self.conn self.conn = None @use_default_bucket def bucket_exists(self, bucket): try: logging.debug(f"head_bucket bucketname {bucket}") self.conn.head_bucket(Bucket=bucket) exists = True except ClientError: logging.exception(f"head_bucket error {bucket}") exists = False return exists def health(self): bucket = self.bucket fnm = "txtxtxtxt1" fnm, binary = f"{self.prefix_path}/{fnm}" if self.prefix_path else fnm, b"_t@@@1" if not self.bucket_exists(bucket): self.conn.create_bucket(Bucket=bucket) logging.debug(f"create bucket {bucket} ********") r = self.conn.upload_fileobj(BytesIO(binary), bucket, fnm) return r def get_properties(self, bucket, key): return {} def list(self, bucket, dir, recursive=True): return [] @use_prefix_path @use_default_bucket def put(self, bucket, fnm, binary, tenant_id=None): logging.debug(f"bucket name {bucket}; filename :{fnm}:") for _ in range(1): try: if not self.bucket_exists(bucket): self.conn.create_bucket(Bucket=bucket) logging.info(f"create bucket {bucket} ********") r = self.conn.upload_fileobj(BytesIO(binary), bucket, fnm) return r except Exception: logging.exception(f"Fail put {bucket}/{fnm}") self.__open__() time.sleep(1) @use_prefix_path @use_default_bucket def rm(self, bucket, fnm, tenant_id=None): try: self.conn.delete_object(Bucket=bucket, Key=fnm) except Exception: logging.exception(f"Fail rm {bucket}/{fnm}") @use_prefix_path @use_default_bucket def get(self, bucket, fnm, tenant_id=None): for _ in range(1): try: r = self.conn.get_object(Bucket=bucket, Key=fnm) object_data = r['Body'].read() return object_data except Exception: logging.exception(f"fail get {bucket}/{fnm}") self.__open__() time.sleep(1) return None @use_prefix_path @use_default_bucket def obj_exist(self, bucket, fnm, tenant_id=None): try: if self.conn.head_object(Bucket=bucket, Key=fnm): return True except ClientError as e: if e.response['Error']['Code'] == '404': return False else: raise @use_prefix_path @use_default_bucket def get_presigned_url(self, bucket, fnm, expires, tenant_id=None): for _ in range(10): try: r = self.conn.generate_presigned_url('get_object', Params={'Bucket': bucket, 'Key': fnm}, ExpiresIn=expires) return r except Exception: logging.exception(f"fail get url {bucket}/{fnm}") self.__open__() time.sleep(1) return None
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/redis_conn.py
rag/utils/redis_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import asyncio import logging import json import uuid import valkey as redis from common.decorator import singleton from common import settings from valkey.lock import Lock REDIS = {} try: REDIS = settings.decrypt_database_config(name="redis") except Exception: try: REDIS = settings.get_base_config("redis", {}) except Exception: REDIS = {} class RedisMsg: def __init__(self, consumer, queue_name, group_name, msg_id, message): self.__consumer = consumer self.__queue_name = queue_name self.__group_name = group_name self.__msg_id = msg_id self.__message = json.loads(message["message"]) def ack(self): try: self.__consumer.xack(self.__queue_name, self.__group_name, self.__msg_id) return True except Exception as e: logging.warning("[EXCEPTION]ack" + str(self.__queue_name) + "||" + str(e)) return False def get_message(self): return self.__message def get_msg_id(self): return self.__msg_id @singleton class RedisDB: lua_delete_if_equal = None lua_token_bucket = None LUA_DELETE_IF_EQUAL_SCRIPT = """ local current_value = redis.call('get', KEYS[1]) if current_value and current_value == ARGV[1] then redis.call('del', KEYS[1]) return 1 end return 0 """ LUA_TOKEN_BUCKET_SCRIPT = """ -- KEYS[1] = rate limit key -- ARGV[1] = capacity -- ARGV[2] = rate -- ARGV[3] = now -- ARGV[4] = cost local key = KEYS[1] local capacity = tonumber(ARGV[1]) local rate = tonumber(ARGV[2]) local now = tonumber(ARGV[3]) local cost = tonumber(ARGV[4]) local data = redis.call("HMGET", key, "tokens", "timestamp") local tokens = tonumber(data[1]) local last_ts = tonumber(data[2]) if tokens == nil then tokens = capacity last_ts = now end local delta = math.max(0, now - last_ts) tokens = math.min(capacity, tokens + delta * rate) if tokens < cost then return {0, tokens} end tokens = tokens - cost redis.call("HMSET", key, "tokens", tokens, "timestamp", now ) redis.call("EXPIRE", key, math.ceil(capacity / rate * 2)) return {1, tokens} """ def __init__(self): self.REDIS = None self.config = REDIS self.__open__() def register_scripts(self) -> None: cls = self.__class__ client = self.REDIS cls.lua_delete_if_equal = client.register_script(cls.LUA_DELETE_IF_EQUAL_SCRIPT) cls.lua_token_bucket = client.register_script(cls.LUA_TOKEN_BUCKET_SCRIPT) def __open__(self): try: conn_params = { "host": self.config["host"].split(":")[0], "port": int(self.config.get("host", ":6379").split(":")[1]), "db": int(self.config.get("db", 1)), "decode_responses": True, } username = self.config.get("username") if username: conn_params["username"] = username password = self.config.get("password") if password: conn_params["password"] = password self.REDIS = redis.StrictRedis(**conn_params) self.register_scripts() except Exception as e: logging.warning(f"Redis can't be connected. Error: {str(e)}") return self.REDIS def health(self): self.REDIS.ping() a, b = "xx", "yy" self.REDIS.set(a, b, 3) if self.REDIS.get(a) == b: return True return False def info(self): info = self.REDIS.info() return { 'redis_version': info["redis_version"], 'server_mode': info["server_mode"] if "server_mode" in info else info.get("redis_mode", ""), 'used_memory': info["used_memory_human"], 'total_system_memory': info["total_system_memory_human"], 'mem_fragmentation_ratio': info["mem_fragmentation_ratio"], 'connected_clients': info["connected_clients"], 'blocked_clients': info["blocked_clients"], 'instantaneous_ops_per_sec': info["instantaneous_ops_per_sec"], 'total_commands_processed': info["total_commands_processed"] } def is_alive(self): return self.REDIS is not None def exist(self, k): if not self.REDIS: return None try: return self.REDIS.exists(k) except Exception as e: logging.warning("RedisDB.exist " + str(k) + " got exception: " + str(e)) self.__open__() def get(self, k): if not self.REDIS: return None try: return self.REDIS.get(k) except Exception as e: logging.warning("RedisDB.get " + str(k) + " got exception: " + str(e)) self.__open__() def set_obj(self, k, obj, exp=3600): try: self.REDIS.set(k, json.dumps(obj, ensure_ascii=False), exp) return True except Exception as e: logging.warning("RedisDB.set_obj " + str(k) + " got exception: " + str(e)) self.__open__() return False def set(self, k, v, exp=3600): try: self.REDIS.set(k, v, exp) return True except Exception as e: logging.warning("RedisDB.set " + str(k) + " got exception: " + str(e)) self.__open__() return False def sadd(self, key: str, member: str): try: self.REDIS.sadd(key, member) return True except Exception as e: logging.warning("RedisDB.sadd " + str(key) + " got exception: " + str(e)) self.__open__() return False def srem(self, key: str, member: str): try: self.REDIS.srem(key, member) return True except Exception as e: logging.warning("RedisDB.srem " + str(key) + " got exception: " + str(e)) self.__open__() return False def smembers(self, key: str): try: res = self.REDIS.smembers(key) return res except Exception as e: logging.warning( "RedisDB.smembers " + str(key) + " got exception: " + str(e) ) self.__open__() return None def zadd(self, key: str, member: str, score: float): try: self.REDIS.zadd(key, {member: score}) return True except Exception as e: logging.warning("RedisDB.zadd " + str(key) + " got exception: " + str(e)) self.__open__() return False def zcount(self, key: str, min: float, max: float): try: res = self.REDIS.zcount(key, min, max) return res except Exception as e: logging.warning("RedisDB.zcount " + str(key) + " got exception: " + str(e)) self.__open__() return 0 def zpopmin(self, key: str, count: int): try: res = self.REDIS.zpopmin(key, count) return res except Exception as e: logging.warning("RedisDB.zpopmin " + str(key) + " got exception: " + str(e)) self.__open__() return None def zrangebyscore(self, key: str, min: float, max: float): try: res = self.REDIS.zrangebyscore(key, min, max) return res except Exception as e: logging.warning( "RedisDB.zrangebyscore " + str(key) + " got exception: " + str(e) ) self.__open__() return None def zremrangebyscore(self, key: str, min: float, max: float): try: res = self.REDIS.zremrangebyscore(key, min, max) return res except Exception as e: logging.warning( f"RedisDB.zremrangebyscore {key} got exception: {e}" ) self.__open__() return 0 def incrby(self, key: str, increment: int): return self.REDIS.incrby(key, increment) def decrby(self, key: str, decrement: int): return self.REDIS.decrby(key, decrement) def generate_auto_increment_id(self, key_prefix: str = "id_generator", namespace: str = "default", increment: int = 1, ensure_minimum: int | None = None) -> int: redis_key = f"{key_prefix}:{namespace}" try: # Use pipeline for atomicity pipe = self.REDIS.pipeline() # Check if key exists pipe.exists(redis_key) # Get/Increment if ensure_minimum is not None: # Ensure minimum value pipe.get(redis_key) results = pipe.execute() if results[0] == 0: # Key doesn't exist start_id = max(1, ensure_minimum) pipe.set(redis_key, start_id) pipe.execute() return start_id else: current = int(results[1]) if current < ensure_minimum: pipe.set(redis_key, ensure_minimum) pipe.execute() return ensure_minimum # Increment operation next_id = self.REDIS.incrby(redis_key, increment) # If it's the first time, set a reasonable initial value if next_id == increment: self.REDIS.set(redis_key, 1 + increment) return 1 + increment return next_id except Exception as e: logging.warning("RedisDB.generate_auto_increment_id got exception: " + str(e)) self.__open__() return -1 def transaction(self, key, value, exp=3600): try: pipeline = self.REDIS.pipeline(transaction=True) pipeline.set(key, value, exp, nx=True) pipeline.execute() return True except Exception as e: logging.warning( "RedisDB.transaction " + str(key) + " got exception: " + str(e) ) self.__open__() return False def queue_product(self, queue, message) -> bool: for _ in range(3): try: payload = {"message": json.dumps(message)} self.REDIS.xadd(queue, payload) return True except Exception as e: logging.exception( "RedisDB.queue_product " + str(queue) + " got exception: " + str(e) ) self.__open__() return False def queue_consumer(self, queue_name, group_name, consumer_name, msg_id=b">") -> RedisMsg: """https://redis.io/docs/latest/commands/xreadgroup/""" for _ in range(3): try: try: group_info = self.REDIS.xinfo_groups(queue_name) if not any(gi["name"] == group_name for gi in group_info): self.REDIS.xgroup_create(queue_name, group_name, id="0", mkstream=True) except redis.exceptions.ResponseError as e: if "no such key" in str(e).lower(): self.REDIS.xgroup_create(queue_name, group_name, id="0", mkstream=True) elif "busygroup" in str(e).lower(): logging.warning("Group already exists, continue.") pass else: raise args = { "groupname": group_name, "consumername": consumer_name, "count": 1, "block": 5, "streams": {queue_name: msg_id}, } messages = self.REDIS.xreadgroup(**args) if not messages: return None stream, element_list = messages[0] if not element_list: return None msg_id, payload = element_list[0] res = RedisMsg(self.REDIS, queue_name, group_name, msg_id, payload) return res except Exception as e: if str(e) == 'no such key': pass else: logging.exception( "RedisDB.queue_consumer " + str(queue_name) + " got exception: " + str(e) ) self.__open__() return None def get_unacked_iterator(self, queue_names: list[str], group_name, consumer_name): try: for queue_name in queue_names: try: group_info = self.REDIS.xinfo_groups(queue_name) except Exception as e: if str(e) == 'no such key': logging.warning(f"RedisDB.get_unacked_iterator queue {queue_name} doesn't exist") continue if not any(gi["name"] == group_name for gi in group_info): logging.warning(f"RedisDB.get_unacked_iterator queue {queue_name} group {group_name} doesn't exist") continue current_min = 0 while True: payload = self.queue_consumer(queue_name, group_name, consumer_name, current_min) if not payload: break current_min = payload.get_msg_id() logging.info(f"RedisDB.get_unacked_iterator {queue_name} {consumer_name} {current_min}") yield payload except Exception: logging.exception( "RedisDB.get_unacked_iterator got exception: " ) self.__open__() def get_pending_msg(self, queue, group_name): try: messages = self.REDIS.xpending_range(queue, group_name, '-', '+', 10) return messages except Exception as e: if 'No such key' not in (str(e) or ''): logging.warning( "RedisDB.get_pending_msg " + str(queue) + " got exception: " + str(e) ) return [] def requeue_msg(self, queue: str, group_name: str, msg_id: str): for _ in range(3): try: messages = self.REDIS.xrange(queue, msg_id, msg_id) if messages: self.REDIS.xadd(queue, messages[0][1]) self.REDIS.xack(queue, group_name, msg_id) except Exception as e: logging.warning( "RedisDB.get_pending_msg " + str(queue) + " got exception: " + str(e) ) self.__open__() def queue_info(self, queue, group_name) -> dict | None: for _ in range(3): try: groups = self.REDIS.xinfo_groups(queue) for group in groups: if group["name"] == group_name: return group except Exception as e: logging.warning( "RedisDB.queue_info " + str(queue) + " got exception: " + str(e) ) self.__open__() return None def delete_if_equal(self, key: str, expected_value: str) -> bool: """ Do following atomically: Delete a key if its value is equals to the given one, do nothing otherwise. """ return bool(self.lua_delete_if_equal(keys=[key], args=[expected_value], client=self.REDIS)) def delete(self, key) -> bool: try: self.REDIS.delete(key) return True except Exception as e: logging.warning("RedisDB.delete " + str(key) + " got exception: " + str(e)) self.__open__() return False REDIS_CONN = RedisDB() class RedisDistributedLock: def __init__(self, lock_key, lock_value=None, timeout=10, blocking_timeout=1): self.lock_key = lock_key if lock_value: self.lock_value = lock_value else: self.lock_value = str(uuid.uuid4()) self.timeout = timeout self.lock = Lock(REDIS_CONN.REDIS, lock_key, timeout=timeout, blocking_timeout=blocking_timeout) def acquire(self): REDIS_CONN.delete_if_equal(self.lock_key, self.lock_value) return self.lock.acquire(token=self.lock_value) async def spin_acquire(self): REDIS_CONN.delete_if_equal(self.lock_key, self.lock_value) while True: if self.lock.acquire(token=self.lock_value): break await asyncio.sleep(10) def release(self): REDIS_CONN.delete_if_equal(self.lock_key, self.lock_value)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/encrypted_storage.py
rag/utils/encrypted_storage.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging from common.crypto_utils import CryptoUtil # from common.decorator import singleton class EncryptedStorageWrapper: """Encrypted storage wrapper that wraps existing storage implementations to provide transparent encryption""" def __init__(self, storage_impl, algorithm="aes-256-cbc", key=None, iv=None): """ Initialize encrypted storage wrapper Args: storage_impl: Original storage implementation instance algorithm: Encryption algorithm, default is aes-256-cbc key: Encryption key, uses RAGFLOW_CRYPTO_KEY environment variable if None iv: Initialization vector, automatically generated if None """ self.storage_impl = storage_impl self.crypto = CryptoUtil(algorithm=algorithm, key=key, iv=iv) self.encryption_enabled = True # Check if storage implementation has required methods # todo: Consider abstracting a storage base class to ensure these methods exist required_methods = ["put", "get", "rm", "obj_exist", "health"] for method in required_methods: if not hasattr(storage_impl, method): raise AttributeError(f"Storage implementation missing required method: {method}") logging.info(f"EncryptedStorageWrapper initialized with algorithm: {algorithm}") def put(self, bucket, fnm, binary, tenant_id=None): """ Encrypt and store data Args: bucket: Bucket name fnm: File name binary: Original binary data tenant_id: Tenant ID (optional) Returns: Storage result """ if not self.encryption_enabled: return self.storage_impl.put(bucket, fnm, binary, tenant_id) try: encrypted_binary = self.crypto.encrypt(binary) return self.storage_impl.put(bucket, fnm, encrypted_binary, tenant_id) except Exception as e: logging.exception(f"Failed to encrypt and store data: {bucket}/{fnm}, error: {str(e)}") raise def get(self, bucket, fnm, tenant_id=None): """ Retrieve and decrypt data Args: bucket: Bucket name fnm: File name tenant_id: Tenant ID (optional) Returns: Decrypted binary data """ try: # Get encrypted data encrypted_binary = self.storage_impl.get(bucket, fnm, tenant_id) if encrypted_binary is None: return None if not self.encryption_enabled: return encrypted_binary # Decrypt data decrypted_binary = self.crypto.decrypt(encrypted_binary) return decrypted_binary except Exception as e: logging.exception(f"Failed to get and decrypt data: {bucket}/{fnm}, error: {str(e)}") raise def rm(self, bucket, fnm, tenant_id=None): """ Delete data (same as original storage implementation, no decryption needed) Args: bucket: Bucket name fnm: File name tenant_id: Tenant ID (optional) Returns: Deletion result """ return self.storage_impl.rm(bucket, fnm, tenant_id) def obj_exist(self, bucket, fnm, tenant_id=None): """ Check if object exists (same as original storage implementation, no decryption needed) Args: bucket: Bucket name fnm: File name tenant_id: Tenant ID (optional) Returns: Whether the object exists """ return self.storage_impl.obj_exist(bucket, fnm, tenant_id) def health(self): """ Health check (uses the original storage implementation's method) Returns: Health check result """ return self.storage_impl.health() def bucket_exists(self, bucket): """ Check if bucket exists (if the original storage implementation has this method) Args: bucket: Bucket name Returns: Whether the bucket exists """ if hasattr(self.storage_impl, "bucket_exists"): return self.storage_impl.bucket_exists(bucket) return False def get_presigned_url(self, bucket, fnm, expires, tenant_id=None): """ Get presigned URL (if the original storage implementation has this method) Args: bucket: Bucket name fnm: File name expires: Expiration time tenant_id: Tenant ID (optional) Returns: Presigned URL """ if hasattr(self.storage_impl, "get_presigned_url"): return self.storage_impl.get_presigned_url(bucket, fnm, expires, tenant_id) return None def scan(self, bucket, fnm, tenant_id=None): """ Scan objects (if the original storage implementation has this method) Args: bucket: Bucket name fnm: File name prefix tenant_id: Tenant ID (optional) Returns: Scan results """ if hasattr(self.storage_impl, "scan"): return self.storage_impl.scan(bucket, fnm, tenant_id) return None def copy(self, src_bucket, src_path, dest_bucket, dest_path): """ Copy object (if the original storage implementation has this method) Args: src_bucket: Source bucket name src_path: Source file path dest_bucket: Destination bucket name dest_path: Destination file path Returns: Copy result """ if hasattr(self.storage_impl, "copy"): return self.storage_impl.copy(src_bucket, src_path, dest_bucket, dest_path) return False def move(self, src_bucket, src_path, dest_bucket, dest_path): """ Move object (if the original storage implementation has this method) Args: src_bucket: Source bucket name src_path: Source file path dest_bucket: Destination bucket name dest_path: Destination file path Returns: Move result """ if hasattr(self.storage_impl, "move"): return self.storage_impl.move(src_bucket, src_path, dest_bucket, dest_path) return False def remove_bucket(self, bucket): """ Remove bucket (if the original storage implementation has this method) Args: bucket: Bucket name Returns: Remove result """ if hasattr(self.storage_impl, "remove_bucket"): return self.storage_impl.remove_bucket(bucket) return False def enable_encryption(self): """Enable encryption""" self.encryption_enabled = True logging.info("Encryption enabled") def disable_encryption(self): """Disable encryption""" self.encryption_enabled = False logging.info("Encryption disabled") # Create singleton wrapper function def create_encrypted_storage(storage_impl, algorithm=None, key=None, encryption_enabled=True): """ Create singleton instance of encrypted storage wrapper Args: storage_impl: Original storage implementation instance algorithm: Encryption algorithm, uses environment variable RAGFLOW_CRYPTO_ALGORITHM or default if None key: Encryption key, uses environment variable RAGFLOW_CRYPTO_KEY if None encryption_enabled: Whether to enable encryption functionality Returns: Encrypted storage wrapper instance """ wrapper = EncryptedStorageWrapper(storage_impl, algorithm=algorithm, key=key) wrapper.encryption_enabled = encryption_enabled if encryption_enabled: logging.info("Encryption enabled in storage wrapper") else: logging.info("Encryption disabled in storage wrapper") return wrapper
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/minio_conn.py
rag/utils/minio_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import time from minio import Minio from minio.commonconfig import CopySource from minio.error import S3Error from io import BytesIO from common.decorator import singleton from common import settings @singleton class RAGFlowMinio: def __init__(self): self.conn = None # Use `or None` to convert empty strings to None, ensuring single-bucket # mode is truly disabled when not configured self.bucket = settings.MINIO.get('bucket', None) or None self.prefix_path = settings.MINIO.get('prefix_path', None) or None self.__open__() @staticmethod def use_default_bucket(method): def wrapper(self, bucket, *args, **kwargs): # If there is a default bucket, use the default bucket # but preserve the original bucket identifier so it can be # used as a path prefix inside the physical/default bucket. original_bucket = bucket actual_bucket = self.bucket if self.bucket else bucket if self.bucket: # pass original identifier forward for use by other decorators kwargs['_orig_bucket'] = original_bucket return method(self, actual_bucket, *args, **kwargs) return wrapper @staticmethod def use_prefix_path(method): def wrapper(self, bucket, fnm, *args, **kwargs): # If a default MINIO bucket is configured, the use_default_bucket # decorator will have replaced the `bucket` arg with the physical # bucket name and forwarded the original identifier as `_orig_bucket`. # Prefer that original identifier when constructing the key path so # objects are stored under <physical-bucket>/<identifier>/... orig_bucket = kwargs.pop('_orig_bucket', None) if self.prefix_path: # If a prefix_path is configured, include it and then the identifier if orig_bucket: fnm = f"{self.prefix_path}/{orig_bucket}/{fnm}" else: fnm = f"{self.prefix_path}/{fnm}" else: # No prefix_path configured. If orig_bucket exists and the # physical bucket equals configured default, use orig_bucket as a path. if orig_bucket and bucket == self.bucket: fnm = f"{orig_bucket}/{fnm}" return method(self, bucket, fnm, *args, **kwargs) return wrapper def __open__(self): try: if self.conn: self.__close__() except Exception: pass try: self.conn = Minio(settings.MINIO["host"], access_key=settings.MINIO["user"], secret_key=settings.MINIO["password"], secure=False ) except Exception: logging.exception( "Fail to connect %s " % settings.MINIO["host"]) def __close__(self): del self.conn self.conn = None def health(self): bucket = self.bucket if self.bucket else "ragflow-bucket" fnm = "_health_check" if self.prefix_path: fnm = f"{self.prefix_path}/{fnm}" binary = b"_t@@@1" # Don't try to create bucket - it should already exist # if not self.conn.bucket_exists(bucket): # self.conn.make_bucket(bucket) r = self.conn.put_object(bucket, fnm, BytesIO(binary), len(binary) ) return r @use_default_bucket @use_prefix_path def put(self, bucket, fnm, binary, tenant_id=None): for _ in range(3): try: # Note: bucket must already exist - we don't have permission to create buckets if not self.bucket and not self.conn.bucket_exists(bucket): self.conn.make_bucket(bucket) r = self.conn.put_object(bucket, fnm, BytesIO(binary), len(binary) ) return r except Exception: logging.exception(f"Fail to put {bucket}/{fnm}:") self.__open__() time.sleep(1) @use_default_bucket @use_prefix_path def rm(self, bucket, fnm, tenant_id=None): try: self.conn.remove_object(bucket, fnm) except Exception: logging.exception(f"Fail to remove {bucket}/{fnm}:") @use_default_bucket @use_prefix_path def get(self, bucket, filename, tenant_id=None): for _ in range(1): try: r = self.conn.get_object(bucket, filename) return r.read() except Exception: logging.exception(f"Fail to get {bucket}/{filename}") self.__open__() time.sleep(1) return @use_default_bucket @use_prefix_path def obj_exist(self, bucket, filename, tenant_id=None): try: if not self.conn.bucket_exists(bucket): return False if self.conn.stat_object(bucket, filename): return True else: return False except S3Error as e: if e.code in ["NoSuchKey", "NoSuchBucket", "ResourceNotFound"]: return False except Exception: logging.exception(f"obj_exist {bucket}/{filename} got exception") return False @use_default_bucket def bucket_exists(self, bucket): try: if not self.conn.bucket_exists(bucket): return False else: return True except S3Error as e: if e.code in ["NoSuchKey", "NoSuchBucket", "ResourceNotFound"]: return False except Exception: logging.exception(f"bucket_exist {bucket} got exception") return False @use_default_bucket @use_prefix_path def get_presigned_url(self, bucket, fnm, expires, tenant_id=None): for _ in range(10): try: return self.conn.get_presigned_url("GET", bucket, fnm, expires) except Exception: logging.exception(f"Fail to get_presigned {bucket}/{fnm}:") self.__open__() time.sleep(1) return @use_default_bucket def remove_bucket(self, bucket, **kwargs): orig_bucket = kwargs.pop('_orig_bucket', None) try: if self.bucket: # Single bucket mode: remove objects with prefix prefix = "" if self.prefix_path: prefix = f"{self.prefix_path}/" if orig_bucket: prefix += f"{orig_bucket}/" # List objects with prefix objects_to_delete = self.conn.list_objects(bucket, prefix=prefix, recursive=True) for obj in objects_to_delete: self.conn.remove_object(bucket, obj.object_name) # Do NOT remove the physical bucket else: if self.conn.bucket_exists(bucket): objects_to_delete = self.conn.list_objects(bucket, recursive=True) for obj in objects_to_delete: self.conn.remove_object(bucket, obj.object_name) self.conn.remove_bucket(bucket) except Exception: logging.exception(f"Fail to remove bucket {bucket}") def _resolve_bucket_and_path(self, bucket, fnm): if self.bucket: if self.prefix_path: fnm = f"{self.prefix_path}/{bucket}/{fnm}" else: fnm = f"{bucket}/{fnm}" bucket = self.bucket elif self.prefix_path: fnm = f"{self.prefix_path}/{fnm}" return bucket, fnm def copy(self, src_bucket, src_path, dest_bucket, dest_path): try: src_bucket, src_path = self._resolve_bucket_and_path(src_bucket, src_path) dest_bucket, dest_path = self._resolve_bucket_and_path(dest_bucket, dest_path) if not self.conn.bucket_exists(dest_bucket): self.conn.make_bucket(dest_bucket) try: self.conn.stat_object(src_bucket, src_path) except Exception as e: logging.exception(f"Source object not found: {src_bucket}/{src_path}, {e}") return False self.conn.copy_object( dest_bucket, dest_path, CopySource(src_bucket, src_path), ) return True except Exception: logging.exception(f"Fail to copy {src_bucket}/{src_path} -> {dest_bucket}/{dest_path}") return False def move(self, src_bucket, src_path, dest_bucket, dest_path): try: if self.copy(src_bucket, src_path, dest_bucket, dest_path): self.rm(src_bucket, src_path) return True else: logging.error(f"Copy failed, move aborted: {src_bucket}/{src_path}") return False except Exception: logging.exception(f"Fail to move {src_bucket}/{src_path} -> {dest_bucket}/{dest_path}") return False
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/azure_sas_conn.py
rag/utils/azure_sas_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import os import time from io import BytesIO from common.decorator import singleton from azure.storage.blob import ContainerClient from common import settings @singleton class RAGFlowAzureSasBlob: def __init__(self): self.conn = None self.container_url = os.getenv('CONTAINER_URL', settings.AZURE["container_url"]) self.sas_token = os.getenv('SAS_TOKEN', settings.AZURE["sas_token"]) self.__open__() def __open__(self): try: if self.conn: self.__close__() except Exception: pass try: self.conn = ContainerClient.from_container_url(self.container_url + "?" + self.sas_token) except Exception: logging.exception("Fail to connect %s " % self.container_url) def __close__(self): del self.conn self.conn = None def health(self): _bucket, fnm, binary = "txtxtxtxt1", "txtxtxtxt1", b"_t@@@1" return self.conn.upload_blob(name=fnm, data=BytesIO(binary), length=len(binary)) def put(self, bucket, fnm, binary): for _ in range(3): try: return self.conn.upload_blob(name=fnm, data=BytesIO(binary), length=len(binary)) except Exception: logging.exception(f"Fail put {bucket}/{fnm}") self.__open__() time.sleep(1) def rm(self, bucket, fnm): try: self.conn.delete_blob(fnm) except Exception: logging.exception(f"Fail rm {bucket}/{fnm}") def get(self, bucket, fnm): for _ in range(1): try: r = self.conn.download_blob(fnm) return r.read() except Exception: logging.exception(f"fail get {bucket}/{fnm}") self.__open__() time.sleep(1) return def obj_exist(self, bucket, fnm): try: return self.conn.get_blob_client(fnm).exists() except Exception: logging.exception(f"Fail put {bucket}/{fnm}") return False def get_presigned_url(self, bucket, fnm, expires): for _ in range(10): try: return self.conn.get_presigned_url("GET", bucket, fnm, expires) except Exception: logging.exception(f"fail get {bucket}/{fnm}") self.__open__() time.sleep(1) return
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/es_conn.py
rag/utils/es_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import re import json import time import copy from elasticsearch_dsl import UpdateByQuery, Q, Search from elastic_transport import ConnectionTimeout from common.decorator import singleton from common.doc_store.doc_store_base import MatchTextExpr, OrderByExpr, MatchExpr, MatchDenseExpr, FusionExpr from common.doc_store.es_conn_base import ESConnectionBase from common.float_utils import get_float from common.constants import PAGERANK_FLD, TAG_FLD ATTEMPT_TIME = 2 @singleton class ESConnection(ESConnectionBase): """ CRUD operations """ def search( self, select_fields: list[str], highlight_fields: list[str], condition: dict, match_expressions: list[MatchExpr], order_by: OrderByExpr, offset: int, limit: int, index_names: str | list[str], knowledgebase_ids: list[str], agg_fields: list[str] | None = None, rank_feature: dict | None = None ): """ Refers to https://www.elastic.co/guide/en/elasticsearch/reference/current/query-dsl.html """ if isinstance(index_names, str): index_names = index_names.split(",") assert isinstance(index_names, list) and len(index_names) > 0 assert "_id" not in condition bool_query = Q("bool", must=[]) condition["kb_id"] = knowledgebase_ids for k, v in condition.items(): if k == "available_int": if v == 0: bool_query.filter.append(Q("range", available_int={"lt": 1})) else: bool_query.filter.append( Q("bool", must_not=Q("range", available_int={"lt": 1}))) continue if not v: continue if isinstance(v, list): bool_query.filter.append(Q("terms", **{k: v})) elif isinstance(v, str) or isinstance(v, int): bool_query.filter.append(Q("term", **{k: v})) else: raise Exception( f"Condition `{str(k)}={str(v)}` value type is {str(type(v))}, expected to be int, str or list.") s = Search() vector_similarity_weight = 0.5 for m in match_expressions: if isinstance(m, FusionExpr) and m.method == "weighted_sum" and "weights" in m.fusion_params: assert len(match_expressions) == 3 and isinstance(match_expressions[0], MatchTextExpr) and isinstance( match_expressions[1], MatchDenseExpr) and isinstance( match_expressions[2], FusionExpr) weights = m.fusion_params["weights"] vector_similarity_weight = get_float(weights.split(",")[1]) for m in match_expressions: if isinstance(m, MatchTextExpr): minimum_should_match = m.extra_options.get("minimum_should_match", 0.0) if isinstance(minimum_should_match, float): minimum_should_match = str(int(minimum_should_match * 100)) + "%" bool_query.must.append(Q("query_string", fields=m.fields, type="best_fields", query=m.matching_text, minimum_should_match=minimum_should_match, boost=1)) bool_query.boost = 1.0 - vector_similarity_weight elif isinstance(m, MatchDenseExpr): assert (bool_query is not None) similarity = 0.0 if "similarity" in m.extra_options: similarity = m.extra_options["similarity"] s = s.knn(m.vector_column_name, m.topn, m.topn * 2, query_vector=list(m.embedding_data), filter=bool_query.to_dict(), similarity=similarity, ) if bool_query and rank_feature: for fld, sc in rank_feature.items(): if fld != PAGERANK_FLD: fld = f"{TAG_FLD}.{fld}" bool_query.should.append(Q("rank_feature", field=fld, linear={}, boost=sc)) if bool_query: s = s.query(bool_query) for field in highlight_fields: s = s.highlight(field) if order_by: orders = list() for field, order in order_by.fields: order = "asc" if order == 0 else "desc" if field in ["page_num_int", "top_int"]: order_info = {"order": order, "unmapped_type": "float", "mode": "avg", "numeric_type": "double"} elif field.endswith("_int") or field.endswith("_flt"): order_info = {"order": order, "unmapped_type": "float"} else: order_info = {"order": order, "unmapped_type": "text"} orders.append({field: order_info}) s = s.sort(*orders) if agg_fields: for fld in agg_fields: s.aggs.bucket(f'aggs_{fld}', 'terms', field=fld, size=1000000) if limit > 0: s = s[offset:offset + limit] q = s.to_dict() self.logger.debug(f"ESConnection.search {str(index_names)} query: " + json.dumps(q)) for i in range(ATTEMPT_TIME): try: # print(json.dumps(q, ensure_ascii=False)) res = self.es.search(index=index_names, body=q, timeout="600s", # search_type="dfs_query_then_fetch", track_total_hits=True, _source=True) if str(res.get("timed_out", "")).lower() == "true": raise Exception("Es Timeout.") self.logger.debug(f"ESConnection.search {str(index_names)} res: " + str(res)) return res except ConnectionTimeout: self.logger.exception("ES request timeout") self._connect() continue except Exception as e: self.logger.exception(f"ESConnection.search {str(index_names)} query: " + str(q) + str(e)) raise e self.logger.error(f"ESConnection.search timeout for {ATTEMPT_TIME} times!") raise Exception("ESConnection.search timeout.") def insert(self, documents: list[dict], index_name: str, knowledgebase_id: str = None) -> list[str]: # Refers to https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-bulk.html operations = [] for d in documents: assert "_id" not in d assert "id" in d d_copy = copy.deepcopy(d) d_copy["kb_id"] = knowledgebase_id meta_id = d_copy.pop("id", "") operations.append( {"index": {"_index": index_name, "_id": meta_id}}) operations.append(d_copy) res = [] for _ in range(ATTEMPT_TIME): try: res = [] r = self.es.bulk(index=index_name, operations=operations, refresh=False, timeout="60s") if re.search(r"False", str(r["errors"]), re.IGNORECASE): return res for item in r["items"]: for action in ["create", "delete", "index", "update"]: if action in item and "error" in item[action]: res.append(str(item[action]["_id"]) + ":" + str(item[action]["error"])) return res except ConnectionTimeout: self.logger.exception("ES request timeout") time.sleep(3) self._connect() continue except Exception as e: res.append(str(e)) self.logger.warning("ESConnection.insert got exception: " + str(e)) return res def update(self, condition: dict, new_value: dict, index_name: str, knowledgebase_id: str) -> bool: doc = copy.deepcopy(new_value) doc.pop("id", None) condition["kb_id"] = knowledgebase_id if "id" in condition and isinstance(condition["id"], str): # update specific single document chunk_id = condition["id"] for i in range(ATTEMPT_TIME): for k in doc.keys(): if "feas" != k.split("_")[-1]: continue try: self.es.update(index=index_name, id=chunk_id, script=f"ctx._source.remove(\"{k}\");") except Exception: self.logger.exception( f"ESConnection.update(index={index_name}, id={chunk_id}, doc={json.dumps(condition, ensure_ascii=False)}) got exception") try: self.es.update(index=index_name, id=chunk_id, doc=doc) return True except Exception as e: self.logger.exception( f"ESConnection.update(index={index_name}, id={chunk_id}, doc={json.dumps(condition, ensure_ascii=False)}) got exception: " + str( e)) break return False # update unspecific maybe-multiple documents bool_query = Q("bool") for k, v in condition.items(): if not isinstance(k, str) or not v: continue if k == "exists": bool_query.filter.append(Q("exists", field=v)) continue if isinstance(v, list): bool_query.filter.append(Q("terms", **{k: v})) elif isinstance(v, str) or isinstance(v, int): bool_query.filter.append(Q("term", **{k: v})) else: raise Exception( f"Condition `{str(k)}={str(v)}` value type is {str(type(v))}, expected to be int, str or list.") scripts = [] params = {} for k, v in new_value.items(): if k == "remove": if isinstance(v, str): scripts.append(f"ctx._source.remove('{v}');") if isinstance(v, dict): for kk, vv in v.items(): scripts.append(f"int i=ctx._source.{kk}.indexOf(params.p_{kk});ctx._source.{kk}.remove(i);") params[f"p_{kk}"] = vv continue if k == "add": if isinstance(v, dict): for kk, vv in v.items(): scripts.append(f"ctx._source.{kk}.add(params.pp_{kk});") params[f"pp_{kk}"] = vv.strip() continue if (not isinstance(k, str) or not v) and k != "available_int": continue if isinstance(v, str): v = re.sub(r"(['\n\r]|\\.)", " ", v) params[f"pp_{k}"] = v scripts.append(f"ctx._source.{k}=params.pp_{k};") elif isinstance(v, int) or isinstance(v, float): scripts.append(f"ctx._source.{k}={v};") elif isinstance(v, list): scripts.append(f"ctx._source.{k}=params.pp_{k};") params[f"pp_{k}"] = json.dumps(v, ensure_ascii=False) else: raise Exception( f"newValue `{str(k)}={str(v)}` value type is {str(type(v))}, expected to be int, str.") ubq = UpdateByQuery( index=index_name).using( self.es).query(bool_query) ubq = ubq.script(source="".join(scripts), params=params) ubq = ubq.params(refresh=True) ubq = ubq.params(slices=5) ubq = ubq.params(conflicts="proceed") for _ in range(ATTEMPT_TIME): try: _ = ubq.execute() return True except ConnectionTimeout: self.logger.exception("ES request timeout") time.sleep(3) self._connect() continue except Exception as e: self.logger.error("ESConnection.update got exception: " + str(e) + "\n".join(scripts)) break return False def delete(self, condition: dict, index_name: str, knowledgebase_id: str) -> int: assert "_id" not in condition condition["kb_id"] = knowledgebase_id if "id" in condition: chunk_ids = condition["id"] if not isinstance(chunk_ids, list): chunk_ids = [chunk_ids] if not chunk_ids: # when chunk_ids is empty, delete all qry = Q("match_all") else: qry = Q("ids", values=chunk_ids) else: qry = Q("bool") for k, v in condition.items(): if k == "exists": qry.filter.append(Q("exists", field=v)) elif k == "must_not": if isinstance(v, dict): for kk, vv in v.items(): if kk == "exists": qry.must_not.append(Q("exists", field=vv)) elif isinstance(v, list): qry.must.append(Q("terms", **{k: v})) elif isinstance(v, str) or isinstance(v, int): qry.must.append(Q("term", **{k: v})) else: raise Exception("Condition value must be int, str or list.") self.logger.debug("ESConnection.delete query: " + json.dumps(qry.to_dict())) for _ in range(ATTEMPT_TIME): try: res = self.es.delete_by_query( index=index_name, body=Search().query(qry).to_dict(), refresh=True) return res["deleted"] except ConnectionTimeout: self.logger.exception("ES request timeout") time.sleep(3) self._connect() continue except Exception as e: self.logger.warning("ESConnection.delete got exception: " + str(e)) if re.search(r"(not_found)", str(e), re.IGNORECASE): return 0 return 0 """ Helper functions for search result """ def get_fields(self, res, fields: list[str]) -> dict[str, dict]: res_fields = {} if not fields: return {} for d in self._get_source(res): m = {n: d.get(n) for n in fields if d.get(n) is not None} for n, v in m.items(): if isinstance(v, list): m[n] = v continue if n == "available_int" and isinstance(v, (int, float)): m[n] = v continue if not isinstance(v, str): m[n] = str(m[n]) # if n.find("tks") > 0: # m[n] = remove_redundant_spaces(m[n]) if m: res_fields[d["id"]] = m return res_fields
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/__init__.py
rag/utils/__init__.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/s3_conn.py
rag/utils/s3_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import boto3 from botocore.exceptions import ClientError from botocore.config import Config import time from io import BytesIO from common.decorator import singleton from common import settings @singleton class RAGFlowS3: def __init__(self): self.conn = None self.s3_config = settings.S3 self.access_key = self.s3_config.get('access_key', None) self.secret_key = self.s3_config.get('secret_key', None) self.session_token = self.s3_config.get('session_token', None) self.region_name = self.s3_config.get('region_name', None) self.endpoint_url = self.s3_config.get('endpoint_url', None) self.signature_version = self.s3_config.get('signature_version', None) self.addressing_style = self.s3_config.get('addressing_style', None) self.bucket = self.s3_config.get('bucket', None) self.prefix_path = self.s3_config.get('prefix_path', None) self.__open__() @staticmethod def use_default_bucket(method): def wrapper(self, bucket, *args, **kwargs): # If there is a default bucket, use the default bucket actual_bucket = self.bucket if self.bucket else bucket return method(self, actual_bucket, *args, **kwargs) return wrapper @staticmethod def use_prefix_path(method): def wrapper(self, bucket, fnm, *args, **kwargs): # If the prefix path is set, use the prefix path. # The bucket passed from the upstream call is # used as the file prefix. This is especially useful when you're using the default bucket if self.prefix_path: fnm = f"{self.prefix_path}/{bucket}/{fnm}" return method(self, bucket, fnm, *args, **kwargs) return wrapper def __open__(self): try: if self.conn: self.__close__() except Exception: pass try: s3_params = {} config_kwargs = {} # if not set ak/sk, boto3 s3 client would try several ways to do the authentication # see doc: https://boto3.amazonaws.com/v1/documentation/api/latest/guide/credentials.html#configuring-credentials if self.access_key and self.secret_key: s3_params = { 'aws_access_key_id': self.access_key, 'aws_secret_access_key': self.secret_key, 'aws_session_token': self.session_token, } if self.region_name: s3_params['region_name'] = self.region_name if self.endpoint_url: s3_params['endpoint_url'] = self.endpoint_url # Configure signature_version and addressing_style through Config object if self.signature_version: config_kwargs['signature_version'] = self.signature_version if self.addressing_style: config_kwargs['s3'] = {'addressing_style': self.addressing_style} if config_kwargs: s3_params['config'] = Config(**config_kwargs) self.conn = [boto3.client('s3', **s3_params)] except Exception: logging.exception(f"Fail to connect at region {self.region_name} or endpoint {self.endpoint_url}") def __close__(self): del self.conn[0] self.conn = None @use_default_bucket def bucket_exists(self, bucket, *args, **kwargs): try: logging.debug(f"head_bucket bucketname {bucket}") self.conn[0].head_bucket(Bucket=bucket) exists = True except ClientError: logging.exception(f"head_bucket error {bucket}") exists = False return exists def health(self): bucket = self.bucket fnm = "txtxtxtxt1" fnm, binary = f"{self.prefix_path}/{fnm}" if self.prefix_path else fnm, b"_t@@@1" if not self.bucket_exists(bucket): self.conn[0].create_bucket(Bucket=bucket) logging.debug(f"create bucket {bucket} ********") r = self.conn[0].upload_fileobj(BytesIO(binary), bucket, fnm) return r def get_properties(self, bucket, key): return {} def list(self, bucket, dir, recursive=True): return [] @use_prefix_path @use_default_bucket def put(self, bucket, fnm, binary, *args, **kwargs): logging.debug(f"bucket name {bucket}; filename :{fnm}:") for _ in range(1): try: if not self.bucket_exists(bucket): self.conn[0].create_bucket(Bucket=bucket) logging.info(f"create bucket {bucket} ********") r = self.conn[0].upload_fileobj(BytesIO(binary), bucket, fnm) return r except Exception: logging.exception(f"Fail put {bucket}/{fnm}") self.__open__() time.sleep(1) @use_prefix_path @use_default_bucket def rm(self, bucket, fnm, *args, **kwargs): try: self.conn[0].delete_object(Bucket=bucket, Key=fnm) except Exception: logging.exception(f"Fail rm {bucket}/{fnm}") @use_prefix_path @use_default_bucket def get(self, bucket, fnm, *args, **kwargs): for _ in range(1): try: r = self.conn[0].get_object(Bucket=bucket, Key=fnm) object_data = r['Body'].read() return object_data except Exception: logging.exception(f"fail get {bucket}/{fnm}") self.__open__() time.sleep(1) return None @use_prefix_path @use_default_bucket def obj_exist(self, bucket, fnm, *args, **kwargs): try: if self.conn[0].head_object(Bucket=bucket, Key=fnm): return True except ClientError as e: if e.response['Error']['Code'] == '404': return False else: raise @use_prefix_path @use_default_bucket def get_presigned_url(self, bucket, fnm, expires, *args, **kwargs): for _ in range(10): try: r = self.conn[0].generate_presigned_url('get_object', Params={'Bucket': bucket, 'Key': fnm}, ExpiresIn=expires) return r except Exception: logging.exception(f"fail get url {bucket}/{fnm}") self.__open__() time.sleep(1) return None @use_default_bucket def rm_bucket(self, bucket, *args, **kwargs): for conn in self.conn: try: if not conn.bucket_exists(bucket): continue for o in conn.list_objects_v2(Bucket=bucket): conn.delete_object(bucket, o.object_name) conn.delete_bucket(Bucket=bucket) return except Exception as e: logging.error(f"Fail rm {bucket}: " + str(e))
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/opendal_conn.py
rag/utils/opendal_conn.py
import opendal import logging import pymysql import re from urllib.parse import quote_plus from common.config_utils import get_base_config from common.decorator import singleton CREATE_TABLE_SQL = """ CREATE TABLE IF NOT EXISTS `{}` ( `key` VARCHAR(255) PRIMARY KEY, `value` LONGBLOB, `created_at` TIMESTAMP DEFAULT CURRENT_TIMESTAMP, `updated_at` TIMESTAMP DEFAULT CURRENT_TIMESTAMP ON UPDATE CURRENT_TIMESTAMP ); """ SET_MAX_ALLOWED_PACKET_SQL = """ SET GLOBAL max_allowed_packet={} """ def get_opendal_config(): try: opendal_config = get_base_config('opendal', {}) if opendal_config.get("scheme", "mysql") == 'mysql': mysql_config = get_base_config('mysql', {}) max_packet = mysql_config.get("max_allowed_packet", 134217728) kwargs = { "scheme": "mysql", "host": mysql_config.get("host", "127.0.0.1"), "port": str(mysql_config.get("port", 3306)), "user": mysql_config.get("user", "root"), "password": mysql_config.get("password", ""), "database": mysql_config.get("name", "test_open_dal"), "table": opendal_config.get("config", {}).get("oss_table", "opendal_storage"), "max_allowed_packet": str(max_packet) } kwargs[ "connection_string"] = f"mysql://{kwargs['user']}:{quote_plus(kwargs['password'])}@{kwargs['host']}:{kwargs['port']}/{kwargs['database']}?max_allowed_packet={max_packet}" else: scheme = opendal_config.get("scheme") config_data = opendal_config.get("config", {}) kwargs = {"scheme": scheme, **config_data} # Only include non-sensitive keys in logs. Do NOT # add 'password' or any key containing embedded credentials # (like 'connection_string'). safe_log_info = { "scheme": kwargs.get("scheme"), "host": kwargs.get("host"), "port": kwargs.get("port"), "database": kwargs.get("database"), "table": kwargs.get("table"), # indicate presence of credentials without logging them "has_credentials": any(k in kwargs for k in ("password", "connection_string")), } logging.info("Loaded OpenDAL configuration (non sensitive fields only): %s", safe_log_info) # For safety, explicitly remove sensitive keys from kwargs after use if "password" in kwargs: del kwargs["password"] if "connection_string" in kwargs: del kwargs["connection_string"] return kwargs except Exception as e: logging.error("Failed to load OpenDAL configuration from yaml: %s", str(e)) raise @singleton class OpenDALStorage: def __init__(self): self._kwargs = get_opendal_config() self._scheme = self._kwargs.get('scheme', 'mysql') if self._scheme == 'mysql': self.init_db_config() self.init_opendal_mysql_table() self._operator = opendal.Operator(**self._kwargs) logging.info("OpenDALStorage initialized successfully") def health(self): bucket, fnm, binary = "txtxtxtxt1", "txtxtxtxt1", b"_t@@@1" return self._operator.write(f"{bucket}/{fnm}", binary) def put(self, bucket, fnm, binary, tenant_id=None): self._operator.write(f"{bucket}/{fnm}", binary) def get(self, bucket, fnm, tenant_id=None): return self._operator.read(f"{bucket}/{fnm}") def rm(self, bucket, fnm, tenant_id=None): self._operator.delete(f"{bucket}/{fnm}") self._operator.__init__() def scan(self, bucket, fnm, tenant_id=None): return self._operator.scan(f"{bucket}/{fnm}") def obj_exist(self, bucket, fnm, tenant_id=None): return self._operator.exists(f"{bucket}/{fnm}") def init_db_config(self): try: conn = pymysql.connect( host=self._kwargs['host'], port=int(self._kwargs['port']), user=self._kwargs['user'], password=self._kwargs['password'], database=self._kwargs['database'] ) cursor = conn.cursor() max_packet = self._kwargs.get('max_allowed_packet', 4194304) # Default to 4MB if not specified # Ensure max_packet is a valid integer to prevent SQL injection cursor.execute(SET_MAX_ALLOWED_PACKET_SQL.format(int(max_packet))) conn.commit() cursor.close() conn.close() logging.info(f"Database configuration initialized with max_allowed_packet={max_packet}") except Exception as e: logging.error(f"Failed to initialize database configuration: {str(e)}") raise def init_opendal_mysql_table(self): table_name = self._kwargs['table'] # Validate table name to prevent SQL injection if not re.match(r'^[a-zA-Z0-9_]+$', table_name): raise ValueError(f"Invalid table name: {table_name}") conn = pymysql.connect( host=self._kwargs['host'], port=int(self._kwargs['port']), user=self._kwargs['user'], password=self._kwargs['password'], database=self._kwargs['database'] ) cursor = conn.cursor() cursor.execute(CREATE_TABLE_SQL.format(table_name)) conn.commit() cursor.close() conn.close() logging.info(f"Table `{table_name}` initialized.")
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/base64_image.py
rag/utils/base64_image.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import asyncio import base64 import logging from functools import partial from io import BytesIO from PIL import Image test_image_base64 = "iVBORw0KGgoAAAANSUhEUgAAAGQAAABkCAIAAAD/gAIDAAAA6ElEQVR4nO3QwQ3AIBDAsIP9d25XIC+EZE8QZc18w5l9O+AlZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBWYFZgVmBT+IYAHHLHkdEgAAAABJRU5ErkJggg==" test_image = base64.b64decode(test_image_base64) async def image2id(d: dict, storage_put_func: partial, objname: str, bucket: str = "imagetemps"): import logging from io import BytesIO from rag.svr.task_executor import minio_limiter if "image" not in d: return if not d["image"]: del d["image"] return def encode_image(): with BytesIO() as buf: img = d["image"] if isinstance(img, bytes): buf.write(img) buf.seek(0) return buf.getvalue() if img.mode in ("RGBA", "P"): img = img.convert("RGB") try: img.save(buf, format="JPEG") except OSError as e: logging.warning(f"Saving image exception: {e}") return None buf.seek(0) return buf.getvalue() jpeg_binary = await asyncio.to_thread(encode_image) if jpeg_binary is None: del d["image"] return async with minio_limiter: await asyncio.to_thread( lambda: storage_put_func(bucket=bucket, fnm=objname, binary=jpeg_binary) ) d["img_id"] = f"{bucket}-{objname}" if not isinstance(d["image"], bytes): d["image"].close() del d["image"] def id2image(image_id: str | None, storage_get_func: partial): if not image_id: return arr = image_id.split("-") if len(arr) != 2: return bkt, nm = image_id.split("-") try: blob = storage_get_func(bucket=bkt, fnm=nm) if not blob: return return Image.open(BytesIO(blob)) except Exception as e: logging.exception(e)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/opensearch_conn.py
rag/utils/opensearch_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import re import json import time import os import copy from opensearchpy import OpenSearch, NotFoundError from opensearchpy import UpdateByQuery, Q, Search, Index from opensearchpy import ConnectionTimeout from common.decorator import singleton from common.file_utils import get_project_base_directory from common.doc_store.doc_store_base import DocStoreConnection, MatchExpr, OrderByExpr, MatchTextExpr, MatchDenseExpr, \ FusionExpr from rag.nlp import is_english, rag_tokenizer from common.constants import PAGERANK_FLD, TAG_FLD from common import settings ATTEMPT_TIME = 2 logger = logging.getLogger('ragflow.opensearch_conn') @singleton class OSConnection(DocStoreConnection): def __init__(self): self.info = {} logger.info(f"Use OpenSearch {settings.OS['hosts']} as the doc engine.") for _ in range(ATTEMPT_TIME): try: self.os = OpenSearch( settings.OS["hosts"].split(","), http_auth=(settings.OS["username"], settings.OS[ "password"]) if "username" in settings.OS and "password" in settings.OS else None, verify_certs=False, timeout=600 ) if self.os: self.info = self.os.info() break except Exception as e: logger.warning(f"{str(e)}. Waiting OpenSearch {settings.OS['hosts']} to be healthy.") time.sleep(5) if not self.os.ping(): msg = f"OpenSearch {settings.OS['hosts']} is unhealthy in 120s." logger.error(msg) raise Exception(msg) v = self.info.get("version", {"number": "2.18.0"}) v = v["number"].split(".")[0] if int(v) < 2: msg = f"OpenSearch version must be greater than or equal to 2, current version: {v}" logger.error(msg) raise Exception(msg) fp_mapping = os.path.join(get_project_base_directory(), "conf", "os_mapping.json") if not os.path.exists(fp_mapping): msg = f"OpenSearch mapping file not found at {fp_mapping}" logger.error(msg) raise Exception(msg) self.mapping = json.load(open(fp_mapping, "r")) logger.info(f"OpenSearch {settings.OS['hosts']} is healthy.") """ Database operations """ def db_type(self) -> str: return "opensearch" def health(self) -> dict: health_dict = dict(self.os.cluster.health()) health_dict["type"] = "opensearch" return health_dict """ Table operations """ def create_idx(self, indexName: str, knowledgebaseId: str, vectorSize: int): if self.index_exist(indexName, knowledgebaseId): return True try: from opensearchpy.client import IndicesClient return IndicesClient(self.os).create(index=indexName, body=self.mapping) except Exception: logger.exception("OSConnection.createIndex error %s" % (indexName)) def delete_idx(self, indexName: str, knowledgebaseId: str): if len(knowledgebaseId) > 0: # The index need to be alive after any kb deletion since all kb under this tenant are in one index. return try: self.os.indices.delete(index=indexName, allow_no_indices=True) except NotFoundError: pass except Exception: logger.exception("OSConnection.deleteIdx error %s" % (indexName)) def index_exist(self, indexName: str, knowledgebaseId: str = None) -> bool: s = Index(indexName, self.os) for i in range(ATTEMPT_TIME): try: return s.exists() except Exception as e: logger.exception("OSConnection.indexExist got exception") if str(e).find("Timeout") > 0 or str(e).find("Conflict") > 0: continue break return False """ CRUD operations """ def search( self, selectFields: list[str], highlightFields: list[str], condition: dict, matchExprs: list[MatchExpr], orderBy: OrderByExpr, offset: int, limit: int, indexNames: str | list[str], knowledgebaseIds: list[str], aggFields: list[str] = [], rank_feature: dict | None = None ): """ Refers to https://github.com/opensearch-project/opensearch-py/blob/main/guides/dsl.md """ use_knn = False if isinstance(indexNames, str): indexNames = indexNames.split(",") assert isinstance(indexNames, list) and len(indexNames) > 0 assert "_id" not in condition bqry = Q("bool", must=[]) condition["kb_id"] = knowledgebaseIds for k, v in condition.items(): if k == "available_int": if v == 0: bqry.filter.append(Q("range", available_int={"lt": 1})) else: bqry.filter.append( Q("bool", must_not=Q("range", available_int={"lt": 1}))) continue if not v: continue if isinstance(v, list): bqry.filter.append(Q("terms", **{k: v})) elif isinstance(v, str) or isinstance(v, int): bqry.filter.append(Q("term", **{k: v})) else: raise Exception( f"Condition `{str(k)}={str(v)}` value type is {str(type(v))}, expected to be int, str or list.") s = Search() vector_similarity_weight = 0.5 for m in matchExprs: if isinstance(m, FusionExpr) and m.method == "weighted_sum" and "weights" in m.fusion_params: assert len(matchExprs) == 3 and isinstance(matchExprs[0], MatchTextExpr) and isinstance(matchExprs[1], MatchDenseExpr) and isinstance( matchExprs[2], FusionExpr) weights = m.fusion_params["weights"] vector_similarity_weight = float(weights.split(",")[1]) knn_query = {} for m in matchExprs: if isinstance(m, MatchTextExpr): minimum_should_match = m.extra_options.get("minimum_should_match", 0.0) if isinstance(minimum_should_match, float): minimum_should_match = str(int(minimum_should_match * 100)) + "%" bqry.must.append(Q("query_string", fields=m.fields, type="best_fields", query=m.matching_text, minimum_should_match=minimum_should_match, boost=1)) bqry.boost = 1.0 - vector_similarity_weight # Elasticsearch has the encapsulation of KNN_search in python sdk # while the Python SDK for OpenSearch does not provide encapsulation for KNN_search, # the following codes implement KNN_search in OpenSearch using DSL # Besides, Opensearch's DSL for KNN_search query syntax differs from that in Elasticsearch, I also made some adaptions for it elif isinstance(m, MatchDenseExpr): assert (bqry is not None) similarity = 0.0 if "similarity" in m.extra_options: similarity = m.extra_options["similarity"] use_knn = True vector_column_name = m.vector_column_name knn_query[vector_column_name] = {} knn_query[vector_column_name]["vector"] = list(m.embedding_data) knn_query[vector_column_name]["k"] = m.topn knn_query[vector_column_name]["filter"] = bqry.to_dict() knn_query[vector_column_name]["boost"] = similarity if bqry and rank_feature: for fld, sc in rank_feature.items(): if fld != PAGERANK_FLD: fld = f"{TAG_FLD}.{fld}" bqry.should.append(Q("rank_feature", field=fld, linear={}, boost=sc)) if bqry: s = s.query(bqry) for field in highlightFields: s = s.highlight(field, force_source=True, no_match_size=30, require_field_match=False) if orderBy: orders = list() for field, order in orderBy.fields: order = "asc" if order == 0 else "desc" if field in ["page_num_int", "top_int"]: order_info = {"order": order, "unmapped_type": "float", "mode": "avg", "numeric_type": "double"} elif field.endswith("_int") or field.endswith("_flt"): order_info = {"order": order, "unmapped_type": "float"} else: order_info = {"order": order, "unmapped_type": "text"} orders.append({field: order_info}) s = s.sort(*orders) for fld in aggFields: s.aggs.bucket(f'aggs_{fld}', 'terms', field=fld, size=1000000) if limit > 0: s = s[offset:offset + limit] q = s.to_dict() logger.debug(f"OSConnection.search {str(indexNames)} query: " + json.dumps(q)) if use_knn: del q["query"] q["query"] = {"knn": knn_query} for i in range(ATTEMPT_TIME): try: res = self.os.search(index=indexNames, body=q, timeout=600, # search_type="dfs_query_then_fetch", track_total_hits=True, _source=True) if str(res.get("timed_out", "")).lower() == "true": raise Exception("OpenSearch Timeout.") logger.debug(f"OSConnection.search {str(indexNames)} res: " + str(res)) return res except Exception as e: logger.exception(f"OSConnection.search {str(indexNames)} query: " + str(q)) if str(e).find("Timeout") > 0: continue raise e logger.error(f"OSConnection.search timeout for {ATTEMPT_TIME} times!") raise Exception("OSConnection.search timeout.") def get(self, chunkId: str, indexName: str, knowledgebaseIds: list[str]) -> dict | None: for i in range(ATTEMPT_TIME): try: res = self.os.get(index=(indexName), id=chunkId, _source=True, ) if str(res.get("timed_out", "")).lower() == "true": raise Exception("Es Timeout.") chunk = res["_source"] chunk["id"] = chunkId return chunk except NotFoundError: return None except Exception as e: logger.exception(f"OSConnection.get({chunkId}) got exception") if str(e).find("Timeout") > 0: continue raise e logger.error(f"OSConnection.get timeout for {ATTEMPT_TIME} times!") raise Exception("OSConnection.get timeout.") def insert(self, documents: list[dict], indexName: str, knowledgebaseId: str = None) -> list[str]: # Refers to https://opensearch.org/docs/latest/api-reference/document-apis/bulk/ operations = [] for d in documents: assert "_id" not in d assert "id" in d d_copy = copy.deepcopy(d) meta_id = d_copy.pop("id", "") operations.append( {"index": {"_index": indexName, "_id": meta_id}}) operations.append(d_copy) res = [] for _ in range(ATTEMPT_TIME): try: res = [] r = self.os.bulk(index=(indexName), body=operations, refresh=False, timeout=60) if re.search(r"False", str(r["errors"]), re.IGNORECASE): return res for item in r["items"]: for action in ["create", "delete", "index", "update"]: if action in item and "error" in item[action]: res.append(str(item[action]["_id"]) + ":" + str(item[action]["error"])) return res except Exception as e: res.append(str(e)) logger.warning("OSConnection.insert got exception: " + str(e)) res = [] if re.search(r"(Timeout|time out)", str(e), re.IGNORECASE): res.append(str(e)) time.sleep(3) continue return res def update(self, condition: dict, newValue: dict, indexName: str, knowledgebaseId: str) -> bool: doc = copy.deepcopy(newValue) doc.pop("id", None) if "id" in condition and isinstance(condition["id"], str): # update specific single document chunkId = condition["id"] for i in range(ATTEMPT_TIME): try: self.os.update(index=indexName, id=chunkId, body={"doc": doc}) return True except Exception as e: logger.exception( f"OSConnection.update(index={indexName}, id={id}, doc={json.dumps(condition, ensure_ascii=False)}) got exception") if re.search(r"(timeout|connection)", str(e).lower()): continue break return False # update unspecific maybe-multiple documents bqry = Q("bool") for k, v in condition.items(): if not isinstance(k, str) or not v: continue if k == "exists": bqry.filter.append(Q("exists", field=v)) continue if isinstance(v, list): bqry.filter.append(Q("terms", **{k: v})) elif isinstance(v, str) or isinstance(v, int): bqry.filter.append(Q("term", **{k: v})) else: raise Exception( f"Condition `{str(k)}={str(v)}` value type is {str(type(v))}, expected to be int, str or list.") scripts = [] params = {} for k, v in newValue.items(): if k == "remove": if isinstance(v, str): scripts.append(f"ctx._source.remove('{v}');") if isinstance(v, dict): for kk, vv in v.items(): scripts.append(f"int i=ctx._source.{kk}.indexOf(params.p_{kk});ctx._source.{kk}.remove(i);") params[f"p_{kk}"] = vv continue if k == "add": if isinstance(v, dict): for kk, vv in v.items(): scripts.append(f"ctx._source.{kk}.add(params.pp_{kk});") params[f"pp_{kk}"] = vv.strip() continue if (not isinstance(k, str) or not v) and k != "available_int": continue if isinstance(v, str): v = re.sub(r"(['\n\r]|\\.)", " ", v) params[f"pp_{k}"] = v scripts.append(f"ctx._source.{k}=params.pp_{k};") elif isinstance(v, int) or isinstance(v, float): scripts.append(f"ctx._source.{k}={v};") elif isinstance(v, list): scripts.append(f"ctx._source.{k}=params.pp_{k};") params[f"pp_{k}"] = json.dumps(v, ensure_ascii=False) else: raise Exception( f"newValue `{str(k)}={str(v)}` value type is {str(type(v))}, expected to be int, str.") ubq = UpdateByQuery( index=indexName).using( self.os).query(bqry) ubq = ubq.script(source="".join(scripts), params=params) ubq = ubq.params(refresh=True) ubq = ubq.params(slices=5) ubq = ubq.params(conflicts="proceed") for _ in range(ATTEMPT_TIME): try: _ = ubq.execute() return True except Exception as e: logger.error("OSConnection.update got exception: " + str(e) + "\n".join(scripts)) if re.search(r"(timeout|connection|conflict)", str(e).lower()): continue break return False def delete(self, condition: dict, indexName: str, knowledgebaseId: str) -> int: qry = None assert "_id" not in condition if "id" in condition: chunk_ids = condition["id"] if not isinstance(chunk_ids, list): chunk_ids = [chunk_ids] if not chunk_ids: # when chunk_ids is empty, delete all qry = Q("match_all") else: qry = Q("ids", values=chunk_ids) else: qry = Q("bool") for k, v in condition.items(): if k == "exists": qry.filter.append(Q("exists", field=v)) elif k == "must_not": if isinstance(v, dict): for kk, vv in v.items(): if kk == "exists": qry.must_not.append(Q("exists", field=vv)) elif isinstance(v, list): qry.must.append(Q("terms", **{k: v})) elif isinstance(v, str) or isinstance(v, int): qry.must.append(Q("term", **{k: v})) else: raise Exception("Condition value must be int, str or list.") logger.debug("OSConnection.delete query: " + json.dumps(qry.to_dict())) for _ in range(ATTEMPT_TIME): try: # print(Search().query(qry).to_dict(), flush=True) res = self.os.delete_by_query( index=indexName, body=Search().query(qry).to_dict(), refresh=True) return res["deleted"] except Exception as e: logger.warning("OSConnection.delete got exception: " + str(e)) if re.search(r"(timeout|connection)", str(e).lower()): time.sleep(3) continue if re.search(r"(not_found)", str(e), re.IGNORECASE): return 0 return 0 """ Helper functions for search result """ def get_total(self, res): if isinstance(res["hits"]["total"], type({})): return res["hits"]["total"]["value"] return res["hits"]["total"] def get_doc_ids(self, res): return [d["_id"] for d in res["hits"]["hits"]] def __getSource(self, res): rr = [] for d in res["hits"]["hits"]: d["_source"]["id"] = d["_id"] d["_source"]["_score"] = d["_score"] rr.append(d["_source"]) return rr def get_fields(self, res, fields: list[str]) -> dict[str, dict]: res_fields = {} if not fields: return {} for d in self.__getSource(res): m = {n: d.get(n) for n in fields if d.get(n) is not None} for n, v in m.items(): if isinstance(v, list): m[n] = v continue if not isinstance(v, str): m[n] = str(m[n]) # if n.find("tks") > 0: # m[n] = remove_redundant_spaces(m[n]) if m: res_fields[d["id"]] = m return res_fields def get_highlight(self, res, keywords: list[str], fieldnm: str): ans = {} for d in res["hits"]["hits"]: hlts = d.get("highlight") if not hlts: continue txt = "...".join([a for a in list(hlts.items())[0][1]]) if not is_english(txt.split()): ans[d["_id"]] = txt continue txt = d["_source"][fieldnm] txt = re.sub(r"[\r\n]", " ", txt, flags=re.IGNORECASE | re.MULTILINE) txts = [] for t in re.split(r"[.?!;\n]", txt): for w in keywords: t = re.sub(r"(^|[ .?/'\"\(\)!,:;-])(%s)([ .?/'\"\(\)!,:;-])" % re.escape(w), r"\1<em>\2</em>\3", t, flags=re.IGNORECASE | re.MULTILINE) if not re.search(r"<em>[^<>]+</em>", t, flags=re.IGNORECASE | re.MULTILINE): continue txts.append(t) ans[d["_id"]] = "...".join(txts) if txts else "...".join([a for a in list(hlts.items())[0][1]]) return ans def get_aggregation(self, res, fieldnm: str): agg_field = "aggs_" + fieldnm if "aggregations" not in res or agg_field not in res["aggregations"]: return list() bkts = res["aggregations"][agg_field]["buckets"] return [(b["key"], b["doc_count"]) for b in bkts] """ SQL """ def sql(self, sql: str, fetch_size: int, format: str): logger.debug(f"OSConnection.sql get sql: {sql}") sql = re.sub(r"[ `]+", " ", sql) sql = sql.replace("%", "") replaces = [] for r in re.finditer(r" ([a-z_]+_l?tks)( like | ?= ?)'([^']+)'", sql): fld, v = r.group(1), r.group(3) match = " MATCH({}, '{}', 'operator=OR;minimum_should_match=30%') ".format( fld, rag_tokenizer.fine_grained_tokenize(rag_tokenizer.tokenize(v))) replaces.append( ("{}{}'{}'".format( r.group(1), r.group(2), r.group(3)), match)) for p, r in replaces: sql = sql.replace(p, r, 1) logger.debug(f"OSConnection.sql to os: {sql}") for i in range(ATTEMPT_TIME): try: res = self.os.sql.query(body={"query": sql, "fetch_size": fetch_size}, format=format, request_timeout="2s") return res except ConnectionTimeout: logger.exception("OSConnection.sql timeout") continue except Exception: logger.exception("OSConnection.sql got exception") return None logger.error(f"OSConnection.sql timeout for {ATTEMPT_TIME} times!") return None
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/azure_spn_conn.py
rag/utils/azure_spn_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import os import time from common.decorator import singleton from azure.identity import ClientSecretCredential, AzureAuthorityHosts from azure.storage.filedatalake import FileSystemClient from common import settings @singleton class RAGFlowAzureSpnBlob: def __init__(self): self.conn = None self.account_url = os.getenv('ACCOUNT_URL', settings.AZURE["account_url"]) self.client_id = os.getenv('CLIENT_ID', settings.AZURE["client_id"]) self.secret = os.getenv('SECRET', settings.AZURE["secret"]) self.tenant_id = os.getenv('TENANT_ID', settings.AZURE["tenant_id"]) self.container_name = os.getenv('CONTAINER_NAME', settings.AZURE["container_name"]) self.__open__() def __open__(self): try: if self.conn: self.__close__() except Exception: pass try: credentials = ClientSecretCredential(tenant_id=self.tenant_id, client_id=self.client_id, client_secret=self.secret, authority=AzureAuthorityHosts.AZURE_CHINA) self.conn = FileSystemClient(account_url=self.account_url, file_system_name=self.container_name, credential=credentials) except Exception: logging.exception("Fail to connect %s" % self.account_url) def __close__(self): del self.conn self.conn = None def health(self): _bucket, fnm, binary = "txtxtxtxt1", "txtxtxtxt1", b"_t@@@1" f = self.conn.create_file(fnm) f.append_data(binary, offset=0, length=len(binary)) return f.flush_data(len(binary)) def put(self, bucket, fnm, binary): for _ in range(3): try: f = self.conn.create_file(fnm) f.append_data(binary, offset=0, length=len(binary)) return f.flush_data(len(binary)) except Exception: logging.exception(f"Fail put {bucket}/{fnm}") self.__open__() time.sleep(1) return None return None def rm(self, bucket, fnm): try: self.conn.delete_file(fnm) except Exception: logging.exception(f"Fail rm {bucket}/{fnm}") def get(self, bucket, fnm): for _ in range(1): try: client = self.conn.get_file_client(fnm) r = client.download_file() return r.read() except Exception: logging.exception(f"fail get {bucket}/{fnm}") self.__open__() time.sleep(1) return None def obj_exist(self, bucket, fnm): try: client = self.conn.get_file_client(fnm) return client.exists() except Exception: logging.exception(f"Fail put {bucket}/{fnm}") return False def get_presigned_url(self, bucket, fnm, expires): for _ in range(10): try: return self.conn.get_presigned_url("GET", bucket, fnm, expires) except Exception: logging.exception(f"fail get {bucket}/{fnm}") self.__open__() time.sleep(1) return None
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/gcs_conn.py
rag/utils/gcs_conn.py
# Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import time import datetime from io import BytesIO from google.cloud import storage from google.api_core.exceptions import NotFound from common.decorator import singleton from common import settings @singleton class RAGFlowGCS: def __init__(self): self.client = None self.bucket_name = None self.__open__() def __open__(self): try: if self.client: self.client = None except Exception: pass try: self.client = storage.Client() self.bucket_name = settings.GCS["bucket"] except Exception: logging.exception("Fail to connect to GCS") def _get_blob_path(self, folder, filename): """Helper to construct the path: folder/filename""" if not folder: return filename return f"{folder}/{filename}" def health(self): folder, fnm, binary = "ragflow-health", "health_check", b"_t@@@1" try: bucket_obj = self.client.bucket(self.bucket_name) if not bucket_obj.exists(): logging.error(f"Health check failed: Main bucket '{self.bucket_name}' does not exist.") return False blob_path = self._get_blob_path(folder, fnm) blob = bucket_obj.blob(blob_path) blob.upload_from_file(BytesIO(binary), content_type='application/octet-stream') return True except Exception as e: logging.exception(f"Health check failed: {e}") return False def put(self, bucket, fnm, binary, tenant_id=None): # RENAMED PARAMETER: bucket_name -> bucket (to match interface) for _ in range(3): try: bucket_obj = self.client.bucket(self.bucket_name) blob_path = self._get_blob_path(bucket, fnm) blob = bucket_obj.blob(blob_path) blob.upload_from_file(BytesIO(binary), content_type='application/octet-stream') return True except NotFound: logging.error(f"Fail to put: Main bucket {self.bucket_name} does not exist.") return False except Exception: logging.exception(f"Fail to put {bucket}/{fnm}:") self.__open__() time.sleep(1) return False def rm(self, bucket, fnm, tenant_id=None): # RENAMED PARAMETER: bucket_name -> bucket try: bucket_obj = self.client.bucket(self.bucket_name) blob_path = self._get_blob_path(bucket, fnm) blob = bucket_obj.blob(blob_path) blob.delete() except NotFound: pass except Exception: logging.exception(f"Fail to remove {bucket}/{fnm}:") def get(self, bucket, filename, tenant_id=None): # RENAMED PARAMETER: bucket_name -> bucket for _ in range(1): try: bucket_obj = self.client.bucket(self.bucket_name) blob_path = self._get_blob_path(bucket, filename) blob = bucket_obj.blob(blob_path) return blob.download_as_bytes() except NotFound: logging.warning(f"File not found {bucket}/{filename} in {self.bucket_name}") return None except Exception: logging.exception(f"Fail to get {bucket}/{filename}") self.__open__() time.sleep(1) return None def obj_exist(self, bucket, filename, tenant_id=None): # RENAMED PARAMETER: bucket_name -> bucket try: bucket_obj = self.client.bucket(self.bucket_name) blob_path = self._get_blob_path(bucket, filename) blob = bucket_obj.blob(blob_path) return blob.exists() except Exception: logging.exception(f"obj_exist {bucket}/{filename} got exception") return False def bucket_exists(self, bucket): # RENAMED PARAMETER: bucket_name -> bucket try: bucket_obj = self.client.bucket(self.bucket_name) return bucket_obj.exists() except Exception: logging.exception(f"bucket_exist check for {self.bucket_name} got exception") return False def get_presigned_url(self, bucket, fnm, expires, tenant_id=None): # RENAMED PARAMETER: bucket_name -> bucket for _ in range(10): try: bucket_obj = self.client.bucket(self.bucket_name) blob_path = self._get_blob_path(bucket, fnm) blob = bucket_obj.blob(blob_path) expiration = expires if isinstance(expires, int): expiration = datetime.timedelta(seconds=expires) url = blob.generate_signed_url( version="v4", expiration=expiration, method="GET" ) return url except Exception: logging.exception(f"Fail to get_presigned {bucket}/{fnm}:") self.__open__() time.sleep(1) return None def remove_bucket(self, bucket): # RENAMED PARAMETER: bucket_name -> bucket try: bucket_obj = self.client.bucket(self.bucket_name) prefix = f"{bucket}/" blobs = list(self.client.list_blobs(self.bucket_name, prefix=prefix)) if blobs: bucket_obj.delete_blobs(blobs) except Exception: logging.exception(f"Fail to remove virtual bucket (folder) {bucket}") def copy(self, src_bucket, src_path, dest_bucket, dest_path): # RENAMED PARAMETERS to match original interface try: bucket_obj = self.client.bucket(self.bucket_name) src_blob_path = self._get_blob_path(src_bucket, src_path) dest_blob_path = self._get_blob_path(dest_bucket, dest_path) src_blob = bucket_obj.blob(src_blob_path) if not src_blob.exists(): logging.error(f"Source object not found: {src_blob_path}") return False bucket_obj.copy_blob(src_blob, bucket_obj, dest_blob_path) return True except NotFound: logging.error(f"Copy failed: Main bucket {self.bucket_name} does not exist.") return False except Exception: logging.exception(f"Fail to copy {src_bucket}/{src_path} -> {dest_bucket}/{dest_path}") return False def move(self, src_bucket, src_path, dest_bucket, dest_path): try: if self.copy(src_bucket, src_path, dest_bucket, dest_path): self.rm(src_bucket, src_path) return True else: logging.error(f"Copy failed, move aborted: {src_bucket}/{src_path}") return False except Exception: logging.exception(f"Fail to move {src_bucket}/{src_path} -> {dest_bucket}/{dest_path}") return False
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/utils/ob_conn.py
rag/utils/ob_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json import logging import os import re import threading import time from typing import Any, Optional import numpy as np from elasticsearch_dsl import Q, Search from pydantic import BaseModel from pymysql.converters import escape_string from pyobvector import ObVecClient, FtsIndexParam, FtsParser, ARRAY, VECTOR from pyobvector.client import ClusterVersionException from pyobvector.client.hybrid_search import HybridSearch from pyobvector.util import ObVersion from sqlalchemy import text, Column, String, Integer, JSON, Double, Row, Table from sqlalchemy.dialects.mysql import LONGTEXT, TEXT from sqlalchemy.sql.type_api import TypeEngine from common import settings from common.constants import PAGERANK_FLD, TAG_FLD from common.decorator import singleton from common.float_utils import get_float from common.doc_store.doc_store_base import DocStoreConnection, MatchExpr, OrderByExpr, FusionExpr, MatchTextExpr, \ MatchDenseExpr from rag.nlp import rag_tokenizer ATTEMPT_TIME = 2 OB_QUERY_TIMEOUT = int(os.environ.get("OB_QUERY_TIMEOUT", "100_000_000")) logger = logging.getLogger('ragflow.ob_conn') column_order_id = Column("_order_id", Integer, nullable=True, comment="chunk order id for maintaining sequence") column_group_id = Column("group_id", String(256), nullable=True, comment="group id for external retrieval") column_definitions: list[Column] = [ Column("id", String(256), primary_key=True, comment="chunk id"), Column("kb_id", String(256), nullable=False, index=True, comment="knowledge base id"), Column("doc_id", String(256), nullable=True, index=True, comment="document id"), Column("docnm_kwd", String(256), nullable=True, comment="document name"), Column("doc_type_kwd", String(256), nullable=True, comment="document type"), Column("title_tks", String(256), nullable=True, comment="title tokens"), Column("title_sm_tks", String(256), nullable=True, comment="fine-grained (small) title tokens"), Column("content_with_weight", LONGTEXT, nullable=True, comment="the original content"), Column("content_ltks", LONGTEXT, nullable=True, comment="long text tokens derived from content_with_weight"), Column("content_sm_ltks", LONGTEXT, nullable=True, comment="fine-grained (small) tokens derived from content_ltks"), Column("pagerank_fea", Integer, nullable=True, comment="page rank priority, usually set in kb level"), Column("important_kwd", ARRAY(String(256)), nullable=True, comment="keywords"), Column("important_tks", TEXT, nullable=True, comment="keyword tokens"), Column("question_kwd", ARRAY(String(1024)), nullable=True, comment="questions"), Column("question_tks", TEXT, nullable=True, comment="question tokens"), Column("tag_kwd", ARRAY(String(256)), nullable=True, comment="tags"), Column("tag_feas", JSON, nullable=True, comment="tag features used for 'rank_feature', format: [tag -> relevance score]"), Column("available_int", Integer, nullable=False, index=True, server_default="1", comment="status of availability, 0 for unavailable, 1 for available"), Column("create_time", String(19), nullable=True, comment="creation time in YYYY-MM-DD HH:MM:SS format"), Column("create_timestamp_flt", Double, nullable=True, comment="creation timestamp in float format"), Column("img_id", String(128), nullable=True, comment="image id"), Column("position_int", ARRAY(ARRAY(Integer)), nullable=True, comment="position"), Column("page_num_int", ARRAY(Integer), nullable=True, comment="page number"), Column("top_int", ARRAY(Integer), nullable=True, comment="rank from the top"), Column("knowledge_graph_kwd", String(256), nullable=True, index=True, comment="knowledge graph chunk type"), Column("source_id", ARRAY(String(256)), nullable=True, comment="source document id"), Column("entity_kwd", String(256), nullable=True, comment="entity name"), Column("entity_type_kwd", String(256), nullable=True, index=True, comment="entity type"), Column("from_entity_kwd", String(256), nullable=True, comment="the source entity of this edge"), Column("to_entity_kwd", String(256), nullable=True, comment="the target entity of this edge"), Column("weight_int", Integer, nullable=True, comment="the weight of this edge"), Column("weight_flt", Double, nullable=True, comment="the weight of community report"), Column("entities_kwd", ARRAY(String(256)), nullable=True, comment="node ids of entities"), Column("rank_flt", Double, nullable=True, comment="rank of this entity"), Column("removed_kwd", String(256), nullable=True, index=True, server_default="'N'", comment="whether it has been deleted"), Column("metadata", JSON, nullable=True, comment="metadata for this chunk"), Column("extra", JSON, nullable=True, comment="extra information of non-general chunk"), column_order_id, column_group_id, ] column_names: list[str] = [col.name for col in column_definitions] column_types: dict[str, TypeEngine] = {col.name: col.type for col in column_definitions} array_columns: list[str] = [col.name for col in column_definitions if isinstance(col.type, ARRAY)] vector_column_pattern = re.compile(r"q_(?P<vector_size>\d+)_vec") index_columns: list[str] = [ "kb_id", "doc_id", "available_int", "knowledge_graph_kwd", "entity_type_kwd", "removed_kwd", ] fts_columns_origin: list[str] = [ "docnm_kwd^10", "content_with_weight", "important_tks^20", "question_tks^20", ] fts_columns_tks: list[str] = [ "title_tks^10", "title_sm_tks^5", "important_tks^20", "question_tks^20", "content_ltks^2", "content_sm_ltks", ] index_name_template = "ix_%s_%s" fulltext_index_name_template = "fts_idx_%s" # MATCH AGAINST: https://www.oceanbase.com/docs/common-oceanbase-database-cn-1000000002017607 fulltext_search_template = "MATCH (%s) AGAINST ('%s' IN NATURAL LANGUAGE MODE)" # cosine_distance: https://www.oceanbase.com/docs/common-oceanbase-database-cn-1000000002012938 vector_search_template = "cosine_distance(%s, '%s')" class SearchResult(BaseModel): total: int chunks: list[dict] def get_column_value(column_name: str, value: Any) -> Any: if column_name in column_types: column_type = column_types[column_name] if isinstance(column_type, String): return str(value) elif isinstance(column_type, Integer): return int(value) elif isinstance(column_type, Double): return float(value) elif isinstance(column_type, ARRAY) or isinstance(column_type, JSON): if isinstance(value, str): try: return json.loads(value) except json.JSONDecodeError: return value else: return value else: raise ValueError(f"Unsupported column type for column '{column_name}': {column_type}") elif vector_column_pattern.match(column_name): if isinstance(value, str): try: return json.loads(value) except json.JSONDecodeError: return value else: return value elif column_name == "_score": return float(value) else: raise ValueError(f"Unknown column '{column_name}' with value '{value}'.") def get_default_value(column_name: str) -> Any: if column_name == "available_int": return 1 elif column_name == "removed_kwd": return "N" elif column_name == "_order_id": return 0 else: return None def get_value_str(value: Any) -> str: if isinstance(value, str): cleaned_str = value.replace('\\', '\\\\') cleaned_str = cleaned_str.replace('\n', '\\n') cleaned_str = cleaned_str.replace('\r', '\\r') cleaned_str = cleaned_str.replace('\t', '\\t') return f"'{escape_string(cleaned_str)}'" elif isinstance(value, bool): return "true" if value else "false" elif value is None: return "NULL" elif isinstance(value, (list, dict)): json_str = json.dumps(value, ensure_ascii=False) return f"'{escape_string(json_str)}'" else: return str(value) def get_metadata_filter_expression(metadata_filtering_conditions: dict) -> str: """ Convert metadata filtering conditions to MySQL JSON path expression. Args: metadata_filtering_conditions: dict with 'conditions' and 'logical_operator' keys Returns: MySQL JSON path expression string """ if not metadata_filtering_conditions: return "" conditions = metadata_filtering_conditions.get("conditions", []) logical_operator = metadata_filtering_conditions.get("logical_operator", "and").upper() if not conditions: return "" if logical_operator not in ["AND", "OR"]: raise ValueError(f"Unsupported logical operator: {logical_operator}. Only 'and' and 'or' are supported.") metadata_filters = [] for condition in conditions: name = condition.get("name") comparison_operator = condition.get("comparison_operator") value = condition.get("value") if not all([name, comparison_operator]): continue expr = f"JSON_EXTRACT(metadata, '$.{name}')" value_str = get_value_str(value) if value else "" # Convert comparison operator to MySQL JSON path syntax if comparison_operator == "is": # JSON_EXTRACT(metadata, '$.field_name') = 'value' metadata_filters.append(f"{expr} = {value_str}") elif comparison_operator == "is not": metadata_filters.append(f"{expr} != {value_str}") elif comparison_operator == "contains": metadata_filters.append(f"JSON_CONTAINS({expr}, {value_str})") elif comparison_operator == "not contains": metadata_filters.append(f"NOT JSON_CONTAINS({expr}, {value_str})") elif comparison_operator == "start with": metadata_filters.append(f"{expr} LIKE CONCAT({value_str}, '%')") elif comparison_operator == "end with": metadata_filters.append(f"{expr} LIKE CONCAT('%', {value_str})") elif comparison_operator == "empty": metadata_filters.append(f"({expr} IS NULL OR {expr} = '' OR {expr} = '[]' OR {expr} = '{{}}')") elif comparison_operator == "not empty": metadata_filters.append(f"({expr} IS NOT NULL AND {expr} != '' AND {expr} != '[]' AND {expr} != '{{}}')") # Number operators elif comparison_operator == "=": metadata_filters.append(f"CAST({expr} AS DECIMAL(20,10)) = {value_str}") elif comparison_operator == "≠": metadata_filters.append(f"CAST({expr} AS DECIMAL(20,10)) != {value_str}") elif comparison_operator == ">": metadata_filters.append(f"CAST({expr} AS DECIMAL(20,10)) > {value_str}") elif comparison_operator == "<": metadata_filters.append(f"CAST({expr} AS DECIMAL(20,10)) < {value_str}") elif comparison_operator == "≥": metadata_filters.append(f"CAST({expr} AS DECIMAL(20,10)) >= {value_str}") elif comparison_operator == "≤": metadata_filters.append(f"CAST({expr} AS DECIMAL(20,10)) <= {value_str}") # Time operators elif comparison_operator == "before": metadata_filters.append(f"CAST({expr} AS DATETIME) < {value_str}") elif comparison_operator == "after": metadata_filters.append(f"CAST({expr} AS DATETIME) > {value_str}") else: logger.warning(f"Unsupported comparison operator: {comparison_operator}") continue if not metadata_filters: return "" return f"({f' {logical_operator} '.join(metadata_filters)})" def get_filters(condition: dict) -> list[str]: filters: list[str] = [] for k, v in condition.items(): if not v: continue if k == "exists": filters.append(f"{v} IS NOT NULL") elif k == "must_not" and isinstance(v, dict) and "exists" in v: filters.append(f"{v.get('exists')} IS NULL") elif k == "metadata_filtering_conditions": # Handle metadata filtering conditions metadata_filter = get_metadata_filter_expression(v) if metadata_filter: filters.append(metadata_filter) elif k in array_columns: if isinstance(v, list): array_filters = [] for vv in v: array_filters.append(f"array_contains({k}, {get_value_str(vv)})") array_filter = " OR ".join(array_filters) filters.append(f"({array_filter})") else: filters.append(f"array_contains({k}, {get_value_str(v)})") elif isinstance(v, list): values: list[str] = [] for item in v: values.append(get_value_str(item)) value = ", ".join(values) filters.append(f"{k} IN ({value})") else: filters.append(f"{k} = {get_value_str(v)}") return filters def _try_with_lock(lock_name: str, process_func, check_func, timeout: int = None): if not timeout: timeout = int(os.environ.get("OB_DDL_TIMEOUT", "60")) if not check_func(): from rag.utils.redis_conn import RedisDistributedLock lock = RedisDistributedLock(lock_name) if lock.acquire(): logger.info(f"acquired lock success: {lock_name}, start processing.") try: process_func() return finally: lock.release() if not check_func(): logger.info(f"Waiting for process complete for {lock_name} on other task executors.") time.sleep(1) count = 1 while count < timeout and not check_func(): count += 1 time.sleep(1) if count >= timeout and not check_func(): raise Exception(f"Timeout to wait for process complete for {lock_name}.") @singleton class OBConnection(DocStoreConnection): def __init__(self): scheme: str = settings.OB.get("scheme") ob_config = settings.OB.get("config", {}) if scheme and scheme.lower() == "mysql": mysql_config = settings.get_base_config("mysql", {}) logger.info("Use MySQL scheme to create OceanBase connection.") host = mysql_config.get("host", "localhost") port = mysql_config.get("port", 2881) self.username = mysql_config.get("user", "root@test") self.password = mysql_config.get("password", "infini_rag_flow") max_connections = mysql_config.get("max_connections", 300) else: logger.info("Use customized config to create OceanBase connection.") host = ob_config.get("host", "localhost") port = ob_config.get("port", 2881) self.username = ob_config.get("user", "root@test") self.password = ob_config.get("password", "infini_rag_flow") max_connections = ob_config.get("max_connections", 300) self.db_name = ob_config.get("db_name", "test") self.uri = f"{host}:{port}" logger.info(f"Use OceanBase '{self.uri}' as the doc engine.") # Set the maximum number of connections that can be created above the pool_size. # By default, this is half of max_connections, but at least 10. # This allows the pool to handle temporary spikes in demand without exhausting resources. max_overflow = int(os.environ.get("OB_MAX_OVERFLOW", max(max_connections // 2, 10))) # Set the number of seconds to wait before giving up when trying to get a connection from the pool. # Default is 30 seconds, but can be overridden with the OB_POOL_TIMEOUT environment variable. pool_timeout = int(os.environ.get("OB_POOL_TIMEOUT", "30")) for _ in range(ATTEMPT_TIME): try: self.client = ObVecClient( uri=self.uri, user=self.username, password=self.password, db_name=self.db_name, pool_pre_ping=True, pool_recycle=3600, pool_size=max_connections, max_overflow=max_overflow, pool_timeout=pool_timeout, ) break except Exception as e: logger.warning(f"{str(e)}. Waiting OceanBase {self.uri} to be healthy.") time.sleep(5) if self.client is None: msg = f"OceanBase {self.uri} connection failed after {ATTEMPT_TIME} attempts." logger.error(msg) raise Exception(msg) self._load_env_vars() self._check_ob_version() self._try_to_update_ob_query_timeout() self.es = None if self.enable_hybrid_search: try: self.es = HybridSearch( uri=self.uri, user=self.username, password=self.password, db_name=self.db_name, pool_pre_ping=True, pool_recycle=3600, pool_size=max_connections, max_overflow=max_overflow, pool_timeout=pool_timeout, ) logger.info("OceanBase Hybrid Search feature is enabled") except ClusterVersionException as e: logger.info("Failed to initialize HybridSearch client, fallback to use SQL", exc_info=e) self.es = None if self.es is not None and self.search_original_content: logger.info("HybridSearch is enabled, forcing search_original_content to False") self.search_original_content = False # Determine which columns to use for full-text search dynamically: # If HybridSearch is enabled (self.es is not None), we must use tokenized columns (fts_columns_tks) # for compatibility and performance with HybridSearch. Otherwise, we use the original content columns # (fts_columns_origin), which may be controlled by an environment variable. self.fulltext_search_columns = fts_columns_origin if self.search_original_content else fts_columns_tks self._table_exists_cache: set[str] = set() self._table_exists_cache_lock = threading.RLock() logger.info(f"OceanBase {self.uri} is healthy.") def _check_ob_version(self): try: res = self.client.perform_raw_text_sql("SELECT OB_VERSION() FROM DUAL").fetchone() version_str = res[0] if res else None logger.info(f"OceanBase {self.uri} version is {version_str}") except Exception as e: raise Exception(f"Failed to get OceanBase version from {self.uri}, error: {str(e)}") if not version_str: raise Exception(f"Failed to get OceanBase version from {self.uri}.") ob_version = ObVersion.from_db_version_string(version_str) if ob_version < ObVersion.from_db_version_nums(4, 3, 5, 1): raise Exception( f"The version of OceanBase needs to be higher than or equal to 4.3.5.1, current version is {version_str}" ) def _try_to_update_ob_query_timeout(self): try: val = self._get_variable_value("ob_query_timeout") if val and int(val) >= OB_QUERY_TIMEOUT: return except Exception as e: logger.warning("Failed to get 'ob_query_timeout' variable: %s", str(e)) try: self.client.perform_raw_text_sql(f"SET GLOBAL ob_query_timeout={OB_QUERY_TIMEOUT}") logger.info("Set GLOBAL variable 'ob_query_timeout' to %d.", OB_QUERY_TIMEOUT) # refresh connection pool to ensure 'ob_query_timeout' has taken effect self.client.engine.dispose() if self.es is not None: self.es.engine.dispose() logger.info("Disposed all connections in engine pool to refresh connection pool") except Exception as e: logger.warning(f"Failed to set 'ob_query_timeout' variable: {str(e)}") def _load_env_vars(self): def is_true(var: str, default: str) -> bool: return os.getenv(var, default).lower() in ['true', '1', 'yes', 'y'] self.enable_fulltext_search = is_true('ENABLE_FULLTEXT_SEARCH', 'true') logger.info(f"ENABLE_FULLTEXT_SEARCH={self.enable_fulltext_search}") self.use_fulltext_hint = is_true('USE_FULLTEXT_HINT', 'true') logger.info(f"USE_FULLTEXT_HINT={self.use_fulltext_hint}") self.search_original_content = is_true("SEARCH_ORIGINAL_CONTENT", 'true') logger.info(f"SEARCH_ORIGINAL_CONTENT={self.search_original_content}") self.enable_hybrid_search = is_true('ENABLE_HYBRID_SEARCH', 'false') logger.info(f"ENABLE_HYBRID_SEARCH={self.enable_hybrid_search}") self.use_fulltext_first_fusion_search = is_true('USE_FULLTEXT_FIRST_FUSION_SEARCH', 'true') logger.info(f"USE_FULLTEXT_FIRST_FUSION_SEARCH={self.use_fulltext_first_fusion_search}") """ Database operations """ def db_type(self) -> str: return "oceanbase" def health(self) -> dict: return { "uri": self.uri, "version_comment": self._get_variable_value("version_comment") } def _get_variable_value(self, var_name: str) -> Any: rows = self.client.perform_raw_text_sql(f"SHOW VARIABLES LIKE '{var_name}'") for row in rows: return row[1] raise Exception(f"Variable '{var_name}' not found.") def _check_table_exists_cached(self, table_name: str) -> bool: """ Check table existence with cache to reduce INFORMATION_SCHEMA queries under high concurrency. Only caches when table exists. Does not cache when table does not exist. Thread-safe implementation: read operations are lock-free (GIL-protected), write operations are protected by RLock to ensure cache consistency. Args: table_name: Table name Returns: Whether the table exists with all required indexes and columns """ if table_name in self._table_exists_cache: return True try: if not self.client.check_table_exists(table_name): return False for column_name in index_columns: if not self._index_exists(table_name, index_name_template % (table_name, column_name)): return False for fts_column in self.fulltext_search_columns: column_name = fts_column.split("^")[0] if not self._index_exists(table_name, fulltext_index_name_template % column_name): return False for column in [column_order_id, column_group_id]: if not self._column_exist(table_name, column.name): return False except Exception as e: raise Exception(f"OBConnection._check_table_exists_cached error: {str(e)}") with self._table_exists_cache_lock: if table_name not in self._table_exists_cache: self._table_exists_cache.add(table_name) return True """ Table operations """ def create_idx(self, indexName: str, knowledgebaseId: str, vectorSize: int): vector_field_name = f"q_{vectorSize}_vec" vector_index_name = f"{vector_field_name}_idx" try: _try_with_lock( lock_name=f"ob_create_table_{indexName}", check_func=lambda: self.client.check_table_exists(indexName), process_func=lambda: self._create_table(indexName), ) for column_name in index_columns: _try_with_lock( lock_name=f"ob_add_idx_{indexName}_{column_name}", check_func=lambda: self._index_exists(indexName, index_name_template % (indexName, column_name)), process_func=lambda: self._add_index(indexName, column_name), ) for fts_column in self.fulltext_search_columns: column_name = fts_column.split("^")[0] _try_with_lock( lock_name=f"ob_add_fulltext_idx_{indexName}_{column_name}", check_func=lambda: self._index_exists(indexName, fulltext_index_name_template % column_name), process_func=lambda: self._add_fulltext_index(indexName, column_name), ) _try_with_lock( lock_name=f"ob_add_vector_column_{indexName}_{vector_field_name}", check_func=lambda: self._column_exist(indexName, vector_field_name), process_func=lambda: self._add_vector_column(indexName, vectorSize), ) _try_with_lock( lock_name=f"ob_add_vector_idx_{indexName}_{vector_field_name}", check_func=lambda: self._index_exists(indexName, vector_index_name), process_func=lambda: self._add_vector_index(indexName, vector_field_name), ) # new columns migration for column in [column_order_id, column_group_id]: _try_with_lock( lock_name=f"ob_add_{column.name}_{indexName}", check_func=lambda: self._column_exist(indexName, column.name), process_func=lambda: self._add_column(indexName, column), ) except Exception as e: raise Exception(f"OBConnection.createIndex error: {str(e)}") finally: # always refresh metadata to make sure it contains the latest table structure self.client.refresh_metadata([indexName]) def delete_idx(self, indexName: str, knowledgebaseId: str): if len(knowledgebaseId) > 0: # The index need to be alive after any kb deletion since all kb under this tenant are in one index. return try: if self.client.check_table_exists(table_name=indexName): self.client.drop_table_if_exist(indexName) logger.info(f"Dropped table '{indexName}'.") except Exception as e: raise Exception(f"OBConnection.deleteIndex error: {str(e)}") def index_exist(self, indexName: str, knowledgebaseId: str = None) -> bool: return self._check_table_exists_cached(indexName) def _get_count(self, table_name: str, filter_list: list[str] = None) -> int: where_clause = "WHERE " + " AND ".join(filter_list) if len(filter_list) > 0 else "" (count,) = self.client.perform_raw_text_sql( f"SELECT COUNT(*) FROM {table_name} {where_clause}" ).fetchone() return count def _column_exist(self, table_name: str, column_name: str) -> bool: return self._get_count( table_name="INFORMATION_SCHEMA.COLUMNS", filter_list=[ f"TABLE_SCHEMA = '{self.db_name}'", f"TABLE_NAME = '{table_name}'", f"COLUMN_NAME = '{column_name}'", ]) > 0 def _index_exists(self, table_name: str, index_name: str) -> bool: return self._get_count( table_name="INFORMATION_SCHEMA.STATISTICS", filter_list=[ f"TABLE_SCHEMA = '{self.db_name}'", f"TABLE_NAME = '{table_name}'", f"INDEX_NAME = '{index_name}'", ]) > 0 def _create_table(self, table_name: str): # remove outdated metadata for external changes if table_name in self.client.metadata_obj.tables: self.client.metadata_obj.remove(Table(table_name, self.client.metadata_obj)) table_options = { "mysql_charset": "utf8mb4", "mysql_collate": "utf8mb4_unicode_ci", "mysql_organization": "heap", } self.client.create_table( table_name=table_name, columns=column_definitions, **table_options, ) logger.info(f"Created table '{table_name}'.") def _add_index(self, table_name: str, column_name: str): index_name = index_name_template % (table_name, column_name) self.client.create_index( table_name=table_name, is_vec_index=False, index_name=index_name, column_names=[column_name], ) logger.info(f"Created index '{index_name}' on table '{table_name}'.") def _add_fulltext_index(self, table_name: str, column_name: str): fulltext_index_name = fulltext_index_name_template % column_name self.client.create_fts_idx_with_fts_index_param( table_name=table_name, fts_idx_param=FtsIndexParam( index_name=fulltext_index_name, field_names=[column_name], parser_type=FtsParser.IK, ), ) logger.info(f"Created full text index '{fulltext_index_name}' on table '{table_name}'.") def _add_vector_column(self, table_name: str, vector_size: int): vector_field_name = f"q_{vector_size}_vec" self.client.add_columns( table_name=table_name, columns=[Column(vector_field_name, VECTOR(vector_size), nullable=True)], ) logger.info(f"Added vector column '{vector_field_name}' to table '{table_name}'.") def _add_vector_index(self, table_name: str, vector_field_name: str): vector_index_name = f"{vector_field_name}_idx" self.client.create_index( table_name=table_name, is_vec_index=True, index_name=vector_index_name, column_names=[vector_field_name], vidx_params="distance=cosine, type=hnsw, lib=vsag", ) logger.info( f"Created vector index '{vector_index_name}' on table '{table_name}' with column '{vector_field_name}'." ) def _add_column(self, table_name: str, column: Column): try: self.client.add_columns( table_name=table_name, columns=[column], ) logger.info(f"Added column '{column.name}' to table '{table_name}'.") except Exception as e: logger.warning(f"Failed to add column '{column.name}' to table '{table_name}': {str(e)}") """ CRUD operations """ def search( self, selectFields: list[str], highlightFields: list[str], condition: dict, matchExprs: list[MatchExpr], orderBy: OrderByExpr, offset: int, limit: int, indexNames: str | list[str], knowledgebaseIds: list[str], aggFields: list[str] = [], rank_feature: dict | None = None, **kwargs, ): if isinstance(indexNames, str): indexNames = indexNames.split(",") assert isinstance(indexNames, list) and len(indexNames) > 0 indexNames = list(set(indexNames)) if len(matchExprs) == 3: if not self.enable_fulltext_search: # disable fulltext search in fusion search, which means fallback to vector search matchExprs = [m for m in matchExprs if isinstance(m, MatchDenseExpr)] else: for m in matchExprs: if isinstance(m, FusionExpr): weights = m.fusion_params["weights"] vector_similarity_weight = get_float(weights.split(",")[1]) # skip the search if its weight is zero if vector_similarity_weight <= 0.0: matchExprs = [m for m in matchExprs if isinstance(m, MatchTextExpr)] elif vector_similarity_weight >= 1.0:
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/llm/sequence2txt_model.py
rag/llm/sequence2txt_model.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import base64 import io import json import os import re from abc import ABC import tempfile import requests from openai import OpenAI from openai.lib.azure import AzureOpenAI from common.token_utils import num_tokens_from_string class Base(ABC): def __init__(self, key, model_name, **kwargs): """ Abstract base class constructor. Parameters are not stored; initialization is left to subclasses. """ pass def transcription(self, audio_path, **kwargs): audio_file = open(audio_path, "rb") transcription = self.client.audio.transcriptions.create(model=self.model_name, file=audio_file) return transcription.text.strip(), num_tokens_from_string(transcription.text.strip()) def audio2base64(self, audio): if isinstance(audio, bytes): return base64.b64encode(audio).decode("utf-8") if isinstance(audio, io.BytesIO): return base64.b64encode(audio.getvalue()).decode("utf-8") raise TypeError("The input audio file should be in binary format.") class GPTSeq2txt(Base): _FACTORY_NAME = "OpenAI" def __init__(self, key, model_name="whisper-1", base_url="https://api.openai.com/v1", **kwargs): if not base_url: base_url = "https://api.openai.com/v1" self.client = OpenAI(api_key=key, base_url=base_url) self.model_name = model_name class QWenSeq2txt(Base): _FACTORY_NAME = "Tongyi-Qianwen" def __init__(self, key, model_name="qwen-audio-asr", **kwargs): import dashscope dashscope.api_key = key self.model_name = model_name def transcription(self, audio_path): import dashscope if audio_path.startswith("http"): audio_input = audio_path else: audio_input = f"file://{audio_path}" messages = [ { "role": "system", "content": [{"text": ""}] }, { "role": "user", "content": [{"audio": audio_input}] } ] resp = dashscope.MultiModalConversation.call( model=self.model_name, messages=messages, result_format="message", asr_options={ "enable_lid": True, "enable_itn": False } ) try: text = resp["output"]["choices"][0]["message"].content[0]["text"] except Exception as e: text = "**ERROR**: " + str(e) return text, num_tokens_from_string(text) def stream_transcription(self, audio_path): import dashscope if audio_path.startswith("http"): audio_input = audio_path else: audio_input = f"file://{audio_path}" messages = [ { "role": "system", "content": [{"text": ""}] }, { "role": "user", "content": [{"audio": audio_input}] } ] stream = dashscope.MultiModalConversation.call( model=self.model_name, messages=messages, result_format="message", stream=True, asr_options={ "enable_lid": True, "enable_itn": False } ) full = "" for chunk in stream: try: piece = chunk["output"]["choices"][0]["message"].content[0]["text"] full = piece yield {"event": "delta", "text": piece} except Exception as e: yield {"event": "error", "text": str(e)} yield {"event": "final", "text": full} class AzureSeq2txt(Base): _FACTORY_NAME = "Azure-OpenAI" def __init__(self, key, model_name, lang="Chinese", **kwargs): self.client = AzureOpenAI(api_key=key, azure_endpoint=kwargs["base_url"], api_version="2024-02-01") self.model_name = model_name self.lang = lang class XinferenceSeq2txt(Base): _FACTORY_NAME = "Xinference" def __init__(self, key, model_name="whisper-small", **kwargs): self.base_url = kwargs.get("base_url", None) self.model_name = model_name self.key = key def transcription(self, audio, language="zh", prompt=None, response_format="json", temperature=0.7): if isinstance(audio, str): audio_file = open(audio, "rb") audio_data = audio_file.read() audio_file_name = audio.split("/")[-1] else: audio_data = audio audio_file_name = "audio.wav" payload = {"model": self.model_name, "language": language, "prompt": prompt, "response_format": response_format, "temperature": temperature} files = {"file": (audio_file_name, audio_data, "audio/wav")} try: response = requests.post(f"{self.base_url}/v1/audio/transcriptions", files=files, data=payload) response.raise_for_status() result = response.json() if "text" in result: transcription_text = result["text"].strip() return transcription_text, num_tokens_from_string(transcription_text) else: return "**ERROR**: Failed to retrieve transcription.", 0 except requests.exceptions.RequestException as e: return f"**ERROR**: {str(e)}", 0 class TencentCloudSeq2txt(Base): _FACTORY_NAME = "Tencent Cloud" def __init__(self, key, model_name="16k_zh", base_url="https://asr.tencentcloudapi.com"): from tencentcloud.asr.v20190614 import asr_client from tencentcloud.common import credential key = json.loads(key) sid = key.get("tencent_cloud_sid", "") sk = key.get("tencent_cloud_sk", "") cred = credential.Credential(sid, sk) self.client = asr_client.AsrClient(cred, "") self.model_name = model_name def transcription(self, audio, max_retries=60, retry_interval=5): import time from tencentcloud.asr.v20190614 import models from tencentcloud.common.exception.tencent_cloud_sdk_exception import ( TencentCloudSDKException, ) b64 = self.audio2base64(audio) try: # dispatch disk req = models.CreateRecTaskRequest() params = { "EngineModelType": self.model_name, "ChannelNum": 1, "ResTextFormat": 0, "SourceType": 1, "Data": b64, } req.from_json_string(json.dumps(params)) resp = self.client.CreateRecTask(req) # loop query req = models.DescribeTaskStatusRequest() params = {"TaskId": resp.Data.TaskId} req.from_json_string(json.dumps(params)) retries = 0 while retries < max_retries: resp = self.client.DescribeTaskStatus(req) if resp.Data.StatusStr == "success": text = re.sub(r"\[\d+:\d+\.\d+,\d+:\d+\.\d+\]\s*", "", resp.Data.Result).strip() return text, num_tokens_from_string(text) elif resp.Data.StatusStr == "failed": return ( "**ERROR**: Failed to retrieve speech recognition results.", 0, ) else: time.sleep(retry_interval) retries += 1 return "**ERROR**: Max retries exceeded. Task may still be processing.", 0 except TencentCloudSDKException as e: return "**ERROR**: " + str(e), 0 except Exception as e: return "**ERROR**: " + str(e), 0 class GPUStackSeq2txt(Base): _FACTORY_NAME = "GPUStack" def __init__(self, key, model_name, base_url): if not base_url: raise ValueError("url cannot be None") if base_url.split("/")[-1] != "v1": base_url = os.path.join(base_url, "v1") self.base_url = base_url self.model_name = model_name self.key = key class GiteeSeq2txt(Base): _FACTORY_NAME = "GiteeAI" def __init__(self, key, model_name="whisper-1", base_url="https://ai.gitee.com/v1/", **kwargs): if not base_url: base_url = "https://ai.gitee.com/v1/" self.client = OpenAI(api_key=key, base_url=base_url) self.model_name = model_name class DeepInfraSeq2txt(Base): _FACTORY_NAME = "DeepInfra" def __init__(self, key, model_name, base_url="https://api.deepinfra.com/v1/openai", **kwargs): if not base_url: base_url = "https://api.deepinfra.com/v1/openai" self.client = OpenAI(api_key=key, base_url=base_url) self.model_name = model_name class CometAPISeq2txt(Base): _FACTORY_NAME = "CometAPI" def __init__(self, key, model_name="whisper-1", base_url="https://api.cometapi.com/v1", **kwargs): if not base_url: base_url = "https://api.cometapi.com/v1" self.client = OpenAI(api_key=key, base_url=base_url) self.model_name = model_name class DeerAPISeq2txt(Base): _FACTORY_NAME = "DeerAPI" def __init__(self, key, model_name="whisper-1", base_url="https://api.deerapi.com/v1", **kwargs): if not base_url: base_url = "https://api.deerapi.com/v1" self.client = OpenAI(api_key=key, base_url=base_url) self.model_name = model_name class ZhipuSeq2txt(Base): _FACTORY_NAME = "ZHIPU-AI" def __init__(self, key, model_name="glm-asr", base_url="https://open.bigmodel.cn/api/paas/v4", **kwargs): if not base_url: base_url = "https://open.bigmodel.cn/api/paas/v4" self.base_url = base_url self.api_key = key self.model_name = model_name self.gen_conf = kwargs.get("gen_conf", {}) self.stream = kwargs.get("stream", False) def _convert_to_wav(self, input_path): ext = os.path.splitext(input_path)[1].lower() if ext in [".wav", ".mp3"]: return input_path fd, out_path = tempfile.mkstemp(suffix=".wav") os.close(fd) try: import ffmpeg import imageio_ffmpeg as ffmpeg_exe ffmpeg_path = ffmpeg_exe.get_ffmpeg_exe() ( ffmpeg .input(input_path) .output(out_path, ar=16000, ac=1) .overwrite_output() .run(cmd=ffmpeg_path,quiet=True) ) return out_path except Exception as e: raise RuntimeError(f"audio convert failed: {e}") def transcription(self, audio_path): payload = { "model": self.model_name, "temperature": str(self.gen_conf.get("temperature", 0.75)) or "0.75", "stream": self.stream, } headers = {"Authorization": f"Bearer {self.api_key}"} converted = self._convert_to_wav(audio_path) with open(converted, "rb") as audio_file: files = {"file": audio_file} try: response = requests.post( url=f"{self.base_url}/audio/transcriptions", data=payload, files=files, headers=headers, ) body = response.json() if response.status_code == 200: full_content = body["text"] return full_content, num_tokens_from_string(full_content) else: error = body["error"] return f"**ERROR**: code: {error['code']}, message: {error['message']}", 0 except Exception as e: return "**ERROR**: " + str(e), 0
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/llm/tts_model.py
rag/llm/tts_model.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import _thread as thread import base64 import hashlib import hmac import json import queue import re import ssl import time from abc import ABC from datetime import datetime from time import mktime from typing import Annotated, Literal from urllib.parse import urlencode from wsgiref.handlers import format_date_time import httpx import ormsgpack import requests import websocket from pydantic import BaseModel, conint from common.token_utils import num_tokens_from_string class ServeReferenceAudio(BaseModel): audio: bytes text: str class ServeTTSRequest(BaseModel): text: str chunk_length: Annotated[int, conint(ge=100, le=300, strict=True)] = 200 # Audio format format: Literal["wav", "pcm", "mp3"] = "mp3" mp3_bitrate: Literal[64, 128, 192] = 128 # References audios for in-context learning references: list[ServeReferenceAudio] = [] # Reference id # For example, if you want use https://fish.audio/m/7f92f8afb8ec43bf81429cc1c9199cb1/ # Just pass 7f92f8afb8ec43bf81429cc1c9199cb1 reference_id: str | None = None # Normalize text for en & zh, this increase stability for numbers normalize: bool = True # Balance mode will reduce latency to 300ms, but may decrease stability latency: Literal["normal", "balanced"] = "normal" class Base(ABC): def __init__(self, key, model_name, base_url, **kwargs): """ Abstract base class constructor. Parameters are not stored; subclasses should handle their own initialization. """ pass def tts(self, audio): pass def normalize_text(self, text): return re.sub(r"(\*\*|##\d+\$\$|#)", "", text) class FishAudioTTS(Base): _FACTORY_NAME = "Fish Audio" def __init__(self, key, model_name, base_url="https://api.fish.audio/v1/tts"): if not base_url: base_url = "https://api.fish.audio/v1/tts" key = json.loads(key) self.headers = { "api-key": key.get("fish_audio_ak"), "content-type": "application/msgpack", } self.ref_id = key.get("fish_audio_refid") self.base_url = base_url def tts(self, text): from http import HTTPStatus text = self.normalize_text(text) request = ServeTTSRequest(text=text, reference_id=self.ref_id) with httpx.Client() as client: try: with client.stream( method="POST", url=self.base_url, content=ormsgpack.packb(request, option=ormsgpack.OPT_SERIALIZE_PYDANTIC), headers=self.headers, timeout=None, ) as response: if response.status_code == HTTPStatus.OK: for chunk in response.iter_bytes(): yield chunk else: response.raise_for_status() yield num_tokens_from_string(text) except httpx.HTTPStatusError as e: raise RuntimeError(f"**ERROR**: {e}") class QwenTTS(Base): _FACTORY_NAME = "Tongyi-Qianwen" def __init__(self, key, model_name, base_url=""): import dashscope self.model_name = model_name dashscope.api_key = key def tts(self, text): from collections import deque from dashscope.api_entities.dashscope_response import SpeechSynthesisResponse from dashscope.audio.tts import ResultCallback, SpeechSynthesisResult, SpeechSynthesizer class Callback(ResultCallback): def __init__(self) -> None: self.dque = deque() def _run(self): while True: if not self.dque: time.sleep(0) continue val = self.dque.popleft() if val: yield val else: break def on_open(self): pass def on_complete(self): self.dque.append(None) def on_error(self, response: SpeechSynthesisResponse): raise RuntimeError(str(response)) def on_close(self): pass def on_event(self, result: SpeechSynthesisResult): if result.get_audio_frame() is not None: self.dque.append(result.get_audio_frame()) text = self.normalize_text(text) callback = Callback() SpeechSynthesizer.call(model=self.model_name, text=text, callback=callback, format="mp3") try: for data in callback._run(): yield data yield num_tokens_from_string(text) except Exception as e: raise RuntimeError(f"**ERROR**: {e}") class OpenAITTS(Base): _FACTORY_NAME = "OpenAI" def __init__(self, key, model_name="tts-1", base_url="https://api.openai.com/v1"): if not base_url: base_url = "https://api.openai.com/v1" self.api_key = key self.model_name = model_name self.base_url = base_url self.headers = {"Authorization": f"Bearer {self.api_key}", "Content-Type": "application/json"} def tts(self, text, voice="alloy"): text = self.normalize_text(text) payload = {"model": self.model_name, "voice": voice, "input": text} response = requests.post(f"{self.base_url}/audio/speech", headers=self.headers, json=payload, stream=True) if response.status_code != 200: raise Exception(f"**Error**: {response.status_code}, {response.text}") for chunk in response.iter_content(): if chunk: yield chunk class SparkTTS(Base): _FACTORY_NAME = "XunFei Spark" STATUS_FIRST_FRAME = 0 STATUS_CONTINUE_FRAME = 1 STATUS_LAST_FRAME = 2 def __init__(self, key, model_name, base_url=""): key = json.loads(key) self.APPID = key.get("spark_app_id", "xxxxxxx") self.APISecret = key.get("spark_api_secret", "xxxxxxx") self.APIKey = key.get("spark_api_key", "xxxxxx") self.model_name = model_name self.CommonArgs = {"app_id": self.APPID} self.audio_queue = queue.Queue() # 用来存储音频数据 # 生成url def create_url(self): url = "wss://tts-api.xfyun.cn/v2/tts" now = datetime.now() date = format_date_time(mktime(now.timetuple())) signature_origin = "host: " + "ws-api.xfyun.cn" + "\n" signature_origin += "date: " + date + "\n" signature_origin += "GET " + "/v2/tts " + "HTTP/1.1" signature_sha = hmac.new(self.APISecret.encode("utf-8"), signature_origin.encode("utf-8"), digestmod=hashlib.sha256).digest() signature_sha = base64.b64encode(signature_sha).decode(encoding="utf-8") authorization_origin = 'api_key="%s", algorithm="%s", headers="%s", signature="%s"' % (self.APIKey, "hmac-sha256", "host date request-line", signature_sha) authorization = base64.b64encode(authorization_origin.encode("utf-8")).decode(encoding="utf-8") v = {"authorization": authorization, "date": date, "host": "ws-api.xfyun.cn"} url = url + "?" + urlencode(v) return url def tts(self, text): BusinessArgs = {"aue": "lame", "sfl": 1, "auf": "audio/L16;rate=16000", "vcn": self.model_name, "tte": "utf8"} Data = {"status": 2, "text": base64.b64encode(text.encode("utf-8")).decode("utf-8")} CommonArgs = {"app_id": self.APPID} audio_queue = self.audio_queue model_name = self.model_name class Callback: def __init__(self): self.audio_queue = audio_queue def on_message(self, ws, message): message = json.loads(message) code = message["code"] sid = message["sid"] audio = message["data"]["audio"] audio = base64.b64decode(audio) status = message["data"]["status"] if status == 2: ws.close() if code != 0: errMsg = message["message"] raise Exception(f"sid:{sid} call error:{errMsg} code:{code}") else: self.audio_queue.put(audio) def on_error(self, ws, error): raise Exception(error) def on_close(self, ws, close_status_code, close_msg): self.audio_queue.put(None) # None is terminator def on_open(self, ws): def run(*args): d = {"common": CommonArgs, "business": BusinessArgs, "data": Data} ws.send(json.dumps(d)) thread.start_new_thread(run, ()) wsUrl = self.create_url() websocket.enableTrace(False) a = Callback() ws = websocket.WebSocketApp(wsUrl, on_open=a.on_open, on_error=a.on_error, on_close=a.on_close, on_message=a.on_message) status_code = 0 ws.run_forever(sslopt={"cert_reqs": ssl.CERT_NONE}) while True: audio_chunk = self.audio_queue.get() if audio_chunk is None: if status_code == 0: raise Exception(f"Fail to access model({model_name}) using the provided credentials. **ERROR**: Invalid APPID, API Secret, or API Key.") else: break status_code = 1 yield audio_chunk class XinferenceTTS(Base): _FACTORY_NAME = "Xinference" def __init__(self, key, model_name, **kwargs): self.base_url = kwargs.get("base_url", None) self.model_name = model_name self.headers = {"accept": "application/json", "Content-Type": "application/json"} def tts(self, text, voice="中文女", stream=True): payload = {"model": self.model_name, "input": text, "voice": voice} response = requests.post(f"{self.base_url}/v1/audio/speech", headers=self.headers, json=payload, stream=stream) if response.status_code != 200: raise Exception(f"**Error**: {response.status_code}, {response.text}") for chunk in response.iter_content(chunk_size=1024): if chunk: yield chunk class OllamaTTS(Base): def __init__(self, key, model_name="ollama-tts", base_url="https://api.ollama.ai/v1"): if not base_url: base_url = "https://api.ollama.ai/v1" self.model_name = model_name self.base_url = base_url self.headers = {"Content-Type": "application/json"} if key and key != "x": self.headers["Authorization"] = f"Bearer {key}" def tts(self, text, voice="standard-voice"): payload = {"model": self.model_name, "voice": voice, "input": text} response = requests.post(f"{self.base_url}/audio/tts", headers=self.headers, json=payload, stream=True) if response.status_code != 200: raise Exception(f"**Error**: {response.status_code}, {response.text}") for chunk in response.iter_content(): if chunk: yield chunk class GPUStackTTS(Base): _FACTORY_NAME = "GPUStack" def __init__(self, key, model_name, **kwargs): self.base_url = kwargs.get("base_url", None) self.api_key = key self.model_name = model_name self.headers = {"accept": "application/json", "Content-Type": "application/json", "Authorization": f"Bearer {self.api_key}"} def tts(self, text, voice="Chinese Female", stream=True): payload = {"model": self.model_name, "input": text, "voice": voice} response = requests.post(f"{self.base_url}/v1/audio/speech", headers=self.headers, json=payload, stream=stream) if response.status_code != 200: raise Exception(f"**Error**: {response.status_code}, {response.text}") for chunk in response.iter_content(chunk_size=1024): if chunk: yield chunk class SILICONFLOWTTS(Base): _FACTORY_NAME = "SILICONFLOW" def __init__(self, key, model_name="FunAudioLLM/CosyVoice2-0.5B", base_url="https://api.siliconflow.cn/v1"): if not base_url: base_url = "https://api.siliconflow.cn/v1" self.api_key = key self.model_name = model_name self.base_url = base_url self.headers = {"Authorization": f"Bearer {self.api_key}", "Content-Type": "application/json"} def tts(self, text, voice="anna"): text = self.normalize_text(text) payload = { "model": self.model_name, "input": text, "voice": f"{self.model_name}:{voice}", "response_format": "mp3", "sample_rate": 123, "stream": True, "speed": 1, "gain": 0, } response = requests.post(f"{self.base_url}/audio/speech", headers=self.headers, json=payload) if response.status_code != 200: raise Exception(f"**Error**: {response.status_code}, {response.text}") for chunk in response.iter_content(): if chunk: yield chunk class DeepInfraTTS(OpenAITTS): _FACTORY_NAME = "DeepInfra" def __init__(self, key, model_name, base_url="https://api.deepinfra.com/v1/openai", **kwargs): if not base_url: base_url = "https://api.deepinfra.com/v1/openai" super().__init__(key, model_name, base_url, **kwargs) class CometAPITTS(OpenAITTS): _FACTORY_NAME = "CometAPI" def __init__(self, key, model_name, base_url="https://api.cometapi.com/v1", **kwargs): if not base_url: base_url = "https://api.cometapi.com/v1" super().__init__(key, model_name, base_url, **kwargs) class DeerAPITTS(OpenAITTS): _FACTORY_NAME = "DeerAPI" def __init__(self, key, model_name, base_url="https://api.deerapi.com/v1", **kwargs): if not base_url: base_url = "https://api.deerapi.com/v1" super().__init__(key, model_name, base_url, **kwargs)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/llm/embedding_model.py
rag/llm/embedding_model.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json import os import threading from abc import ABC from urllib.parse import urljoin import dashscope import google.generativeai as genai import numpy as np import requests from ollama import Client from openai import OpenAI from zhipuai import ZhipuAI from common.log_utils import log_exception from common.token_utils import num_tokens_from_string, truncate, total_token_count_from_response from common import settings import logging import base64 class Base(ABC): def __init__(self, key, model_name, **kwargs): """ Constructor for abstract base class. Parameters are accepted for interface consistency but are not stored. Subclasses should implement their own initialization as needed. """ pass def encode(self, texts: list): raise NotImplementedError("Please implement encode method!") def encode_queries(self, text: str): raise NotImplementedError("Please implement encode method!") class BuiltinEmbed(Base): _FACTORY_NAME = "Builtin" MAX_TOKENS = {"Qwen/Qwen3-Embedding-0.6B": 30000, "BAAI/bge-m3": 8000, "BAAI/bge-small-en-v1.5": 500} _model = None _model_name = "" _max_tokens = 500 _model_lock = threading.Lock() def __init__(self, key, model_name, **kwargs): logging.info(f"Initialize BuiltinEmbed according to settings.EMBEDDING_CFG: {settings.EMBEDDING_CFG}") embedding_cfg = settings.EMBEDDING_CFG if not BuiltinEmbed._model and "tei-" in os.getenv("COMPOSE_PROFILES", ""): with BuiltinEmbed._model_lock: BuiltinEmbed._model_name = settings.EMBEDDING_MDL BuiltinEmbed._max_tokens = BuiltinEmbed.MAX_TOKENS.get(settings.EMBEDDING_MDL, 500) BuiltinEmbed._model = HuggingFaceEmbed(embedding_cfg["api_key"], settings.EMBEDDING_MDL, base_url=embedding_cfg["base_url"]) self._model = BuiltinEmbed._model self._model_name = BuiltinEmbed._model_name self._max_tokens = BuiltinEmbed._max_tokens def encode(self, texts: list): batch_size = 16 # TEI is able to auto truncate inputs according to https://github.com/huggingface/text-embeddings-inference. token_count = 0 ress = None for i in range(0, len(texts), batch_size): embeddings, token_count_delta = self._model.encode(texts[i : i + batch_size]) token_count += token_count_delta if ress is None: ress = embeddings else: ress = np.concatenate((ress, embeddings), axis=0) return ress, token_count def encode_queries(self, text: str): return self._model.encode_queries(text) class OpenAIEmbed(Base): _FACTORY_NAME = "OpenAI" def __init__(self, key, model_name="text-embedding-ada-002", base_url="https://api.openai.com/v1"): if not base_url: base_url = "https://api.openai.com/v1" self.client = OpenAI(api_key=key, base_url=base_url) self.model_name = model_name def encode(self, texts: list): # OpenAI requires batch size <=16 batch_size = 16 texts = [truncate(t, 8191) for t in texts] ress = [] total_tokens = 0 for i in range(0, len(texts), batch_size): res = self.client.embeddings.create(input=texts[i : i + batch_size], model=self.model_name, encoding_format="float", extra_body={"drop_params": True}) try: ress.extend([d.embedding for d in res.data]) total_tokens += total_token_count_from_response(res) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") return np.array(ress), total_tokens def encode_queries(self, text): res = self.client.embeddings.create(input=[truncate(text, 8191)], model=self.model_name, encoding_format="float",extra_body={"drop_params": True}) try: return np.array(res.data[0].embedding), total_token_count_from_response(res) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") class LocalAIEmbed(Base): _FACTORY_NAME = "LocalAI" def __init__(self, key, model_name, base_url): if not base_url: raise ValueError("Local embedding model url cannot be None") base_url = urljoin(base_url, "v1") self.client = OpenAI(api_key="empty", base_url=base_url) self.model_name = model_name.split("___")[0] def encode(self, texts: list): batch_size = 16 ress = [] for i in range(0, len(texts), batch_size): res = self.client.embeddings.create(input=texts[i : i + batch_size], model=self.model_name) try: ress.extend([d.embedding for d in res.data]) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") # local embedding for LmStudio donot count tokens return np.array(ress), 1024 def encode_queries(self, text): embds, cnt = self.encode([text]) return np.array(embds[0]), cnt class AzureEmbed(OpenAIEmbed): _FACTORY_NAME = "Azure-OpenAI" def __init__(self, key, model_name, **kwargs): from openai.lib.azure import AzureOpenAI api_key = json.loads(key).get("api_key", "") api_version = json.loads(key).get("api_version", "2024-02-01") self.client = AzureOpenAI(api_key=api_key, azure_endpoint=kwargs["base_url"], api_version=api_version) self.model_name = model_name class BaiChuanEmbed(OpenAIEmbed): _FACTORY_NAME = "BaiChuan" def __init__(self, key, model_name="Baichuan-Text-Embedding", base_url="https://api.baichuan-ai.com/v1"): if not base_url: base_url = "https://api.baichuan-ai.com/v1" super().__init__(key, model_name, base_url) class QWenEmbed(Base): _FACTORY_NAME = "Tongyi-Qianwen" def __init__(self, key, model_name="text_embedding_v2", **kwargs): self.key = key self.model_name = model_name def encode(self, texts: list): import time import dashscope batch_size = 4 res = [] token_count = 0 texts = [truncate(t, 2048) for t in texts] for i in range(0, len(texts), batch_size): retry_max = 5 resp = dashscope.TextEmbedding.call(model=self.model_name, input=texts[i : i + batch_size], api_key=self.key, text_type="document") while (resp["output"] is None or resp["output"].get("embeddings") is None) and retry_max > 0: time.sleep(10) resp = dashscope.TextEmbedding.call(model=self.model_name, input=texts[i : i + batch_size], api_key=self.key, text_type="document") retry_max -= 1 if retry_max == 0 and (resp["output"] is None or resp["output"].get("embeddings") is None): if resp.get("message"): log_exception(ValueError(f"Retry_max reached, calling embedding model failed: {resp['message']}")) else: log_exception(ValueError("Retry_max reached, calling embedding model failed")) raise try: embds = [[] for _ in range(len(resp["output"]["embeddings"]))] for e in resp["output"]["embeddings"]: embds[e["text_index"]] = e["embedding"] res.extend(embds) token_count += total_token_count_from_response(resp) except Exception as _e: log_exception(_e, resp) raise return np.array(res), token_count def encode_queries(self, text): resp = dashscope.TextEmbedding.call(model=self.model_name, input=text[:2048], api_key=self.key, text_type="query") try: return np.array(resp["output"]["embeddings"][0]["embedding"]), total_token_count_from_response(resp) except Exception as _e: log_exception(_e, resp) raise Exception(f"Error: {resp}") class ZhipuEmbed(Base): _FACTORY_NAME = "ZHIPU-AI" def __init__(self, key, model_name="embedding-2", **kwargs): self.client = ZhipuAI(api_key=key) self.model_name = model_name def encode(self, texts: list): arr = [] tks_num = 0 MAX_LEN = -1 if self.model_name.lower() == "embedding-2": MAX_LEN = 512 if self.model_name.lower() == "embedding-3": MAX_LEN = 3072 if MAX_LEN > 0: texts = [truncate(t, MAX_LEN) for t in texts] for txt in texts: res = self.client.embeddings.create(input=txt, model=self.model_name) try: arr.append(res.data[0].embedding) tks_num += total_token_count_from_response(res) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") return np.array(arr), tks_num def encode_queries(self, text): res = self.client.embeddings.create(input=text, model=self.model_name) try: return np.array(res.data[0].embedding), total_token_count_from_response(res) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") class OllamaEmbed(Base): _FACTORY_NAME = "Ollama" _special_tokens = ["<|endoftext|>"] def __init__(self, key, model_name, **kwargs): self.client = Client(host=kwargs["base_url"]) if not key or key == "x" else Client(host=kwargs["base_url"], headers={"Authorization": f"Bearer {key}"}) self.model_name = model_name self.keep_alive = kwargs.get("ollama_keep_alive", int(os.environ.get("OLLAMA_KEEP_ALIVE", -1))) def encode(self, texts: list): arr = [] tks_num = 0 for txt in texts: # remove special tokens if they exist base on regex in one request for token in OllamaEmbed._special_tokens: txt = txt.replace(token, "") res = self.client.embeddings(prompt=txt, model=self.model_name, options={"use_mmap": True}, keep_alive=self.keep_alive) try: arr.append(res["embedding"]) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") tks_num += 128 return np.array(arr), tks_num def encode_queries(self, text): # remove special tokens if they exist for token in OllamaEmbed._special_tokens: text = text.replace(token, "") res = self.client.embeddings(prompt=text, model=self.model_name, options={"use_mmap": True}, keep_alive=self.keep_alive) try: return np.array(res["embedding"]), 128 except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") class XinferenceEmbed(Base): _FACTORY_NAME = "Xinference" def __init__(self, key, model_name="", base_url=""): base_url = urljoin(base_url, "v1") self.client = OpenAI(api_key=key, base_url=base_url) self.model_name = model_name def encode(self, texts: list): batch_size = 16 ress = [] total_tokens = 0 for i in range(0, len(texts), batch_size): res = None try: res = self.client.embeddings.create(input=texts[i : i + batch_size], model=self.model_name) ress.extend([d.embedding for d in res.data]) total_tokens += total_token_count_from_response(res) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") return np.array(ress), total_tokens def encode_queries(self, text): res = None try: res = self.client.embeddings.create(input=[text], model=self.model_name) return np.array(res.data[0].embedding), total_token_count_from_response(res) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") class YoudaoEmbed(Base): _FACTORY_NAME = "Youdao" _client = None def __init__(self, key=None, model_name="maidalun1020/bce-embedding-base_v1", **kwargs): pass def encode(self, texts: list): batch_size = 10 res = [] token_count = 0 for t in texts: token_count += num_tokens_from_string(t) for i in range(0, len(texts), batch_size): embds = YoudaoEmbed._client.encode(texts[i : i + batch_size]) res.extend(embds) return np.array(res), token_count def encode_queries(self, text): embds = YoudaoEmbed._client.encode([text]) return np.array(embds[0]), num_tokens_from_string(text) class JinaMultiVecEmbed(Base): _FACTORY_NAME = "Jina" def __init__(self, key, model_name="jina-embeddings-v4", base_url="https://api.jina.ai/v1/embeddings"): self.base_url = "https://api.jina.ai/v1/embeddings" self.headers = {"Content-Type": "application/json", "Authorization": f"Bearer {key}"} self.model_name = model_name def encode(self, texts: list[str|bytes], task="retrieval.passage"): batch_size = 16 ress = [] token_count = 0 input = [] for text in texts: if isinstance(text, str): input.append({"text": text}) elif isinstance(text, bytes): img_b64s = None try: base64.b64decode(text, validate=True) img_b64s = text.decode('utf8') except Exception: img_b64s = base64.b64encode(text).decode('utf8') input.append({"image": img_b64s}) # base64 encoded image for i in range(0, len(texts), batch_size): data = {"model": self.model_name, "input": input[i : i + batch_size]} if "v4" in self.model_name: data["return_multivector"] = True if "v3" in self.model_name or "v4" in self.model_name: data['task'] = task data['truncate'] = True response = requests.post(self.base_url, headers=self.headers, json=data) try: res = response.json() for d in res['data']: if data.get("return_multivector", False): # v4 token_embs = np.asarray(d['embeddings'], dtype=np.float32) chunk_emb = token_embs.mean(axis=0) else: # v2/v3 chunk_emb = np.asarray(d['embedding'], dtype=np.float32) ress.append(chunk_emb) token_count += total_token_count_from_response(res) except Exception as _e: log_exception(_e, response) raise Exception(f"Error: {response}") return np.array(ress), token_count def encode_queries(self, text): embds, cnt = self.encode([text], task="retrieval.query") return np.array(embds[0]), cnt class MistralEmbed(Base): _FACTORY_NAME = "Mistral" def __init__(self, key, model_name="mistral-embed", base_url=None): from mistralai.client import MistralClient self.client = MistralClient(api_key=key) self.model_name = model_name def encode(self, texts: list): import time import random texts = [truncate(t, 8196) for t in texts] batch_size = 16 ress = [] token_count = 0 for i in range(0, len(texts), batch_size): retry_max = 5 while retry_max > 0: try: res = self.client.embeddings(input=texts[i : i + batch_size], model=self.model_name) ress.extend([d.embedding for d in res.data]) token_count += total_token_count_from_response(res) break except Exception as _e: if retry_max == 1: log_exception(_e) delay = random.uniform(20, 60) time.sleep(delay) retry_max -= 1 return np.array(ress), token_count def encode_queries(self, text): import time import random retry_max = 5 while retry_max > 0: try: res = self.client.embeddings(input=[truncate(text, 8196)], model=self.model_name) return np.array(res.data[0].embedding), total_token_count_from_response(res) except Exception as _e: if retry_max == 1: log_exception(_e) delay = random.randint(20, 60) time.sleep(delay) retry_max -= 1 class BedrockEmbed(Base): _FACTORY_NAME = "Bedrock" def __init__(self, key, model_name, **kwargs): import boto3 # `key` protocol (backend stores as JSON string in `api_key`): # - Must decode into a dict. # - Required: `auth_mode`, `bedrock_region`. # - Supported auth modes: # - "access_key_secret": requires `bedrock_ak` + `bedrock_sk`. # - "iam_role": requires `aws_role_arn` and assumes role via STS. # - else: treated as "assume_role" (default AWS credential chain). key = json.loads(key) mode = key.get("auth_mode") if not mode: logging.error("Bedrock auth_mode is not provided in the key") raise ValueError("Bedrock auth_mode must be provided in the key") self.bedrock_region = key.get("bedrock_region") self.model_name = model_name self.is_amazon = self.model_name.split(".")[0] == "amazon" self.is_cohere = self.model_name.split(".")[0] == "cohere" if mode == "access_key_secret": self.bedrock_ak = key.get("bedrock_ak") self.bedrock_sk = key.get("bedrock_sk") self.client = boto3.client(service_name="bedrock-runtime", region_name=self.bedrock_region, aws_access_key_id=self.bedrock_ak, aws_secret_access_key=self.bedrock_sk) elif mode == "iam_role": self.aws_role_arn = key.get("aws_role_arn") sts_client = boto3.client("sts", region_name=self.bedrock_region) resp = sts_client.assume_role(RoleArn=self.aws_role_arn, RoleSessionName="BedrockSession") creds = resp["Credentials"] self.client = boto3.client( service_name="bedrock-runtime", aws_access_key_id=creds["AccessKeyId"], aws_secret_access_key=creds["SecretAccessKey"], aws_session_token=creds["SessionToken"], ) else: # assume_role self.client = boto3.client("bedrock-runtime", region_name=self.bedrock_region) def encode(self, texts: list): texts = [truncate(t, 8196) for t in texts] embeddings = [] token_count = 0 for text in texts: if self.is_amazon: body = {"inputText": text} elif self.is_cohere: body = {"texts": [text], "input_type": "search_document"} response = self.client.invoke_model(modelId=self.model_name, body=json.dumps(body)) try: model_response = json.loads(response["body"].read()) embeddings.extend([model_response["embedding"]]) token_count += num_tokens_from_string(text) except Exception as _e: log_exception(_e, response) return np.array(embeddings), token_count def encode_queries(self, text): embeddings = [] token_count = num_tokens_from_string(text) if self.is_amazon: body = {"inputText": truncate(text, 8196)} elif self.is_cohere: body = {"texts": [truncate(text, 8196)], "input_type": "search_query"} response = self.client.invoke_model(modelId=self.model_name, body=json.dumps(body)) try: model_response = json.loads(response["body"].read()) embeddings.extend(model_response["embedding"]) except Exception as _e: log_exception(_e, response) return np.array(embeddings), token_count class GeminiEmbed(Base): _FACTORY_NAME = "Gemini" def __init__(self, key, model_name="models/text-embedding-004", **kwargs): self.key = key self.model_name = "models/" + model_name def encode(self, texts: list): texts = [truncate(t, 2048) for t in texts] token_count = sum(num_tokens_from_string(text) for text in texts) genai.configure(api_key=self.key) batch_size = 16 ress = [] for i in range(0, len(texts), batch_size): result = genai.embed_content(model=self.model_name, content=texts[i : i + batch_size], task_type="retrieval_document", title="Embedding of single string") try: ress.extend(result["embedding"]) except Exception as _e: log_exception(_e, result) raise Exception(f"Error: {result}") return np.array(ress), token_count def encode_queries(self, text): genai.configure(api_key=self.key) result = genai.embed_content(model=self.model_name, content=truncate(text, 2048), task_type="retrieval_document", title="Embedding of single string") token_count = num_tokens_from_string(text) try: return np.array(result["embedding"]), token_count except Exception as _e: log_exception(_e, result) raise Exception(f"Error: {result}") class NvidiaEmbed(Base): _FACTORY_NAME = "NVIDIA" def __init__(self, key, model_name, base_url="https://integrate.api.nvidia.com/v1/embeddings"): if not base_url: base_url = "https://integrate.api.nvidia.com/v1/embeddings" self.api_key = key self.base_url = base_url self.headers = { "accept": "application/json", "Content-Type": "application/json", "authorization": f"Bearer {self.api_key}", } self.model_name = model_name if model_name == "nvidia/embed-qa-4": self.base_url = "https://ai.api.nvidia.com/v1/retrieval/nvidia/embeddings" self.model_name = "NV-Embed-QA" if model_name == "snowflake/arctic-embed-l": self.base_url = "https://ai.api.nvidia.com/v1/retrieval/snowflake/arctic-embed-l/embeddings" def encode(self, texts: list): batch_size = 16 ress = [] token_count = 0 for i in range(0, len(texts), batch_size): payload = { "input": texts[i : i + batch_size], "input_type": "query", "model": self.model_name, "encoding_format": "float", "truncate": "END", } response = requests.post(self.base_url, headers=self.headers, json=payload) try: res = response.json() ress.extend([d["embedding"] for d in res["data"]]) token_count += total_token_count_from_response(res) except Exception as _e: log_exception(_e, response) raise Exception(f"Error: {response}") return np.array(ress), token_count def encode_queries(self, text): embds, cnt = self.encode([text]) return np.array(embds[0]), cnt class LmStudioEmbed(LocalAIEmbed): _FACTORY_NAME = "LM-Studio" def __init__(self, key, model_name, base_url): if not base_url: raise ValueError("Local llm url cannot be None") base_url = urljoin(base_url, "v1") self.client = OpenAI(api_key="lm-studio", base_url=base_url) self.model_name = model_name class OpenAI_APIEmbed(OpenAIEmbed): _FACTORY_NAME = ["VLLM", "OpenAI-API-Compatible"] def __init__(self, key, model_name, base_url): if not base_url: raise ValueError("url cannot be None") base_url = urljoin(base_url, "v1") self.client = OpenAI(api_key=key, base_url=base_url) self.model_name = model_name.split("___")[0] class CoHereEmbed(Base): _FACTORY_NAME = "Cohere" def __init__(self, key, model_name, base_url=None): from cohere import Client self.client = Client(api_key=key) self.model_name = model_name def encode(self, texts: list): batch_size = 16 ress = [] token_count = 0 for i in range(0, len(texts), batch_size): res = self.client.embed( texts=texts[i : i + batch_size], model=self.model_name, input_type="search_document", embedding_types=["float"], ) try: ress.extend([d for d in res.embeddings.float]) token_count += total_token_count_from_response(res) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") return np.array(ress), token_count def encode_queries(self, text): res = self.client.embed( texts=[text], model=self.model_name, input_type="search_query", embedding_types=["float"], ) try: return np.array(res.embeddings.float[0]), int(total_token_count_from_response(res)) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") class TogetherAIEmbed(OpenAIEmbed): _FACTORY_NAME = "TogetherAI" def __init__(self, key, model_name, base_url="https://api.together.xyz/v1"): if not base_url: base_url = "https://api.together.xyz/v1" super().__init__(key, model_name, base_url=base_url) class PerfXCloudEmbed(OpenAIEmbed): _FACTORY_NAME = "PerfXCloud" def __init__(self, key, model_name, base_url="https://cloud.perfxlab.cn/v1"): if not base_url: base_url = "https://cloud.perfxlab.cn/v1" super().__init__(key, model_name, base_url) class UpstageEmbed(OpenAIEmbed): _FACTORY_NAME = "Upstage" def __init__(self, key, model_name, base_url="https://api.upstage.ai/v1/solar"): if not base_url: base_url = "https://api.upstage.ai/v1/solar" super().__init__(key, model_name, base_url) class SILICONFLOWEmbed(Base): _FACTORY_NAME = "SILICONFLOW" def __init__(self, key, model_name, base_url="https://api.siliconflow.cn/v1/embeddings"): if not base_url: base_url = "https://api.siliconflow.cn/v1/embeddings" self.headers = { "accept": "application/json", "content-type": "application/json", "authorization": f"Bearer {key}", } self.base_url = base_url self.model_name = model_name def encode(self, texts: list): batch_size = 16 ress = [] token_count = 0 for i in range(0, len(texts), batch_size): texts_batch = texts[i : i + batch_size] if self.model_name in ["BAAI/bge-large-zh-v1.5", "BAAI/bge-large-en-v1.5"]: # limit 512, 340 is almost safe texts_batch = [" " if not text.strip() else truncate(text, 256) for text in texts_batch] else: texts_batch = [" " if not text.strip() else text for text in texts_batch] payload = { "model": self.model_name, "input": texts_batch, "encoding_format": "float", } response = requests.post(self.base_url, json=payload, headers=self.headers) try: res = response.json() ress.extend([d["embedding"] for d in res["data"]]) token_count += total_token_count_from_response(res) except Exception as _e: log_exception(_e, response) raise Exception(f"Error: {response}") return np.array(ress), token_count def encode_queries(self, text): payload = { "model": self.model_name, "input": text, "encoding_format": "float", } response = requests.post(self.base_url, json=payload, headers=self.headers) try: res = response.json() return np.array(res["data"][0]["embedding"]), total_token_count_from_response(res) except Exception as _e: log_exception(_e, response) raise Exception(f"Error: {response}") class ReplicateEmbed(Base): _FACTORY_NAME = "Replicate" def __init__(self, key, model_name, base_url=None): from replicate.client import Client self.model_name = model_name self.client = Client(api_token=key) def encode(self, texts: list): batch_size = 16 token_count = sum([num_tokens_from_string(text) for text in texts]) ress = [] for i in range(0, len(texts), batch_size): res = self.client.run(self.model_name, input={"texts": texts[i : i + batch_size]}) ress.extend(res) return np.array(ress), token_count def encode_queries(self, text): res = self.client.embed(self.model_name, input={"texts": [text]}) return np.array(res), num_tokens_from_string(text) class BaiduYiyanEmbed(Base): _FACTORY_NAME = "BaiduYiyan" def __init__(self, key, model_name, base_url=None): import qianfan key = json.loads(key) ak = key.get("yiyan_ak", "") sk = key.get("yiyan_sk", "") self.client = qianfan.Embedding(ak=ak, sk=sk) self.model_name = model_name def encode(self, texts: list, batch_size=16): res = self.client.do(model=self.model_name, texts=texts).body try: return ( np.array([r["embedding"] for r in res["data"]]), total_token_count_from_response(res), ) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") def encode_queries(self, text): res = self.client.do(model=self.model_name, texts=[text]).body try: return ( np.array([r["embedding"] for r in res["data"]]), total_token_count_from_response(res), ) except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") class VoyageEmbed(Base): _FACTORY_NAME = "Voyage AI" def __init__(self, key, model_name, base_url=None): import voyageai self.client = voyageai.Client(api_key=key) self.model_name = model_name def encode(self, texts: list): batch_size = 16 ress = [] token_count = 0 for i in range(0, len(texts), batch_size): res = self.client.embed(texts=texts[i : i + batch_size], model=self.model_name, input_type="document") try: ress.extend(res.embeddings) token_count += res.total_tokens except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") return np.array(ress), token_count def encode_queries(self, text): res = self.client.embed(texts=text, model=self.model_name, input_type="query") try: return np.array(res.embeddings)[0], res.total_tokens except Exception as _e: log_exception(_e, res) raise Exception(f"Error: {res}") class HuggingFaceEmbed(Base): _FACTORY_NAME = "HuggingFace" def __init__(self, key, model_name, base_url=None, **kwargs): if not model_name: raise ValueError("Model name cannot be None") self.key = key
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/llm/cv_model.py
rag/llm/cv_model.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import asyncio import base64 import json import logging import os import re import tempfile from abc import ABC from copy import deepcopy from io import BytesIO from pathlib import Path from urllib.parse import urljoin import requests from openai import OpenAI, AsyncOpenAI from openai.lib.azure import AzureOpenAI, AsyncAzureOpenAI from common.token_utils import num_tokens_from_string, total_token_count_from_response from rag.nlp import is_english from rag.prompts.generator import vision_llm_describe_prompt class Base(ABC): def __init__(self, **kwargs): # Configure retry parameters self.max_retries = kwargs.get("max_retries", int(os.environ.get("LLM_MAX_RETRIES", 5))) self.base_delay = kwargs.get("retry_interval", float(os.environ.get("LLM_BASE_DELAY", 2.0))) self.max_rounds = kwargs.get("max_rounds", 5) self.is_tools = False self.tools = [] self.toolcall_sessions = {} self.extra_body = None def describe(self, image): raise NotImplementedError("Please implement encode method!") def describe_with_prompt(self, image, prompt=None): raise NotImplementedError("Please implement encode method!") def _form_history(self, system, history, images=None): hist = [] if system: hist.append({"role": "system", "content": system}) for h in history: if images and h["role"] == "user": h["content"] = self._image_prompt(h["content"], images) images = [] hist.append(h) return hist def _image_prompt(self, text, images): if not images: return text if isinstance(images, str) or "bytes" in type(images).__name__: images = [images] pmpt = [{"type": "text", "text": text}] for img in images: pmpt.append({"type": "image_url", "image_url": {"url": img if isinstance(img, str) and img.startswith("data:") else f"data:image/png;base64,{img}"}}) return pmpt async def async_chat(self, system, history, gen_conf, images=None, **kwargs): try: response = await self.async_client.chat.completions.create( model=self.model_name, messages=self._form_history(system, history, images), extra_body=self.extra_body, ) return response.choices[0].message.content.strip(), response.usage.total_tokens except Exception as e: return "**ERROR**: " + str(e), 0 async def async_chat_streamly(self, system, history, gen_conf, images=None, **kwargs): ans = "" tk_count = 0 try: response = await self.async_client.chat.completions.create( model=self.model_name, messages=self._form_history(system, history, images), stream=True, extra_body=self.extra_body, ) async for resp in response: if not resp.choices[0].delta.content: continue delta = resp.choices[0].delta.content ans = delta if resp.choices[0].finish_reason == "length": ans += "...\nFor the content length reason, it stopped, continue?" if is_english([ans]) else "······\n由于长度的原因,回答被截断了,要继续吗?" if resp.choices[0].finish_reason == "stop": tk_count += resp.usage.total_tokens yield ans except Exception as e: yield ans + "\n**ERROR**: " + str(e) yield tk_count @staticmethod def image2base64_rawvalue(self, image): # Return a base64 string without data URL header if isinstance(image, bytes): b64 = base64.b64encode(image).decode("utf-8") return b64 if isinstance(image, BytesIO): data = image.getvalue() b64 = base64.b64encode(data).decode("utf-8") return b64 with BytesIO() as buffered: try: image.save(buffered, format="JPEG") except Exception: # reset buffer before saving PNG buffered.seek(0) buffered.truncate() image.save(buffered, format="PNG") data = buffered.getvalue() b64 = base64.b64encode(data).decode("utf-8") return b64 @staticmethod def image2base64(image): # Return a data URL with the correct MIME to avoid provider mismatches if isinstance(image, bytes): # Best-effort magic number sniffing mime = "image/png" if len(image) >= 2 and image[0] == 0xFF and image[1] == 0xD8: mime = "image/jpeg" b64 = base64.b64encode(image).decode("utf-8") return f"data:{mime};base64,{b64}" if isinstance(image, BytesIO): data = image.getvalue() mime = "image/png" if len(data) >= 2 and data[0] == 0xFF and data[1] == 0xD8: mime = "image/jpeg" b64 = base64.b64encode(data).decode("utf-8") return f"data:{mime};base64,{b64}" with BytesIO() as buffered: fmt = "jpeg" try: image.save(buffered, format="JPEG") except Exception: # reset buffer before saving PNG buffered.seek(0) buffered.truncate() image.save(buffered, format="PNG") fmt = "png" data = buffered.getvalue() b64 = base64.b64encode(data).decode("utf-8") mime = f"image/{fmt}" return f"data:{mime};base64,{b64}" def prompt(self, b64): return [ { "role": "user", "content": self._image_prompt( "请用中文详细描述一下图中的内容,比如时间,地点,人物,事情,人物心情等,如果有数据请提取出数据。" if self.lang.lower() == "chinese" else "Please describe the content of this picture, like where, when, who, what happen. If it has number data, please extract them out.", b64, ), } ] def vision_llm_prompt(self, b64, prompt=None): return [{"role": "user", "content": self._image_prompt(prompt if prompt else vision_llm_describe_prompt(), b64)}] class GptV4(Base): _FACTORY_NAME = "OpenAI" def __init__(self, key, model_name="gpt-4-vision-preview", lang="Chinese", base_url="https://api.openai.com/v1", **kwargs): if not base_url: base_url = "https://api.openai.com/v1" self.api_key = key self.client = OpenAI(api_key=key, base_url=base_url) self.async_client = AsyncOpenAI(api_key=key, base_url=base_url) self.model_name = model_name self.lang = lang super().__init__(**kwargs) def describe(self, image): b64 = self.image2base64(image) res = self.client.chat.completions.create( model=self.model_name, messages=self.prompt(b64), extra_body=self.extra_body ) return res.choices[0].message.content.strip(), total_token_count_from_response(res) def describe_with_prompt(self, image, prompt=None): b64 = self.image2base64(image) res = self.client.chat.completions.create( model=self.model_name, messages=self.vision_llm_prompt(b64, prompt), extra_body=self.extra_body, ) return res.choices[0].message.content.strip(), total_token_count_from_response(res) class AzureGptV4(GptV4): _FACTORY_NAME = "Azure-OpenAI" def __init__(self, key, model_name, lang="Chinese", **kwargs): api_key = json.loads(key).get("api_key", "") api_version = json.loads(key).get("api_version", "2024-02-01") self.client = AzureOpenAI(api_key=api_key, azure_endpoint=kwargs["base_url"], api_version=api_version) self.async_client = AsyncAzureOpenAI(api_key=api_key, azure_endpoint=kwargs["base_url"], api_version=api_version) self.model_name = model_name self.lang = lang Base.__init__(self, **kwargs) class xAICV(GptV4): _FACTORY_NAME = "xAI" def __init__(self, key, model_name="grok-3", lang="Chinese", base_url=None, **kwargs): if not base_url: base_url = "https://api.x.ai/v1" super().__init__(key, model_name, lang=lang, base_url=base_url, **kwargs) class QWenCV(GptV4): _FACTORY_NAME = "Tongyi-Qianwen" def __init__(self, key, model_name="qwen-vl-chat-v1", lang="Chinese", base_url=None, **kwargs): if not base_url: base_url = "https://dashscope.aliyuncs.com/compatible-mode/v1" super().__init__(key, model_name, lang=lang, base_url=base_url, **kwargs) async def async_chat(self, system, history, gen_conf, images=None, video_bytes=None, filename="", **kwargs): if video_bytes: try: summary, summary_num_tokens = self._process_video(video_bytes, filename) return summary, summary_num_tokens except Exception as e: return "**ERROR**: " + str(e), 0 return "**ERROR**: Method chat not supported yet.", 0 def _process_video(self, video_bytes, filename): from dashscope import MultiModalConversation video_suffix = Path(filename).suffix or ".mp4" with tempfile.NamedTemporaryFile(delete=False, suffix=video_suffix) as tmp: tmp.write(video_bytes) tmp_path = tmp.name video_path = f"file://{tmp_path}" messages = [ { "role": "user", "content": [ { "video": video_path, "fps": 2, }, { "text": "Please summarize this video in proper sentences.", }, ], } ] def call_api(): response = MultiModalConversation.call( api_key=self.api_key, model=self.model_name, messages=messages, ) if response.get("message"): raise Exception(response["message"]) summary = response["output"]["choices"][0]["message"].content[0]["text"] return summary, num_tokens_from_string(summary) try: return call_api() except Exception as e1: import dashscope dashscope.base_http_api_url = "https://dashscope-intl.aliyuncs.com/api/v1" try: return call_api() except Exception as e2: raise RuntimeError(f"Both default and intl endpoint failed.\nFirst error: {e1}\nSecond error: {e2}") class HunyuanCV(GptV4): _FACTORY_NAME = "Tencent Hunyuan" def __init__(self, key, model_name, lang="Chinese", base_url=None, **kwargs): if not base_url: base_url = "https://api.hunyuan.cloud.tencent.com/v1" super().__init__(key, model_name, lang=lang, base_url=base_url, **kwargs) class Zhipu4V(GptV4): _FACTORY_NAME = "ZHIPU-AI" def __init__(self, key, model_name="glm-4v", lang="Chinese", **kwargs): self.client = OpenAI(api_key=key, base_url="https://open.bigmodel.cn/api/paas/v4/") self.async_client = AsyncOpenAI(api_key=key, base_url="https://open.bigmodel.cn/api/paas/v4/") self.model_name = model_name self.lang = lang Base.__init__(self, **kwargs) def _clean_conf(self, gen_conf): if "max_tokens" in gen_conf: del gen_conf["max_tokens"] gen_conf = self._clean_conf_plealty(gen_conf) return gen_conf def _clean_conf_plealty(self, gen_conf): if "presence_penalty" in gen_conf: del gen_conf["presence_penalty"] if "frequency_penalty" in gen_conf: del gen_conf["frequency_penalty"] return gen_conf def _request(self, msg, stream, gen_conf={}): response = requests.post( self.base_url, json={"model": self.model_name, "messages": msg, "stream": stream, **gen_conf}, headers={ "Authorization": f"Bearer {self.api_key}", "Content-Type": "application/json", }, ) return response.json() async def async_chat(self, system, history, gen_conf, images=None, **kwargs): if system and history and history[0].get("role") != "system": history.insert(0, {"role": "system", "content": system}) gen_conf = self._clean_conf(gen_conf) logging.info(json.dumps(history, ensure_ascii=False, indent=2)) response = await self.async_client.chat.completions.create(model=self.model_name, messages=self._form_history(system, history, images), stream=False, **gen_conf) content = response.choices[0].message.content.strip() cleaned = re.sub(r"<\|(begin_of_box|end_of_box)\|>", "", content).strip() return cleaned, total_token_count_from_response(response) async def async_chat_streamly(self, system, history, gen_conf, images=None, **kwargs): from rag.llm.chat_model import LENGTH_NOTIFICATION_CN, LENGTH_NOTIFICATION_EN from rag.nlp import is_chinese if system and history and history[0].get("role") != "system": history.insert(0, {"role": "system", "content": system}) gen_conf = self._clean_conf(gen_conf) ans = "" tk_count = 0 try: logging.info(json.dumps(history, ensure_ascii=False, indent=2)) response = await self.async_client.chat.completions.create(model=self.model_name, messages=self._form_history(system, history, images), stream=True, **gen_conf) async for resp in response: if not resp.choices[0].delta.content: continue delta = resp.choices[0].delta.content ans = delta if resp.choices[0].finish_reason == "length": if is_chinese(ans): ans += LENGTH_NOTIFICATION_CN else: ans += LENGTH_NOTIFICATION_EN tk_count = total_token_count_from_response(resp) if resp.choices[0].finish_reason == "stop": tk_count = total_token_count_from_response(resp) yield ans except Exception as e: yield ans + "\n**ERROR**: " + str(e) yield tk_count def describe(self, image): return self.describe_with_prompt(image) def describe_with_prompt(self, image, prompt=None): b64 = self.image2base64(image) if prompt is None: prompt = "Describe this image." # Chat messages messages = [{"role": "user", "content": [{"type": "image_url", "image_url": {"url": b64}}, {"type": "text", "text": prompt}]}] resp = self.client.chat.completions.create(model=self.model_name, messages=messages, stream=False) content = resp.choices[0].message.content.strip() cleaned = re.sub(r"<\|(begin_of_box|end_of_box)\|>", "", content).strip() return cleaned, num_tokens_from_string(cleaned) class StepFunCV(GptV4): _FACTORY_NAME = "StepFun" def __init__(self, key, model_name="step-1v-8k", lang="Chinese", base_url="https://api.stepfun.com/v1", **kwargs): if not base_url: base_url = "https://api.stepfun.com/v1" self.client = OpenAI(api_key=key, base_url=base_url) self.async_client = AsyncOpenAI(api_key=key, base_url=base_url) self.model_name = model_name self.lang = lang Base.__init__(self, **kwargs) class VolcEngineCV(GptV4): _FACTORY_NAME = "VolcEngine" def __init__(self, key, model_name, lang="Chinese", base_url="https://ark.cn-beijing.volces.com/api/v3", **kwargs): if not base_url: base_url = "https://ark.cn-beijing.volces.com/api/v3" ark_api_key = json.loads(key).get("ark_api_key", "") self.client = OpenAI(api_key=ark_api_key, base_url=base_url) self.async_client = AsyncOpenAI(api_key=ark_api_key, base_url=base_url) self.model_name = json.loads(key).get("ep_id", "") + json.loads(key).get("endpoint_id", "") self.lang = lang Base.__init__(self, **kwargs) class LmStudioCV(GptV4): _FACTORY_NAME = "LM-Studio" def __init__(self, key, model_name, lang="Chinese", base_url="", **kwargs): if not base_url: raise ValueError("Local llm url cannot be None") base_url = urljoin(base_url, "v1") self.client = OpenAI(api_key="lm-studio", base_url=base_url) self.async_client = AsyncOpenAI(api_key="lm-studio", base_url=base_url) self.model_name = model_name self.lang = lang Base.__init__(self, **kwargs) class OpenAI_APICV(GptV4): _FACTORY_NAME = ["VLLM", "OpenAI-API-Compatible"] def __init__(self, key, model_name, lang="Chinese", base_url="", **kwargs): if not base_url: raise ValueError("url cannot be None") base_url = urljoin(base_url, "v1") self.client = OpenAI(api_key=key, base_url=base_url) self.async_client = AsyncOpenAI(api_key=key, base_url=base_url) self.model_name = model_name.split("___")[0] self.lang = lang Base.__init__(self, **kwargs) class TogetherAICV(GptV4): _FACTORY_NAME = "TogetherAI" def __init__(self, key, model_name, lang="Chinese", base_url="https://api.together.xyz/v1", **kwargs): if not base_url: base_url = "https://api.together.xyz/v1" super().__init__(key, model_name, lang, base_url, **kwargs) class YiCV(GptV4): _FACTORY_NAME = "01.AI" def __init__(self, key, model_name, lang="Chinese", base_url="https://api.lingyiwanwu.com/v1", **kwargs): if not base_url: base_url = "https://api.lingyiwanwu.com/v1" super().__init__(key, model_name, lang, base_url, **kwargs) class SILICONFLOWCV(GptV4): _FACTORY_NAME = "SILICONFLOW" def __init__(self, key, model_name, lang="Chinese", base_url="https://api.siliconflow.cn/v1", **kwargs): if not base_url: base_url = "https://api.siliconflow.cn/v1" super().__init__(key, model_name, lang, base_url, **kwargs) class OpenRouterCV(GptV4): _FACTORY_NAME = "OpenRouter" def __init__(self, key, model_name, lang="Chinese", base_url="https://openrouter.ai/api/v1", **kwargs): if not base_url: base_url = "https://openrouter.ai/api/v1" api_key = json.loads(key).get("api_key", "") self.client = OpenAI(api_key=api_key, base_url=base_url) self.async_client = AsyncOpenAI(api_key=api_key, base_url=base_url) self.model_name = model_name self.lang = lang Base.__init__(self, **kwargs) provider_order = json.loads(key).get("provider_order", "") self.extra_body = {} if provider_order: def _to_order_list(x): if x is None: return [] if isinstance(x, str): return [s.strip() for s in x.split(",") if s.strip()] if isinstance(x, (list, tuple)): return [str(s).strip() for s in x if str(s).strip()] return [] provider_cfg = {} provider_order = _to_order_list(provider_order) provider_cfg["order"] = provider_order provider_cfg["allow_fallbacks"] = False self.extra_body["provider"] = provider_cfg class LocalAICV(GptV4): _FACTORY_NAME = "LocalAI" def __init__(self, key, model_name, base_url, lang="Chinese", **kwargs): if not base_url: raise ValueError("Local cv model url cannot be None") base_url = urljoin(base_url, "v1") self.client = OpenAI(api_key="empty", base_url=base_url) self.async_client = AsyncOpenAI(api_key="empty", base_url=base_url) self.model_name = model_name.split("___")[0] self.lang = lang Base.__init__(self, **kwargs) class XinferenceCV(GptV4): _FACTORY_NAME = "Xinference" def __init__(self, key, model_name="", lang="Chinese", base_url="", **kwargs): base_url = urljoin(base_url, "v1") self.client = OpenAI(api_key=key, base_url=base_url) self.async_client = AsyncOpenAI(api_key=key, base_url=base_url) self.model_name = model_name self.lang = lang Base.__init__(self, **kwargs) class GPUStackCV(GptV4): _FACTORY_NAME = "GPUStack" def __init__(self, key, model_name, lang="Chinese", base_url="", **kwargs): if not base_url: raise ValueError("Local llm url cannot be None") base_url = urljoin(base_url, "v1") self.client = OpenAI(api_key=key, base_url=base_url) self.async_client = AsyncOpenAI(api_key=key, base_url=base_url) self.model_name = model_name self.lang = lang Base.__init__(self, **kwargs) class LocalCV(Base): _FACTORY_NAME = "Local" def __init__(self, key, model_name="glm-4v", lang="Chinese", **kwargs): pass def describe(self, image): return "", 0 class OllamaCV(Base): _FACTORY_NAME = "Ollama" def __init__(self, key, model_name, lang="Chinese", **kwargs): from ollama import Client self.client = Client(host=kwargs["base_url"]) self.model_name = model_name self.lang = lang self.keep_alive = kwargs.get("ollama_keep_alive", int(os.environ.get("OLLAMA_KEEP_ALIVE", -1))) Base.__init__(self, **kwargs) def _clean_img(self, img): if not isinstance(img, str): return img # remove the header like "data/*;base64," if img.startswith("data:") and ";base64," in img: img = img.split(";base64,")[1] return img def _clean_conf(self, gen_conf): options = {} if "temperature" in gen_conf: options["temperature"] = gen_conf["temperature"] if "top_p" in gen_conf: options["top_k"] = gen_conf["top_p"] if "presence_penalty" in gen_conf: options["presence_penalty"] = gen_conf["presence_penalty"] if "frequency_penalty" in gen_conf: options["frequency_penalty"] = gen_conf["frequency_penalty"] return options def _form_history(self, system, history, images=None): hist = deepcopy(history) if system and hist[0]["role"] == "user": hist.insert(0, {"role": "system", "content": system}) if not images: return hist temp_images = [] for img in images: temp_images.append(self._clean_img(img)) for his in hist: if his["role"] == "user": his["images"] = temp_images break return hist def describe(self, image): prompt = self.prompt("") try: response = self.client.generate( model=self.model_name, prompt=prompt[0]["content"], images=[image], ) ans = response["response"].strip() return ans, 128 except Exception as e: return "**ERROR**: " + str(e), 0 def describe_with_prompt(self, image, prompt=None): vision_prompt = self.vision_llm_prompt("", prompt) if prompt else self.vision_llm_prompt("") try: response = self.client.generate( model=self.model_name, prompt=vision_prompt[0]["content"], images=[image], ) ans = response["response"].strip() return ans, 128 except Exception as e: return "**ERROR**: " + str(e), 0 async def async_chat(self, system, history, gen_conf, images=None, **kwargs): try: response = await asyncio.to_thread(self.client.chat, model=self.model_name, messages=self._form_history(system, history, images), options=self._clean_conf(gen_conf), keep_alive=self.keep_alive) ans = response["message"]["content"].strip() return ans, response["eval_count"] + response.get("prompt_eval_count", 0) except Exception as e: return "**ERROR**: " + str(e), 0 async def async_chat_streamly(self, system, history, gen_conf, images=None, **kwargs): ans = "" try: response = await asyncio.to_thread(self.client.chat, model=self.model_name, messages=self._form_history(system, history, images), stream=True, options=self._clean_conf(gen_conf), keep_alive=self.keep_alive) for resp in response: if resp["done"]: yield resp.get("prompt_eval_count", 0) + resp.get("eval_count", 0) ans = resp["message"]["content"] yield ans except Exception as e: yield ans + "\n**ERROR**: " + str(e) yield 0 class GeminiCV(Base): _FACTORY_NAME = "Gemini" def __init__(self, key, model_name="gemini-1.0-pro-vision-latest", lang="Chinese", **kwargs): from google import genai self.api_key = key self.model_name = model_name self.client = genai.Client(api_key=key) self.lang = lang Base.__init__(self, **kwargs) logging.info(f"[GeminiCV] Initialized with model={self.model_name} lang={self.lang}") def _image_to_part(self, image): from google.genai import types if isinstance(image, str) and image.startswith("data:") and ";base64," in image: header, b64data = image.split(",", 1) mime = header.split(":", 1)[1].split(";", 1)[0] data = base64.b64decode(b64data) else: data_url = self.image2base64(image) header, b64data = data_url.split(",", 1) mime = header.split(":", 1)[1].split(";", 1)[0] data = base64.b64decode(b64data) return types.Part( inline_data=types.Blob( mime_type=mime, data=data, ) ) def _form_history(self, system, history, images=None): from google.genai import types contents = [] images = images or [] system_len = len(system) if isinstance(system, str) else 0 history_len = len(history) if history else 0 images_len = len(images) logging.info(f"[GeminiCV] _form_history called: system_len={system_len} history_len={history_len} images_len={images_len}") image_parts = [] for img in images: try: image_parts.append(self._image_to_part(img)) except Exception: continue remaining_history = history or [] if system or remaining_history: parts = [] if system: parts.append(types.Part(text=system)) if remaining_history: first = remaining_history[0] parts.append(types.Part(text=first.get("content", ""))) remaining_history = remaining_history[1:] parts.extend(image_parts) contents.append(types.Content(role="user", parts=parts)) elif image_parts: contents.append(types.Content(role="user", parts=image_parts)) role_map = {"user": "user", "assistant": "model", "system": "user"} for h in remaining_history: role = role_map.get(h.get("role"), "user") contents.append( types.Content( role=role, parts=[types.Part(text=h.get("content", ""))], ) ) return contents def describe(self, image): from google.genai import types prompt = ( "请用中文详细描述一下图中的内容,比如时间,地点,人物,事情,人物心情等,如果有数据请提取出数据。" if self.lang.lower() == "chinese" else "Please describe the content of this picture, like where, when, who, what happen. If it has number data, please extract them out." ) contents = [ types.Content( role="user", parts=[ types.Part(text=prompt), self._image_to_part(image), ], ) ] res = self.client.models.generate_content( model=self.model_name, contents=contents, ) return res.text, total_token_count_from_response(res) def describe_with_prompt(self, image, prompt=None): from google.genai import types vision_prompt = prompt if prompt else vision_llm_describe_prompt() contents = [ types.Content( role="user", parts=[ types.Part(text=vision_prompt), self._image_to_part(image), ], ) ] res = self.client.models.generate_content( model=self.model_name, contents=contents, ) return res.text, total_token_count_from_response(res) async def async_chat(self, system, history, gen_conf, images=None, video_bytes=None, filename="", **kwargs): if video_bytes: try: size = len(video_bytes) if video_bytes else 0 logging.info(f"[GeminiCV] async_chat called with video: filename={filename} size={size}") summary, summary_num_tokens = await asyncio.to_thread(self._process_video, video_bytes, filename) return summary, summary_num_tokens except Exception as e: logging.info(f"[GeminiCV] async_chat video error: {e}") return "**ERROR**: " + str(e), 0 from google.genai import types history_len = len(history) if history else 0 images_len = len(images) if images else 0 logging.info(f"[GeminiCV] async_chat called: history_len={history_len} images_len={images_len} gen_conf={gen_conf}") generation_config = types.GenerateContentConfig( temperature=gen_conf.get("temperature", 0.3), top_p=gen_conf.get("top_p", 0.7), ) try: response = await self.client.aio.models.generate_content( model=self.model_name, contents=self._form_history(system, history, images), config=generation_config, ) ans = response.text logging.info("[GeminiCV] async_chat completed") return ans, total_token_count_from_response(response) except Exception as e: logging.warning(f"[GeminiCV] async_chat error: {e}") return "**ERROR**: " + str(e), 0 async def async_chat_streamly(self, system, history, gen_conf, images=None, **kwargs): ans = "" response = None try: from google.genai import types generation_config = types.GenerateContentConfig( temperature=gen_conf.get("temperature", 0.3), top_p=gen_conf.get("top_p", 0.7), ) history_len = len(history) if history else 0 images_len = len(images) if images else 0 logging.info(f"[GeminiCV] async_chat_streamly called: history_len={history_len} images_len={images_len} gen_conf={gen_conf}") response_stream = await self.client.aio.models.generate_content_stream( model=self.model_name, contents=self._form_history(system, history, images),
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/llm/chat_model.py
rag/llm/chat_model.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import asyncio import json import logging import os import random import re import time from abc import ABC from copy import deepcopy from urllib.parse import urljoin import json_repair import litellm import openai from openai import AsyncOpenAI, OpenAI from strenum import StrEnum from common.token_utils import num_tokens_from_string, total_token_count_from_response from rag.llm import FACTORY_DEFAULT_BASE_URL, LITELLM_PROVIDER_PREFIX, SupportedLiteLLMProvider from rag.nlp import is_chinese, is_english # Error message constants class LLMErrorCode(StrEnum): ERROR_RATE_LIMIT = "RATE_LIMIT_EXCEEDED" ERROR_AUTHENTICATION = "AUTH_ERROR" ERROR_INVALID_REQUEST = "INVALID_REQUEST" ERROR_SERVER = "SERVER_ERROR" ERROR_TIMEOUT = "TIMEOUT" ERROR_CONNECTION = "CONNECTION_ERROR" ERROR_MODEL = "MODEL_ERROR" ERROR_MAX_ROUNDS = "ERROR_MAX_ROUNDS" ERROR_CONTENT_FILTER = "CONTENT_FILTERED" ERROR_QUOTA = "QUOTA_EXCEEDED" ERROR_MAX_RETRIES = "MAX_RETRIES_EXCEEDED" ERROR_GENERIC = "GENERIC_ERROR" class ReActMode(StrEnum): FUNCTION_CALL = "function_call" REACT = "react" ERROR_PREFIX = "**ERROR**" LENGTH_NOTIFICATION_CN = "······\n由于大模型的上下文窗口大小限制,回答已经被大模型截断。" LENGTH_NOTIFICATION_EN = "...\nThe answer is truncated by your chosen LLM due to its limitation on context length." class Base(ABC): def __init__(self, key, model_name, base_url, **kwargs): timeout = int(os.environ.get("LLM_TIMEOUT_SECONDS", 600)) self.client = OpenAI(api_key=key, base_url=base_url, timeout=timeout) self.async_client = AsyncOpenAI(api_key=key, base_url=base_url, timeout=timeout) self.model_name = model_name # Configure retry parameters self.max_retries = kwargs.get("max_retries", int(os.environ.get("LLM_MAX_RETRIES", 5))) self.base_delay = kwargs.get("retry_interval", float(os.environ.get("LLM_BASE_DELAY", 2.0))) self.max_rounds = kwargs.get("max_rounds", 5) self.is_tools = False self.tools = [] self.toolcall_sessions = {} def _get_delay(self): return self.base_delay * random.uniform(10, 150) def _classify_error(self, error): error_str = str(error).lower() keywords_mapping = [ (["quota", "capacity", "credit", "billing", "balance", "欠费"], LLMErrorCode.ERROR_QUOTA), (["rate limit", "429", "tpm limit", "too many requests", "requests per minute"], LLMErrorCode.ERROR_RATE_LIMIT), (["auth", "key", "apikey", "401", "forbidden", "permission"], LLMErrorCode.ERROR_AUTHENTICATION), (["invalid", "bad request", "400", "format", "malformed", "parameter"], LLMErrorCode.ERROR_INVALID_REQUEST), (["server", "503", "502", "504", "500", "unavailable"], LLMErrorCode.ERROR_SERVER), (["timeout", "timed out"], LLMErrorCode.ERROR_TIMEOUT), (["connect", "network", "unreachable", "dns"], LLMErrorCode.ERROR_CONNECTION), (["filter", "content", "policy", "blocked", "safety", "inappropriate"], LLMErrorCode.ERROR_CONTENT_FILTER), (["model", "not found", "does not exist", "not available"], LLMErrorCode.ERROR_MODEL), (["max rounds"], LLMErrorCode.ERROR_MODEL), ] for words, code in keywords_mapping: if re.search("({})".format("|".join(words)), error_str): return code return LLMErrorCode.ERROR_GENERIC def _clean_conf(self, gen_conf): model_name_lower = (self.model_name or "").lower() # gpt-5 and gpt-5.1 endpoints have inconsistent parameter support, clear custom generation params to prevent unexpected issues if "gpt-5" in model_name_lower: gen_conf = {} return gen_conf if "max_tokens" in gen_conf: del gen_conf["max_tokens"] allowed_conf = { "temperature", "max_completion_tokens", "top_p", "stream", "stream_options", "stop", "n", "presence_penalty", "frequency_penalty", "functions", "function_call", "logit_bias", "user", "response_format", "seed", "tools", "tool_choice", "logprobs", "top_logprobs", "extra_headers", } gen_conf = {k: v for k, v in gen_conf.items() if k in allowed_conf} return gen_conf async def _async_chat_streamly(self, history, gen_conf, **kwargs): logging.info("[HISTORY STREAMLY]" + json.dumps(history, ensure_ascii=False, indent=4)) reasoning_start = False request_kwargs = {"model": self.model_name, "messages": history, "stream": True, **gen_conf} stop = kwargs.get("stop") if stop: request_kwargs["stop"] = stop response = await self.async_client.chat.completions.create(**request_kwargs) async for resp in response: if not resp.choices: continue if not resp.choices[0].delta.content: resp.choices[0].delta.content = "" if kwargs.get("with_reasoning", True) and hasattr(resp.choices[0].delta, "reasoning_content") and resp.choices[0].delta.reasoning_content: ans = "" if not reasoning_start: reasoning_start = True ans = "<think>" ans += resp.choices[0].delta.reasoning_content + "</think>" else: reasoning_start = False ans = resp.choices[0].delta.content tol = total_token_count_from_response(resp) if not tol: tol = num_tokens_from_string(resp.choices[0].delta.content) finish_reason = resp.choices[0].finish_reason if hasattr(resp.choices[0], "finish_reason") else "" if finish_reason == "length": if is_chinese(ans): ans += LENGTH_NOTIFICATION_CN else: ans += LENGTH_NOTIFICATION_EN yield ans, tol async def async_chat_streamly(self, system, history, gen_conf: dict = {}, **kwargs): if system and history and history[0].get("role") != "system": history.insert(0, {"role": "system", "content": system}) gen_conf = self._clean_conf(gen_conf) ans = "" total_tokens = 0 for attempt in range(self.max_retries + 1): try: async for delta_ans, tol in self._async_chat_streamly(history, gen_conf, **kwargs): ans = delta_ans total_tokens += tol yield ans yield total_tokens return except Exception as e: e = await self._exceptions_async(e, attempt) if e: yield e yield total_tokens return def _length_stop(self, ans): if is_chinese([ans]): return ans + LENGTH_NOTIFICATION_CN return ans + LENGTH_NOTIFICATION_EN @property def _retryable_errors(self) -> set[str]: return { LLMErrorCode.ERROR_RATE_LIMIT, LLMErrorCode.ERROR_SERVER, } def _should_retry(self, error_code: str) -> bool: return error_code in self._retryable_errors def _exceptions(self, e, attempt) -> str | None: logging.exception("OpenAI chat_with_tools") # Classify the error error_code = self._classify_error(e) if attempt == self.max_retries: error_code = LLMErrorCode.ERROR_MAX_RETRIES if self._should_retry(error_code): delay = self._get_delay() logging.warning(f"Error: {error_code}. Retrying in {delay:.2f} seconds... (Attempt {attempt + 1}/{self.max_retries})") time.sleep(delay) return None msg = f"{ERROR_PREFIX}: {error_code} - {str(e)}" logging.error(f"sync base giving up: {msg}") return msg async def _exceptions_async(self, e, attempt): logging.exception("OpenAI async completion") error_code = self._classify_error(e) if attempt == self.max_retries: error_code = LLMErrorCode.ERROR_MAX_RETRIES if self._should_retry(error_code): delay = self._get_delay() logging.warning(f"Error: {error_code}. Retrying in {delay:.2f} seconds... (Attempt {attempt + 1}/{self.max_retries})") await asyncio.sleep(delay) return None msg = f"{ERROR_PREFIX}: {error_code} - {str(e)}" logging.error(f"async base giving up: {msg}") return msg def _verbose_tool_use(self, name, args, res): return "<tool_call>" + json.dumps({"name": name, "args": args, "result": res}, ensure_ascii=False, indent=2) + "</tool_call>" def _append_history(self, hist, tool_call, tool_res): hist.append( { "role": "assistant", "tool_calls": [ { "index": tool_call.index, "id": tool_call.id, "function": { "name": tool_call.function.name, "arguments": tool_call.function.arguments, }, "type": "function", }, ], } ) try: if isinstance(tool_res, dict): tool_res = json.dumps(tool_res, ensure_ascii=False) finally: hist.append({"role": "tool", "tool_call_id": tool_call.id, "content": str(tool_res)}) return hist def bind_tools(self, toolcall_session, tools): if not (toolcall_session and tools): return self.is_tools = True self.toolcall_session = toolcall_session self.tools = tools async def async_chat_with_tools(self, system: str, history: list, gen_conf: dict = {}): gen_conf = self._clean_conf(gen_conf) if system and history and history[0].get("role") != "system": history.insert(0, {"role": "system", "content": system}) ans = "" tk_count = 0 hist = deepcopy(history) for attempt in range(self.max_retries + 1): history = deepcopy(hist) try: for _ in range(self.max_rounds + 1): logging.info(f"{self.tools=}") response = await self.async_client.chat.completions.create(model=self.model_name, messages=history, tools=self.tools, tool_choice="auto", **gen_conf) tk_count += total_token_count_from_response(response) if any([not response.choices, not response.choices[0].message]): raise Exception(f"500 response structure error. Response: {response}") if not hasattr(response.choices[0].message, "tool_calls") or not response.choices[0].message.tool_calls: if hasattr(response.choices[0].message, "reasoning_content") and response.choices[0].message.reasoning_content: ans += "<think>" + response.choices[0].message.reasoning_content + "</think>" ans += response.choices[0].message.content if response.choices[0].finish_reason == "length": ans = self._length_stop(ans) return ans, tk_count for tool_call in response.choices[0].message.tool_calls: logging.info(f"Response {tool_call=}") name = tool_call.function.name try: args = json_repair.loads(tool_call.function.arguments) tool_response = await asyncio.to_thread(self.toolcall_session.tool_call, name, args) history = self._append_history(history, tool_call, tool_response) ans += self._verbose_tool_use(name, args, tool_response) except Exception as e: logging.exception(msg=f"Wrong JSON argument format in LLM tool call response: {tool_call}") history.append({"role": "tool", "tool_call_id": tool_call.id, "content": f"Tool call error: \n{tool_call}\nException:\n" + str(e)}) ans += self._verbose_tool_use(name, {}, str(e)) logging.warning(f"Exceed max rounds: {self.max_rounds}") history.append({"role": "user", "content": f"Exceed max rounds: {self.max_rounds}"}) response, token_count = await self._async_chat(history, gen_conf) ans += response tk_count += token_count return ans, tk_count except Exception as e: e = await self._exceptions_async(e, attempt) if e: return e, tk_count assert False, "Shouldn't be here." async def async_chat_streamly_with_tools(self, system: str, history: list, gen_conf: dict = {}): gen_conf = self._clean_conf(gen_conf) tools = self.tools if system and history and history[0].get("role") != "system": history.insert(0, {"role": "system", "content": system}) total_tokens = 0 hist = deepcopy(history) for attempt in range(self.max_retries + 1): history = deepcopy(hist) try: for _ in range(self.max_rounds + 1): reasoning_start = False logging.info(f"{tools=}") response = await self.async_client.chat.completions.create(model=self.model_name, messages=history, stream=True, tools=tools, tool_choice="auto", **gen_conf) final_tool_calls = {} answer = "" async for resp in response: if not hasattr(resp, "choices") or not resp.choices: continue delta = resp.choices[0].delta if hasattr(delta, "tool_calls") and delta.tool_calls: for tool_call in delta.tool_calls: index = tool_call.index if index not in final_tool_calls: if not tool_call.function.arguments: tool_call.function.arguments = "" final_tool_calls[index] = tool_call else: final_tool_calls[index].function.arguments += tool_call.function.arguments or "" continue if not hasattr(delta, "content") or delta.content is None: delta.content = "" if hasattr(delta, "reasoning_content") and delta.reasoning_content: ans = "" if not reasoning_start: reasoning_start = True ans = "<think>" ans += delta.reasoning_content + "</think>" yield ans else: reasoning_start = False answer += delta.content yield delta.content tol = total_token_count_from_response(resp) if not tol: total_tokens += num_tokens_from_string(delta.content) else: total_tokens = tol finish_reason = getattr(resp.choices[0], "finish_reason", "") if finish_reason == "length": yield self._length_stop("") if answer: yield total_tokens return for tool_call in final_tool_calls.values(): name = tool_call.function.name try: args = json_repair.loads(tool_call.function.arguments) yield self._verbose_tool_use(name, args, "Begin to call...") tool_response = await asyncio.to_thread(self.toolcall_session.tool_call, name, args) history = self._append_history(history, tool_call, tool_response) yield self._verbose_tool_use(name, args, tool_response) except Exception as e: logging.exception(msg=f"Wrong JSON argument format in LLM tool call response: {tool_call}") history.append({"role": "tool", "tool_call_id": tool_call.id, "content": f"Tool call error: \n{tool_call}\nException:\n" + str(e)}) yield self._verbose_tool_use(name, {}, str(e)) logging.warning(f"Exceed max rounds: {self.max_rounds}") history.append({"role": "user", "content": f"Exceed max rounds: {self.max_rounds}"}) response = await self.async_client.chat.completions.create(model=self.model_name, messages=history, stream=True, tools=tools, tool_choice="auto", **gen_conf) async for resp in response: if not hasattr(resp, "choices") or not resp.choices: continue delta = resp.choices[0].delta if not hasattr(delta, "content") or delta.content is None: continue tol = total_token_count_from_response(resp) if not tol: total_tokens += num_tokens_from_string(delta.content) else: total_tokens = tol yield delta.content yield total_tokens return except Exception as e: e = await self._exceptions_async(e, attempt) if e: logging.error(f"async_chat_streamly failed: {e}") yield e yield total_tokens return assert False, "Shouldn't be here." async def _async_chat(self, history, gen_conf, **kwargs): logging.info("[HISTORY]" + json.dumps(history, ensure_ascii=False, indent=2)) if self.model_name.lower().find("qwq") >= 0: logging.info(f"[INFO] {self.model_name} detected as reasoning model, using async_chat_streamly") final_ans = "" tol_token = 0 async for delta, tol in self._async_chat_streamly(history, gen_conf, with_reasoning=False, **kwargs): if delta.startswith("<think>") or delta.endswith("</think>"): continue final_ans += delta tol_token = tol if len(final_ans.strip()) == 0: final_ans = "**ERROR**: Empty response from reasoning model" return final_ans.strip(), tol_token if self.model_name.lower().find("qwen3") >= 0: kwargs["extra_body"] = {"enable_thinking": False} response = await self.async_client.chat.completions.create(model=self.model_name, messages=history, **gen_conf, **kwargs) if not response.choices or not response.choices[0].message or not response.choices[0].message.content: return "", 0 ans = response.choices[0].message.content.strip() if response.choices[0].finish_reason == "length": ans = self._length_stop(ans) return ans, total_token_count_from_response(response) async def async_chat(self, system, history, gen_conf={}, **kwargs): if system and history and history[0].get("role") != "system": history.insert(0, {"role": "system", "content": system}) gen_conf = self._clean_conf(gen_conf) for attempt in range(self.max_retries + 1): try: return await self._async_chat(history, gen_conf, **kwargs) except Exception as e: e = await self._exceptions_async(e, attempt) if e: return e, 0 assert False, "Shouldn't be here." class XinferenceChat(Base): _FACTORY_NAME = "Xinference" def __init__(self, key=None, model_name="", base_url="", **kwargs): if not base_url: raise ValueError("Local llm url cannot be None") base_url = urljoin(base_url, "v1") super().__init__(key, model_name, base_url, **kwargs) class HuggingFaceChat(Base): _FACTORY_NAME = "HuggingFace" def __init__(self, key=None, model_name="", base_url="", **kwargs): if not base_url: raise ValueError("Local llm url cannot be None") base_url = urljoin(base_url, "v1") super().__init__(key, model_name.split("___")[0], base_url, **kwargs) class ModelScopeChat(Base): _FACTORY_NAME = "ModelScope" def __init__(self, key=None, model_name="", base_url="", **kwargs): if not base_url: raise ValueError("Local llm url cannot be None") base_url = urljoin(base_url, "v1") super().__init__(key, model_name.split("___")[0], base_url, **kwargs) class BaiChuanChat(Base): _FACTORY_NAME = "BaiChuan" def __init__(self, key, model_name="Baichuan3-Turbo", base_url="https://api.baichuan-ai.com/v1", **kwargs): if not base_url: base_url = "https://api.baichuan-ai.com/v1" super().__init__(key, model_name, base_url, **kwargs) @staticmethod def _format_params(params): return { "temperature": params.get("temperature", 0.3), "top_p": params.get("top_p", 0.85), } def _clean_conf(self, gen_conf): return { "temperature": gen_conf.get("temperature", 0.3), "top_p": gen_conf.get("top_p", 0.85), } def _chat(self, history, gen_conf={}, **kwargs): response = self.client.chat.completions.create( model=self.model_name, messages=history, extra_body={"tools": [{"type": "web_search", "web_search": {"enable": True, "search_mode": "performance_first"}}]}, **gen_conf, ) ans = response.choices[0].message.content.strip() if response.choices[0].finish_reason == "length": if is_chinese([ans]): ans += LENGTH_NOTIFICATION_CN else: ans += LENGTH_NOTIFICATION_EN return ans, total_token_count_from_response(response) def chat_streamly(self, system, history, gen_conf={}, **kwargs): if system and history and history[0].get("role") != "system": history.insert(0, {"role": "system", "content": system}) if "max_tokens" in gen_conf: del gen_conf["max_tokens"] ans = "" total_tokens = 0 try: response = self.client.chat.completions.create( model=self.model_name, messages=history, extra_body={"tools": [{"type": "web_search", "web_search": {"enable": True, "search_mode": "performance_first"}}]}, stream=True, **self._format_params(gen_conf), ) for resp in response: if not resp.choices: continue if not resp.choices[0].delta.content: resp.choices[0].delta.content = "" ans = resp.choices[0].delta.content tol = total_token_count_from_response(resp) if not tol: total_tokens += num_tokens_from_string(resp.choices[0].delta.content) else: total_tokens = tol if resp.choices[0].finish_reason == "length": if is_chinese([ans]): ans += LENGTH_NOTIFICATION_CN else: ans += LENGTH_NOTIFICATION_EN yield ans except Exception as e: yield ans + "\n**ERROR**: " + str(e) yield total_tokens class LocalAIChat(Base): _FACTORY_NAME = "LocalAI" def __init__(self, key, model_name, base_url=None, **kwargs): super().__init__(key, model_name, base_url=base_url, **kwargs) if not base_url: raise ValueError("Local llm url cannot be None") base_url = urljoin(base_url, "v1") self.client = OpenAI(api_key="empty", base_url=base_url) self.model_name = model_name.split("___")[0] class LocalLLM(Base): def __init__(self, key, model_name, base_url=None, **kwargs): super().__init__(key, model_name, base_url=base_url, **kwargs) from jina import Client self.client = Client(port=12345, protocol="grpc", asyncio=True) def _prepare_prompt(self, system, history, gen_conf): from rag.svr.jina_server import Prompt if system and history and history[0].get("role") != "system": history.insert(0, {"role": "system", "content": system}) return Prompt(message=history, gen_conf=gen_conf) def _stream_response(self, endpoint, prompt): from rag.svr.jina_server import Generation answer = "" try: res = self.client.stream_doc(on=endpoint, inputs=prompt, return_type=Generation) loop = asyncio.get_event_loop() try: while True: answer = loop.run_until_complete(res.__anext__()).text yield answer except StopAsyncIteration: pass except Exception as e: yield answer + "\n**ERROR**: " + str(e) yield num_tokens_from_string(answer) def chat(self, system, history, gen_conf={}, **kwargs): if "max_tokens" in gen_conf: del gen_conf["max_tokens"] prompt = self._prepare_prompt(system, history, gen_conf) chat_gen = self._stream_response("/chat", prompt) ans = next(chat_gen) total_tokens = next(chat_gen) return ans, total_tokens def chat_streamly(self, system, history, gen_conf={}, **kwargs): if "max_tokens" in gen_conf: del gen_conf["max_tokens"] prompt = self._prepare_prompt(system, history, gen_conf) return self._stream_response("/stream", prompt) class VolcEngineChat(Base): _FACTORY_NAME = "VolcEngine" def __init__(self, key, model_name, base_url="https://ark.cn-beijing.volces.com/api/v3", **kwargs): """ Since do not want to modify the original database fields, and the VolcEngine authentication method is quite special, Assemble ark_api_key, ep_id into api_key, store it as a dictionary type, and parse it for use model_name is for display only """ base_url = base_url if base_url else "https://ark.cn-beijing.volces.com/api/v3" ark_api_key = json.loads(key).get("ark_api_key", "") model_name = json.loads(key).get("ep_id", "") + json.loads(key).get("endpoint_id", "") super().__init__(ark_api_key, model_name, base_url, **kwargs) class MistralChat(Base): _FACTORY_NAME = "Mistral" def __init__(self, key, model_name, base_url=None, **kwargs): super().__init__(key, model_name, base_url=base_url, **kwargs) from mistralai.client import MistralClient self.client = MistralClient(api_key=key) self.model_name = model_name def _clean_conf(self, gen_conf): for k in list(gen_conf.keys()): if k not in ["temperature", "top_p", "max_tokens"]: del gen_conf[k] return gen_conf def _chat(self, history, gen_conf={}, **kwargs): gen_conf = self._clean_conf(gen_conf) response = self.client.chat(model=self.model_name, messages=history, **gen_conf) ans = response.choices[0].message.content if response.choices[0].finish_reason == "length": if is_chinese(ans): ans += LENGTH_NOTIFICATION_CN else: ans += LENGTH_NOTIFICATION_EN return ans, total_token_count_from_response(response) def chat_streamly(self, system, history, gen_conf={}, **kwargs): if system and history and history[0].get("role") != "system": history.insert(0, {"role": "system", "content": system}) gen_conf = self._clean_conf(gen_conf) ans = "" total_tokens = 0 try: response = self.client.chat_stream(model=self.model_name, messages=history, **gen_conf, **kwargs) for resp in response: if not resp.choices or not resp.choices[0].delta.content: continue ans = resp.choices[0].delta.content total_tokens += 1 if resp.choices[0].finish_reason == "length": if is_chinese(ans): ans += LENGTH_NOTIFICATION_CN else: ans += LENGTH_NOTIFICATION_EN yield ans except openai.APIError as e: yield ans + "\n**ERROR**: " + str(e) yield total_tokens class LmStudioChat(Base): _FACTORY_NAME = "LM-Studio" def __init__(self, key, model_name, base_url, **kwargs): if not base_url: raise ValueError("Local llm url cannot be None") base_url = urljoin(base_url, "v1") super().__init__(key, model_name, base_url, **kwargs) self.client = OpenAI(api_key="lm-studio", base_url=base_url) self.model_name = model_name class OpenAI_APIChat(Base): _FACTORY_NAME = ["VLLM", "OpenAI-API-Compatible"] def __init__(self, key, model_name, base_url, **kwargs): if not base_url: raise ValueError("url cannot be None") model_name = model_name.split("___")[0] super().__init__(key, model_name, base_url, **kwargs) class LeptonAIChat(Base): _FACTORY_NAME = "LeptonAI" def __init__(self, key, model_name, base_url=None, **kwargs): if not base_url: base_url = urljoin("https://" + model_name + ".lepton.run", "api/v1") super().__init__(key, model_name, base_url, **kwargs) class ReplicateChat(Base): _FACTORY_NAME = "Replicate" def __init__(self, key, model_name, base_url=None, **kwargs): super().__init__(key, model_name, base_url=base_url, **kwargs) from replicate.client import Client self.model_name = model_name self.client = Client(api_token=key) def _chat(self, history, gen_conf={}, **kwargs): system = history[0]["content"] if history and history[0]["role"] == "system" else "" prompt = "\n".join([item["role"] + ":" + item["content"] for item in history[-5:] if item["role"] != "system"]) response = self.client.run( self.model_name, input={"system_prompt": system, "prompt": prompt, **gen_conf}, ) ans = "".join(response) return ans, num_tokens_from_string(ans) def chat_streamly(self, system, history, gen_conf={}, **kwargs): if "max_tokens" in gen_conf: del gen_conf["max_tokens"] prompt = "\n".join([item["role"] + ":" + item["content"] for item in history[-5:]]) ans = "" try: response = self.client.run( self.model_name, input={"system_prompt": system, "prompt": prompt, **gen_conf}, ) for resp in response: ans = resp yield ans except Exception as e: yield ans + "\n**ERROR**: " + str(e)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/rag/llm/__init__.py
rag/llm/__init__.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # AFTER UPDATING THIS FILE, PLEASE ENSURE THAT docs/references/supported_models.mdx IS ALSO UPDATED for consistency! # import importlib import inspect from strenum import StrEnum class SupportedLiteLLMProvider(StrEnum): Tongyi_Qianwen = "Tongyi-Qianwen" Dashscope = "Dashscope" Bedrock = "Bedrock" Moonshot = "Moonshot" xAI = "xAI" DeepInfra = "DeepInfra" Groq = "Groq" Cohere = "Cohere" Gemini = "Gemini" DeepSeek = "DeepSeek" Nvidia = "NVIDIA" TogetherAI = "TogetherAI" Anthropic = "Anthropic" Ollama = "Ollama" LongCat = "LongCat" CometAPI = "CometAPI" SILICONFLOW = "SILICONFLOW" OpenRouter = "OpenRouter" StepFun = "StepFun" PPIO = "PPIO" PerfXCloud = "PerfXCloud" Upstage = "Upstage" NovitaAI = "NovitaAI" Lingyi_AI = "01.AI" GiteeAI = "GiteeAI" AI_302 = "302.AI" JiekouAI = "Jiekou.AI" ZHIPU_AI = "ZHIPU-AI" MiniMax = "MiniMax" DeerAPI = "DeerAPI" GPUStack = "GPUStack" OpenAI = "OpenAI" Azure_OpenAI = "Azure-OpenAI" FACTORY_DEFAULT_BASE_URL = { SupportedLiteLLMProvider.Tongyi_Qianwen: "https://dashscope.aliyuncs.com/compatible-mode/v1", SupportedLiteLLMProvider.Dashscope: "https://dashscope.aliyuncs.com/compatible-mode/v1", SupportedLiteLLMProvider.Moonshot: "https://api.moonshot.cn/v1", SupportedLiteLLMProvider.Ollama: "", SupportedLiteLLMProvider.LongCat: "https://api.longcat.chat/openai", SupportedLiteLLMProvider.CometAPI: "https://api.cometapi.com/v1", SupportedLiteLLMProvider.SILICONFLOW: "https://api.siliconflow.cn/v1", SupportedLiteLLMProvider.OpenRouter: "https://openrouter.ai/api/v1", SupportedLiteLLMProvider.StepFun: "https://api.stepfun.com/v1", SupportedLiteLLMProvider.PPIO: "https://api.ppinfra.com/v3/openai", SupportedLiteLLMProvider.PerfXCloud: "https://cloud.perfxlab.cn/v1", SupportedLiteLLMProvider.Upstage: "https://api.upstage.ai/v1/solar", SupportedLiteLLMProvider.NovitaAI: "https://api.novita.ai/v3/openai", SupportedLiteLLMProvider.Lingyi_AI: "https://api.lingyiwanwu.com/v1", SupportedLiteLLMProvider.GiteeAI: "https://ai.gitee.com/v1/", SupportedLiteLLMProvider.AI_302: "https://api.302.ai/v1", SupportedLiteLLMProvider.Anthropic: "https://api.anthropic.com/", SupportedLiteLLMProvider.JiekouAI: "https://api.jiekou.ai/openai", SupportedLiteLLMProvider.ZHIPU_AI: "https://open.bigmodel.cn/api/paas/v4", SupportedLiteLLMProvider.MiniMax: "https://api.minimaxi.com/v1", SupportedLiteLLMProvider.DeerAPI: "https://api.deerapi.com/v1", SupportedLiteLLMProvider.OpenAI: "https://api.openai.com/v1", } LITELLM_PROVIDER_PREFIX = { SupportedLiteLLMProvider.Tongyi_Qianwen: "dashscope/", SupportedLiteLLMProvider.Dashscope: "dashscope/", SupportedLiteLLMProvider.Bedrock: "bedrock/", SupportedLiteLLMProvider.Moonshot: "moonshot/", SupportedLiteLLMProvider.xAI: "xai/", SupportedLiteLLMProvider.DeepInfra: "deepinfra/", SupportedLiteLLMProvider.Groq: "groq/", SupportedLiteLLMProvider.Cohere: "", # don't need a prefix SupportedLiteLLMProvider.Gemini: "gemini/", SupportedLiteLLMProvider.DeepSeek: "deepseek/", SupportedLiteLLMProvider.Nvidia: "nvidia_nim/", SupportedLiteLLMProvider.TogetherAI: "together_ai/", SupportedLiteLLMProvider.Anthropic: "", # don't need a prefix SupportedLiteLLMProvider.Ollama: "ollama_chat/", SupportedLiteLLMProvider.LongCat: "openai/", SupportedLiteLLMProvider.CometAPI: "openai/", SupportedLiteLLMProvider.SILICONFLOW: "openai/", SupportedLiteLLMProvider.OpenRouter: "openai/", SupportedLiteLLMProvider.StepFun: "openai/", SupportedLiteLLMProvider.PPIO: "openai/", SupportedLiteLLMProvider.PerfXCloud: "openai/", SupportedLiteLLMProvider.Upstage: "openai/", SupportedLiteLLMProvider.NovitaAI: "openai/", SupportedLiteLLMProvider.Lingyi_AI: "openai/", SupportedLiteLLMProvider.GiteeAI: "openai/", SupportedLiteLLMProvider.AI_302: "openai/", SupportedLiteLLMProvider.JiekouAI: "openai/", SupportedLiteLLMProvider.ZHIPU_AI: "openai/", SupportedLiteLLMProvider.MiniMax: "openai/", SupportedLiteLLMProvider.DeerAPI: "openai/", SupportedLiteLLMProvider.GPUStack: "openai/", SupportedLiteLLMProvider.OpenAI: "openai/", SupportedLiteLLMProvider.Azure_OpenAI: "azure/", } ChatModel = globals().get("ChatModel", {}) CvModel = globals().get("CvModel", {}) EmbeddingModel = globals().get("EmbeddingModel", {}) RerankModel = globals().get("RerankModel", {}) Seq2txtModel = globals().get("Seq2txtModel", {}) TTSModel = globals().get("TTSModel", {}) OcrModel = globals().get("OcrModel", {}) MODULE_MAPPING = { "chat_model": ChatModel, "cv_model": CvModel, "embedding_model": EmbeddingModel, "rerank_model": RerankModel, "sequence2txt_model": Seq2txtModel, "tts_model": TTSModel, "ocr_model": OcrModel, } package_name = __name__ for module_name, mapping_dict in MODULE_MAPPING.items(): full_module_name = f"{package_name}.{module_name}" module = importlib.import_module(full_module_name) base_class = None lite_llm_base_class = None for name, obj in inspect.getmembers(module): if inspect.isclass(obj): if name == "Base": base_class = obj elif name == "LiteLLMBase": lite_llm_base_class = obj assert hasattr(obj, "_FACTORY_NAME"), "LiteLLMbase should have _FACTORY_NAME field." if hasattr(obj, "_FACTORY_NAME"): if isinstance(obj._FACTORY_NAME, list): for factory_name in obj._FACTORY_NAME: mapping_dict[factory_name] = obj else: mapping_dict[obj._FACTORY_NAME] = obj if base_class is not None: for _, obj in inspect.getmembers(module): if inspect.isclass(obj) and issubclass(obj, base_class) and obj is not base_class and hasattr(obj, "_FACTORY_NAME"): if isinstance(obj._FACTORY_NAME, list): for factory_name in obj._FACTORY_NAME: mapping_dict[factory_name] = obj else: mapping_dict[obj._FACTORY_NAME] = obj __all__ = [ "ChatModel", "CvModel", "EmbeddingModel", "RerankModel", "Seq2txtModel", "TTSModel", "OcrModel", ]
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false