repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_message_app/test_forget_message.py
test/testcases/test_web_api/test_message_app/test_forget_message.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import random import pytest from test_web_api.common import forget_message, list_memory_message, get_message_content from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = forget_message(invalid_auth, "empty_memory_id", 0) assert res["code"] == expected_code, res assert res["message"] == expected_message, res @pytest.mark.usefixtures("add_memory_with_5_raw_message_func") class TestForgetMessage: @pytest.mark.p1 def test_forget_message(self, WebApiAuth): memory_id = self.memory_id list_res = list_memory_message(WebApiAuth, memory_id) assert list_res["code"] == 0, list_res assert len(list_res["data"]["messages"]["message_list"]) > 0 message = random.choice(list_res["data"]["messages"]["message_list"]) res = forget_message(WebApiAuth, memory_id, message["message_id"]) assert res["code"] == 0, res forgot_message_res = get_message_content(WebApiAuth, memory_id, message["message_id"]) assert forgot_message_res["code"] == 0, forgot_message_res assert forgot_message_res["data"]["forget_at"] not in ["-", ""], forgot_message_res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_message_app/test_add_message.py
test/testcases/test_web_api/test_message_app/test_add_message.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import time import uuid import pytest from test_web_api.common import list_memory_message, add_message from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = add_message(invalid_auth, {}) assert res["code"] == expected_code, res assert res["message"] == expected_message, res @pytest.mark.usefixtures("add_empty_raw_type_memory") class TestAddRawMessage: @pytest.mark.p1 def test_add_raw_message(self, WebApiAuth): memory_id = self.memory_id agent_id = uuid.uuid4().hex session_id = uuid.uuid4().hex message_payload = { "memory_id": [memory_id], "agent_id": agent_id, "session_id": session_id, "user_id": "", "user_input": "what is pineapple?", "agent_response": """ A pineapple is a tropical fruit known for its sweet, tangy flavor and distinctive, spiky appearance. Here are the key facts: Scientific Name: Ananas comosus Physical Description: It has a tough, spiky, diamond-patterned outer skin (rind) that is usually green, yellow, or brownish. Inside, the juicy yellow flesh surrounds a fibrous core. Growth: Unlike most fruits, pineapples do not grow on trees. They grow from a central stem as a composite fruit, meaning they are formed from many individual berries that fuse together around the core. They grow on a short, leafy plant close to the ground. Uses: Pineapples are eaten fresh, cooked, grilled, juiced, or canned. They are a popular ingredient in desserts, fruit salads, savory dishes (like pizzas or ham glazes), smoothies, and cocktails. Nutrition: They are a good source of Vitamin C, manganese, and contain an enzyme called bromelain, which aids in digestion and can tenderize meat. Symbolism: The pineapple is a traditional symbol of hospitality and welcome in many cultures. Are you asking about the fruit itself, or its use in a specific context? """ } add_res = add_message(WebApiAuth, message_payload) assert add_res["code"] == 0, add_res time.sleep(2) # make sure refresh to index before search message_res = list_memory_message(WebApiAuth, memory_id, params={"agent_id": agent_id, "keywords": session_id}) assert message_res["code"] == 0, message_res assert message_res["data"]["messages"]["total_count"] > 0 for message in message_res["data"]["messages"]["message_list"]: assert message["agent_id"] == agent_id, message assert message["session_id"] == session_id, message @pytest.mark.usefixtures("add_empty_multiple_type_memory") class TestAddMultipleTypeMessage: @pytest.mark.p1 def test_add_multiple_type_message(self, WebApiAuth): memory_id = self.memory_id agent_id = uuid.uuid4().hex session_id = uuid.uuid4().hex message_payload = { "memory_id": [memory_id], "agent_id": agent_id, "session_id": session_id, "user_id": "", "user_input": "what is pineapple?", "agent_response": """ A pineapple is a tropical fruit known for its sweet, tangy flavor and distinctive, spiky appearance. Here are the key facts: Scientific Name: Ananas comosus Physical Description: It has a tough, spiky, diamond-patterned outer skin (rind) that is usually green, yellow, or brownish. Inside, the juicy yellow flesh surrounds a fibrous core. Growth: Unlike most fruits, pineapples do not grow on trees. They grow from a central stem as a composite fruit, meaning they are formed from many individual berries that fuse together around the core. They grow on a short, leafy plant close to the ground. Uses: Pineapples are eaten fresh, cooked, grilled, juiced, or canned. They are a popular ingredient in desserts, fruit salads, savory dishes (like pizzas or ham glazes), smoothies, and cocktails. Nutrition: They are a good source of Vitamin C, manganese, and contain an enzyme called bromelain, which aids in digestion and can tenderize meat. Symbolism: The pineapple is a traditional symbol of hospitality and welcome in many cultures. Are you asking about the fruit itself, or its use in a specific context? """ } add_res = add_message(WebApiAuth, message_payload) assert add_res["code"] == 0, add_res time.sleep(2) # make sure refresh to index before search message_res = list_memory_message(WebApiAuth, memory_id, params={"agent_id": agent_id, "keywords": session_id}) assert message_res["code"] == 0, message_res assert message_res["data"]["messages"]["total_count"] > 0 for message in message_res["data"]["messages"]["message_list"]: assert message["agent_id"] == agent_id, message assert message["session_id"] == session_id, message @pytest.mark.usefixtures("add_2_multiple_type_memory") class TestAddToMultipleMemory: @pytest.mark.p1 def test_add_to_multiple_memory(self, WebApiAuth): memory_ids = self.memory_ids agent_id = uuid.uuid4().hex session_id = uuid.uuid4().hex message_payload = { "memory_id": memory_ids, "agent_id": agent_id, "session_id": session_id, "user_id": "", "user_input": "what is pineapple?", "agent_response": """ A pineapple is a tropical fruit known for its sweet, tangy flavor and distinctive, spiky appearance. Here are the key facts: Scientific Name: Ananas comosus Physical Description: It has a tough, spiky, diamond-patterned outer skin (rind) that is usually green, yellow, or brownish. Inside, the juicy yellow flesh surrounds a fibrous core. Growth: Unlike most fruits, pineapples do not grow on trees. They grow from a central stem as a composite fruit, meaning they are formed from many individual berries that fuse together around the core. They grow on a short, leafy plant close to the ground. Uses: Pineapples are eaten fresh, cooked, grilled, juiced, or canned. They are a popular ingredient in desserts, fruit salads, savory dishes (like pizzas or ham glazes), smoothies, and cocktails. Nutrition: They are a good source of Vitamin C, manganese, and contain an enzyme called bromelain, which aids in digestion and can tenderize meat. Symbolism: The pineapple is a traditional symbol of hospitality and welcome in many cultures. Are you asking about the fruit itself, or its use in a specific context? """ } add_res = add_message(WebApiAuth, message_payload) assert add_res["code"] == 0, add_res time.sleep(2) # make sure refresh to index before search for memory_id in memory_ids: message_res = list_memory_message(WebApiAuth, memory_id, params={"agent_id": agent_id, "keywords": session_id}) assert message_res["code"] == 0, message_res assert message_res["data"]["messages"]["total_count"] > 0 for message in message_res["data"]["messages"]["message_list"]: assert message["agent_id"] == agent_id, message assert message["session_id"] == session_id, message
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_message_app/test_list_message.py
test/testcases/test_web_api/test_message_app/test_list_message.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import os import random import pytest from test_web_api.common import list_memory_message from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = list_memory_message(invalid_auth, "") assert res["code"] == expected_code, res assert res["message"] == expected_message, res @pytest.mark.usefixtures("add_memory_with_5_raw_message_func") class TestMessageList: @pytest.mark.p1 def test_params_unset(self, WebApiAuth): memory_id = self.memory_id res = list_memory_message(WebApiAuth, memory_id, params=None) assert res["code"] == 0, res assert len(res["data"]["messages"]["message_list"]) == 5, res @pytest.mark.p1 def test_params_empty(self, WebApiAuth): memory_id = self.memory_id res = list_memory_message(WebApiAuth, memory_id, params={}) assert res["code"] == 0, res assert len(res["data"]["messages"]["message_list"]) == 5, res @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_page_size", [ ({"page": 1, "page_size": 10}, 5), ({"page": 2, "page_size": 10}, 0), ({"page": 1, "page_size": 2}, 2), ({"page": 3, "page_size": 2}, 1), ({"page": 5, "page_size": 10}, 0), ], ids=["normal_first_page", "beyond_max_page", "normal_last_partial_page", "normal_middle_page", "full_data_single_page"], ) def test_page_size(self, WebApiAuth, params, expected_page_size): # have added 5 messages in fixture memory_id = self.memory_id res = list_memory_message(WebApiAuth, memory_id, params=params) assert res["code"] == 0, res assert len(res["data"]["messages"]["message_list"]) == expected_page_size, res @pytest.mark.p2 def test_filter_agent_id(self, WebApiAuth): memory_id = self.memory_id agent_ids = self.agent_ids agent_id = random.choice(agent_ids) res = list_memory_message(WebApiAuth, memory_id, params={"agent_id": agent_id}) assert res["code"] == 0, res for message in res["data"]["messages"]["message_list"]: assert message["agent_id"] == agent_id, message @pytest.mark.p2 @pytest.mark.skipif(os.getenv("DOC_ENGINE") == "infinity", reason="Not support.") def test_search_keyword(self, WebApiAuth): memory_id = self.memory_id session_ids = self.session_ids session_id = random.choice(session_ids) slice_start = random.randint(0, len(session_id) - 2) slice_end = random.randint(slice_start + 1, len(session_id) - 1) keyword = session_id[slice_start:slice_end] res = list_memory_message(WebApiAuth, memory_id, params={"keywords": keyword}) assert res["code"] == 0, res assert len(res["data"]["messages"]["message_list"]) > 0, res for message in res["data"]["messages"]["message_list"]: assert keyword in message["session_id"], message
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_message_app/conftest.py
test/testcases/test_web_api/test_message_app/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import time import uuid import pytest import random from test_web_api.common import create_memory, list_memory, add_message, delete_memory @pytest.fixture(scope="class") def add_empty_raw_type_memory(request, WebApiAuth): def cleanup(): memory_list_res = list_memory(WebApiAuth) exist_memory_ids = [memory["id"] for memory in memory_list_res["data"]["memory_list"]] for _memory_id in exist_memory_ids: delete_memory(WebApiAuth, _memory_id) request.addfinalizer(cleanup) payload = { "name": "test_memory_0", "memory_type": ["raw"], "embd_id": "BAAI/bge-small-en-v1.5@Builtin", "llm_id": "glm-4-flash@ZHIPU-AI" } res = create_memory(WebApiAuth, payload) memory_id = res["data"]["id"] request.cls.memory_id = memory_id request.cls.memory_type = payload["memory_type"] return memory_id @pytest.fixture(scope="class") def add_empty_multiple_type_memory(request, WebApiAuth): def cleanup(): memory_list_res = list_memory(WebApiAuth) exist_memory_ids = [memory["id"] for memory in memory_list_res["data"]["memory_list"]] for _memory_id in exist_memory_ids: delete_memory(WebApiAuth, _memory_id) request.addfinalizer(cleanup) payload = { "name": "test_memory_0", "memory_type": ["raw"] + random.choices(["semantic", "episodic", "procedural"], k=random.randint(1, 3)), "embd_id": "BAAI/bge-small-en-v1.5@Builtin", "llm_id": "glm-4-flash@ZHIPU-AI" } res = create_memory(WebApiAuth, payload) memory_id = res["data"]["id"] request.cls.memory_id = memory_id request.cls.memory_type = payload["memory_type"] return memory_id @pytest.fixture(scope="class") def add_2_multiple_type_memory(request, WebApiAuth): def cleanup(): memory_list_res = list_memory(WebApiAuth) exist_memory_ids = [memory["id"] for memory in memory_list_res["data"]["memory_list"]] for _memory_id in exist_memory_ids: delete_memory(WebApiAuth, _memory_id) request.addfinalizer(cleanup) memory_ids = [] for i in range(2): payload = { "name": f"test_memory_{i}", "memory_type": ["raw"] + random.choices(["semantic", "episodic", "procedural"], k=random.randint(1, 3)), "embd_id": "BAAI/bge-small-en-v1.5@Builtin", "llm_id": "glm-4-flash@ZHIPU-AI" } res = create_memory(WebApiAuth, payload) memory_ids.append(res["data"]["id"]) request.cls.memory_ids = memory_ids return memory_ids @pytest.fixture(scope="class") def add_memory_with_multiple_type_message_func(request, WebApiAuth): def cleanup(): memory_list_res = list_memory(WebApiAuth) exist_memory_ids = [memory["id"] for memory in memory_list_res["data"]["memory_list"]] for _memory_id in exist_memory_ids: delete_memory(WebApiAuth, _memory_id) request.addfinalizer(cleanup) payload = { "name": "test_memory_0", "memory_type": ["raw"] + random.choices(["semantic", "episodic", "procedural"], k=random.randint(1, 3)), "embd_id": "BAAI/bge-small-en-v1.5@Builtin", "llm_id": "glm-4-flash@ZHIPU-AI" } res = create_memory(WebApiAuth, payload) memory_id = res["data"]["id"] agent_id = uuid.uuid4().hex message_payload = { "memory_id": [memory_id], "agent_id": agent_id, "session_id": uuid.uuid4().hex, "user_id": "", "user_input": "what is coriander?", "agent_response": """ Coriander is a versatile herb with two main edible parts, and its name can refer to both: 1. Leaves and Stems (often called Cilantro or Fresh Coriander): These are the fresh, green, fragrant leaves and tender stems of the plant Coriandrum sativum. They have a bright, citrusy, and sometimes pungent flavor. Cilantro is widely used as a garnish or key ingredient in cuisines like Mexican, Indian, Thai, and Middle Eastern. 2. Seeds (called Coriander Seeds): These are the dried, golden-brown seeds of the same plant. When ground, they become coriander powder. The seeds have a warm, nutty, floral, and slightly citrusy taste, completely different from the fresh leaves. They are a fundamental spice in curries, stews, pickles, and baking. Key Point of Confusion: The naming differs by region. In North America, "coriander" typically refers to the seeds, while "cilantro" refers to the fresh leaves. In the UK, Europe, and many other parts of the world, "coriander" refers to the fresh herb, and the seeds are called "coriander seeds." """ } add_message(WebApiAuth, message_payload) request.cls.memory_id = memory_id request.cls.agent_id = agent_id time.sleep(2) # make sure refresh to index before search return memory_id @pytest.fixture(scope="class") def add_memory_with_5_raw_message_func(request, WebApiAuth): def cleanup(): memory_list_res = list_memory(WebApiAuth) exist_memory_ids = [memory["id"] for memory in memory_list_res["data"]["memory_list"]] for _memory_id in exist_memory_ids: delete_memory(WebApiAuth, _memory_id) request.addfinalizer(cleanup) payload = { "name": "test_memory_1", "memory_type": ["raw"], "embd_id": "BAAI/bge-small-en-v1.5@Builtin", "llm_id": "glm-4-flash@ZHIPU-AI" } res = create_memory(WebApiAuth, payload) memory_id = res["data"]["id"] agent_ids = [uuid.uuid4().hex for _ in range(2)] session_ids = [uuid.uuid4().hex for _ in range(5)] for i in range(5): message_payload = { "memory_id": [memory_id], "agent_id": agent_ids[i % 2], "session_id": session_ids[i], "user_id": "", "user_input": "what is coriander?", "agent_response": """ Coriander is a versatile herb with two main edible parts, and its name can refer to both: 1. Leaves and Stems (often called Cilantro or Fresh Coriander): These are the fresh, green, fragrant leaves and tender stems of the plant Coriandrum sativum. They have a bright, citrusy, and sometimes pungent flavor. Cilantro is widely used as a garnish or key ingredient in cuisines like Mexican, Indian, Thai, and Middle Eastern. 2. Seeds (called Coriander Seeds): These are the dried, golden-brown seeds of the same plant. When ground, they become coriander powder. The seeds have a warm, nutty, floral, and slightly citrusy taste, completely different from the fresh leaves. They are a fundamental spice in curries, stews, pickles, and baking. Key Point of Confusion: The naming differs by region. In North America, "coriander" typically refers to the seeds, while "cilantro" refers to the fresh leaves. In the UK, Europe, and many other parts of the world, "coriander" refers to the fresh herb, and the seeds are called "coriander seeds." """ } add_message(WebApiAuth, message_payload) request.cls.memory_id = memory_id request.cls.agent_ids = agent_ids request.cls.session_ids = session_ids time.sleep(2) # make sure refresh to index before search return memory_id
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_message_app/test_search_message.py
test/testcases/test_web_api/test_message_app/test_search_message.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from test_web_api.common import search_message, list_memory_message from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = search_message(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res @pytest.mark.usefixtures("add_memory_with_multiple_type_message_func") class TestSearchMessage: @pytest.mark.p1 def test_query(self, WebApiAuth): memory_id = self.memory_id list_res = list_memory_message(WebApiAuth, memory_id) assert list_res["code"] == 0, list_res assert list_res["data"]["messages"]["total_count"] > 0 query = "Coriander is a versatile herb with two main edible parts. What's its name can refer to?" res = search_message(WebApiAuth, {"memory_id": memory_id, "query": query}) assert res["code"] == 0, res assert len(res["data"]) > 0 @pytest.mark.p2 def test_query_with_agent_filter(self, WebApiAuth): memory_id = self.memory_id list_res = list_memory_message(WebApiAuth, memory_id) assert list_res["code"] == 0, list_res assert list_res["data"]["messages"]["total_count"] > 0 agent_id = self.agent_id query = "Coriander is a versatile herb with two main edible parts. What's its name can refer to?" res = search_message(WebApiAuth, {"memory_id": memory_id, "query": query, "agent_id": agent_id}) assert res["code"] == 0, res assert len(res["data"]) > 0 for message in res["data"]: assert message["agent_id"] == agent_id, message @pytest.mark.p2 def test_query_with_not_default_params(self, WebApiAuth): memory_id = self.memory_id list_res = list_memory_message(WebApiAuth, memory_id) assert list_res["code"] == 0, list_res assert list_res["data"]["messages"]["total_count"] > 0 query = "Coriander is a versatile herb with two main edible parts. What's its name can refer to?" params = { "similarity_threshold": 0.1, "keywords_similarity_weight": 0.6, "top_n": 4 } res = search_message(WebApiAuth, {"memory_id": memory_id, "query": query, **params}) assert res["code"] == 0, res assert len(res["data"]) > 0 assert len(res["data"]) <= params["top_n"]
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_document_app/test_rm_documents.py
test/testcases/test_web_api/test_document_app/test_rm_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import bulk_upload_documents, delete_document, list_documents from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = delete_document(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestDocumentsDeletion: @pytest.mark.p1 @pytest.mark.parametrize( "payload, expected_code, expected_message, remaining", [ (None, 101, "required argument are missing: doc_id; ", 3), ({"doc_id": ""}, 109, "No authorization.", 3), ({"doc_id": "invalid_id"}, 109, "No authorization.", 3), ({"doc_id": "\n!?。;!?\"'"}, 109, "No authorization.", 3), ("not json", 101, "required argument are missing: doc_id; ", 3), (lambda r: {"doc_id": r[0]}, 0, "", 2), ], ) def test_basic_scenarios(self, WebApiAuth, add_documents_func, payload, expected_code, expected_message, remaining): kb_id, document_ids = add_documents_func if callable(payload): payload = payload(document_ids) res = delete_document(WebApiAuth, payload) assert res["code"] == expected_code, res if res["code"] != 0: assert res["message"] == expected_message, res res = list_documents(WebApiAuth, {"kb_id": kb_id}) assert len(res["data"]["docs"]) == remaining, res assert res["data"]["total"] == remaining, res @pytest.mark.p2 def test_repeated_deletion(self, WebApiAuth, add_documents_func): _, document_ids = add_documents_func for doc_id in document_ids: res = delete_document(WebApiAuth, {"doc_id": doc_id}) assert res["code"] == 0, res for doc_id in document_ids: res = delete_document(WebApiAuth, {"doc_id": doc_id}) assert res["code"] == 109, res assert res["message"] == "No authorization.", res @pytest.mark.p3 def test_concurrent_deletion(WebApiAuth, add_dataset, tmp_path): count = 100 kb_id = add_dataset document_ids = bulk_upload_documents(WebApiAuth, kb_id, count, tmp_path) with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(delete_document, WebApiAuth, {"doc_id": document_ids[i]}) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures), responses @pytest.mark.p3 def test_delete_100(WebApiAuth, add_dataset, tmp_path): documents_num = 100 kb_id = add_dataset document_ids = bulk_upload_documents(WebApiAuth, kb_id, documents_num, tmp_path) res = list_documents(WebApiAuth, {"kb_id": kb_id}) assert res["data"]["total"] == documents_num, res for doc_id in document_ids: res = delete_document(WebApiAuth, {"doc_id": doc_id}) assert res["code"] == 0, res res = list_documents(WebApiAuth, {"kb_id": kb_id}) assert res["data"]["total"] == 0, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_document_app/test_create_document.py
test/testcases/test_web_api/test_document_app/test_create_document.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import string from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import create_document, list_kbs from configs import DOCUMENT_NAME_LIMIT, INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth from utils.file_utils import create_txt_file @pytest.mark.p1 @pytest.mark.usefixtures("clear_datasets") class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = create_document(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestDocumentCreate: @pytest.mark.p3 def test_filename_empty(self, WebApiAuth, add_dataset_func): kb_id = add_dataset_func payload = {"name": "", "kb_id": kb_id} res = create_document(WebApiAuth, payload) assert res["code"] == 101, res assert res["message"] == "File name can't be empty.", res @pytest.mark.p2 def test_filename_max_length(self, WebApiAuth, add_dataset_func, tmp_path): kb_id = add_dataset_func fp = create_txt_file(tmp_path / f"{'a' * (DOCUMENT_NAME_LIMIT - 4)}.txt") res = create_document(WebApiAuth, {"name": fp.name, "kb_id": kb_id}) assert res["code"] == 0, res assert res["data"]["name"] == fp.name, res @pytest.mark.p2 def test_invalid_kb_id(self, WebApiAuth): res = create_document(WebApiAuth, {"name": "ragflow_test.txt", "kb_id": "invalid_kb_id"}) assert res["code"] == 102, res assert res["message"] == "Can't find this dataset!", res @pytest.mark.p3 def test_filename_special_characters(self, WebApiAuth, add_dataset_func): kb_id = add_dataset_func illegal_chars = '<>:"/\\|?*' translation_table = str.maketrans({char: "_" for char in illegal_chars}) safe_filename = string.punctuation.translate(translation_table) filename = f"{safe_filename}.txt" res = create_document(WebApiAuth, {"name": filename, "kb_id": kb_id}) assert res["code"] == 0, res assert res["data"]["kb_id"] == kb_id, res assert res["data"]["name"] == filename, f"Expected: {filename}, Got: {res['data']['name']}" @pytest.mark.p3 def test_concurrent_upload(self, WebApiAuth, add_dataset_func): kb_id = add_dataset_func count = 20 filenames = [f"ragflow_test_{i}.txt" for i in range(count)] with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(create_document, WebApiAuth, {"name": name, "kb_id": kb_id}) for name in filenames] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures), responses res = list_kbs(WebApiAuth, {"id": kb_id}) assert res["data"]["kbs"][0]["doc_num"] == count, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_document_app/test_upload_documents.py
test/testcases/test_web_api/test_document_app/test_upload_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import string from concurrent.futures import ThreadPoolExecutor, as_completed import pytest import requests from common import DOCUMENT_APP_URL, list_kbs, upload_documents from configs import DOCUMENT_NAME_LIMIT, HOST_ADDRESS, INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth from requests_toolbelt import MultipartEncoder from utils.file_utils import create_txt_file @pytest.mark.p1 @pytest.mark.usefixtures("clear_datasets") class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = upload_documents(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestDocumentsUpload: @pytest.mark.p1 def test_valid_single_upload(self, WebApiAuth, add_dataset_func, tmp_path): kb_id = add_dataset_func fp = create_txt_file(tmp_path / "ragflow_test.txt") res = upload_documents(WebApiAuth, {"kb_id": kb_id}, [fp]) assert res["code"] == 0, res assert res["data"][0]["kb_id"] == kb_id, res assert res["data"][0]["name"] == fp.name, res @pytest.mark.p1 @pytest.mark.parametrize( "generate_test_files", [ "docx", "excel", "ppt", "image", "pdf", "txt", "md", "json", "eml", "html", ], indirect=True, ) def test_file_type_validation(self, WebApiAuth, add_dataset_func, generate_test_files, request): kb_id = add_dataset_func fp = generate_test_files[request.node.callspec.params["generate_test_files"]] res = upload_documents(WebApiAuth, {"kb_id": kb_id}, [fp]) assert res["code"] == 0, res assert res["data"][0]["kb_id"] == kb_id, res assert res["data"][0]["name"] == fp.name, res @pytest.mark.p2 @pytest.mark.parametrize( "file_type", ["exe", "unknown"], ) def test_unsupported_file_type(self, WebApiAuth, add_dataset_func, tmp_path, file_type): kb_id = add_dataset_func fp = tmp_path / f"ragflow_test.{file_type}" fp.touch() res = upload_documents(WebApiAuth, {"kb_id": kb_id}, [fp]) assert res["code"] == 500, res assert res["message"] == f"ragflow_test.{file_type}: This type of file has not been supported yet!", res @pytest.mark.p2 def test_missing_file(self, WebApiAuth, add_dataset_func): kb_id = add_dataset_func res = upload_documents(WebApiAuth, {"kb_id": kb_id}) assert res["code"] == 101, res assert res["message"] == "No file part!", res @pytest.mark.p3 def test_empty_file(self, WebApiAuth, add_dataset_func, tmp_path): kb_id = add_dataset_func fp = tmp_path / "empty.txt" fp.touch() res = upload_documents(WebApiAuth, {"kb_id": kb_id}, [fp]) assert res["code"] == 0, res assert res["data"][0]["size"] == 0, res @pytest.mark.p3 def test_filename_empty(self, WebApiAuth, add_dataset_func, tmp_path): kb_id = add_dataset_func fp = create_txt_file(tmp_path / "ragflow_test.txt") url = f"{HOST_ADDRESS}{DOCUMENT_APP_URL}/upload" fields = [("file", ("", fp.open("rb"))), ("kb_id", kb_id)] m = MultipartEncoder(fields=fields) res = requests.post( url=url, headers={"Content-Type": m.content_type}, auth=WebApiAuth, data=m, ) assert res.json()["code"] == 101, res assert res.json()["message"] == "No file selected!", res @pytest.mark.p2 def test_filename_exceeds_max_length(self, WebApiAuth, add_dataset_func, tmp_path): kb_id = add_dataset_func fp = create_txt_file(tmp_path / f"{'a' * (DOCUMENT_NAME_LIMIT - 4)}.txt") res = upload_documents(WebApiAuth, {"kb_id": kb_id}, [fp]) assert res["code"] == 0, res assert res["data"][0]["name"] == fp.name, res @pytest.mark.p2 def test_invalid_kb_id(self, WebApiAuth, tmp_path): fp = create_txt_file(tmp_path / "ragflow_test.txt") res = upload_documents(WebApiAuth, {"kb_id": "invalid_kb_id"}, [fp]) assert res["code"] == 100, res assert res["message"] == """LookupError("Can't find this dataset!")""", res @pytest.mark.p2 def test_duplicate_files(self, WebApiAuth, add_dataset_func, tmp_path): kb_id = add_dataset_func fp = create_txt_file(tmp_path / "ragflow_test.txt") res = upload_documents(WebApiAuth, {"kb_id": kb_id}, [fp, fp]) assert res["code"] == 0, res assert len(res["data"]) == 2, res for i in range(len(res["data"])): assert res["data"][i]["kb_id"] == kb_id, res expected_name = fp.name if i != 0: expected_name = f"{fp.stem}({i}){fp.suffix}" assert res["data"][i]["name"] == expected_name, res @pytest.mark.p3 def test_filename_special_characters(self, WebApiAuth, add_dataset_func, tmp_path): kb_id = add_dataset_func illegal_chars = '<>:"/\\|?*' translation_table = str.maketrans({char: "_" for char in illegal_chars}) safe_filename = string.punctuation.translate(translation_table) fp = tmp_path / f"{safe_filename}.txt" fp.write_text("Sample text content") res = upload_documents(WebApiAuth, {"kb_id": kb_id}, [fp]) assert res["code"] == 0, res assert len(res["data"]) == 1, res assert res["data"][0]["kb_id"] == kb_id, res assert res["data"][0]["name"] == fp.name, res @pytest.mark.p1 def test_multiple_files(self, WebApiAuth, add_dataset_func, tmp_path): kb_id = add_dataset_func expected_document_count = 20 fps = [] for i in range(expected_document_count): fp = create_txt_file(tmp_path / f"ragflow_test_{i}.txt") fps.append(fp) res = upload_documents(WebApiAuth, {"kb_id": kb_id}, fps) assert res["code"] == 0, res res = list_kbs(WebApiAuth) assert res["data"]["kbs"][0]["doc_num"] == expected_document_count, res @pytest.mark.p3 def test_concurrent_upload(self, WebApiAuth, add_dataset_func, tmp_path): kb_id = add_dataset_func count = 20 fps = [] for i in range(count): fp = create_txt_file(tmp_path / f"ragflow_test_{i}.txt") fps.append(fp) with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(upload_documents, WebApiAuth, {"kb_id": kb_id}, fps[i : i + 1]) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures), responses res = list_kbs(WebApiAuth) assert res["data"]["kbs"][0]["doc_num"] == count, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_document_app/test_list_documents.py
test/testcases/test_web_api/test_document_app/test_list_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import list_documents from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth from utils import is_sorted @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = list_documents(invalid_auth, {"kb_id": "dataset_id"}) assert res["code"] == expected_code assert res["message"] == expected_message class TestDocumentsList: @pytest.mark.p1 def test_default(self, WebApiAuth, add_documents): kb_id, _ = add_documents res = list_documents(WebApiAuth, {"kb_id": kb_id}) assert res["code"] == 0 assert len(res["data"]["docs"]) == 5 assert res["data"]["total"] == 5 @pytest.mark.p3 @pytest.mark.parametrize( "kb_id, expected_code, expected_message", [ ("", 101, 'Lack of "KB ID"'), ("invalid_dataset_id", 103, "Only owner of dataset authorized for this operation."), ], ) def test_invalid_dataset_id(self, WebApiAuth, kb_id, expected_code, expected_message): res = list_documents(WebApiAuth, {"kb_id": kb_id}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page": None, "page_size": 2}, 0, 5, ""), ({"page": 0, "page_size": 2}, 0, 5, ""), ({"page": 2, "page_size": 2}, 0, 2, ""), ({"page": 3, "page_size": 2}, 0, 1, ""), ({"page": "3", "page_size": 2}, 0, 1, ""), pytest.param({"page": -1, "page_size": 2}, 100, 0, "1064", marks=pytest.mark.skip(reason="issues/5851")), pytest.param({"page": "a", "page_size": 2}, 100, 0, """ValueError("invalid literal for int() with base 10: 'a'")""", marks=pytest.mark.skip(reason="issues/5851")), ], ) def test_page(self, WebApiAuth, add_documents, params, expected_code, expected_page_size, expected_message): kb_id, _ = add_documents res = list_documents(WebApiAuth, {"kb_id": kb_id, **params}) assert res["code"] == expected_code, res if expected_code == 0: assert len(res["data"]["docs"]) == expected_page_size, res assert res["data"]["total"] == 5, res else: assert res["message"] == expected_message, res @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page_size": None}, 0, 5, ""), ({"page_size": 0}, 0, 5, ""), ({"page_size": 1}, 0, 5, ""), ({"page_size": 6}, 0, 5, ""), ({"page_size": "1"}, 0, 5, ""), pytest.param({"page_size": -1}, 100, 0, "1064", marks=pytest.mark.skip(reason="issues/5851")), pytest.param({"page_size": "a"}, 100, 0, """ValueError("invalid literal for int() with base 10: 'a'")""", marks=pytest.mark.skip(reason="issues/5851")), ], ) def test_page_size(self, WebApiAuth, add_documents, params, expected_code, expected_page_size, expected_message): kb_id, _ = add_documents res = list_documents(WebApiAuth, {"kb_id": kb_id, **params}) assert res["code"] == expected_code, res if expected_code == 0: assert len(res["data"]["docs"]) == expected_page_size, res else: assert res["message"] == expected_message, res @pytest.mark.p3 @pytest.mark.parametrize( "params, expected_code, assertions, expected_message", [ ({"orderby": None}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"orderby": "create_time"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"orderby": "update_time"}, 0, lambda r: (is_sorted(r["data"]["docs"], "update_time", True)), ""), pytest.param({"orderby": "name", "desc": "False"}, 0, lambda r: (is_sorted(r["data"]["docs"], "name", False)), "", marks=pytest.mark.skip(reason="issues/5851")), pytest.param({"orderby": "unknown"}, 102, 0, "orderby should be create_time or update_time", marks=pytest.mark.skip(reason="issues/5851")), ], ) def test_orderby(self, WebApiAuth, add_documents, params, expected_code, assertions, expected_message): kb_id, _ = add_documents res = list_documents(WebApiAuth, {"kb_id": kb_id, **params}) assert res["code"] == expected_code, res if expected_code == 0: if callable(assertions): assert assertions(res) else: assert res["message"] == expected_message, res @pytest.mark.p3 @pytest.mark.parametrize( "params, expected_code, assertions, expected_message", [ ({"desc": None}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"desc": "true"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"desc": "True"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"desc": True}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), pytest.param({"desc": "false"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", False)), "", marks=pytest.mark.skip(reason="issues/5851")), ({"desc": "False"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", False)), ""), ({"desc": False}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", False)), ""), ({"desc": "False", "orderby": "update_time"}, 0, lambda r: (is_sorted(r["data"]["docs"], "update_time", False)), ""), pytest.param({"desc": "unknown"}, 102, 0, "desc should be true or false", marks=pytest.mark.skip(reason="issues/5851")), ], ) def test_desc(self, WebApiAuth, add_documents, params, expected_code, assertions, expected_message): kb_id, _ = add_documents res = list_documents(WebApiAuth, {"kb_id": kb_id, **params}) assert res["code"] == expected_code, res if expected_code == 0: if callable(assertions): assert assertions(res) else: assert res["message"] == expected_message, res @pytest.mark.p2 @pytest.mark.parametrize( "params, expected_num", [ ({"keywords": None}, 5), ({"keywords": ""}, 5), ({"keywords": "0"}, 1), ({"keywords": "ragflow_test_upload"}, 5), ({"keywords": "unknown"}, 0), ], ) def test_keywords(self, WebApiAuth, add_documents, params, expected_num): kb_id, _ = add_documents res = list_documents(WebApiAuth, {"kb_id": kb_id, **params}) assert res["code"] == 0, res assert len(res["data"]["docs"]) == expected_num, res assert res["data"]["total"] == expected_num, res @pytest.mark.p3 def test_concurrent_list(self, WebApiAuth, add_documents): kb_id, _ = add_documents count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(list_documents, WebApiAuth, {"kb_id": kb_id}) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures), responses
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_document_app/conftest.py
test/testcases/test_web_api/test_document_app/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import bulk_upload_documents, delete_document, list_documents @pytest.fixture(scope="function") def add_document_func(request, WebApiAuth, add_dataset, ragflow_tmp_dir): def cleanup(): res = list_documents(WebApiAuth, {"kb_id": dataset_id}) for doc in res["data"]["docs"]: delete_document(WebApiAuth, {"doc_id": doc["id"]}) request.addfinalizer(cleanup) dataset_id = add_dataset return dataset_id, bulk_upload_documents(WebApiAuth, dataset_id, 1, ragflow_tmp_dir)[0] @pytest.fixture(scope="class") def add_documents(request, WebApiAuth, add_dataset, ragflow_tmp_dir): def cleanup(): res = list_documents(WebApiAuth, {"kb_id": dataset_id}) for doc in res["data"]["docs"]: delete_document(WebApiAuth, {"doc_id": doc["id"]}) request.addfinalizer(cleanup) dataset_id = add_dataset return dataset_id, bulk_upload_documents(WebApiAuth, dataset_id, 5, ragflow_tmp_dir) @pytest.fixture(scope="function") def add_documents_func(request, WebApiAuth, add_dataset_func, ragflow_tmp_dir): def cleanup(): res = list_documents(WebApiAuth, {"kb_id": dataset_id}) for doc in res["data"]["docs"]: delete_document(WebApiAuth, {"doc_id": doc["id"]}) request.addfinalizer(cleanup) dataset_id = add_dataset_func return dataset_id, bulk_upload_documents(WebApiAuth, dataset_id, 3, ragflow_tmp_dir)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_document_app/test_paser_documents.py
test/testcases/test_web_api/test_document_app/test_paser_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import bulk_upload_documents, list_documents, parse_documents from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth from utils import wait_for @wait_for(30, 1, "Document parsing timeout") def condition(_auth, _kb_id, _document_ids=None): res = list_documents(_auth, {"kb_id": _kb_id}) target_docs = res["data"]["docs"] if _document_ids is None: for doc in target_docs: if doc["run"] != "3": return False return True target_ids = set(_document_ids) for doc in target_docs: if doc["id"] in target_ids: if doc.get("run") != "3": return False return True def validate_document_parse_done(auth, _kb_id, _document_ids): res = list_documents(auth, {"kb_id": _kb_id}) for doc in res["data"]["docs"]: if doc["id"] not in _document_ids: continue assert doc["run"] == "3" assert len(doc["process_begin_at"]) > 0 assert doc["process_duration"] > 0 assert doc["progress"] > 0 assert "Task done" in doc["progress_msg"] def validate_document_parse_cancel(auth, _kb_id, _document_ids): res = list_documents(auth, {"kb_id": _kb_id}) for doc in res["data"]["docs"]: if doc["id"] not in _document_ids: continue assert doc["run"] == "2" assert len(doc["process_begin_at"]) > 0 assert doc["progress"] == 0.0 @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = parse_documents(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestDocumentsParse: @pytest.mark.parametrize( "payload, expected_code, expected_message", [ pytest.param(None, 101, "required argument are missing: doc_ids, run; ", marks=pytest.mark.skip), pytest.param({"doc_ids": [], "run": "1"}, 0, "", marks=pytest.mark.p1), pytest.param({"doc_ids": ["invalid_id"], "run": "1"}, 109, "No authorization.", marks=pytest.mark.p3), pytest.param({"doc_ids": ["\n!?。;!?\"'"], "run": "1"}, 109, "No authorization.", marks=pytest.mark.p3), pytest.param("not json", 101, "required argument are missing: doc_ids, run; ", marks=pytest.mark.skip), pytest.param(lambda r: {"doc_ids": r[:1], "run": "1"}, 0, "", marks=pytest.mark.p1), pytest.param(lambda r: {"doc_ids": r, "run": "1"}, 0, "", marks=pytest.mark.p1), ], ) def test_basic_scenarios(self, WebApiAuth, add_documents_func, payload, expected_code, expected_message): kb_id, document_ids = add_documents_func if callable(payload): payload = payload(document_ids) res = parse_documents(WebApiAuth, payload) assert res["code"] == expected_code, res if expected_code == 0: condition(WebApiAuth, kb_id, payload["doc_ids"]) validate_document_parse_done(WebApiAuth, kb_id, payload["doc_ids"]) else: assert res["message"] == expected_message, res @pytest.mark.parametrize( "payload", [ pytest.param(lambda r: {"doc_ids": ["invalid_id"] + r, "run": "1"}, marks=pytest.mark.p3), pytest.param(lambda r: {"doc_ids": r[:1] + ["invalid_id"] + r[1:3], "run": "1"}, marks=pytest.mark.p1), pytest.param(lambda r: {"doc_ids": r + ["invalid_id"], "run": "1"}, marks=pytest.mark.p3), ], ) def test_parse_partial_invalid_document_id(self, WebApiAuth, add_documents_func, payload): _, document_ids = add_documents_func if callable(payload): payload = payload(document_ids) res = parse_documents(WebApiAuth, payload) assert res["code"] == 109, res assert res["message"] == "No authorization.", res @pytest.mark.p3 def test_repeated_parse(self, WebApiAuth, add_documents_func): kb_id, document_ids = add_documents_func res = parse_documents(WebApiAuth, {"doc_ids": document_ids, "run": "1"}) assert res["code"] == 0, res condition(WebApiAuth, kb_id, document_ids) res = parse_documents(WebApiAuth, {"doc_ids": document_ids, "run": "1"}) assert res["code"] == 0, res @pytest.mark.p3 def test_duplicate_parse(self, WebApiAuth, add_documents_func): kb_id, document_ids = add_documents_func res = parse_documents(WebApiAuth, {"doc_ids": document_ids + document_ids, "run": "1"}) assert res["code"] == 0, res assert res["message"] == "success", res condition(WebApiAuth, kb_id, document_ids) validate_document_parse_done(WebApiAuth, kb_id, document_ids) @pytest.mark.p3 def test_parse_100_files(WebApiAuth, add_dataset_func, tmp_path): @wait_for(100, 1, "Document parsing timeout") def condition(_auth, _kb_id, _document_num): res = list_documents(_auth, {"kb_id": _kb_id, "page_size": _document_num}) for doc in res["data"]["docs"]: if doc["run"] != "3": return False return True document_num = 100 kb_id = add_dataset_func document_ids = bulk_upload_documents(WebApiAuth, kb_id, document_num, tmp_path) res = parse_documents(WebApiAuth, {"doc_ids": document_ids, "run": "1"}) assert res["code"] == 0, res condition(WebApiAuth, kb_id, document_num) validate_document_parse_done(WebApiAuth, kb_id, document_ids) @pytest.mark.p3 def test_concurrent_parse(WebApiAuth, add_dataset_func, tmp_path): @wait_for(120, 1, "Document parsing timeout") def condition(_auth, _kb_id, _document_num): res = list_documents(_auth, {"kb_id": _kb_id, "page_size": _document_num}) for doc in res["data"]["docs"]: if doc["run"] != "3": return False return True count = 100 kb_id = add_dataset_func document_ids = bulk_upload_documents(WebApiAuth, kb_id, count, tmp_path) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( parse_documents, WebApiAuth, {"doc_ids": [document_ids[i]], "run": "1"}, ) for i in range(count) ] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) condition(WebApiAuth, kb_id, count) validate_document_parse_done(WebApiAuth, kb_id, document_ids) # @pytest.mark.skip class TestDocumentsParseStop: @pytest.mark.parametrize( "payload, expected_code, expected_message", [ pytest.param(None, 101, "required argument are missing: doc_ids, run; ", marks=pytest.mark.skip), pytest.param({"doc_ids": [], "run": "2"}, 0, "", marks=pytest.mark.p1), pytest.param({"doc_ids": ["invalid_id"], "run": "2"}, 109, "No authorization.", marks=pytest.mark.p3), pytest.param({"doc_ids": ["\n!?。;!?\"'"], "run": "2"}, 109, "No authorization.", marks=pytest.mark.p3), pytest.param("not json", 101, "required argument are missing: doc_ids, run; ", marks=pytest.mark.skip), pytest.param(lambda r: {"doc_ids": r[:1], "run": "2"}, 0, "", marks=pytest.mark.p1), pytest.param(lambda r: {"doc_ids": r, "run": "2"}, 0, "", marks=pytest.mark.p1), ], ) def test_basic_scenarios(self, WebApiAuth, add_documents_func, payload, expected_code, expected_message): @wait_for(10, 1, "Document parsing timeout") def condition(_auth, _kb_id, _doc_ids): res = list_documents(_auth, {"kb_id": _kb_id}) for doc in res["data"]["docs"]: if doc["id"] in _doc_ids: if doc["run"] != "3": return False return True kb_id, document_ids = add_documents_func parse_documents(WebApiAuth, {"doc_ids": document_ids, "run": "1"}) if callable(payload): payload = payload(document_ids) res = parse_documents(WebApiAuth, payload) assert res["code"] == expected_code, res if expected_code == 0: completed_document_ids = list(set(document_ids) - set(payload["doc_ids"])) condition(WebApiAuth, kb_id, completed_document_ids) validate_document_parse_cancel(WebApiAuth, kb_id, payload["doc_ids"]) validate_document_parse_done(WebApiAuth, kb_id, completed_document_ids) else: assert res["message"] == expected_message, res @pytest.mark.skip @pytest.mark.parametrize( "payload", [ lambda r: {"doc_ids": ["invalid_id"] + r, "run": "2"}, lambda r: {"doc_ids": r[:1] + ["invalid_id"] + r[1:3], "run": "2"}, lambda r: {"doc_ids": r + ["invalid_id"], "run": "2"}, ], ) def test_stop_parse_partial_invalid_document_id(self, WebApiAuth, add_documents_func, payload): kb_id, document_ids = add_documents_func parse_documents(WebApiAuth, {"doc_ids": document_ids, "run": "1"}) if callable(payload): payload = payload(document_ids) res = parse_documents(WebApiAuth, payload) assert res["code"] == 109, res assert res["message"] == "No authorization.", res validate_document_parse_cancel(WebApiAuth, kb_id, document_ids)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_kb_app/test_create_kb.py
test/testcases/test_web_api/test_kb_app/test_create_kb.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import create_kb from configs import DATASET_NAME_LIMIT, INVALID_API_TOKEN from hypothesis import example, given, settings from libs.auth import RAGFlowWebApiAuth from utils.hypothesis_utils import valid_names @pytest.mark.usefixtures("clear_datasets") class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ids=["empty_auth", "invalid_api_token"], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = create_kb(invalid_auth, {"name": "auth_test"}) assert res["code"] == expected_code, res assert res["message"] == expected_message, res @pytest.mark.usefixtures("clear_datasets") class TestCapability: @pytest.mark.p3 def test_create_kb_1k(self, WebApiAuth): for i in range(1_000): payload = {"name": f"dataset_{i}"} res = create_kb(WebApiAuth, payload) assert res["code"] == 0, f"Failed to create dataset {i}" @pytest.mark.p3 def test_create_kb_concurrent(self, WebApiAuth): count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(create_kb, WebApiAuth, {"name": f"dataset_{i}"}) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.usefixtures("clear_datasets") class TestDatasetCreate: @pytest.mark.p1 @given(name=valid_names()) @example("a" * 128) @settings(max_examples=20) def test_name(self, WebApiAuth, name): res = create_kb(WebApiAuth, {"name": name}) assert res["code"] == 0, res @pytest.mark.p2 @pytest.mark.parametrize( "name, expected_message", [ ("", "Dataset name can't be empty."), (" ", "Dataset name can't be empty."), ("a" * (DATASET_NAME_LIMIT + 1), "Dataset name length is 129 which is large than 128"), (0, "Dataset name must be string."), (None, "Dataset name must be string."), ], ids=["empty_name", "space_name", "too_long_name", "invalid_name", "None_name"], ) def test_name_invalid(self, WebApiAuth, name, expected_message): payload = {"name": name} res = create_kb(WebApiAuth, payload) assert res["code"] == 102, res assert expected_message in res["message"], res @pytest.mark.p3 def test_name_duplicated(self, WebApiAuth): name = "duplicated_name" payload = {"name": name} res = create_kb(WebApiAuth, payload) assert res["code"] == 0, res res = create_kb(WebApiAuth, payload) assert res["code"] == 0, res @pytest.mark.p3 def test_name_case_insensitive(self, WebApiAuth): name = "CaseInsensitive" payload = {"name": name.upper()} res = create_kb(WebApiAuth, payload) assert res["code"] == 0, res payload = {"name": name.lower()} res = create_kb(WebApiAuth, payload) assert res["code"] == 0, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_kb_app/test_detail_kb.py
test/testcases/test_web_api/test_kb_app/test_detail_kb.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import ( detail_kb, ) from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = detail_kb(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestDatasetsDetail: @pytest.mark.p1 def test_kb_id(self, WebApiAuth, add_dataset): kb_id = add_dataset payload = {"kb_id": kb_id} res = detail_kb(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["name"] == "kb_0" @pytest.mark.p2 def test_id_wrong_uuid(self, WebApiAuth): payload = {"kb_id": "d94a8dc02c9711f0930f7fbc369eab6d"} res = detail_kb(WebApiAuth, payload) assert res["code"] == 103, res assert "Only owner of dataset authorized for this operation." in res["message"], res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_kb_app/test_list_kbs.py
test/testcases/test_web_api/test_kb_app/test_list_kbs.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import list_kbs from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth from utils import is_sorted class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = list_kbs(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestCapability: @pytest.mark.p3 def test_concurrent_list(self, WebApiAuth): count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(list_kbs, WebApiAuth) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.usefixtures("add_datasets") class TestDatasetsList: @pytest.mark.p1 def test_params_unset(self, WebApiAuth): res = list_kbs(WebApiAuth, None) assert res["code"] == 0, res assert len(res["data"]["kbs"]) == 5, res @pytest.mark.p2 def test_params_empty(self, WebApiAuth): res = list_kbs(WebApiAuth, {}) assert res["code"] == 0, res assert len(res["data"]["kbs"]) == 5, res @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_page_size", [ ({"page": 2, "page_size": 2}, 2), ({"page": 3, "page_size": 2}, 1), ({"page": 4, "page_size": 2}, 0), ({"page": "2", "page_size": 2}, 2), ({"page": 1, "page_size": 10}, 5), ], ids=["normal_middle_page", "normal_last_partial_page", "beyond_max_page", "string_page_number", "full_data_single_page"], ) def test_page(self, WebApiAuth, params, expected_page_size): res = list_kbs(WebApiAuth, params) assert res["code"] == 0, res assert len(res["data"]["kbs"]) == expected_page_size, res @pytest.mark.skip @pytest.mark.p2 @pytest.mark.parametrize( "params, expected_code, expected_message", [ ({"page": 0}, 101, "Input should be greater than or equal to 1"), ({"page": "a"}, 101, "Input should be a valid integer, unable to parse string as an integer"), ], ids=["page_0", "page_a"], ) def test_page_invalid(self, WebApiAuth, params, expected_code, expected_message): res = list_kbs(WebApiAuth, params=params) assert res["code"] == expected_code, res assert expected_message in res["message"], res @pytest.mark.p2 def test_page_none(self, WebApiAuth): params = {"page": None} res = list_kbs(WebApiAuth, params) assert res["code"] == 0, res assert len(res["data"]["kbs"]) == 5, res @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_page_size", [ ({"page": 1, "page_size": 1}, 1), ({"page": 1, "page_size": 3}, 3), ({"page": 1, "page_size": 5}, 5), ({"page": 1, "page_size": 6}, 5), ({"page": 1, "page_size": "1"}, 1), ], ids=["min_valid_page_size", "medium_page_size", "page_size_equals_total", "page_size_exceeds_total", "string_type_page_size"], ) def test_page_size(self, WebApiAuth, params, expected_page_size): res = list_kbs(WebApiAuth, params) assert res["code"] == 0, res assert len(res["data"]["kbs"]) == expected_page_size, res @pytest.mark.skip @pytest.mark.p2 @pytest.mark.parametrize( "params, expected_code, expected_message", [ ({"page_size": 0}, 101, "Input should be greater than or equal to 1"), ({"page_size": "a"}, 101, "Input should be a valid integer, unable to parse string as an integer"), ], ) def test_page_size_invalid(self, WebApiAuth, params, expected_code, expected_message): res = list_kbs(WebApiAuth, params) assert res["code"] == expected_code, res assert expected_message in res["message"], res @pytest.mark.p2 def test_page_size_none(self, WebApiAuth): params = {"page_size": None} res = list_kbs(WebApiAuth, params) assert res["code"] == 0, res assert len(res["data"]["kbs"]) == 5, res @pytest.mark.p2 @pytest.mark.parametrize( "params, assertions", [ ({"orderby": "update_time"}, lambda r: (is_sorted(r["data"]["kbs"], "update_time", True))), ], ids=["orderby_update_time"], ) def test_orderby(self, WebApiAuth, params, assertions): res = list_kbs(WebApiAuth, params) assert res["code"] == 0, res if callable(assertions): assert assertions(res), res @pytest.mark.p2 @pytest.mark.parametrize( "params, assertions", [ ({"desc": "True"}, lambda r: (is_sorted(r["data"]["kbs"], "update_time", True))), ({"desc": "False"}, lambda r: (is_sorted(r["data"]["kbs"], "update_time", False))), ], ids=["desc=True", "desc=False"], ) def test_desc(self, WebApiAuth, params, assertions): res = list_kbs(WebApiAuth, params) assert res["code"] == 0, res if callable(assertions): assert assertions(res), res @pytest.mark.p2 @pytest.mark.parametrize( "params, expected_page_size", [ ({"parser_id": "naive"}, 5), ({"parser_id": "qa"}, 0), ], ids=["naive", "dqa"], ) def test_parser_id(self, WebApiAuth, params, expected_page_size): res = list_kbs(WebApiAuth, params) assert res["code"] == 0, res assert len(res["data"]["kbs"]) == expected_page_size, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_kb_app/test_rm_kb.py
test/testcases/test_web_api/test_kb_app/test_rm_kb.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import ( list_kbs, rm_kb, ) from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = rm_kb(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestDatasetsDelete: @pytest.mark.p1 def test_kb_id(self, WebApiAuth, add_datasets_func): kb_ids = add_datasets_func payload = {"kb_id": kb_ids[0]} res = rm_kb(WebApiAuth, payload) assert res["code"] == 0, res res = list_kbs(WebApiAuth) assert len(res["data"]["kbs"]) == 2, res @pytest.mark.p2 @pytest.mark.usefixtures("add_dataset_func") def test_id_wrong_uuid(self, WebApiAuth): payload = {"kb_id": "d94a8dc02c9711f0930f7fbc369eab6d"} res = rm_kb(WebApiAuth, payload) assert res["code"] == 109, res assert "No authorization." in res["message"], res res = list_kbs(WebApiAuth) assert len(res["data"]["kbs"]) == 1, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_kb_app/test_update_kb.py
test/testcases/test_web_api/test_kb_app/test_update_kb.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import os from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import update_kb from configs import DATASET_NAME_LIMIT, INVALID_API_TOKEN from hypothesis import HealthCheck, example, given, settings from libs.auth import RAGFlowWebApiAuth from utils import encode_avatar from utils.file_utils import create_image_file from utils.hypothesis_utils import valid_names class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ids=["empty_auth", "invalid_api_token"], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = update_kb(invalid_auth, "dataset_id") assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestCapability: @pytest.mark.p3 def test_update_dateset_concurrent(self, WebApiAuth, add_dataset_func): dataset_id = add_dataset_func count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( update_kb, WebApiAuth, { "kb_id": dataset_id, "name": f"dataset_{i}", "description": "", "parser_id": "naive", }, ) for i in range(count) ] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) class TestDatasetUpdate: @pytest.mark.p3 def test_dataset_id_not_uuid(self, WebApiAuth): payload = {"name": "not uuid", "description": "", "parser_id": "naive", "kb_id": "not_uuid"} res = update_kb(WebApiAuth, payload) assert res["code"] == 109, res assert "No authorization." in res["message"], res @pytest.mark.p1 @given(name=valid_names()) @example("a" * 128) @settings(max_examples=20, suppress_health_check=[HealthCheck.function_scoped_fixture]) def test_name(self, WebApiAuth, add_dataset_func, name): dataset_id = add_dataset_func payload = {"name": name, "description": "", "parser_id": "naive", "kb_id": dataset_id} res = update_kb(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["name"] == name, res @pytest.mark.p2 @pytest.mark.parametrize( "name, expected_message", [ ("", "Dataset name can't be empty."), (" ", "Dataset name can't be empty."), ("a" * (DATASET_NAME_LIMIT + 1), "Dataset name length is 129 which is large than 128"), (0, "Dataset name must be string."), (None, "Dataset name must be string."), ], ids=["empty_name", "space_name", "too_long_name", "invalid_name", "None_name"], ) def test_name_invalid(self, WebApiAuth, add_dataset_func, name, expected_message): kb_id = add_dataset_func payload = {"name": name, "description": "", "parser_id": "naive", "kb_id": kb_id} res = update_kb(WebApiAuth, payload) assert res["code"] == 102, res assert expected_message in res["message"], res @pytest.mark.p3 def test_name_duplicated(self, WebApiAuth, add_datasets_func): kb_id = add_datasets_func[0] name = "kb_1" payload = {"name": name, "description": "", "parser_id": "naive", "kb_id": kb_id} res = update_kb(WebApiAuth, payload) assert res["code"] == 102, res assert res["message"] == "Duplicated dataset name.", res @pytest.mark.p3 def test_name_case_insensitive(self, WebApiAuth, add_datasets_func): kb_id = add_datasets_func[0] name = "KB_1" payload = {"name": name, "description": "", "parser_id": "naive", "kb_id": kb_id} res = update_kb(WebApiAuth, payload) assert res["code"] == 102, res assert res["message"] == "Duplicated dataset name.", res @pytest.mark.p2 def test_avatar(self, WebApiAuth, add_dataset_func, tmp_path): kb_id = add_dataset_func fn = create_image_file(tmp_path / "ragflow_test.png") payload = { "name": "avatar", "description": "", "parser_id": "naive", "kb_id": kb_id, "avatar": f"data:image/png;base64,{encode_avatar(fn)}", } res = update_kb(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["avatar"] == f"data:image/png;base64,{encode_avatar(fn)}", res @pytest.mark.p2 def test_description(self, WebApiAuth, add_dataset_func): kb_id = add_dataset_func payload = {"name": "description", "description": "description", "parser_id": "naive", "kb_id": kb_id} res = update_kb(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["description"] == "description", res @pytest.mark.p1 @pytest.mark.parametrize( "embedding_model", [ "BAAI/bge-small-en-v1.5@Builtin", "embedding-3@ZHIPU-AI", ], ids=["builtin_baai", "tenant_zhipu"], ) def test_embedding_model(self, WebApiAuth, add_dataset_func, embedding_model): kb_id = add_dataset_func payload = {"name": "embedding_model", "description": "", "parser_id": "naive", "kb_id": kb_id, "embd_id": embedding_model} res = update_kb(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["embd_id"] == embedding_model, res @pytest.mark.p1 @pytest.mark.parametrize( "permission", [ "me", "team", ], ids=["me", "team"], ) def test_permission(self, WebApiAuth, add_dataset_func, permission): kb_id = add_dataset_func payload = {"name": "permission", "description": "", "parser_id": "naive", "kb_id": kb_id, "permission": permission} res = update_kb(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["permission"] == permission.lower().strip(), res @pytest.mark.p1 @pytest.mark.parametrize( "chunk_method", [ "naive", "book", "email", "laws", "manual", "one", "paper", "picture", "presentation", "qa", "table", pytest.param("tag", marks=pytest.mark.skipif(os.getenv("DOC_ENGINE") == "infinity", reason="Infinity does not support parser_id=tag")), ], ids=["naive", "book", "email", "laws", "manual", "one", "paper", "picture", "presentation", "qa", "table", "tag"], ) def test_chunk_method(self, WebApiAuth, add_dataset_func, chunk_method): kb_id = add_dataset_func payload = {"name": "chunk_method", "description": "", "parser_id": chunk_method, "kb_id": kb_id} res = update_kb(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["parser_id"] == chunk_method, res @pytest.mark.p1 @pytest.mark.skipif(os.getenv("DOC_ENGINE") != "infinity", reason="Infinity does not support parser_id=tag") def test_chunk_method_tag_with_infinity(self, WebApiAuth, add_dataset_func): kb_id = add_dataset_func payload = {"name": "chunk_method", "description": "", "parser_id": "tag", "kb_id": kb_id} res = update_kb(WebApiAuth, payload) assert res["code"] == 103, res assert res["message"] == "The chunking method Tag has not been supported by Infinity yet.", res @pytest.mark.skipif(os.getenv("DOC_ENGINE") == "infinity", reason="#8208") @pytest.mark.p2 @pytest.mark.parametrize("pagerank", [0, 50, 100], ids=["min", "mid", "max"]) def test_pagerank(self, WebApiAuth, add_dataset_func, pagerank): kb_id = add_dataset_func payload = {"name": "pagerank", "description": "", "parser_id": "naive", "kb_id": kb_id, "pagerank": pagerank} res = update_kb(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["pagerank"] == pagerank, res @pytest.mark.skipif(os.getenv("DOC_ENGINE") == "infinity", reason="#8208") @pytest.mark.p2 def test_pagerank_set_to_0(self, WebApiAuth, add_dataset_func): kb_id = add_dataset_func payload = {"name": "pagerank", "description": "", "parser_id": "naive", "kb_id": kb_id, "pagerank": 50} res = update_kb(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["pagerank"] == 50, res payload = {"name": "pagerank", "description": "", "parser_id": "naive", "kb_id": kb_id, "pagerank": 0} res = update_kb(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["pagerank"] == 0, res @pytest.mark.skipif(os.getenv("DOC_ENGINE") != "infinity", reason="#8208") @pytest.mark.p2 def test_pagerank_infinity(self, WebApiAuth, add_dataset_func): kb_id = add_dataset_func payload = {"name": "pagerank", "description": "", "parser_id": "naive", "kb_id": kb_id, "pagerank": 50} res = update_kb(WebApiAuth, payload) assert res["code"] == 102, res assert res["message"] == "'pagerank' can only be set when doc_engine is elasticsearch", res @pytest.mark.p1 @pytest.mark.parametrize( "parser_config", [ {"auto_keywords": 0}, {"auto_keywords": 16}, {"auto_keywords": 32}, {"auto_questions": 0}, {"auto_questions": 5}, {"auto_questions": 10}, {"chunk_token_num": 1}, {"chunk_token_num": 1024}, {"chunk_token_num": 2048}, {"delimiter": "\n"}, {"delimiter": " "}, {"html4excel": True}, {"html4excel": False}, {"layout_recognize": "DeepDOC"}, {"layout_recognize": "Plain Text"}, {"tag_kb_ids": ["1", "2"]}, {"topn_tags": 1}, {"topn_tags": 5}, {"topn_tags": 10}, {"filename_embd_weight": 0.1}, {"filename_embd_weight": 0.5}, {"filename_embd_weight": 1.0}, {"task_page_size": 1}, {"task_page_size": None}, {"pages": [[1, 100]]}, {"pages": None}, {"graphrag": {"use_graphrag": True}}, {"graphrag": {"use_graphrag": False}}, {"graphrag": {"entity_types": ["age", "sex", "height", "weight"]}}, {"graphrag": {"method": "general"}}, {"graphrag": {"method": "light"}}, {"graphrag": {"community": True}}, {"graphrag": {"community": False}}, {"graphrag": {"resolution": True}}, {"graphrag": {"resolution": False}}, {"raptor": {"use_raptor": True}}, {"raptor": {"use_raptor": False}}, {"raptor": {"prompt": "Who are you?"}}, {"raptor": {"max_token": 1}}, {"raptor": {"max_token": 1024}}, {"raptor": {"max_token": 2048}}, {"raptor": {"threshold": 0.0}}, {"raptor": {"threshold": 0.5}}, {"raptor": {"threshold": 1.0}}, {"raptor": {"max_cluster": 1}}, {"raptor": {"max_cluster": 512}}, {"raptor": {"max_cluster": 1024}}, {"raptor": {"random_seed": 0}}, ], ids=[ "auto_keywords_min", "auto_keywords_mid", "auto_keywords_max", "auto_questions_min", "auto_questions_mid", "auto_questions_max", "chunk_token_num_min", "chunk_token_num_mid", "chunk_token_num_max", "delimiter", "delimiter_space", "html4excel_true", "html4excel_false", "layout_recognize_DeepDOC", "layout_recognize_navie", "tag_kb_ids", "topn_tags_min", "topn_tags_mid", "topn_tags_max", "filename_embd_weight_min", "filename_embd_weight_mid", "filename_embd_weight_max", "task_page_size_min", "task_page_size_None", "pages", "pages_none", "graphrag_true", "graphrag_false", "graphrag_entity_types", "graphrag_method_general", "graphrag_method_light", "graphrag_community_true", "graphrag_community_false", "graphrag_resolution_true", "graphrag_resolution_false", "raptor_true", "raptor_false", "raptor_prompt", "raptor_max_token_min", "raptor_max_token_mid", "raptor_max_token_max", "raptor_threshold_min", "raptor_threshold_mid", "raptor_threshold_max", "raptor_max_cluster_min", "raptor_max_cluster_mid", "raptor_max_cluster_max", "raptor_random_seed_min", ], ) def test_parser_config(self, WebApiAuth, add_dataset_func, parser_config): kb_id = add_dataset_func payload = {"name": "parser_config", "description": "", "parser_id": "naive", "kb_id": kb_id, "parser_config": parser_config} res = update_kb(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["parser_config"] == parser_config, res @pytest.mark.p2 @pytest.mark.parametrize( "payload", [ {"id": "id"}, {"tenant_id": "e57c1966f99211efb41e9e45646e0111"}, {"created_by": "created_by"}, {"create_date": "Tue, 11 Mar 2025 13:37:23 GMT"}, {"create_time": 1741671443322}, {"update_date": "Tue, 11 Mar 2025 13:37:23 GMT"}, {"update_time": 1741671443339}, ], ) def test_field_unsupported(self, WebApiAuth, add_dataset_func, payload): kb_id = add_dataset_func full_payload = {"name": "field_unsupported", "description": "", "parser_id": "naive", "kb_id": kb_id, **payload} res = update_kb(WebApiAuth, full_payload) assert res["code"] == 101, res assert "isn't allowed" in res["message"], res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_kb_app/conftest.py
test/testcases/test_web_api/test_kb_app/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import batch_create_datasets from libs.auth import RAGFlowWebApiAuth from pytest import FixtureRequest from ragflow_sdk import RAGFlow @pytest.fixture(scope="class") def add_datasets(request: FixtureRequest, client: RAGFlow, WebApiAuth: RAGFlowWebApiAuth) -> list[str]: def cleanup(): client.delete_datasets(ids=None) request.addfinalizer(cleanup) return batch_create_datasets(WebApiAuth, 5) @pytest.fixture(scope="function") def add_datasets_func(request: FixtureRequest, client: RAGFlow, WebApiAuth: RAGFlowWebApiAuth) -> list[str]: def cleanup(): client.delete_datasets(ids=None) request.addfinalizer(cleanup) return batch_create_datasets(WebApiAuth, 3)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_dialog_app/test_get_dialog.py
test/testcases/test_web_api/test_dialog_app/test_get_dialog.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import create_dialog, get_dialog from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth @pytest.mark.usefixtures("clear_dialogs") class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ids=["empty_auth", "invalid_api_token"], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message, add_dialog_func): _, dialog_id = add_dialog_func res = get_dialog(invalid_auth, {"dialog_id": dialog_id}) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestDialogGet: @pytest.mark.p1 def test_get_existing_dialog(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func res = get_dialog(WebApiAuth, {"dialog_id": dialog_id}) assert res["code"] == 0, res data = res["data"] assert data["id"] == dialog_id, res assert "name" in data, res assert "description" in data, res assert "kb_ids" in data, res assert "kb_names" in data, res assert "prompt_config" in data, res assert "llm_setting" in data, res assert "top_n" in data, res assert "top_k" in data, res assert "similarity_threshold" in data, res assert "vector_similarity_weight" in data, res @pytest.mark.p1 def test_get_dialog_with_kb_names(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func res = get_dialog(WebApiAuth, {"dialog_id": dialog_id}) assert res["code"] == 0, res data = res["data"] assert isinstance(data["kb_ids"], list), res assert isinstance(data["kb_names"], list), res assert len(data["kb_ids"]) == len(data["kb_names"]), res @pytest.mark.p2 def test_get_nonexistent_dialog(self, WebApiAuth): fake_dialog_id = "nonexistent_dialog_id" res = get_dialog(WebApiAuth, {"dialog_id": fake_dialog_id}) assert res["code"] == 102, res assert "Dialog not found" in res["message"], res @pytest.mark.p2 def test_get_dialog_missing_id(self, WebApiAuth): res = get_dialog(WebApiAuth, {}) assert res["code"] == 100, res assert res["message"] == "<BadRequestKeyError '400: Bad Request'>", res @pytest.mark.p2 def test_get_dialog_empty_id(self, WebApiAuth): res = get_dialog(WebApiAuth, {"dialog_id": ""}) assert res["code"] == 102, res @pytest.mark.p2 def test_get_dialog_invalid_id_format(self, WebApiAuth): res = get_dialog(WebApiAuth, {"dialog_id": "invalid_format"}) assert res["code"] == 102, res @pytest.mark.p3 def test_get_dialog_data_structure(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func res = get_dialog(WebApiAuth, {"dialog_id": dialog_id}) assert res["code"] == 0, res data = res["data"] required_fields = [ "id", "name", "description", "kb_ids", "kb_names", "prompt_config", "llm_setting", "top_n", "top_k", "similarity_threshold", "vector_similarity_weight", "create_time", "update_time", ] for field in required_fields: assert field in data, f"Missing field: {field}" assert isinstance(data["id"], str), res assert isinstance(data["name"], str), res assert isinstance(data["kb_ids"], list), res assert isinstance(data["kb_names"], list), res assert isinstance(data["prompt_config"], dict), res assert isinstance(data["top_n"], int), res assert isinstance(data["top_k"], int), res assert isinstance(data["similarity_threshold"], (int, float)), res assert isinstance(data["vector_similarity_weight"], (int, float)), res @pytest.mark.p3 def test_get_dialog_prompt_config_structure(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func res = get_dialog(WebApiAuth, {"dialog_id": dialog_id}) assert res["code"] == 0, res prompt_config = res["data"]["prompt_config"] assert "system" in prompt_config, res assert "parameters" in prompt_config, res assert isinstance(prompt_config["system"], str), res assert isinstance(prompt_config["parameters"], list), res @pytest.mark.p3 def test_get_dialog_with_multiple_kbs(self, WebApiAuth, add_dataset_func): dataset_id1 = add_dataset_func dataset_id2 = add_dataset_func payload = { "name": "multi_kb_dialog", "kb_ids": [dataset_id1, dataset_id2], "prompt_config": {"system": "You are a helpful assistant with knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}]}, } create_res = create_dialog(WebApiAuth, payload) assert create_res["code"] == 0, create_res dialog_id = create_res["data"]["id"] res = get_dialog(WebApiAuth, {"dialog_id": dialog_id}) assert res["code"] == 0, res data = res["data"] assert len(data["kb_ids"]) == 2, res assert len(data["kb_names"]) == 2, res assert dataset_id1 in data["kb_ids"], res assert dataset_id2 in data["kb_ids"], res @pytest.mark.p3 def test_get_dialog_with_invalid_kb(self, WebApiAuth): payload = { "name": "invalid_kb_dialog", "kb_ids": ["invalid_kb_id"], "prompt_config": {"system": "You are a helpful assistant with knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}]}, } create_res = create_dialog(WebApiAuth, payload) assert create_res["code"] == 0, create_res dialog_id = create_res["data"]["id"] res = get_dialog(WebApiAuth, {"dialog_id": dialog_id}) assert res["code"] == 0, res data = res["data"] assert len(data["kb_ids"]) == 0, res assert len(data["kb_names"]) == 0, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_dialog_app/test_dialog_edge_cases.py
test/testcases/test_web_api/test_dialog_app/test_dialog_edge_cases.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import create_dialog, delete_dialog, get_dialog, update_dialog @pytest.mark.usefixtures("clear_dialogs") class TestDialogEdgeCases: @pytest.mark.p2 def test_create_dialog_with_tavily_api_key(self, WebApiAuth): """Test creating dialog with Tavily API key instead of dataset""" payload = { "name": "tavily_dialog", "prompt_config": {"system": "You are a helpful assistant. Use this knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}], "tavily_api_key": "test_tavily_key"}, } res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res @pytest.mark.skip @pytest.mark.p2 def test_create_dialog_with_different_embedding_models(self, WebApiAuth): """Test creating dialog with knowledge bases that have different embedding models""" # This test would require creating datasets with different embedding models # For now, we'll test the error case with a mock scenario payload = { "name": "mixed_embedding_dialog", "kb_ids": ["kb_with_model_a", "kb_with_model_b"], "prompt_config": {"system": "You are a helpful assistant with knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}]}, } res = create_dialog(WebApiAuth, payload) # This should fail due to different embedding models assert res["code"] == 102, res assert "Datasets use different embedding models" in res["message"], res @pytest.mark.p2 def test_create_dialog_with_extremely_long_system_prompt(self, WebApiAuth): """Test creating dialog with very long system prompt""" long_prompt = "You are a helpful assistant. " * 1000 payload = {"name": "long_prompt_dialog", "prompt_config": {"system": long_prompt, "parameters": []}} res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res @pytest.mark.p2 def test_create_dialog_with_unicode_characters(self, WebApiAuth): """Test creating dialog with Unicode characters in various fields""" payload = { "name": "Unicode测试对话🤖", "description": "测试Unicode字符支持 with émojis 🚀🌟", "icon": "🤖", "prompt_config": {"system": "你是一个有用的助手。You are helpful. Vous êtes utile. 🌍", "parameters": []}, } res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["name"] == "Unicode测试对话🤖", res assert res["data"]["description"] == "测试Unicode字符支持 with émojis 🚀🌟", res @pytest.mark.p2 def test_create_dialog_with_extreme_parameter_values(self, WebApiAuth): """Test creating dialog with extreme parameter values""" payload = { "name": "extreme_params_dialog", "top_n": 0, "top_k": 1, "similarity_threshold": 0.0, "vector_similarity_weight": 1.0, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}, } res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["top_n"] == 0, res assert res["data"]["top_k"] == 1, res assert res["data"]["similarity_threshold"] == 0.0, res assert res["data"]["vector_similarity_weight"] == 1.0, res @pytest.mark.p2 def test_create_dialog_with_negative_parameter_values(self, WebApiAuth): """Test creating dialog with negative parameter values""" payload = { "name": "negative_params_dialog", "top_n": -1, "top_k": -100, "similarity_threshold": -0.5, "vector_similarity_weight": -0.3, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}, } res = create_dialog(WebApiAuth, payload) assert res["code"] in [0, 102], res @pytest.mark.p2 def test_update_dialog_with_empty_kb_ids(self, WebApiAuth, add_dialog_func): """Test updating dialog to remove all knowledge bases""" dataset_id, dialog_id = add_dialog_func payload = {"dialog_id": dialog_id, "kb_ids": [], "prompt_config": {"system": "You are a helpful assistant without knowledge.", "parameters": []}} res = update_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["kb_ids"] == [], res @pytest.mark.p2 def test_update_dialog_with_null_values(self, WebApiAuth, add_dialog_func): """Test updating dialog with null/None values""" dataset_id, dialog_id = add_dialog_func payload = {"dialog_id": dialog_id, "description": None, "icon": None, "rerank_id": None, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = update_dialog(WebApiAuth, payload) assert res["code"] == 0, res @pytest.mark.p3 def test_dialog_with_complex_prompt_parameters(self, WebApiAuth, add_dataset_func): """Test dialog with complex prompt parameter configurations""" payload = { "name": "complex_params_dialog", "prompt_config": { "system": "You are {role} assistant. Use {knowledge} and consider {context}. Optional: {optional_param}", "parameters": [{"key": "role", "optional": False}, {"key": "knowledge", "optional": True}, {"key": "context", "optional": False}, {"key": "optional_param", "optional": True}], }, "kb_ids": [add_dataset_func], } res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res @pytest.mark.p3 def test_dialog_with_malformed_prompt_parameters(self, WebApiAuth): """Test dialog with malformed prompt parameter configurations""" payload = { "name": "malformed_params_dialog", "prompt_config": { "system": "You are a helpful assistant.", "parameters": [ { "key": "", "optional": False, }, {"optional": True}, { "key": "valid_param", }, ], }, } res = create_dialog(WebApiAuth, payload) assert res["code"] in [0, 102], res @pytest.mark.p3 def test_dialog_operations_with_special_ids(self, WebApiAuth): """Test dialog operations with special ID formats""" special_ids = [ "00000000-0000-0000-0000-000000000000", "ffffffff-ffff-ffff-ffff-ffffffffffff", "12345678-1234-1234-1234-123456789abc", ] for special_id in special_ids: res = get_dialog(WebApiAuth, {"dialog_id": special_id}) assert res["code"] == 102, f"Should fail for ID: {special_id}" res = delete_dialog(WebApiAuth, {"dialog_ids": [special_id]}) assert res["code"] == 103, f"Should fail for ID: {special_id}" @pytest.mark.p3 def test_dialog_with_extremely_large_llm_settings(self, WebApiAuth): """Test dialog with very large LLM settings""" large_llm_setting = { "model": "gpt-4", "temperature": 0.7, "max_tokens": 999999, "custom_param_" + "x" * 1000: "large_value_" + "y" * 1000, } payload = {"name": "large_llm_settings_dialog", "llm_setting": large_llm_setting, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res @pytest.mark.p3 def test_concurrent_dialog_operations(self, WebApiAuth, add_dialog_func): """Test concurrent operations on the same dialog""" from concurrent.futures import ThreadPoolExecutor, as_completed _, dialog_id = add_dialog_func def update_operation(i): payload = {"dialog_id": dialog_id, "name": f"concurrent_update_{i}", "prompt_config": {"system": f"You are assistant number {i}.", "parameters": []}} return update_dialog(WebApiAuth, payload) with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(update_operation, i) for i in range(10)] responses = [future.result() for future in as_completed(futures)] successful_updates = sum(1 for response in responses if response["code"] == 0) assert successful_updates > 0, "No updates succeeded" res = get_dialog(WebApiAuth, {"dialog_id": dialog_id}) assert res["code"] == 0, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_dialog_app/test_delete_dialogs.py
test/testcases/test_web_api/test_dialog_app/test_delete_dialogs.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import batch_create_dialogs, create_dialog, delete_dialog, list_dialogs from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth @pytest.mark.usefixtures("clear_dialogs") class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ids=["empty_auth", "invalid_api_token"], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message, add_dialog_func): _, dialog_id = add_dialog_func payload = {"dialog_ids": [dialog_id]} res = delete_dialog(invalid_auth, payload) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestDialogDelete: @pytest.mark.p1 def test_delete_single_dialog(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 1, res payload = {"dialog_ids": [dialog_id]} res = delete_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"] is True, res res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 0, res @pytest.mark.p1 def test_delete_multiple_dialogs(self, WebApiAuth, add_dialogs_func): _, dialog_ids = add_dialogs_func res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 5, res payload = {"dialog_ids": dialog_ids} res = delete_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"] is True, res res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 0, res @pytest.mark.p1 def test_delete_partial_dialogs(self, WebApiAuth, add_dialogs_func): _, dialog_ids = add_dialogs_func dialogs_to_delete = dialog_ids[:3] payload = {"dialog_ids": dialogs_to_delete} res = delete_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"] is True, res res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 2, res remaining_ids = [dialog["id"] for dialog in res["data"]] for dialog_id in dialog_ids[3:]: assert dialog_id in remaining_ids, res @pytest.mark.p2 def test_delete_nonexistent_dialog(self, WebApiAuth): fake_dialog_id = "nonexistent_dialog_id" payload = {"dialog_ids": [fake_dialog_id]} res = delete_dialog(WebApiAuth, payload) assert res["code"] == 103, res assert "Only owner of dialog authorized for this operation." in res["message"], res @pytest.mark.p2 def test_delete_empty_dialog_ids(self, WebApiAuth): payload = {"dialog_ids": []} res = delete_dialog(WebApiAuth, payload) assert res["code"] == 0, res @pytest.mark.p2 def test_delete_missing_dialog_ids(self, WebApiAuth): payload = {} res = delete_dialog(WebApiAuth, payload) assert res["code"] == 101, res assert res["message"] == "required argument are missing: dialog_ids; ", res @pytest.mark.p2 def test_delete_invalid_dialog_ids_format(self, WebApiAuth): payload = {"dialog_ids": "not_a_list"} res = delete_dialog(WebApiAuth, payload) assert res["code"] == 103, res assert res["message"] == "Only owner of dialog authorized for this operation.", res @pytest.mark.p2 def test_delete_mixed_valid_invalid_dialogs(self, WebApiAuth, add_dialog_func): _, valid_dialog_id = add_dialog_func invalid_dialog_id = "nonexistent_dialog_id" payload = {"dialog_ids": [valid_dialog_id, invalid_dialog_id]} res = delete_dialog(WebApiAuth, payload) assert res["code"] == 103, res assert res["message"] == "Only owner of dialog authorized for this operation.", res res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 1, res @pytest.mark.p3 def test_delete_dialog_concurrent(self, WebApiAuth, add_dialogs_func): _, dialog_ids = add_dialogs_func count = len(dialog_ids) with ThreadPoolExecutor(max_workers=3) as executor: futures = [executor.submit(delete_dialog, WebApiAuth, {"dialog_ids": [dialog_id]}) for dialog_id in dialog_ids] responses = [future.result() for future in as_completed(futures)] successful_deletions = sum(1 for response in responses if response["code"] == 0) assert successful_deletions > 0, "No dialogs were successfully deleted" res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == count - successful_deletions, res @pytest.mark.p3 def test_delete_dialog_idempotent(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func payload = {"dialog_ids": [dialog_id]} res = delete_dialog(WebApiAuth, payload) assert res["code"] == 0, res res = delete_dialog(WebApiAuth, payload) assert res["code"] == 0, res @pytest.mark.p3 def test_delete_large_batch_dialogs(self, WebApiAuth, add_document): dataset_id, _ = add_document dialog_ids = batch_create_dialogs(WebApiAuth, 50, [dataset_id]) assert len(dialog_ids) == 50, "Failed to create 50 dialogs" payload = {"dialog_ids": dialog_ids} res = delete_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"] is True, res res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 0, res @pytest.mark.p3 def test_delete_dialog_with_special_characters(self, WebApiAuth): payload = {"name": "Dialog with 特殊字符 and émojis 🤖", "description": "Test dialog with special characters", "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} create_res = create_dialog(WebApiAuth, payload) assert create_res["code"] == 0, create_res dialog_id = create_res["data"]["id"] delete_payload = {"dialog_ids": [dialog_id]} res = delete_dialog(WebApiAuth, delete_payload) assert res["code"] == 0, res assert res["data"] is True, res res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 0, res @pytest.mark.p3 def test_delete_dialog_preserves_other_user_dialogs(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func payload = {"dialog_ids": [dialog_id]} res = delete_dialog(WebApiAuth, payload) assert res["code"] == 0, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_dialog_app/test_create_dialog.py
test/testcases/test_web_api/test_dialog_app/test_create_dialog.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import create_dialog from configs import CHAT_ASSISTANT_NAME_LIMIT, INVALID_API_TOKEN from hypothesis import example, given, settings from libs.auth import RAGFlowWebApiAuth from utils.hypothesis_utils import valid_names @pytest.mark.usefixtures("clear_dialogs") class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ids=["empty_auth", "invalid_api_token"], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): payload = {"name": "auth_test", "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = create_dialog(invalid_auth, payload) assert res["code"] == expected_code, res assert res["message"] == expected_message, res @pytest.mark.usefixtures("clear_dialogs") class TestCapability: @pytest.mark.p3 def test_create_dialog_100(self, WebApiAuth): for i in range(100): payload = {"name": f"dialog_{i}", "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, f"Failed to create dialog {i}" @pytest.mark.p3 def test_create_dialog_concurrent(self, WebApiAuth): count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(create_dialog, WebApiAuth, {"name": f"dialog_{i}", "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}}) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.usefixtures("clear_dialogs") class TestDialogCreate: @pytest.mark.p1 @given(name=valid_names()) @example("a" * CHAT_ASSISTANT_NAME_LIMIT) @settings(max_examples=20) def test_name(self, WebApiAuth, name): payload = {"name": name, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res @pytest.mark.p2 @pytest.mark.parametrize( "name, expected_code, expected_message", [ ("", 102, "Dialog name can't be empty."), (" ", 102, "Dialog name can't be empty."), ("a" * (CHAT_ASSISTANT_NAME_LIMIT + 1), 102, "Dialog name length is 256 which is larger than 255"), (0, 102, "Dialog name must be string."), (None, 102, "Dialog name must be string."), ], ids=["empty_name", "space_name", "too_long_name", "invalid_name", "None_name"], ) def test_name_invalid(self, WebApiAuth, name, expected_code, expected_message): payload = {"name": name, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = create_dialog(WebApiAuth, payload) assert res["code"] == expected_code, res assert res["message"] == expected_message, res @pytest.mark.p1 def test_prompt_config_required(self, WebApiAuth): payload = {"name": "test_dialog"} res = create_dialog(WebApiAuth, payload) assert res["code"] == 101, res assert res["message"] == "required argument are missing: prompt_config; ", res @pytest.mark.p1 def test_prompt_config_with_knowledge_no_kb(self, WebApiAuth): payload = {"name": "test_dialog", "prompt_config": {"system": "You are a helpful assistant. Use this knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}]}} res = create_dialog(WebApiAuth, payload) assert res["code"] == 102, res assert "Please remove `{knowledge}` in system prompt" in res["message"], res @pytest.mark.p1 def test_prompt_config_parameter_not_used(self, WebApiAuth): payload = {"name": "test_dialog", "prompt_config": {"system": "You are a helpful assistant.", "parameters": [{"key": "unused_param", "optional": False}]}} res = create_dialog(WebApiAuth, payload) assert res["code"] == 102, res assert "Parameter 'unused_param' is not used" in res["message"], res @pytest.mark.p1 def test_create_with_kb_ids(self, WebApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = { "name": "test_dialog_with_kb", "kb_ids": [dataset_id], "prompt_config": {"system": "You are a helpful assistant. Use this knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}]}, } res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["kb_ids"] == [dataset_id], res @pytest.mark.p2 def test_create_with_all_parameters(self, WebApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = { "name": "comprehensive_dialog", "description": "A comprehensive test dialog", "icon": "🤖", "kb_ids": [dataset_id], "top_n": 10, "top_k": 2048, "rerank_id": "", "similarity_threshold": 0.2, "vector_similarity_weight": 0.5, "llm_setting": {"model": "gpt-4", "temperature": 0.8, "max_tokens": 1000}, "prompt_config": {"system": "You are a helpful assistant. Use this knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}]}, } res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res data = res["data"] assert data["name"] == "comprehensive_dialog", res assert data["description"] == "A comprehensive test dialog", res assert data["icon"] == "🤖", res assert data["kb_ids"] == [dataset_id], res assert data["top_n"] == 10, res assert data["top_k"] == 2048, res assert data["similarity_threshold"] == 0.2, res assert data["vector_similarity_weight"] == 0.5, res @pytest.mark.p3 def test_name_duplicated(self, WebApiAuth): name = "duplicated_dialog" payload = {"name": name, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res @pytest.mark.p2 def test_optional_parameters(self, WebApiAuth): payload = { "name": "test_optional_params", "prompt_config": {"system": "You are a helpful assistant. Optional param: {optional_param}", "parameters": [{"key": "optional_param", "optional": True}]}, } res = create_dialog(WebApiAuth, payload) assert res["code"] == 0, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_dialog_app/conftest.py
test/testcases/test_web_api/test_dialog_app/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import batch_create_dialogs, delete_dialogs @pytest.fixture(scope="function") def add_dialog_func(request, WebApiAuth, add_dataset_func): def cleanup(): delete_dialogs(WebApiAuth) request.addfinalizer(cleanup) dataset_id = add_dataset_func return dataset_id, batch_create_dialogs(WebApiAuth, 1, [dataset_id])[0] @pytest.fixture(scope="class") def add_dialogs(request, WebApiAuth, add_dataset): def cleanup(): delete_dialogs(WebApiAuth) request.addfinalizer(cleanup) dataset_id = add_dataset return dataset_id, batch_create_dialogs(WebApiAuth, 5, [dataset_id]) @pytest.fixture(scope="function") def add_dialogs_func(request, WebApiAuth, add_dataset_func): def cleanup(): delete_dialogs(WebApiAuth) request.addfinalizer(cleanup) dataset_id = add_dataset_func return dataset_id, batch_create_dialogs(WebApiAuth, 5, [dataset_id])
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_dialog_app/test_list_dialogs.py
test/testcases/test_web_api/test_dialog_app/test_list_dialogs.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import batch_create_dialogs, create_dialog, list_dialogs from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth @pytest.mark.usefixtures("clear_dialogs") class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ids=["empty_auth", "invalid_api_token"], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = list_dialogs(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestDialogList: @pytest.mark.p1 @pytest.mark.usefixtures("add_dialogs_func") def test_list_empty_dialogs(self, WebApiAuth): res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 5, res @pytest.mark.p1 def test_list_multiple_dialogs(self, WebApiAuth, add_dialogs_func): _, dialog_ids = add_dialogs_func res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 5, res returned_ids = [dialog["id"] for dialog in res["data"]] for dialog_id in dialog_ids: assert dialog_id in returned_ids, res @pytest.mark.p2 @pytest.mark.usefixtures("add_dialogs_func") def test_list_dialogs_data_structure(self, WebApiAuth): res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 5, res dialog = res["data"][0] required_fields = [ "id", "name", "description", "kb_ids", "kb_names", "prompt_config", "llm_setting", "top_n", "top_k", "similarity_threshold", "vector_similarity_weight", "create_time", "update_time", ] for field in required_fields: assert field in dialog, f"Missing field: {field}" assert isinstance(dialog["id"], str), res assert isinstance(dialog["name"], str), res assert isinstance(dialog["kb_ids"], list), res assert isinstance(dialog["kb_names"], list), res assert isinstance(dialog["prompt_config"], dict), res assert isinstance(dialog["top_n"], int), res assert isinstance(dialog["top_k"], int), res assert isinstance(dialog["similarity_threshold"], (int, float)), res assert isinstance(dialog["vector_similarity_weight"], (int, float)), res @pytest.mark.p2 @pytest.mark.usefixtures("add_dialogs_func") def test_list_dialogs_with_kb_names(self, WebApiAuth): res = list_dialogs(WebApiAuth) assert res["code"] == 0, res dialog = res["data"][0] assert isinstance(dialog["kb_ids"], list), res assert isinstance(dialog["kb_names"], list), res assert len(dialog["kb_ids"]) == len(dialog["kb_names"]), res @pytest.mark.p2 @pytest.mark.usefixtures("add_dialogs_func") def test_list_dialogs_ordering(self, WebApiAuth): res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 5, res dialogs = res["data"] for i in range(len(dialogs) - 1): current_time = dialogs[i]["create_time"] next_time = dialogs[i + 1]["create_time"] assert current_time >= next_time, f"Dialogs not properly ordered: {current_time} should be >= {next_time}" @pytest.mark.p3 @pytest.mark.usefixtures("clear_dialogs") def test_list_dialogs_with_invalid_kb(self, WebApiAuth): payload = { "name": "invalid_kb_dialog", "kb_ids": ["invalid_kb_id"], "prompt_config": {"system": "You are a helpful assistant with knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}]}, } create_res = create_dialog(WebApiAuth, payload) assert create_res["code"] == 0, create_res res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 1, res dialog = res["data"][0] assert len(dialog["kb_ids"]) == 0, res assert len(dialog["kb_names"]) == 0, res @pytest.mark.p3 @pytest.mark.usefixtures("clear_dialogs") def test_list_dialogs_with_multiple_kbs(self, WebApiAuth, add_dataset_func): dataset_id1 = add_dataset_func dataset_id2 = add_dataset_func payload = { "name": "multi_kb_dialog", "kb_ids": [dataset_id1, dataset_id2], "prompt_config": {"system": "You are a helpful assistant with knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}]}, } create_res = create_dialog(WebApiAuth, payload) assert create_res["code"] == 0, create_res res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 1, res dialog = res["data"][0] assert len(dialog["kb_ids"]) == 2, res assert len(dialog["kb_names"]) == 2, res assert dataset_id1 in dialog["kb_ids"], res assert dataset_id2 in dialog["kb_ids"], res @pytest.mark.p3 @pytest.mark.usefixtures("add_dialogs_func") def test_list_dialogs_prompt_config_structure(self, WebApiAuth): res = list_dialogs(WebApiAuth) assert res["code"] == 0, res dialog = res["data"][0] prompt_config = dialog["prompt_config"] assert "system" in prompt_config, res assert "parameters" in prompt_config, res assert isinstance(prompt_config["system"], str), res assert isinstance(prompt_config["parameters"], list), res @pytest.mark.p3 @pytest.mark.usefixtures("clear_dialogs") def test_list_dialogs_performance(self, WebApiAuth, add_document): dataset_id, _ = add_document dialog_ids = batch_create_dialogs(WebApiAuth, 100, [dataset_id]) assert len(dialog_ids) == 100, "Failed to create 100 dialogs" res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 100, res returned_ids = [dialog["id"] for dialog in res["data"]] for dialog_id in dialog_ids: assert dialog_id in returned_ids, f"Dialog {dialog_id} not found in list" @pytest.mark.p3 @pytest.mark.usefixtures("clear_dialogs") def test_list_dialogs_with_mixed_kb_states(self, WebApiAuth, add_dataset_func): valid_dataset_id = add_dataset_func payload = { "name": "mixed_kb_dialog", "kb_ids": [valid_dataset_id, "invalid_kb_id"], "prompt_config": {"system": "You are a helpful assistant with knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}]}, } create_res = create_dialog(WebApiAuth, payload) assert create_res["code"] == 0, create_res res = list_dialogs(WebApiAuth) assert res["code"] == 0, res assert len(res["data"]) == 1, res dialog = res["data"][0] assert len(dialog["kb_ids"]) == 1, res assert dialog["kb_ids"][0] == valid_dataset_id, res assert len(dialog["kb_names"]) == 1, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_dialog_app/test_update_dialog.py
test/testcases/test_web_api/test_dialog_app/test_update_dialog.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import update_dialog from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth @pytest.mark.usefixtures("clear_dialogs") class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ids=["empty_auth", "invalid_api_token"], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message, add_dialog_func): _, dialog_id = add_dialog_func payload = {"dialog_id": dialog_id, "name": "updated_name", "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = update_dialog(invalid_auth, payload) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestDialogUpdate: @pytest.mark.p1 def test_update_name(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func new_name = "updated_dialog_name" payload = {"dialog_id": dialog_id, "name": new_name, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = update_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["name"] == new_name, res @pytest.mark.p1 def test_update_description(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func new_description = "Updated description" payload = {"dialog_id": dialog_id, "description": new_description, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = update_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["description"] == new_description, res @pytest.mark.p1 def test_update_prompt_config(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func new_prompt_config = {"system": "You are an updated helpful assistant with {param1}.", "parameters": [{"key": "param1", "optional": False}]} payload = {"dialog_id": dialog_id, "prompt_config": new_prompt_config} res = update_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["prompt_config"]["system"] == new_prompt_config["system"], res @pytest.mark.p1 def test_update_kb_ids(self, WebApiAuth, add_dialog_func, add_dataset_func): _, dialog_id = add_dialog_func new_dataset_id = add_dataset_func payload = { "dialog_id": dialog_id, "kb_ids": [new_dataset_id], "prompt_config": {"system": "You are a helpful assistant with knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}]}, } res = update_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert new_dataset_id in res["data"]["kb_ids"], res @pytest.mark.p1 def test_update_llm_settings(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func new_llm_setting = {"model": "gpt-4", "temperature": 0.9, "max_tokens": 2000} payload = {"dialog_id": dialog_id, "llm_setting": new_llm_setting, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = update_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["llm_setting"]["model"] == "gpt-4", res assert res["data"]["llm_setting"]["temperature"] == 0.9, res @pytest.mark.p1 def test_update_retrieval_settings(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func payload = { "dialog_id": dialog_id, "top_n": 15, "top_k": 4096, "similarity_threshold": 0.3, "vector_similarity_weight": 0.7, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}, } res = update_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["top_n"] == 15, res assert res["data"]["top_k"] == 4096, res assert res["data"]["similarity_threshold"] == 0.3, res assert res["data"]["vector_similarity_weight"] == 0.7, res @pytest.mark.p2 def test_update_nonexistent_dialog(self, WebApiAuth): fake_dialog_id = "nonexistent_dialog_id" payload = {"dialog_id": fake_dialog_id, "name": "updated_name", "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = update_dialog(WebApiAuth, payload) assert res["code"] == 102, res assert "Dialog not found" in res["message"], res @pytest.mark.p2 def test_update_with_invalid_prompt_config(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func payload = {"dialog_id": dialog_id, "prompt_config": {"system": "You are a helpful assistant.", "parameters": [{"key": "unused_param", "optional": False}]}} res = update_dialog(WebApiAuth, payload) assert res["code"] == 102, res assert "Parameter 'unused_param' is not used" in res["message"], res @pytest.mark.p2 def test_update_with_knowledge_but_no_kb(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func payload = {"dialog_id": dialog_id, "kb_ids": [], "prompt_config": {"system": "You are a helpful assistant with knowledge: {knowledge}", "parameters": [{"key": "knowledge", "optional": True}]}} res = update_dialog(WebApiAuth, payload) assert res["code"] == 102, res assert "Please remove `{knowledge}` in system prompt" in res["message"], res @pytest.mark.p2 def test_update_icon(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func new_icon = "🚀" payload = {"dialog_id": dialog_id, "icon": new_icon, "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = update_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["icon"] == new_icon, res @pytest.mark.p2 def test_update_rerank_id(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func payload = {"dialog_id": dialog_id, "rerank_id": "test_rerank_model", "prompt_config": {"system": "You are a helpful assistant.", "parameters": []}} res = update_dialog(WebApiAuth, payload) assert res["code"] == 0, res assert res["data"]["rerank_id"] == "test_rerank_model", res @pytest.mark.p3 def test_update_multiple_fields(self, WebApiAuth, add_dialog_func): _, dialog_id = add_dialog_func payload = { "dialog_id": dialog_id, "name": "multi_update_dialog", "description": "Updated with multiple fields", "icon": "🔄", "top_n": 20, "similarity_threshold": 0.4, "prompt_config": {"system": "You are a multi-updated assistant.", "parameters": []}, } res = update_dialog(WebApiAuth, payload) assert res["code"] == 0, res data = res["data"] assert data["name"] == "multi_update_dialog", res assert data["description"] == "Updated with multiple fields", res assert data["icon"] == "🔄", res assert data["top_n"] == 20, res assert data["similarity_threshold"] == 0.4, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_memory_app/test_list_memory.py
test/testcases/test_web_api/test_memory_app/test_list_memory.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from test_web_api.common import list_memory, get_memory_config from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = list_memory(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestCapability: @pytest.mark.p3 def test_capability(self, WebApiAuth): count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(list_memory, WebApiAuth) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.usefixtures("add_memory_func") class TestMemoryList: @pytest.mark.p1 def test_params_unset(self, WebApiAuth): res = list_memory(WebApiAuth, None) assert res["code"] == 0, res @pytest.mark.p1 def test_params_empty(self, WebApiAuth): res = list_memory(WebApiAuth, {}) assert res["code"] == 0, res @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_page_size", [ ({"page": 1, "page_size": 10}, 3), ({"page": 2, "page_size": 10}, 0), ({"page": 1, "page_size": 2}, 2), ({"page": 2, "page_size": 2}, 1), ({"page": 5, "page_size": 10}, 0), ], ids=["normal_first_page", "beyond_max_page", "normal_last_partial_page" , "normal_middle_page", "full_data_single_page"], ) def test_page(self, WebApiAuth, params, expected_page_size): # have added 3 memories in fixture res = list_memory(WebApiAuth, params) assert res["code"] == 0, res assert len(res["data"]["memory_list"]) == expected_page_size, res @pytest.mark.p2 def test_filter_memory_type(self, WebApiAuth): res = list_memory(WebApiAuth, {"memory_type": ["semantic"]}) assert res["code"] == 0, res for memory in res["data"]["memory_list"]: assert "semantic" in memory["memory_type"], res @pytest.mark.p2 def test_filter_multi_memory_type(self, WebApiAuth): res = list_memory(WebApiAuth, {"memory_type": ["episodic", "procedural"]}) assert res["code"] == 0, res for memory in res["data"]["memory_list"]: assert "episodic" in memory["memory_type"] or "procedural" in memory["memory_type"], res @pytest.mark.p2 def test_filter_storage_type(self, WebApiAuth): res = list_memory(WebApiAuth, {"storage_type": "table"}) assert res["code"] == 0, res for memory in res["data"]["memory_list"]: assert memory["storage_type"] == "table", res @pytest.mark.p2 def test_match_keyword(self, WebApiAuth): res = list_memory(WebApiAuth, {"keywords": "s"}) assert res["code"] == 0, res for memory in res["data"]["memory_list"]: assert "s" in memory["name"], res @pytest.mark.p1 def test_get_config(self, WebApiAuth): memory_list = list_memory(WebApiAuth, {}) assert memory_list["code"] == 0, memory_list memory_config = get_memory_config(WebApiAuth, memory_list["data"]["memory_list"][0]["id"]) assert memory_config["code"] == 0, memory_config assert memory_config["data"]["id"] == memory_list["data"]["memory_list"][0]["id"], memory_config for field in ["name", "avatar", "tenant_id", "owner_name", "memory_type", "storage_type", "embd_id", "llm_id", "permissions", "description", "memory_size", "forgetting_policy", "temperature", "system_prompt", "user_prompt"]: assert field in memory_config["data"], memory_config
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_memory_app/test_create_memory.py
test/testcases/test_web_api/test_memory_app/test_create_memory.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import random import re import pytest from test_web_api.common import create_memory from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth from hypothesis import example, given, settings from test.testcases.utils.hypothesis_utils import valid_names class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ids=["empty_auth", "invalid_api_token"] ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = create_memory(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestMemoryCreate: @pytest.mark.p1 @given(name=valid_names()) @example("d" * 128) @settings(max_examples=20) def test_name(self, WebApiAuth, name): payload = { "name": name, "memory_type": ["raw"] + random.choices(["semantic", "episodic", "procedural"], k=random.randint(0, 3)), "embd_id": "BAAI/bge-large-zh-v1.5@SILICONFLOW", "llm_id": "glm-4-flash@ZHIPU-AI" } res = create_memory(WebApiAuth, payload) assert res["code"] == 0, res pattern = rf'^{name}|{name}(?:\((\d+)\))?$' escaped_name = re.escape(res["data"]["name"]) assert re.match(pattern, escaped_name), res @pytest.mark.p2 @pytest.mark.parametrize( "name, expected_message", [ ("", "Memory name cannot be empty or whitespace."), (" ", "Memory name cannot be empty or whitespace."), ("a" * 129, f"Memory name '{'a'*129}' exceeds limit of 128."), ], ids=["empty_name", "space_name", "too_long_name"], ) def test_name_invalid(self, WebApiAuth, name, expected_message): payload = { "name": name, "memory_type": ["raw"] + random.choices(["semantic", "episodic", "procedural"], k=random.randint(0, 3)), "embd_id": "BAAI/bge-large-zh-v1.5@SILICONFLOW", "llm_id": "glm-4-flash@ZHIPU-AI" } res = create_memory(WebApiAuth, payload) assert res["message"] == expected_message, res @pytest.mark.p2 @given(name=valid_names()) def test_type_invalid(self, WebApiAuth, name): payload = { "name": name, "memory_type": ["something"], "embd_id": "BAAI/bge-large-zh-v1.5@SILICONFLOW", "llm_id": "glm-4-flash@ZHIPU-AI" } res = create_memory(WebApiAuth, payload) assert res["message"] == f"Memory type '{ {'something'} }' is not supported.", res @pytest.mark.p3 def test_name_duplicated(self, WebApiAuth): name = "duplicated_name_test" payload = { "name": name, "memory_type": ["raw"] + random.choices(["semantic", "episodic", "procedural"], k=random.randint(0, 3)), "embd_id": "BAAI/bge-large-zh-v1.5@SILICONFLOW", "llm_id": "glm-4-flash@ZHIPU-AI" } res1 = create_memory(WebApiAuth, payload) assert res1["code"] == 0, res1 res2 = create_memory(WebApiAuth, payload) assert res2["code"] == 0, res2
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_memory_app/conftest.py
test/testcases/test_web_api/test_memory_app/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest import random from test_web_api.common import create_memory, list_memory, delete_memory @pytest.fixture(scope="function") def add_memory_func(request, WebApiAuth): def cleanup(): memory_list_res = list_memory(WebApiAuth) exist_memory_ids = [memory["id"] for memory in memory_list_res["data"]["memory_list"]] for memory_id in exist_memory_ids: delete_memory(WebApiAuth, memory_id) request.addfinalizer(cleanup) memory_ids = [] for i in range(3): payload = { "name": f"test_memory_{i}", "memory_type": ["raw"] + random.choices(["semantic", "episodic", "procedural"], k=random.randint(0, 3)), "embd_id": "BAAI/bge-large-zh-v1.5@SILICONFLOW", "llm_id": "glm-4-flash@ZHIPU-AI" } res = create_memory(WebApiAuth, payload) memory_ids.append(res["data"]["id"]) return memory_ids
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_memory_app/test_rm_memory.py
test/testcases/test_web_api/test_memory_app/test_rm_memory.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from test_web_api.common import (list_memory, delete_memory) from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = delete_memory(invalid_auth, "some_memory_id") assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestMemoryDelete: @pytest.mark.p1 def test_memory_id(self, WebApiAuth, add_memory_func): memory_ids = add_memory_func res = delete_memory(WebApiAuth, memory_ids[0]) assert res["code"] == 0, res res = list_memory(WebApiAuth) assert res["data"]["total_count"] == 2, res @pytest.mark.p2 @pytest.mark.usefixtures("add_memory_func") def test_id_wrong_uuid(self, WebApiAuth): res = delete_memory(WebApiAuth, "d94a8dc02c9711f0930f7fbc369eab6d") assert res["code"] == 404, res res = list_memory(WebApiAuth) assert len(res["data"]["memory_list"]) == 3, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_web_api/test_memory_app/test_update_memory.py
test/testcases/test_web_api/test_memory_app/test_update_memory.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from test_web_api.common import update_memory from configs import INVALID_API_TOKEN from libs.auth import RAGFlowWebApiAuth from hypothesis import HealthCheck, example, given, settings from utils import encode_avatar from utils.file_utils import create_image_file from utils.hypothesis_utils import valid_names class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 401, "<Unauthorized '401: Unauthorized'>"), (RAGFlowWebApiAuth(INVALID_API_TOKEN), 401, "<Unauthorized '401: Unauthorized'>"), ], ids=["empty_auth", "invalid_api_token"] ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = update_memory(invalid_auth, "memory_id") assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestMemoryUpdate: @pytest.mark.p1 @given(name=valid_names()) @example("f" * 128) @settings(max_examples=20, suppress_health_check=[HealthCheck.function_scoped_fixture]) def test_name(self, WebApiAuth, add_memory_func, name): memory_ids = add_memory_func payload = {"name": name} res = update_memory(WebApiAuth, memory_ids[0], payload) assert res["code"] == 0, res assert res["data"]["name"] == name, res @pytest.mark.p2 @pytest.mark.parametrize( "name, expected_message", [ ("", "Memory name cannot be empty or whitespace."), (" ", "Memory name cannot be empty or whitespace."), ("a" * 129, f"Memory name '{'a' * 129}' exceeds limit of 128."), ] ) def test_name_invalid(self, WebApiAuth, add_memory_func, name, expected_message): memory_ids = add_memory_func payload = {"name": name} res = update_memory(WebApiAuth, memory_ids[0], payload) assert res["code"] == 101, res assert res["message"] == expected_message, res @pytest.mark.p2 def test_duplicate_name(self, WebApiAuth, add_memory_func): memory_ids = add_memory_func payload = {"name": "Test_Memory"} res = update_memory(WebApiAuth, memory_ids[0], payload) assert res["code"] == 0, res payload = {"name": "Test_Memory"} res = update_memory(WebApiAuth, memory_ids[1], payload) assert res["code"] == 0, res assert res["data"]["name"] == "Test_Memory(1)", res @pytest.mark.p1 def test_avatar(self, WebApiAuth, add_memory_func, tmp_path): memory_ids = add_memory_func fn = create_image_file(tmp_path / "ragflow_test.png") payload = {"avatar": f"data:image/png;base64,{encode_avatar(fn)}"} res = update_memory(WebApiAuth, memory_ids[0], payload) assert res["code"] == 0, res assert res["data"]["avatar"] == f"data:image/png;base64,{encode_avatar(fn)}", res @pytest.mark.p1 def test_description(self, WebApiAuth, add_memory_func): memory_ids = add_memory_func description = "This is a test description." payload = {"description": description} res = update_memory(WebApiAuth, memory_ids[0], payload) assert res["code"] == 0, res assert res["data"]["description"] == description, res @pytest.mark.p1 def test_llm(self, WebApiAuth, add_memory_func): memory_ids = add_memory_func llm_id = "glm-4@ZHIPU-AI" payload = {"llm_id": llm_id} res = update_memory(WebApiAuth, memory_ids[0], payload) assert res["code"] == 0, res assert res["data"]["llm_id"] == llm_id, res @pytest.mark.p1 @pytest.mark.parametrize( "permission", [ "me", "team" ], ids=["me", "team"] ) def test_permission(self, WebApiAuth, add_memory_func, permission): memory_ids = add_memory_func payload = {"permissions": permission} res = update_memory(WebApiAuth, memory_ids[0], payload) assert res["code"] == 0, res assert res["data"]["permissions"] == permission.lower().strip(), res @pytest.mark.p1 def test_memory_size(self, WebApiAuth, add_memory_func): memory_ids = add_memory_func memory_size = 1048576 # 1 MB payload = {"memory_size": memory_size} res = update_memory(WebApiAuth, memory_ids[0], payload) assert res["code"] == 0, res assert res["data"]["memory_size"] == memory_size, res @pytest.mark.p1 def test_temperature(self, WebApiAuth, add_memory_func): memory_ids = add_memory_func temperature = 0.7 payload = {"temperature": temperature} res = update_memory(WebApiAuth, memory_ids[0], payload) assert res["code"] == 0, res assert res["data"]["temperature"] == temperature, res @pytest.mark.p1 def test_system_prompt(self, WebApiAuth, add_memory_func): memory_ids = add_memory_func system_prompt = "This is a system prompt." payload = {"system_prompt": system_prompt} res = update_memory(WebApiAuth, memory_ids[0], payload) assert res["code"] == 0, res assert res["data"]["system_prompt"] == system_prompt, res @pytest.mark.p1 def test_user_prompt(self, WebApiAuth, add_memory_func): memory_ids = add_memory_func user_prompt = "This is a user prompt." payload = {"user_prompt": user_prompt} res = update_memory(WebApiAuth, memory_ids[0], payload) assert res["code"] == 0, res assert res["data"]["user_prompt"] == user_prompt, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/utils/file_utils.py
test/testcases/utils/file_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json from docx import Document # pip install python-docx from openpyxl import Workbook # pip install openpyxl from PIL import Image, ImageDraw # pip install Pillow from pptx import Presentation # pip install python-pptx from reportlab.pdfgen import canvas # pip install reportlab def create_docx_file(path): doc = Document() doc.add_paragraph("This is a test DOCX file.") doc.save(path) return path def create_excel_file(path): wb = Workbook() ws = wb.active ws["A1"] = "Test Excel File" wb.save(path) return path def create_ppt_file(path): prs = Presentation() slide = prs.slides.add_slide(prs.slide_layouts[0]) slide.shapes.title.text = "Test PPT File" prs.save(path) return path def create_image_file(path): img = Image.new("RGB", (100, 100), color="blue") draw = ImageDraw.Draw(img) draw.text((10, 40), "Test", fill="white") img.save(path) return path def create_pdf_file(path): if not isinstance(path, str): path = str(path) c = canvas.Canvas(path) c.drawString(100, 750, "Test PDF File") c.save() return path def create_txt_file(path): with open(path, "w", encoding="utf-8") as f: f.write("This is the content of a test TXT file.") return path def create_md_file(path): md_content = "# Test MD File\n\nThis is a test Markdown file." with open(path, "w", encoding="utf-8") as f: f.write(md_content) return path def create_json_file(path): data = {"message": "This is a test JSON file", "value": 123} with open(path, "w", encoding="utf-8") as f: json.dump(data, f, indent=2) return path def create_eml_file(path): eml_content = ( "From: sender@example.com\n" "To: receiver@example.com\n" "Subject: Test EML File\n\n" "This is a test email content.\n" ) with open(path, "w", encoding="utf-8") as f: f.write(eml_content) return path def create_html_file(path): html_content = ( "<html>\n" "<head><title>Test HTML File</title></head>\n" "<body><h1>This is a test HTML file</h1></body>\n" "</html>" ) with open(path, "w", encoding="utf-8") as f: f.write(html_content) return path
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/utils/__init__.py
test/testcases/utils/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import base64 import functools import hashlib import time from pathlib import Path def encode_avatar(image_path): with Path.open(image_path, "rb") as file: binary_data = file.read() base64_encoded = base64.b64encode(binary_data).decode("utf-8") return base64_encoded def compare_by_hash(file1, file2, algorithm="sha256"): def _calc_hash(file_path): hash_func = hashlib.new(algorithm) with open(file_path, "rb") as f: while chunk := f.read(8192): hash_func.update(chunk) return hash_func.hexdigest() return _calc_hash(file1) == _calc_hash(file2) def wait_for(timeout=10, interval=1, error_msg="Timeout"): def decorator(func): @functools.wraps(func) def wrapper(*args, **kwargs): start_time = time.time() while True: result = func(*args, **kwargs) if result is True: return result elapsed = time.time() - start_time if elapsed > timeout: assert False, error_msg time.sleep(interval) return wrapper return decorator def is_sorted(data, field, descending=True): timestamps = [ds[field] for ds in data] return all(a >= b for a, b in zip(timestamps, timestamps[1:])) if descending else all(a <= b for a, b in zip(timestamps, timestamps[1:]))
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/utils/hypothesis_utils.py
test/testcases/utils/hypothesis_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import hypothesis.strategies as st @st.composite def valid_names(draw): base_chars = "abcdefghijklmnopqrstuvwxyz_" first_char = draw(st.sampled_from([c for c in base_chars if c.isalpha() or c == "_"])) remaining = draw(st.text(alphabet=st.sampled_from(base_chars), min_size=0, max_size=128 - 2)) name = (first_char + remaining)[:128] return name.encode("utf-8").decode("utf-8")
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/common.py
test/testcases/test_http_api/common.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from pathlib import Path import requests from configs import HOST_ADDRESS, VERSION from requests_toolbelt import MultipartEncoder from utils.file_utils import create_txt_file HEADERS = {"Content-Type": "application/json"} DATASETS_API_URL = f"/api/{VERSION}/datasets" FILE_API_URL = f"/api/{VERSION}/datasets/{{dataset_id}}/documents" FILE_CHUNK_API_URL = f"/api/{VERSION}/datasets/{{dataset_id}}/chunks" CHUNK_API_URL = f"/api/{VERSION}/datasets/{{dataset_id}}/documents/{{document_id}}/chunks" CHAT_ASSISTANT_API_URL = f"/api/{VERSION}/chats" SESSION_WITH_CHAT_ASSISTANT_API_URL = f"/api/{VERSION}/chats/{{chat_id}}/sessions" SESSION_WITH_AGENT_API_URL = f"/api/{VERSION}/agents/{{agent_id}}/sessions" # DATASET MANAGEMENT def create_dataset(auth, payload=None, *, headers=HEADERS, data=None): res = requests.post(url=f"{HOST_ADDRESS}{DATASETS_API_URL}", headers=headers, auth=auth, json=payload, data=data) return res.json() def list_datasets(auth, params=None, *, headers=HEADERS): res = requests.get(url=f"{HOST_ADDRESS}{DATASETS_API_URL}", headers=headers, auth=auth, params=params) return res.json() def update_dataset(auth, dataset_id, payload=None, *, headers=HEADERS, data=None): res = requests.put(url=f"{HOST_ADDRESS}{DATASETS_API_URL}/{dataset_id}", headers=headers, auth=auth, json=payload, data=data) return res.json() def delete_datasets(auth, payload=None, *, headers=HEADERS, data=None): res = requests.delete(url=f"{HOST_ADDRESS}{DATASETS_API_URL}", headers=headers, auth=auth, json=payload, data=data) return res.json() def batch_create_datasets(auth, num): ids = [] for i in range(num): res = create_dataset(auth, {"name": f"dataset_{i}"}) ids.append(res["data"]["id"]) return ids # FILE MANAGEMENT WITHIN DATASET def upload_documents(auth, dataset_id, files_path=None): url = f"{HOST_ADDRESS}{FILE_API_URL}".format(dataset_id=dataset_id) if files_path is None: files_path = [] fields = [] file_objects = [] try: for fp in files_path: p = Path(fp) f = p.open("rb") fields.append(("file", (p.name, f))) file_objects.append(f) m = MultipartEncoder(fields=fields) res = requests.post( url=url, headers={"Content-Type": m.content_type}, auth=auth, data=m, ) return res.json() finally: for f in file_objects: f.close() def download_document(auth, dataset_id, document_id, save_path): url = f"{HOST_ADDRESS}{FILE_API_URL}/{document_id}".format(dataset_id=dataset_id) res = requests.get(url=url, auth=auth, stream=True) try: if res.status_code == 200: with open(save_path, "wb") as f: for chunk in res.iter_content(chunk_size=8192): f.write(chunk) finally: res.close() return res def list_documents(auth, dataset_id, params=None): url = f"{HOST_ADDRESS}{FILE_API_URL}".format(dataset_id=dataset_id) res = requests.get(url=url, headers=HEADERS, auth=auth, params=params) return res.json() def update_document(auth, dataset_id, document_id, payload=None): url = f"{HOST_ADDRESS}{FILE_API_URL}/{document_id}".format(dataset_id=dataset_id) res = requests.put(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def delete_documents(auth, dataset_id, payload=None): url = f"{HOST_ADDRESS}{FILE_API_URL}".format(dataset_id=dataset_id) res = requests.delete(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def parse_documents(auth, dataset_id, payload=None): url = f"{HOST_ADDRESS}{FILE_CHUNK_API_URL}".format(dataset_id=dataset_id) res = requests.post(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def stop_parse_documents(auth, dataset_id, payload=None): url = f"{HOST_ADDRESS}{FILE_CHUNK_API_URL}".format(dataset_id=dataset_id) res = requests.delete(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def bulk_upload_documents(auth, dataset_id, num, tmp_path): fps = [] for i in range(num): fp = create_txt_file(tmp_path / f"ragflow_test_upload_{i}.txt") fps.append(fp) res = upload_documents(auth, dataset_id, fps) document_ids = [] for document in res["data"]: document_ids.append(document["id"]) return document_ids # CHUNK MANAGEMENT WITHIN DATASET def add_chunk(auth, dataset_id, document_id, payload=None): url = f"{HOST_ADDRESS}{CHUNK_API_URL}".format(dataset_id=dataset_id, document_id=document_id) res = requests.post(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def list_chunks(auth, dataset_id, document_id, params=None): url = f"{HOST_ADDRESS}{CHUNK_API_URL}".format(dataset_id=dataset_id, document_id=document_id) res = requests.get(url=url, headers=HEADERS, auth=auth, params=params) return res.json() def update_chunk(auth, dataset_id, document_id, chunk_id, payload=None): url = f"{HOST_ADDRESS}{CHUNK_API_URL}/{chunk_id}".format(dataset_id=dataset_id, document_id=document_id) res = requests.put(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def delete_chunks(auth, dataset_id, document_id, payload=None): url = f"{HOST_ADDRESS}{CHUNK_API_URL}".format(dataset_id=dataset_id, document_id=document_id) res = requests.delete(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def retrieval_chunks(auth, payload=None): url = f"{HOST_ADDRESS}/api/v1/retrieval" res = requests.post(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def batch_add_chunks(auth, dataset_id, document_id, num): chunk_ids = [] for i in range(num): res = add_chunk(auth, dataset_id, document_id, {"content": f"chunk test {i}"}) chunk_ids.append(res["data"]["chunk"]["id"]) return chunk_ids # CHAT ASSISTANT MANAGEMENT def create_chat_assistant(auth, payload=None): url = f"{HOST_ADDRESS}{CHAT_ASSISTANT_API_URL}" res = requests.post(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def list_chat_assistants(auth, params=None): url = f"{HOST_ADDRESS}{CHAT_ASSISTANT_API_URL}" res = requests.get(url=url, headers=HEADERS, auth=auth, params=params) return res.json() def update_chat_assistant(auth, chat_assistant_id, payload=None): url = f"{HOST_ADDRESS}{CHAT_ASSISTANT_API_URL}/{chat_assistant_id}" res = requests.put(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def delete_chat_assistants(auth, payload=None): url = f"{HOST_ADDRESS}{CHAT_ASSISTANT_API_URL}" res = requests.delete(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def batch_create_chat_assistants(auth, num): chat_assistant_ids = [] for i in range(num): res = create_chat_assistant(auth, {"name": f"test_chat_assistant_{i}", "dataset_ids": []}) chat_assistant_ids.append(res["data"]["id"]) return chat_assistant_ids # SESSION MANAGEMENT def create_session_with_chat_assistant(auth, chat_assistant_id, payload=None): url = f"{HOST_ADDRESS}{SESSION_WITH_CHAT_ASSISTANT_API_URL}".format(chat_id=chat_assistant_id) res = requests.post(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def list_session_with_chat_assistants(auth, chat_assistant_id, params=None): url = f"{HOST_ADDRESS}{SESSION_WITH_CHAT_ASSISTANT_API_URL}".format(chat_id=chat_assistant_id) res = requests.get(url=url, headers=HEADERS, auth=auth, params=params) return res.json() def update_session_with_chat_assistant(auth, chat_assistant_id, session_id, payload=None): url = f"{HOST_ADDRESS}{SESSION_WITH_CHAT_ASSISTANT_API_URL}/{session_id}".format(chat_id=chat_assistant_id) res = requests.put(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def delete_session_with_chat_assistants(auth, chat_assistant_id, payload=None): url = f"{HOST_ADDRESS}{SESSION_WITH_CHAT_ASSISTANT_API_URL}".format(chat_id=chat_assistant_id) res = requests.delete(url=url, headers=HEADERS, auth=auth, json=payload) return res.json() def batch_add_sessions_with_chat_assistant(auth, chat_assistant_id, num): session_ids = [] for i in range(num): res = create_session_with_chat_assistant(auth, chat_assistant_id, {"name": f"session_with_chat_assistant_{i}"}) session_ids.append(res["data"]["id"]) return session_ids
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/conftest.py
test/testcases/test_http_api/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from time import sleep import pytest from common import ( batch_add_chunks, batch_create_chat_assistants, batch_create_datasets, bulk_upload_documents, delete_chat_assistants, delete_datasets, delete_session_with_chat_assistants, list_documents, parse_documents, ) from libs.auth import RAGFlowHttpApiAuth from utils import wait_for from utils.file_utils import ( create_docx_file, create_eml_file, create_excel_file, create_html_file, create_image_file, create_json_file, create_md_file, create_pdf_file, create_ppt_file, create_txt_file, ) @wait_for(30, 1, "Document parsing timeout") def condition(_auth, _dataset_id): res = list_documents(_auth, _dataset_id) for doc in res["data"]["docs"]: if doc["run"] != "DONE": return False return True @pytest.fixture def generate_test_files(request, tmp_path): file_creators = { "docx": (tmp_path / "ragflow_test.docx", create_docx_file), "excel": (tmp_path / "ragflow_test.xlsx", create_excel_file), "ppt": (tmp_path / "ragflow_test.pptx", create_ppt_file), "image": (tmp_path / "ragflow_test.png", create_image_file), "pdf": (tmp_path / "ragflow_test.pdf", create_pdf_file), "txt": (tmp_path / "ragflow_test.txt", create_txt_file), "md": (tmp_path / "ragflow_test.md", create_md_file), "json": (tmp_path / "ragflow_test.json", create_json_file), "eml": (tmp_path / "ragflow_test.eml", create_eml_file), "html": (tmp_path / "ragflow_test.html", create_html_file), } files = {} for file_type, (file_path, creator_func) in file_creators.items(): if request.param in ["", file_type]: creator_func(file_path) files[file_type] = file_path return files @pytest.fixture(scope="class") def ragflow_tmp_dir(request, tmp_path_factory): class_name = request.cls.__name__ return tmp_path_factory.mktemp(class_name) @pytest.fixture(scope="session") def HttpApiAuth(token): return RAGFlowHttpApiAuth(token) @pytest.fixture(scope="function") def clear_datasets(request, HttpApiAuth): def cleanup(): delete_datasets(HttpApiAuth, {"ids": None}) request.addfinalizer(cleanup) @pytest.fixture(scope="function") def clear_chat_assistants(request, HttpApiAuth): def cleanup(): delete_chat_assistants(HttpApiAuth) request.addfinalizer(cleanup) @pytest.fixture(scope="function") def clear_session_with_chat_assistants(request, HttpApiAuth, add_chat_assistants): def cleanup(): for chat_assistant_id in chat_assistant_ids: delete_session_with_chat_assistants(HttpApiAuth, chat_assistant_id) request.addfinalizer(cleanup) _, _, chat_assistant_ids = add_chat_assistants @pytest.fixture(scope="class") def add_dataset(request, HttpApiAuth): def cleanup(): delete_datasets(HttpApiAuth, {"ids": None}) request.addfinalizer(cleanup) dataset_ids = batch_create_datasets(HttpApiAuth, 1) return dataset_ids[0] @pytest.fixture(scope="function") def add_dataset_func(request, HttpApiAuth): def cleanup(): delete_datasets(HttpApiAuth, {"ids": None}) request.addfinalizer(cleanup) return batch_create_datasets(HttpApiAuth, 1)[0] @pytest.fixture(scope="class") def add_document(HttpApiAuth, add_dataset, ragflow_tmp_dir): dataset_id = add_dataset document_ids = bulk_upload_documents(HttpApiAuth, dataset_id, 1, ragflow_tmp_dir) return dataset_id, document_ids[0] @pytest.fixture(scope="class") def add_chunks(HttpApiAuth, add_document): dataset_id, document_id = add_document parse_documents(HttpApiAuth, dataset_id, {"document_ids": [document_id]}) condition(HttpApiAuth, dataset_id) chunk_ids = batch_add_chunks(HttpApiAuth, dataset_id, document_id, 4) sleep(1) # issues/6487 return dataset_id, document_id, chunk_ids @pytest.fixture(scope="class") def add_chat_assistants(request, HttpApiAuth, add_document): def cleanup(): delete_chat_assistants(HttpApiAuth) request.addfinalizer(cleanup) dataset_id, document_id = add_document parse_documents(HttpApiAuth, dataset_id, {"document_ids": [document_id]}) condition(HttpApiAuth, dataset_id) return dataset_id, document_id, batch_create_chat_assistants(HttpApiAuth, 5)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_chat_assistant_management/test_delete_chat_assistants.py
test/testcases/test_http_api/test_chat_assistant_management/test_delete_chat_assistants.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import batch_create_chat_assistants, delete_chat_assistants, list_chat_assistants from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = delete_chat_assistants(invalid_auth) assert res["code"] == expected_code assert res["message"] == expected_message class TestChatAssistantsDelete: @pytest.mark.parametrize( "payload, expected_code, expected_message, remaining", [ pytest.param(None, 0, "", 0, marks=pytest.mark.p3), pytest.param({"ids": []}, 0, "", 0, marks=pytest.mark.p3), pytest.param({"ids": ["invalid_id"]}, 102, "Assistant(invalid_id) not found.", 5, marks=pytest.mark.p3), pytest.param({"ids": ["\n!?。;!?\"'"]}, 102, """Assistant(\n!?。;!?"\') not found.""", 5, marks=pytest.mark.p3), pytest.param("not json", 100, "AttributeError(\"'str' object has no attribute 'get'\")", 5, marks=pytest.mark.p3), pytest.param(lambda r: {"ids": r[:1]}, 0, "", 4, marks=pytest.mark.p3), pytest.param(lambda r: {"ids": r}, 0, "", 0, marks=pytest.mark.p1), ], ) def test_basic_scenarios(self, HttpApiAuth, add_chat_assistants_func, payload, expected_code, expected_message, remaining): _, _, chat_assistant_ids = add_chat_assistants_func if callable(payload): payload = payload(chat_assistant_ids) res = delete_chat_assistants(HttpApiAuth, payload) assert res["code"] == expected_code if res["code"] != 0: assert res["message"] == expected_message res = list_chat_assistants(HttpApiAuth) assert len(res["data"]) == remaining @pytest.mark.parametrize( "payload", [ pytest.param(lambda r: {"ids": ["invalid_id"] + r}, marks=pytest.mark.p3), pytest.param(lambda r: {"ids": r[:1] + ["invalid_id"] + r[1:5]}, marks=pytest.mark.p1), pytest.param(lambda r: {"ids": r + ["invalid_id"]}, marks=pytest.mark.p3), ], ) def test_delete_partial_invalid_id(self, HttpApiAuth, add_chat_assistants_func, payload): _, _, chat_assistant_ids = add_chat_assistants_func if callable(payload): payload = payload(chat_assistant_ids) res = delete_chat_assistants(HttpApiAuth, payload) assert res["code"] == 0 assert res["data"]["errors"][0] == "Assistant(invalid_id) not found." assert res["data"]["success_count"] == 5 res = list_chat_assistants(HttpApiAuth) assert len(res["data"]) == 0 @pytest.mark.p3 def test_repeated_deletion(self, HttpApiAuth, add_chat_assistants_func): _, _, chat_assistant_ids = add_chat_assistants_func res = delete_chat_assistants(HttpApiAuth, {"ids": chat_assistant_ids}) assert res["code"] == 0 res = delete_chat_assistants(HttpApiAuth, {"ids": chat_assistant_ids}) assert res["code"] == 102 assert "not found" in res["message"] @pytest.mark.p3 def test_duplicate_deletion(self, HttpApiAuth, add_chat_assistants_func): _, _, chat_assistant_ids = add_chat_assistants_func res = delete_chat_assistants(HttpApiAuth, {"ids": chat_assistant_ids + chat_assistant_ids}) assert res["code"] == 0 assert "Duplicate assistant ids" in res["data"]["errors"][0] assert res["data"]["success_count"] == 5 res = list_chat_assistants(HttpApiAuth) assert res["code"] == 0 @pytest.mark.p3 def test_concurrent_deletion(self, HttpApiAuth): count = 100 ids = batch_create_chat_assistants(HttpApiAuth, count) with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(delete_chat_assistants, HttpApiAuth, {"ids": ids[i : i + 1]}) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.p3 def test_delete_10k(self, HttpApiAuth): ids = batch_create_chat_assistants(HttpApiAuth, 1_000) res = delete_chat_assistants(HttpApiAuth, {"ids": ids}) assert res["code"] == 0 res = list_chat_assistants(HttpApiAuth) assert len(res["data"]) == 0
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_chat_assistant_management/conftest.py
test/testcases/test_http_api/test_chat_assistant_management/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import batch_create_chat_assistants, delete_chat_assistants, list_documents, parse_documents from utils import wait_for @wait_for(30, 1, "Document parsing timeout") def condition(_auth, _dataset_id): res = list_documents(_auth, _dataset_id) for doc in res["data"]["docs"]: if doc["run"] != "DONE": return False return True @pytest.fixture(scope="function") def add_chat_assistants_func(request, HttpApiAuth, add_document): def cleanup(): delete_chat_assistants(HttpApiAuth) request.addfinalizer(cleanup) dataset_id, document_id = add_document parse_documents(HttpApiAuth, dataset_id, {"document_ids": [document_id]}) condition(HttpApiAuth, dataset_id) return dataset_id, document_id, batch_create_chat_assistants(HttpApiAuth, 5)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_chat_assistant_management/test_create_chat_assistant.py
test/testcases/test_http_api/test_chat_assistant_management/test_create_chat_assistant.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import create_chat_assistant from configs import CHAT_ASSISTANT_NAME_LIMIT, INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth from utils import encode_avatar from utils.file_utils import create_image_file @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = create_chat_assistant(invalid_auth) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.usefixtures("clear_chat_assistants") class TestChatAssistantCreate: @pytest.mark.p1 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"name": "valid_name"}, 0, ""), pytest.param({"name": "a" * (CHAT_ASSISTANT_NAME_LIMIT + 1)}, 102, "", marks=pytest.mark.skip(reason="issues/")), pytest.param({"name": 1}, 100, "", marks=pytest.mark.skip(reason="issues/")), ({"name": ""}, 102, "`name` is required."), ({"name": "duplicated_name"}, 102, "Duplicated chat name in creating chat."), ({"name": "case insensitive"}, 102, "Duplicated chat name in creating chat."), ], ) def test_name(self, HttpApiAuth, add_chunks, payload, expected_code, expected_message): payload["dataset_ids"] = [] # issues/ if payload["name"] == "duplicated_name": create_chat_assistant(HttpApiAuth, payload) elif payload["name"] == "case insensitive": create_chat_assistant(HttpApiAuth, {"name": payload["name"].upper()}) res = create_chat_assistant(HttpApiAuth, payload) assert res["code"] == expected_code, res if expected_code == 0: assert res["data"]["name"] == payload["name"] else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "dataset_ids, expected_code, expected_message", [ ([], 0, ""), (lambda r: [r], 0, ""), (["invalid_dataset_id"], 102, "You don't own the dataset invalid_dataset_id"), ("invalid_dataset_id", 102, "You don't own the dataset i"), ], ) def test_dataset_ids(self, HttpApiAuth, add_chunks, dataset_ids, expected_code, expected_message): dataset_id, _, _ = add_chunks payload = {"name": "ragflow test"} if callable(dataset_ids): payload["dataset_ids"] = dataset_ids(dataset_id) else: payload["dataset_ids"] = dataset_ids res = create_chat_assistant(HttpApiAuth, payload) assert res["code"] == expected_code, res if expected_code == 0: assert res["data"]["name"] == payload["name"] else: assert res["message"] == expected_message @pytest.mark.p3 def test_avatar(self, HttpApiAuth, tmp_path): fn = create_image_file(tmp_path / "ragflow_test.png") payload = {"name": "avatar_test", "avatar": encode_avatar(fn), "dataset_ids": []} res = create_chat_assistant(HttpApiAuth, payload) assert res["code"] == 0 @pytest.mark.p2 @pytest.mark.parametrize( "llm, expected_code, expected_message", [ ({}, 0, ""), ({"model_name": "glm-4"}, 0, ""), ({"model_name": "unknown"}, 102, "`model_name` unknown doesn't exist"), ({"temperature": 0}, 0, ""), ({"temperature": 1}, 0, ""), pytest.param({"temperature": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"temperature": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"temperature": "a"}, 0, "", marks=pytest.mark.skip), ({"top_p": 0}, 0, ""), ({"top_p": 1}, 0, ""), pytest.param({"top_p": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"top_p": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"top_p": "a"}, 0, "", marks=pytest.mark.skip), ({"presence_penalty": 0}, 0, ""), ({"presence_penalty": 1}, 0, ""), pytest.param({"presence_penalty": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"presence_penalty": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"presence_penalty": "a"}, 0, "", marks=pytest.mark.skip), ({"frequency_penalty": 0}, 0, ""), ({"frequency_penalty": 1}, 0, ""), pytest.param({"frequency_penalty": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"frequency_penalty": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"frequency_penalty": "a"}, 0, "", marks=pytest.mark.skip), ({"max_token": 0}, 0, ""), ({"max_token": 1024}, 0, ""), pytest.param({"max_token": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"max_token": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"max_token": "a"}, 0, "", marks=pytest.mark.skip), pytest.param({"unknown": "unknown"}, 0, "", marks=pytest.mark.skip), ], ) def test_llm(self, HttpApiAuth, add_chunks, llm, expected_code, expected_message): dataset_id, _, _ = add_chunks payload = {"name": "llm_test", "dataset_ids": [dataset_id], "llm": llm} res = create_chat_assistant(HttpApiAuth, payload) assert res["code"] == expected_code if expected_code == 0: if llm: for k, v in llm.items(): assert res["data"]["llm"][k] == v else: assert res["data"]["llm"]["model_name"] == "glm-4-flash@ZHIPU-AI" assert res["data"]["llm"]["temperature"] == 0.1 assert res["data"]["llm"]["top_p"] == 0.3 assert res["data"]["llm"]["presence_penalty"] == 0.4 assert res["data"]["llm"]["frequency_penalty"] == 0.7 assert res["data"]["llm"]["max_tokens"] == 512 else: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "prompt, expected_code, expected_message", [ ({}, 0, ""), ({"similarity_threshold": 0}, 0, ""), ({"similarity_threshold": 1}, 0, ""), pytest.param({"similarity_threshold": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"similarity_threshold": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"similarity_threshold": "a"}, 0, "", marks=pytest.mark.skip), ({"keywords_similarity_weight": 0}, 0, ""), ({"keywords_similarity_weight": 1}, 0, ""), pytest.param({"keywords_similarity_weight": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"keywords_similarity_weight": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"keywords_similarity_weight": "a"}, 0, "", marks=pytest.mark.skip), ({"variables": []}, 0, ""), ({"top_n": 0}, 0, ""), ({"top_n": 1}, 0, ""), pytest.param({"top_n": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"top_n": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"top_n": "a"}, 0, "", marks=pytest.mark.skip), ({"empty_response": "Hello World"}, 0, ""), ({"empty_response": ""}, 0, ""), ({"empty_response": "!@#$%^&*()"}, 0, ""), ({"empty_response": "中文测试"}, 0, ""), pytest.param({"empty_response": 123}, 0, "", marks=pytest.mark.skip), pytest.param({"empty_response": True}, 0, "", marks=pytest.mark.skip), pytest.param({"empty_response": " "}, 0, "", marks=pytest.mark.skip), ({"opener": "Hello World"}, 0, ""), ({"opener": ""}, 0, ""), ({"opener": "!@#$%^&*()"}, 0, ""), ({"opener": "中文测试"}, 0, ""), pytest.param({"opener": 123}, 0, "", marks=pytest.mark.skip), pytest.param({"opener": True}, 0, "", marks=pytest.mark.skip), pytest.param({"opener": " "}, 0, "", marks=pytest.mark.skip), ({"show_quote": True}, 0, ""), ({"show_quote": False}, 0, ""), ({"prompt": "Hello World {knowledge}"}, 0, ""), ({"prompt": "{knowledge}"}, 0, ""), ({"prompt": "!@#$%^&*() {knowledge}"}, 0, ""), ({"prompt": "中文测试 {knowledge}"}, 0, ""), ({"prompt": "Hello World"}, 102, "Parameter 'knowledge' is not used"), ({"prompt": "Hello World", "variables": []}, 0, ""), pytest.param({"prompt": 123}, 100, """AttributeError("\'int\' object has no attribute \'find\'")""", marks=pytest.mark.skip), pytest.param({"prompt": True}, 100, """AttributeError("\'int\' object has no attribute \'find\'")""", marks=pytest.mark.skip), pytest.param({"unknown": "unknown"}, 0, "", marks=pytest.mark.skip), ], ) def test_prompt(self, HttpApiAuth, add_chunks, prompt, expected_code, expected_message): dataset_id, _, _ = add_chunks payload = {"name": "prompt_test", "dataset_ids": [dataset_id], "prompt": prompt} res = create_chat_assistant(HttpApiAuth, payload) assert res["code"] == expected_code if expected_code == 0: if prompt: for k, v in prompt.items(): if k == "keywords_similarity_weight": assert res["data"]["prompt"][k] == 1 - v else: assert res["data"]["prompt"][k] == v else: assert res["data"]["prompt"]["similarity_threshold"] == 0.2 assert res["data"]["prompt"]["keywords_similarity_weight"] == 0.7 assert res["data"]["prompt"]["top_n"] == 6 assert res["data"]["prompt"]["variables"] == [{"key": "knowledge", "optional": False}] assert res["data"]["prompt"]["rerank_model"] == "" assert res["data"]["prompt"]["empty_response"] == "Sorry! No relevant content was found in the knowledge base!" assert res["data"]["prompt"]["opener"] == "Hi! I'm your assistant. What can I do for you?" assert res["data"]["prompt"]["show_quote"] is True assert ( res["data"]["prompt"]["prompt"] == 'You are an intelligent assistant. Please summarize the content of the dataset to answer the question. Please list the data in the dataset and answer in detail. When all dataset content is irrelevant to the question, your answer must include the sentence "The answer you are looking for is not found in the dataset!" Answers need to consider chat history.\n Here is the knowledge base:\n {knowledge}\n The above is the knowledge base.' ) else: assert res["message"] == expected_message class TestChatAssistantCreate2: @pytest.mark.p2 def test_unparsed_document(self, HttpApiAuth, add_document): dataset_id, _ = add_document payload = {"name": "prompt_test", "dataset_ids": [dataset_id]} res = create_chat_assistant(HttpApiAuth, payload) assert res["code"] == 102 assert "doesn't own parsed file" in res["message"]
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_chat_assistant_management/test_update_chat_assistant.py
test/testcases/test_http_api/test_chat_assistant_management/test_update_chat_assistant.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import list_chat_assistants, update_chat_assistant from configs import CHAT_ASSISTANT_NAME_LIMIT, INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth from utils import encode_avatar from utils.file_utils import create_image_file @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = update_chat_assistant(invalid_auth, "chat_assistant_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestChatAssistantUpdate: @pytest.mark.parametrize( "payload, expected_code, expected_message", [ pytest.param({"name": "valid_name"}, 0, "", marks=pytest.mark.p1), pytest.param({"name": "a" * (CHAT_ASSISTANT_NAME_LIMIT + 1)}, 102, "", marks=pytest.mark.skip(reason="issues/")), pytest.param({"name": 1}, 100, "", marks=pytest.mark.skip(reason="issues/")), pytest.param({"name": ""}, 102, "`name` cannot be empty.", marks=pytest.mark.p3), pytest.param({"name": "test_chat_assistant_1"}, 102, "Duplicated chat name in updating chat.", marks=pytest.mark.p3), pytest.param({"name": "TEST_CHAT_ASSISTANT_1"}, 102, "Duplicated chat name in updating chat.", marks=pytest.mark.p3), ], ) def test_name(self, HttpApiAuth, add_chat_assistants_func, payload, expected_code, expected_message): _, _, chat_assistant_ids = add_chat_assistants_func res = update_chat_assistant(HttpApiAuth, chat_assistant_ids[0], payload) assert res["code"] == expected_code, res if expected_code == 0: res = list_chat_assistants(HttpApiAuth, {"id": chat_assistant_ids[0]}) assert res["data"][0]["name"] == payload.get("name") else: assert res["message"] == expected_message @pytest.mark.parametrize( "dataset_ids, expected_code, expected_message", [ pytest.param([], 0, "", marks=pytest.mark.skip(reason="issues/")), pytest.param(lambda r: [r], 0, "", marks=pytest.mark.p1), pytest.param(["invalid_dataset_id"], 102, "You don't own the dataset invalid_dataset_id", marks=pytest.mark.p3), pytest.param("invalid_dataset_id", 102, "You don't own the dataset i", marks=pytest.mark.p3), ], ) def test_dataset_ids(self, HttpApiAuth, add_chat_assistants_func, dataset_ids, expected_code, expected_message): dataset_id, _, chat_assistant_ids = add_chat_assistants_func payload = {"name": "ragflow test"} if callable(dataset_ids): payload["dataset_ids"] = dataset_ids(dataset_id) else: payload["dataset_ids"] = dataset_ids res = update_chat_assistant(HttpApiAuth, chat_assistant_ids[0], payload) assert res["code"] == expected_code, res if expected_code == 0: res = list_chat_assistants(HttpApiAuth, {"id": chat_assistant_ids[0]}) assert res["data"][0]["name"] == payload.get("name") else: assert res["message"] == expected_message @pytest.mark.p3 def test_avatar(self, HttpApiAuth, add_chat_assistants_func, tmp_path): dataset_id, _, chat_assistant_ids = add_chat_assistants_func fn = create_image_file(tmp_path / "ragflow_test.png") payload = {"name": "avatar_test", "avatar": encode_avatar(fn), "dataset_ids": [dataset_id]} res = update_chat_assistant(HttpApiAuth, chat_assistant_ids[0], payload) assert res["code"] == 0 @pytest.mark.p3 @pytest.mark.parametrize( "llm, expected_code, expected_message", [ ({}, 100, "ValueError"), ({"model_name": "glm-4"}, 0, ""), ({"model_name": "unknown"}, 102, "`model_name` unknown doesn't exist"), ({"temperature": 0}, 0, ""), ({"temperature": 1}, 0, ""), pytest.param({"temperature": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"temperature": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"temperature": "a"}, 0, "", marks=pytest.mark.skip), ({"top_p": 0}, 0, ""), ({"top_p": 1}, 0, ""), pytest.param({"top_p": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"top_p": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"top_p": "a"}, 0, "", marks=pytest.mark.skip), ({"presence_penalty": 0}, 0, ""), ({"presence_penalty": 1}, 0, ""), pytest.param({"presence_penalty": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"presence_penalty": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"presence_penalty": "a"}, 0, "", marks=pytest.mark.skip), ({"frequency_penalty": 0}, 0, ""), ({"frequency_penalty": 1}, 0, ""), pytest.param({"frequency_penalty": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"frequency_penalty": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"frequency_penalty": "a"}, 0, "", marks=pytest.mark.skip), ({"max_token": 0}, 0, ""), ({"max_token": 1024}, 0, ""), pytest.param({"max_token": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"max_token": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"max_token": "a"}, 0, "", marks=pytest.mark.skip), pytest.param({"unknown": "unknown"}, 0, "", marks=pytest.mark.skip), ], ) def test_llm(self, HttpApiAuth, add_chat_assistants_func, llm, expected_code, expected_message): dataset_id, _, chat_assistant_ids = add_chat_assistants_func payload = {"name": "llm_test", "dataset_ids": [dataset_id], "llm": llm} res = update_chat_assistant(HttpApiAuth, chat_assistant_ids[0], payload) assert res["code"] == expected_code if expected_code == 0: res = list_chat_assistants(HttpApiAuth, {"id": chat_assistant_ids[0]}) if llm: for k, v in llm.items(): assert res["data"][0]["llm"][k] == v else: assert res["data"][0]["llm"]["model_name"] == "glm-4-flash@ZHIPU-AI" assert res["data"][0]["llm"]["temperature"] == 0.1 assert res["data"][0]["llm"]["top_p"] == 0.3 assert res["data"][0]["llm"]["presence_penalty"] == 0.4 assert res["data"][0]["llm"]["frequency_penalty"] == 0.7 assert res["data"][0]["llm"]["max_tokens"] == 512 else: assert expected_message in res["message"] @pytest.mark.p3 @pytest.mark.parametrize( "prompt, expected_code, expected_message", [ ({}, 100, "ValueError"), ({"similarity_threshold": 0}, 0, ""), ({"similarity_threshold": 1}, 0, ""), pytest.param({"similarity_threshold": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"similarity_threshold": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"similarity_threshold": "a"}, 0, "", marks=pytest.mark.skip), ({"keywords_similarity_weight": 0}, 0, ""), ({"keywords_similarity_weight": 1}, 0, ""), pytest.param({"keywords_similarity_weight": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"keywords_similarity_weight": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"keywords_similarity_weight": "a"}, 0, "", marks=pytest.mark.skip), ({"variables": []}, 0, ""), ({"top_n": 0}, 0, ""), ({"top_n": 1}, 0, ""), pytest.param({"top_n": -1}, 0, "", marks=pytest.mark.skip), pytest.param({"top_n": 10}, 0, "", marks=pytest.mark.skip), pytest.param({"top_n": "a"}, 0, "", marks=pytest.mark.skip), ({"empty_response": "Hello World"}, 0, ""), ({"empty_response": ""}, 0, ""), ({"empty_response": "!@#$%^&*()"}, 0, ""), ({"empty_response": "中文测试"}, 0, ""), pytest.param({"empty_response": 123}, 0, "", marks=pytest.mark.skip), pytest.param({"empty_response": True}, 0, "", marks=pytest.mark.skip), pytest.param({"empty_response": " "}, 0, "", marks=pytest.mark.skip), ({"opener": "Hello World"}, 0, ""), ({"opener": ""}, 0, ""), ({"opener": "!@#$%^&*()"}, 0, ""), ({"opener": "中文测试"}, 0, ""), pytest.param({"opener": 123}, 0, "", marks=pytest.mark.skip), pytest.param({"opener": True}, 0, "", marks=pytest.mark.skip), pytest.param({"opener": " "}, 0, "", marks=pytest.mark.skip), ({"show_quote": True}, 0, ""), ({"show_quote": False}, 0, ""), ({"prompt": "Hello World {knowledge}"}, 0, ""), ({"prompt": "{knowledge}"}, 0, ""), ({"prompt": "!@#$%^&*() {knowledge}"}, 0, ""), ({"prompt": "中文测试 {knowledge}"}, 0, ""), ({"prompt": "Hello World"}, 102, "Parameter 'knowledge' is not used"), ({"prompt": "Hello World", "variables": []}, 0, ""), pytest.param({"prompt": 123}, 100, """AttributeError("\'int\' object has no attribute \'find\'")""", marks=pytest.mark.skip), pytest.param({"prompt": True}, 100, """AttributeError("\'int\' object has no attribute \'find\'")""", marks=pytest.mark.skip), pytest.param({"unknown": "unknown"}, 0, "", marks=pytest.mark.skip), ], ) def test_prompt(self, HttpApiAuth, add_chat_assistants_func, prompt, expected_code, expected_message): dataset_id, _, chat_assistant_ids = add_chat_assistants_func payload = {"name": "prompt_test", "dataset_ids": [dataset_id], "prompt": prompt} res = update_chat_assistant(HttpApiAuth, chat_assistant_ids[0], payload) assert res["code"] == expected_code if expected_code == 0: res = list_chat_assistants(HttpApiAuth, {"id": chat_assistant_ids[0]}) if prompt: for k, v in prompt.items(): if k == "keywords_similarity_weight": assert res["data"][0]["prompt"][k] == 1 - v else: assert res["data"][0]["prompt"][k] == v else: assert res["data"]["prompt"][0]["similarity_threshold"] == 0.2 assert res["data"]["prompt"][0]["keywords_similarity_weight"] == 0.7 assert res["data"]["prompt"][0]["top_n"] == 6 assert res["data"]["prompt"][0]["variables"] == [{"key": "knowledge", "optional": False}] assert res["data"]["prompt"][0]["rerank_model"] == "" assert res["data"]["prompt"][0]["empty_response"] == "Sorry! No relevant content was found in the knowledge base!" assert res["data"]["prompt"][0]["opener"] == "Hi! I'm your assistant. What can I do for you?" assert res["data"]["prompt"][0]["show_quote"] is True assert ( res["data"]["prompt"][0]["prompt"] == 'You are an intelligent assistant. Please summarize the content of the dataset to answer the question. Please list the data in the dataset and answer in detail. When all dataset content is irrelevant to the question, your answer must include the sentence "The answer you are looking for is not found in the dataset!" Answers need to consider chat history.\n Here is the knowledge base:\n {knowledge}\n The above is the knowledge base.' ) else: assert expected_message in res["message"]
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_chat_assistant_management/test_list_chat_assistants.py
test/testcases/test_http_api/test_chat_assistant_management/test_list_chat_assistants.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import delete_datasets, list_chat_assistants from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth from utils import is_sorted @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = list_chat_assistants(invalid_auth) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.usefixtures("add_chat_assistants") class TestChatAssistantsList: @pytest.mark.p1 def test_default(self, HttpApiAuth): res = list_chat_assistants(HttpApiAuth) assert res["code"] == 0 assert len(res["data"]) == 5 @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page": None, "page_size": 2}, 0, 2, ""), ({"page": 0, "page_size": 2}, 0, 2, ""), ({"page": 2, "page_size": 2}, 0, 2, ""), ({"page": 3, "page_size": 2}, 0, 1, ""), ({"page": "3", "page_size": 2}, 0, 1, ""), pytest.param( {"page": -1, "page_size": 2}, 100, 0, "1064", marks=pytest.mark.skip(reason="issues/5851"), ), pytest.param( {"page": "a", "page_size": 2}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip(reason="issues/5851"), ), ], ) def test_page(self, HttpApiAuth, params, expected_code, expected_page_size, expected_message): res = list_chat_assistants(HttpApiAuth, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page_size": None}, 0, 5, ""), ({"page_size": 0}, 0, 0, ""), ({"page_size": 1}, 0, 1, ""), ({"page_size": 6}, 0, 5, ""), ({"page_size": "1"}, 0, 1, ""), pytest.param( {"page_size": -1}, 100, 0, "1064", marks=pytest.mark.skip(reason="issues/5851"), ), pytest.param( {"page_size": "a"}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip(reason="issues/5851"), ), ], ) def test_page_size( self, HttpApiAuth, params, expected_code, expected_page_size, expected_message, ): res = list_chat_assistants(HttpApiAuth, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "params, expected_code, assertions, expected_message", [ ({"orderby": None}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"orderby": "create_time"}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"orderby": "update_time"}, 0, lambda r: (is_sorted(r["data"], "update_time", True)), ""), pytest.param( {"orderby": "name", "desc": "False"}, 0, lambda r: (is_sorted(r["data"], "name", False)), "", marks=pytest.mark.skip(reason="issues/5851"), ), pytest.param( {"orderby": "unknown"}, 102, 0, "orderby should be create_time or update_time", marks=pytest.mark.skip(reason="issues/5851"), ), ], ) def test_orderby( self, HttpApiAuth, params, expected_code, assertions, expected_message, ): res = list_chat_assistants(HttpApiAuth, params=params) assert res["code"] == expected_code if expected_code == 0: if callable(assertions): assert assertions(res) else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "params, expected_code, assertions, expected_message", [ ({"desc": None}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"desc": "true"}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"desc": "True"}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"desc": True}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"desc": "false"}, 0, lambda r: (is_sorted(r["data"], "create_time", False)), ""), ({"desc": "False"}, 0, lambda r: (is_sorted(r["data"], "create_time", False)), ""), ({"desc": False}, 0, lambda r: (is_sorted(r["data"], "create_time", False)), ""), ({"desc": "False", "orderby": "update_time"}, 0, lambda r: (is_sorted(r["data"], "update_time", False)), ""), pytest.param( {"desc": "unknown"}, 102, 0, "desc should be true or false", marks=pytest.mark.skip(reason="issues/5851"), ), ], ) def test_desc( self, HttpApiAuth, params, expected_code, assertions, expected_message, ): res = list_chat_assistants(HttpApiAuth, params=params) assert res["code"] == expected_code if expected_code == 0: if callable(assertions): assert assertions(res) else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_num, expected_message", [ ({"name": None}, 0, 5, ""), ({"name": ""}, 0, 5, ""), ({"name": "test_chat_assistant_1"}, 0, 1, ""), ({"name": "unknown"}, 102, 0, "The chat doesn't exist"), ], ) def test_name(self, HttpApiAuth, params, expected_code, expected_num, expected_message): res = list_chat_assistants(HttpApiAuth, params=params) assert res["code"] == expected_code if expected_code == 0: if params["name"] in [None, ""]: assert len(res["data"]) == expected_num else: assert res["data"][0]["name"] == params["name"] else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "chat_assistant_id, expected_code, expected_num, expected_message", [ (None, 0, 5, ""), ("", 0, 5, ""), (lambda r: r[0], 0, 1, ""), ("unknown", 102, 0, "The chat doesn't exist"), ], ) def test_id( self, HttpApiAuth, add_chat_assistants, chat_assistant_id, expected_code, expected_num, expected_message, ): _, _, chat_assistant_ids = add_chat_assistants if callable(chat_assistant_id): params = {"id": chat_assistant_id(chat_assistant_ids)} else: params = {"id": chat_assistant_id} res = list_chat_assistants(HttpApiAuth, params=params) assert res["code"] == expected_code if expected_code == 0: if params["id"] in [None, ""]: assert len(res["data"]) == expected_num else: assert res["data"][0]["id"] == params["id"] else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "chat_assistant_id, name, expected_code, expected_num, expected_message", [ (lambda r: r[0], "test_chat_assistant_0", 0, 1, ""), (lambda r: r[0], "test_chat_assistant_1", 102, 0, "The chat doesn't exist"), (lambda r: r[0], "unknown", 102, 0, "The chat doesn't exist"), ("id", "chat_assistant_0", 102, 0, "The chat doesn't exist"), ], ) def test_name_and_id( self, HttpApiAuth, add_chat_assistants, chat_assistant_id, name, expected_code, expected_num, expected_message, ): _, _, chat_assistant_ids = add_chat_assistants if callable(chat_assistant_id): params = {"id": chat_assistant_id(chat_assistant_ids), "name": name} else: params = {"id": chat_assistant_id, "name": name} res = list_chat_assistants(HttpApiAuth, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]) == expected_num else: assert res["message"] == expected_message @pytest.mark.p3 def test_concurrent_list(self, HttpApiAuth): count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(list_chat_assistants, HttpApiAuth) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.p3 def test_invalid_params(self, HttpApiAuth): params = {"a": "b"} res = list_chat_assistants(HttpApiAuth, params=params) assert res["code"] == 0 assert len(res["data"]) == 5 @pytest.mark.p2 def test_list_chats_after_deleting_associated_dataset(self, HttpApiAuth, add_chat_assistants): dataset_id, _, _ = add_chat_assistants res = delete_datasets(HttpApiAuth, {"ids": [dataset_id]}) assert res["code"] == 0 res = list_chat_assistants(HttpApiAuth) assert res["code"] == 0 assert len(res["data"]) == 5
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_chunk_management_within_dataset/test_add_chunk.py
test/testcases/test_http_api/test_chunk_management_within_dataset/test_add_chunk.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import add_chunk, delete_documents, list_chunks from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth def validate_chunk_details(dataset_id, document_id, payload, res): chunk = res["data"]["chunk"] assert chunk["dataset_id"] == dataset_id assert chunk["document_id"] == document_id assert chunk["content"] == payload["content"] if "important_keywords" in payload: assert chunk["important_keywords"] == payload["important_keywords"] if "questions" in payload: assert chunk["questions"] == [str(q).strip() for q in payload.get("questions", []) if str(q).strip()] @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = add_chunk(invalid_auth, "dataset_id", "document_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestAddChunk: @pytest.mark.p1 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"content": None}, 100, """TypeError("unsupported operand type(s) for +: \'NoneType\' and \'str\'")"""), ({"content": ""}, 102, "`content` is required"), pytest.param( {"content": 1}, 100, """TypeError("unsupported operand type(s) for +: \'int\' and \'str\'")""", marks=pytest.mark.skip, ), ({"content": "a"}, 0, ""), ({"content": " "}, 102, "`content` is required"), ({"content": "\n!?。;!?\"'"}, 0, ""), ], ) def test_content(self, HttpApiAuth, add_document, payload, expected_code, expected_message): dataset_id, document_id = add_document res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res chunks_count = res["data"]["doc"]["chunk_count"] res = add_chunk(HttpApiAuth, dataset_id, document_id, payload) assert res["code"] == expected_code if expected_code == 0: validate_chunk_details(dataset_id, document_id, payload, res) res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res assert res["data"]["doc"]["chunk_count"] == chunks_count + 1 else: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"content": "chunk test", "important_keywords": ["a", "b", "c"]}, 0, ""), ({"content": "chunk test", "important_keywords": [""]}, 0, ""), ( {"content": "chunk test", "important_keywords": [1]}, 100, "TypeError('sequence item 0: expected str instance, int found')", ), ({"content": "chunk test", "important_keywords": ["a", "a"]}, 0, ""), ({"content": "chunk test", "important_keywords": "abc"}, 102, "`important_keywords` is required to be a list"), ({"content": "chunk test", "important_keywords": 123}, 102, "`important_keywords` is required to be a list"), ], ) def test_important_keywords(self, HttpApiAuth, add_document, payload, expected_code, expected_message): dataset_id, document_id = add_document res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res chunks_count = res["data"]["doc"]["chunk_count"] res = add_chunk(HttpApiAuth, dataset_id, document_id, payload) assert res["code"] == expected_code if expected_code == 0: validate_chunk_details(dataset_id, document_id, payload, res) res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res assert res["data"]["doc"]["chunk_count"] == chunks_count + 1 else: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"content": "chunk test", "questions": ["a", "b", "c"]}, 0, ""), ({"content": "chunk test", "questions": [""]}, 0, ""), ({"content": "chunk test", "questions": [1]}, 100, "TypeError('sequence item 0: expected str instance, int found')"), ({"content": "chunk test", "questions": ["a", "a"]}, 0, ""), ({"content": "chunk test", "questions": "abc"}, 102, "`questions` is required to be a list"), ({"content": "chunk test", "questions": 123}, 102, "`questions` is required to be a list"), ], ) def test_questions(self, HttpApiAuth, add_document, payload, expected_code, expected_message): dataset_id, document_id = add_document res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res chunks_count = res["data"]["doc"]["chunk_count"] res = add_chunk(HttpApiAuth, dataset_id, document_id, payload) assert res["code"] == expected_code if expected_code == 0: validate_chunk_details(dataset_id, document_id, payload, res) if res["code"] != 0: assert False, res res = list_chunks(HttpApiAuth, dataset_id, document_id) assert res["data"]["doc"]["chunk_count"] == chunks_count + 1 else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<NotFound '404: Not Found'>"), ( "invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id.", ), ], ) def test_invalid_dataset_id( self, HttpApiAuth, add_document, dataset_id, expected_code, expected_message, ): _, document_id = add_document res = add_chunk(HttpApiAuth, dataset_id, document_id, {"content": "a"}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "document_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_document_id", 102, "You don't own the document invalid_document_id.", ), ], ) def test_invalid_document_id(self, HttpApiAuth, add_document, document_id, expected_code, expected_message): dataset_id, _ = add_document res = add_chunk(HttpApiAuth, dataset_id, document_id, {"content": "chunk test"}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 def test_repeated_add_chunk(self, HttpApiAuth, add_document): payload = {"content": "chunk test"} dataset_id, document_id = add_document res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res chunks_count = res["data"]["doc"]["chunk_count"] res = add_chunk(HttpApiAuth, dataset_id, document_id, payload) assert res["code"] == 0 validate_chunk_details(dataset_id, document_id, payload, res) res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res assert res["data"]["doc"]["chunk_count"] == chunks_count + 1 res = add_chunk(HttpApiAuth, dataset_id, document_id, payload) assert res["code"] == 0 validate_chunk_details(dataset_id, document_id, payload, res) res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res assert res["data"]["doc"]["chunk_count"] == chunks_count + 2 @pytest.mark.p2 def test_add_chunk_to_deleted_document(self, HttpApiAuth, add_document): dataset_id, document_id = add_document delete_documents(HttpApiAuth, dataset_id, {"ids": [document_id]}) res = add_chunk(HttpApiAuth, dataset_id, document_id, {"content": "chunk test"}) assert res["code"] == 102 assert res["message"] == f"You don't own the document {document_id}." @pytest.mark.skip(reason="issues/6411") def test_concurrent_add_chunk(self, HttpApiAuth, add_document): count = 50 dataset_id, document_id = add_document res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res chunks_count = res["data"]["doc"]["chunk_count"] with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( add_chunk, HttpApiAuth, dataset_id, document_id, {"content": f"chunk test {i}"}, ) for i in range(count) ] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res assert res["data"]["doc"]["chunk_count"] == chunks_count + count
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_chunk_management_within_dataset/test_retrieval_chunks.py
test/testcases/test_http_api/test_chunk_management_within_dataset/test_retrieval_chunks.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import os from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import retrieval_chunks from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = retrieval_chunks(invalid_auth) assert res["code"] == expected_code assert res["message"] == expected_message class TestChunksRetrieval: @pytest.mark.p1 @pytest.mark.parametrize( "payload, expected_code, expected_page_size, expected_message", [ ({"question": "chunk", "dataset_ids": None}, 0, 4, ""), ({"question": "chunk", "document_ids": None}, 102, 0, "`dataset_ids` is required."), ({"question": "chunk", "dataset_ids": None, "document_ids": None}, 0, 4, ""), ({"question": "chunk"}, 102, 0, "`dataset_ids` is required."), ], ) def test_basic_scenarios(self, HttpApiAuth, add_chunks, payload, expected_code, expected_page_size, expected_message): dataset_id, document_id, _ = add_chunks if "dataset_ids" in payload: payload["dataset_ids"] = [dataset_id] if "document_ids" in payload: payload["document_ids"] = [document_id] res = retrieval_chunks(HttpApiAuth, payload) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["chunks"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "payload, expected_code, expected_page_size, expected_message", [ pytest.param( {"page": None, "page_size": 2}, 100, 2, """TypeError("int() argument must be a string, a bytes-like object or a real number, not \'NoneType\'")""", marks=pytest.mark.skip, ), pytest.param( {"page": 0, "page_size": 2}, 100, 0, "ValueError('Search does not support negative slicing.')", marks=pytest.mark.skip, ), ({"page": 2, "page_size": 2}, 0, 2, ""), ({"page": 3, "page_size": 2}, 0, 0, ""), ({"page": "3", "page_size": 2}, 0, 0, ""), pytest.param( {"page": -1, "page_size": 2}, 100, 0, "ValueError('Search does not support negative slicing.')", marks=pytest.mark.skip, ), pytest.param( {"page": "a", "page_size": 2}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip, ), ], ) def test_page(self, HttpApiAuth, add_chunks, payload, expected_code, expected_page_size, expected_message): dataset_id, _, _ = add_chunks payload.update({"question": "chunk", "dataset_ids": [dataset_id]}) res = retrieval_chunks(HttpApiAuth, payload) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["chunks"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "payload, expected_code, expected_page_size, expected_message", [ pytest.param( {"page_size": None}, 100, 0, """TypeError("int() argument must be a string, a bytes-like object or a real number, not \'NoneType\'")""", marks=pytest.mark.skip, ), # ({"page_size": 0}, 0, 0, ""), pytest.param({"page_size": 1}, 0, 1, "", marks=pytest.mark.skip(reason="issues/10692")), ({"page_size": 5}, 0, 4, ""), pytest.param({"page_size": "1"}, 0, 1, "", marks=pytest.mark.skip(reason="issues/10692")), # ({"page_size": -1}, 0, 0, ""), pytest.param( {"page_size": "a"}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip, ), ], ) def test_page_size(self, HttpApiAuth, add_chunks, payload, expected_code, expected_page_size, expected_message): dataset_id, _, _ = add_chunks payload.update({"question": "chunk", "dataset_ids": [dataset_id]}) res = retrieval_chunks(HttpApiAuth, payload) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["chunks"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "payload, expected_code, expected_page_size, expected_message", [ ({"vector_similarity_weight": 0}, 0, 4, ""), ({"vector_similarity_weight": 0.5}, 0, 4, ""), ({"vector_similarity_weight": 10}, 0, 4, ""), pytest.param( {"vector_similarity_weight": "a"}, 100, 0, """ValueError("could not convert string to float: \'a\'")""", marks=pytest.mark.skip, ), ], ) def test_vector_similarity_weight(self, HttpApiAuth, add_chunks, payload, expected_code, expected_page_size, expected_message): dataset_id, _, _ = add_chunks payload.update({"question": "chunk", "dataset_ids": [dataset_id]}) res = retrieval_chunks(HttpApiAuth, payload) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["chunks"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "payload, expected_code, expected_page_size, expected_message", [ ({"top_k": 10}, 0, 4, ""), pytest.param( {"top_k": 1}, 0, 4, "", marks=pytest.mark.skipif(os.getenv("DOC_ENGINE") in ["infinity", "opensearch"], reason="Infinity"), ), pytest.param( {"top_k": 1}, 0, 1, "", marks=pytest.mark.skipif(os.getenv("DOC_ENGINE") in [None, "opensearch", "elasticsearch"], reason="elasticsearch"), ), pytest.param( {"top_k": -1}, 100, 4, "must be greater than 0", marks=pytest.mark.skipif(os.getenv("DOC_ENGINE") in ["infinity", "opensearch"], reason="Infinity"), ), pytest.param( {"top_k": -1}, 100, 4, "3014", marks=pytest.mark.skipif(os.getenv("DOC_ENGINE") in [None, "opensearch", "elasticsearch"], reason="elasticsearch"), ), pytest.param( {"top_k": "a"}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip, ), ], ) def test_top_k(self, HttpApiAuth, add_chunks, payload, expected_code, expected_page_size, expected_message): dataset_id, _, _ = add_chunks payload.update({"question": "chunk", "dataset_ids": [dataset_id]}) res = retrieval_chunks(HttpApiAuth, payload) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["chunks"]) == expected_page_size else: assert expected_message in res["message"] @pytest.mark.skip @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"rerank_id": "BAAI/bge-reranker-v2-m3"}, 0, ""), pytest.param({"rerank_id": "unknown"}, 100, "LookupError('Model(unknown) not authorized')", marks=pytest.mark.skip), ], ) def test_rerank_id(self, HttpApiAuth, add_chunks, payload, expected_code, expected_message): dataset_id, _, _ = add_chunks payload.update({"question": "chunk", "dataset_ids": [dataset_id]}) res = retrieval_chunks(HttpApiAuth, payload) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["chunks"]) > 0 else: assert expected_message in res["message"] @pytest.mark.skip @pytest.mark.parametrize( "payload, expected_code, expected_page_size, expected_message", [ ({"keyword": True}, 0, 5, ""), ({"keyword": "True"}, 0, 5, ""), ({"keyword": False}, 0, 5, ""), ({"keyword": "False"}, 0, 5, ""), ({"keyword": None}, 0, 5, ""), ], ) def test_keyword(self, HttpApiAuth, add_chunks, payload, expected_code, expected_page_size, expected_message): dataset_id, _, _ = add_chunks payload.update({"question": "chunk test", "dataset_ids": [dataset_id]}) res = retrieval_chunks(HttpApiAuth, payload) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["chunks"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "payload, expected_code, expected_highlight, expected_message", [ ({"highlight": True}, 0, True, ""), ({"highlight": "True"}, 0, True, ""), pytest.param({"highlight": False}, 0, False, "", marks=pytest.mark.skip(reason="issues/6648")), ({"highlight": "False"}, 0, False, ""), pytest.param({"highlight": None}, 0, False, "", marks=pytest.mark.skip(reason="issues/6648")), ], ) def test_highlight(self, HttpApiAuth, add_chunks, payload, expected_code, expected_highlight, expected_message): dataset_id, _, _ = add_chunks payload.update({"question": "chunk", "dataset_ids": [dataset_id]}) res = retrieval_chunks(HttpApiAuth, payload) assert res["code"] == expected_code if expected_highlight: for chunk in res["data"]["chunks"]: assert "highlight" in chunk else: for chunk in res["data"]["chunks"]: assert "highlight" not in chunk if expected_code != 0: assert res["message"] == expected_message @pytest.mark.p3 def test_invalid_params(self, HttpApiAuth, add_chunks): dataset_id, _, _ = add_chunks payload = {"question": "chunk", "dataset_ids": [dataset_id], "a": "b"} res = retrieval_chunks(HttpApiAuth, payload) assert res["code"] == 0 assert len(res["data"]["chunks"]) == 4 @pytest.mark.p3 def test_concurrent_retrieval(self, HttpApiAuth, add_chunks): dataset_id, _, _ = add_chunks count = 100 payload = {"question": "chunk", "dataset_ids": [dataset_id]} with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(retrieval_chunks, HttpApiAuth, payload) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_chunk_management_within_dataset/test_update_chunk.py
test/testcases/test_http_api/test_chunk_management_within_dataset/test_update_chunk.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import os from concurrent.futures import ThreadPoolExecutor, as_completed from random import randint import pytest from common import delete_documents, update_chunk from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = update_chunk(invalid_auth, "dataset_id", "document_id", "chunk_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestUpdatedChunk: @pytest.mark.p1 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ pytest.param({"content": None}, 0, "", marks=pytest.mark.skipif(os.getenv("DOC_ENGINE") == "infinity", reason="issues/6509")), pytest.param( {"content": ""}, 100, """APIRequestFailedError(\'Error code: 400, with error text {"error":{"code":"1213","message":"未正常接收到prompt参数。"}}\')""", marks=pytest.mark.skip(reason="issues/6541"), ), pytest.param( {"content": 1}, 100, "TypeError('expected string or bytes-like object')", marks=pytest.mark.skip, ), ({"content": "update chunk"}, 0, ""), pytest.param( {"content": " "}, 100, """APIRequestFailedError(\'Error code: 400, with error text {"error":{"code":"1213","message":"未正常接收到prompt参数。"}}\')""", marks=pytest.mark.skip(reason="issues/6541"), ), ({"content": "\n!?。;!?\"'"}, 0, ""), ], ) def test_content(self, HttpApiAuth, add_chunks, payload, expected_code, expected_message): dataset_id, document_id, chunk_ids = add_chunks res = update_chunk(HttpApiAuth, dataset_id, document_id, chunk_ids[0], payload) assert res["code"] == expected_code if expected_code != 0: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"important_keywords": ["a", "b", "c"]}, 0, ""), ({"important_keywords": [""]}, 0, ""), ({"important_keywords": [1]}, 100, "TypeError('sequence item 0: expected str instance, int found')"), ({"important_keywords": ["a", "a"]}, 0, ""), ({"important_keywords": "abc"}, 102, "`important_keywords` should be a list"), ({"important_keywords": 123}, 102, "`important_keywords` should be a list"), ], ) def test_important_keywords(self, HttpApiAuth, add_chunks, payload, expected_code, expected_message): dataset_id, document_id, chunk_ids = add_chunks res = update_chunk(HttpApiAuth, dataset_id, document_id, chunk_ids[0], payload) assert res["code"] == expected_code if expected_code != 0: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"questions": ["a", "b", "c"]}, 0, ""), ({"questions": [""]}, 0, ""), ({"questions": [1]}, 100, "TypeError('sequence item 0: expected str instance, int found')"), ({"questions": ["a", "a"]}, 0, ""), ({"questions": "abc"}, 102, "`questions` should be a list"), ({"questions": 123}, 102, "`questions` should be a list"), ], ) def test_questions(self, HttpApiAuth, add_chunks, payload, expected_code, expected_message): dataset_id, document_id, chunk_ids = add_chunks res = update_chunk(HttpApiAuth, dataset_id, document_id, chunk_ids[0], payload) assert res["code"] == expected_code if expected_code != 0: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"available": True}, 0, ""), pytest.param({"available": "True"}, 100, """ValueError("invalid literal for int() with base 10: \'True\'")""", marks=pytest.mark.skip), ({"available": 1}, 0, ""), ({"available": False}, 0, ""), pytest.param({"available": "False"}, 100, """ValueError("invalid literal for int() with base 10: \'False\'")""", marks=pytest.mark.skip), ({"available": 0}, 0, ""), ], ) def test_available( self, HttpApiAuth, add_chunks, payload, expected_code, expected_message, ): dataset_id, document_id, chunk_ids = add_chunks res = update_chunk(HttpApiAuth, dataset_id, document_id, chunk_ids[0], payload) assert res["code"] == expected_code if expected_code != 0: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<NotFound '404: Not Found'>"), pytest.param("invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id.", marks=pytest.mark.skipif(os.getenv("DOC_ENGINE") == "infinity", reason="infinity")), pytest.param("invalid_dataset_id", 102, "Can't find this chunk", marks=pytest.mark.skipif(os.getenv("DOC_ENGINE") in [None, "opensearch", "elasticsearch"], reason="elasticsearch")), ], ) def test_invalid_dataset_id(self, HttpApiAuth, add_chunks, dataset_id, expected_code, expected_message): _, document_id, chunk_ids = add_chunks res = update_chunk(HttpApiAuth, dataset_id, document_id, chunk_ids[0]) assert res["code"] == expected_code assert expected_message in res["message"] @pytest.mark.p3 @pytest.mark.parametrize( "document_id, expected_code, expected_message", [ ("", 100, "<NotFound '404: Not Found'>"), ( "invalid_document_id", 102, "You don't own the document invalid_document_id.", ), ], ) def test_invalid_document_id(self, HttpApiAuth, add_chunks, document_id, expected_code, expected_message): dataset_id, _, chunk_ids = add_chunks res = update_chunk(HttpApiAuth, dataset_id, document_id, chunk_ids[0]) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "chunk_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_document_id", 102, "Can't find this chunk invalid_document_id", ), ], ) def test_invalid_chunk_id(self, HttpApiAuth, add_chunks, chunk_id, expected_code, expected_message): dataset_id, document_id, _ = add_chunks res = update_chunk(HttpApiAuth, dataset_id, document_id, chunk_id) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 def test_repeated_update_chunk(self, HttpApiAuth, add_chunks): dataset_id, document_id, chunk_ids = add_chunks res = update_chunk(HttpApiAuth, dataset_id, document_id, chunk_ids[0], {"content": "chunk test 1"}) assert res["code"] == 0 res = update_chunk(HttpApiAuth, dataset_id, document_id, chunk_ids[0], {"content": "chunk test 2"}) assert res["code"] == 0 @pytest.mark.p3 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"unknown_key": "unknown_value"}, 0, ""), ({}, 0, ""), pytest.param(None, 100, """TypeError("argument of type \'NoneType\' is not iterable")""", marks=pytest.mark.skip), ], ) def test_invalid_params(self, HttpApiAuth, add_chunks, payload, expected_code, expected_message): dataset_id, document_id, chunk_ids = add_chunks res = update_chunk(HttpApiAuth, dataset_id, document_id, chunk_ids[0], payload) assert res["code"] == expected_code if expected_code != 0: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.skipif(os.getenv("DOC_ENGINE") == "infinity", reason="issues/6554") def test_concurrent_update_chunk(self, HttpApiAuth, add_chunks): count = 50 dataset_id, document_id, chunk_ids = add_chunks with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( update_chunk, HttpApiAuth, dataset_id, document_id, chunk_ids[randint(0, 3)], {"content": f"update chunk test {i}"}, ) for i in range(count) ] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.p3 def test_update_chunk_to_deleted_document(self, HttpApiAuth, add_chunks): dataset_id, document_id, chunk_ids = add_chunks delete_documents(HttpApiAuth, dataset_id, {"ids": [document_id]}) res = update_chunk(HttpApiAuth, dataset_id, document_id, chunk_ids[0]) assert res["code"] == 102 assert res["message"] in [f"You don't own the document {document_id}.", f"Can't find this chunk {chunk_ids[0]}"]
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_chunk_management_within_dataset/conftest.py
test/testcases/test_http_api/test_chunk_management_within_dataset/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from time import sleep import pytest from common import batch_add_chunks, delete_chunks, list_documents, parse_documents from utils import wait_for @wait_for(30, 1, "Document parsing timeout") def condition(_auth, _dataset_id): res = list_documents(_auth, _dataset_id) for doc in res["data"]["docs"]: if doc["run"] != "DONE": return False return True @pytest.fixture(scope="function") def add_chunks_func(request, HttpApiAuth, add_document): def cleanup(): delete_chunks(HttpApiAuth, dataset_id, document_id, {"chunk_ids": []}) request.addfinalizer(cleanup) dataset_id, document_id = add_document parse_documents(HttpApiAuth, dataset_id, {"document_ids": [document_id]}) condition(HttpApiAuth, dataset_id) chunk_ids = batch_add_chunks(HttpApiAuth, dataset_id, document_id, 4) # issues/6487 sleep(1) return dataset_id, document_id, chunk_ids
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_chunk_management_within_dataset/test_delete_chunks.py
test/testcases/test_http_api/test_chunk_management_within_dataset/test_delete_chunks.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import batch_add_chunks, delete_chunks, list_chunks from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = delete_chunks(invalid_auth, "dataset_id", "document_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestChunksDeletion: @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<NotFound '404: Not Found'>"), ( "invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id.", ), ], ) def test_invalid_dataset_id(self, HttpApiAuth, add_chunks_func, dataset_id, expected_code, expected_message): _, document_id, chunk_ids = add_chunks_func res = delete_chunks(HttpApiAuth, dataset_id, document_id, {"chunk_ids": chunk_ids}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "document_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ("invalid_document_id", 100, """LookupError("Can't find the document with ID invalid_document_id!")"""), ], ) def test_invalid_document_id(self, HttpApiAuth, add_chunks_func, document_id, expected_code, expected_message): dataset_id, _, chunk_ids = add_chunks_func res = delete_chunks(HttpApiAuth, dataset_id, document_id, {"chunk_ids": chunk_ids}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.parametrize( "payload", [ pytest.param(lambda r: {"chunk_ids": ["invalid_id"] + r}, marks=pytest.mark.p3), pytest.param(lambda r: {"chunk_ids": r[:1] + ["invalid_id"] + r[1:4]}, marks=pytest.mark.p1), pytest.param(lambda r: {"chunk_ids": r + ["invalid_id"]}, marks=pytest.mark.p3), ], ) def test_delete_partial_invalid_id(self, HttpApiAuth, add_chunks_func, payload): dataset_id, document_id, chunk_ids = add_chunks_func if callable(payload): payload = payload(chunk_ids) res = delete_chunks(HttpApiAuth, dataset_id, document_id, payload) assert res["code"] == 102 assert res["message"] == "rm_chunk deleted chunks 4, expect 5" res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res assert len(res["data"]["chunks"]) == 1 assert res["data"]["total"] == 1 @pytest.mark.p3 def test_repeated_deletion(self, HttpApiAuth, add_chunks_func): dataset_id, document_id, chunk_ids = add_chunks_func payload = {"chunk_ids": chunk_ids} res = delete_chunks(HttpApiAuth, dataset_id, document_id, payload) assert res["code"] == 0 res = delete_chunks(HttpApiAuth, dataset_id, document_id, payload) assert res["code"] == 102 assert res["message"] == "rm_chunk deleted chunks 0, expect 4" @pytest.mark.p3 def test_duplicate_deletion(self, HttpApiAuth, add_chunks_func): dataset_id, document_id, chunk_ids = add_chunks_func res = delete_chunks(HttpApiAuth, dataset_id, document_id, {"chunk_ids": chunk_ids * 2}) assert res["code"] == 0 assert "Duplicate chunk ids" in res["data"]["errors"][0] assert res["data"]["success_count"] == 4 res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res assert len(res["data"]["chunks"]) == 1 assert res["data"]["total"] == 1 @pytest.mark.p3 def test_concurrent_deletion(self, HttpApiAuth, add_document): count = 100 dataset_id, document_id = add_document chunk_ids = batch_add_chunks(HttpApiAuth, dataset_id, document_id, count) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( delete_chunks, HttpApiAuth, dataset_id, document_id, {"chunk_ids": chunk_ids[i : i + 1]}, ) for i in range(count) ] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.p3 def test_delete_1k(self, HttpApiAuth, add_document): chunks_num = 1_000 dataset_id, document_id = add_document chunk_ids = batch_add_chunks(HttpApiAuth, dataset_id, document_id, chunks_num) # issues/6487 from time import sleep sleep(1) res = delete_chunks(HttpApiAuth, dataset_id, document_id, {"chunk_ids": chunk_ids}) assert res["code"] == 0 res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res assert len(res["data"]["chunks"]) == 0 assert res["data"]["total"] == 0 @pytest.mark.parametrize( "payload, expected_code, expected_message, remaining", [ pytest.param(None, 100, """TypeError("argument of type \'NoneType\' is not iterable")""", 5, marks=pytest.mark.skip), pytest.param({"chunk_ids": ["invalid_id"]}, 102, "rm_chunk deleted chunks 0, expect 1", 5, marks=pytest.mark.p3), pytest.param("not json", 100, """UnboundLocalError("local variable \'duplicate_messages\' referenced before assignment")""", 5, marks=pytest.mark.skip(reason="pull/6376")), pytest.param(lambda r: {"chunk_ids": r[:1]}, 0, "", 4, marks=pytest.mark.p3), pytest.param(lambda r: {"chunk_ids": r}, 0, "", 1, marks=pytest.mark.p1), pytest.param({"chunk_ids": []}, 0, "", 0, marks=pytest.mark.p3), ], ) def test_basic_scenarios( self, HttpApiAuth, add_chunks_func, payload, expected_code, expected_message, remaining, ): dataset_id, document_id, chunk_ids = add_chunks_func if callable(payload): payload = payload(chunk_ids) res = delete_chunks(HttpApiAuth, dataset_id, document_id, payload) assert res["code"] == expected_code if res["code"] != 0: assert res["message"] == expected_message res = list_chunks(HttpApiAuth, dataset_id, document_id) if res["code"] != 0: assert False, res assert len(res["data"]["chunks"]) == remaining assert res["data"]["total"] == remaining
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_chunk_management_within_dataset/test_list_chunks.py
test/testcases/test_http_api/test_chunk_management_within_dataset/test_list_chunks.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import os from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import batch_add_chunks, list_chunks from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = list_chunks(invalid_auth, "dataset_id", "document_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestChunksList: @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page": None, "page_size": 2}, 0, 2, ""), pytest.param({"page": 0, "page_size": 2}, 100, 0, "ValueError('Search does not support negative slicing.')", marks=pytest.mark.skip), ({"page": 2, "page_size": 2}, 0, 2, ""), ({"page": 3, "page_size": 2}, 0, 1, ""), ({"page": "3", "page_size": 2}, 0, 1, ""), pytest.param({"page": -1, "page_size": 2}, 100, 0, "ValueError('Search does not support negative slicing.')", marks=pytest.mark.skip), pytest.param({"page": "a", "page_size": 2}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip), ], ) def test_page(self, HttpApiAuth, add_chunks, params, expected_code, expected_page_size, expected_message): dataset_id, document_id, _ = add_chunks res = list_chunks(HttpApiAuth, dataset_id, document_id, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["chunks"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page_size": None}, 0, 5, ""), pytest.param({"page_size": 0}, 0, 5, ""), ({"page_size": 1}, 0, 1, ""), ({"page_size": 6}, 0, 5, ""), ({"page_size": "1"}, 0, 1, ""), pytest.param({"page_size": -1}, 0, 5, "", marks=pytest.mark.skip), pytest.param({"page_size": "a"}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip), ], ) def test_page_size(self, HttpApiAuth, add_chunks, params, expected_code, expected_page_size, expected_message): dataset_id, document_id, _ = add_chunks res = list_chunks(HttpApiAuth, dataset_id, document_id, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["chunks"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "params, expected_page_size", [ ({"keywords": None}, 5), ({"keywords": ""}, 5), ({"keywords": "1"}, 1), ({"keywords": "chunk"}, 4), pytest.param({"keywords": "ragflow"}, 1, marks=pytest.mark.skipif(os.getenv("DOC_ENGINE") == "infinity", reason="issues/6509")), pytest.param({"keywords": "ragflow"}, 5, marks=pytest.mark.skipif(os.getenv("DOC_ENGINE") != "infinity", reason="issues/6509")), ({"keywords": "unknown"}, 0), ], ) def test_keywords(self, HttpApiAuth, add_chunks, params, expected_page_size): dataset_id, document_id, _ = add_chunks res = list_chunks(HttpApiAuth, dataset_id, document_id, params=params) assert res["code"] == 0 assert len(res["data"]["chunks"]) == expected_page_size @pytest.mark.p1 @pytest.mark.parametrize( "chunk_id, expected_code, expected_page_size, expected_message", [ (None, 0, 5, ""), ("", 0, 5, ""), pytest.param(lambda r: r[0], 0, 1, "", marks=pytest.mark.skipif(os.getenv("DOC_ENGINE") == "infinity", reason="issues/6499")), pytest.param("unknown", 100, 0, """AttributeError("\'NoneType\' object has no attribute \'keys\'")""", marks=pytest.mark.skip), ], ) def test_id( self, HttpApiAuth, add_chunks, chunk_id, expected_code, expected_page_size, expected_message, ): dataset_id, document_id, chunk_ids = add_chunks if callable(chunk_id): params = {"id": chunk_id(chunk_ids)} else: params = {"id": chunk_id} res = list_chunks(HttpApiAuth, dataset_id, document_id, params=params) assert res["code"] == expected_code if expected_code == 0: if params["id"] in [None, ""]: assert len(res["data"]["chunks"]) == expected_page_size else: assert res["data"]["chunks"][0]["id"] == params["id"] else: assert res["message"] == expected_message @pytest.mark.p3 def test_invalid_params(self, HttpApiAuth, add_chunks): dataset_id, document_id, _ = add_chunks params = {"a": "b"} res = list_chunks(HttpApiAuth, dataset_id, document_id, params=params) assert res["code"] == 0 assert len(res["data"]["chunks"]) == 5 @pytest.mark.p3 def test_concurrent_list(self, HttpApiAuth, add_chunks): dataset_id, document_id, _ = add_chunks count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(list_chunks, HttpApiAuth, dataset_id, document_id) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(len(future.result()["data"]["chunks"]) == 5 for future in futures) @pytest.mark.p1 def test_default(self, HttpApiAuth, add_document): dataset_id, document_id = add_document res = list_chunks(HttpApiAuth, dataset_id, document_id) chunks_count = res["data"]["doc"]["chunk_count"] batch_add_chunks(HttpApiAuth, dataset_id, document_id, 31) # issues/6487 from time import sleep sleep(3) res = list_chunks(HttpApiAuth, dataset_id, document_id) assert res["code"] == 0 assert len(res["data"]["chunks"]) == 30 assert res["data"]["doc"]["chunk_count"] == chunks_count + 31 @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<NotFound '404: Not Found'>"), ( "invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id.", ), ], ) def test_invalid_dataset_id(self, HttpApiAuth, add_chunks, dataset_id, expected_code, expected_message): _, document_id, _ = add_chunks res = list_chunks(HttpApiAuth, dataset_id, document_id) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "document_id, expected_code, expected_message", [ ("", 102, "The dataset not own the document chunks."), ( "invalid_document_id", 102, "You don't own the document invalid_document_id.", ), ], ) def test_invalid_document_id(self, HttpApiAuth, add_chunks, document_id, expected_code, expected_message): dataset_id, _, _ = add_chunks res = list_chunks(HttpApiAuth, dataset_id, document_id) assert res["code"] == expected_code assert res["message"] == expected_message
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_session_management/test_list_sessions_with_chat_assistant.py
test/testcases/test_http_api/test_session_management/test_list_sessions_with_chat_assistant.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import delete_chat_assistants, list_session_with_chat_assistants from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth from utils import is_sorted @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = list_session_with_chat_assistants(invalid_auth, "chat_assistant_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestSessionsWithChatAssistantList: @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page": None, "page_size": 2}, 0, 2, ""), pytest.param({"page": 0, "page_size": 2}, 100, 0, "ValueError('Search does not support negative slicing.')", marks=pytest.mark.skip), ({"page": 2, "page_size": 2}, 0, 2, ""), ({"page": 3, "page_size": 2}, 0, 1, ""), ({"page": "3", "page_size": 2}, 0, 1, ""), pytest.param({"page": -1, "page_size": 2}, 100, 0, "ValueError('Search does not support negative slicing.')", marks=pytest.mark.skip), pytest.param({"page": "a", "page_size": 2}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip), ], ) def test_page(self, HttpApiAuth, add_sessions_with_chat_assistant, params, expected_code, expected_page_size, expected_message): chat_assistant_id, _ = add_sessions_with_chat_assistant res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page_size": None}, 0, 5, ""), ({"page_size": 0}, 0, 0, ""), ({"page_size": 1}, 0, 1, ""), ({"page_size": 6}, 0, 5, ""), ({"page_size": "1"}, 0, 1, ""), pytest.param({"page_size": -1}, 0, 5, "", marks=pytest.mark.skip), pytest.param({"page_size": "a"}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip), ], ) def test_page_size(self, HttpApiAuth, add_sessions_with_chat_assistant, params, expected_code, expected_page_size, expected_message): chat_assistant_id, _ = add_sessions_with_chat_assistant res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "params, expected_code, assertions, expected_message", [ ({"orderby": None}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"orderby": "create_time"}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"orderby": "update_time"}, 0, lambda r: (is_sorted(r["data"], "update_time", True)), ""), ({"orderby": "name", "desc": "False"}, 0, lambda r: (is_sorted(r["data"], "name", False)), ""), pytest.param({"orderby": "unknown"}, 102, 0, "orderby should be create_time or update_time", marks=pytest.mark.skip(reason="issues/")), ], ) def test_orderby( self, HttpApiAuth, add_sessions_with_chat_assistant, params, expected_code, assertions, expected_message, ): chat_assistant_id, _ = add_sessions_with_chat_assistant res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, params=params) assert res["code"] == expected_code if expected_code == 0: if callable(assertions): assert assertions(res) else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "params, expected_code, assertions, expected_message", [ ({"desc": None}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"desc": "true"}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"desc": "True"}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"desc": True}, 0, lambda r: (is_sorted(r["data"], "create_time", True)), ""), ({"desc": "false"}, 0, lambda r: (is_sorted(r["data"], "create_time", False)), ""), ({"desc": "False"}, 0, lambda r: (is_sorted(r["data"], "create_time", False)), ""), ({"desc": False}, 0, lambda r: (is_sorted(r["data"], "create_time", False)), ""), ({"desc": "False", "orderby": "update_time"}, 0, lambda r: (is_sorted(r["data"], "update_time", False)), ""), pytest.param({"desc": "unknown"}, 102, 0, "desc should be true or false", marks=pytest.mark.skip(reason="issues/")), ], ) def test_desc( self, HttpApiAuth, add_sessions_with_chat_assistant, params, expected_code, assertions, expected_message, ): chat_assistant_id, _ = add_sessions_with_chat_assistant res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, params=params) assert res["code"] == expected_code if expected_code == 0: if callable(assertions): assert assertions(res) else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_num, expected_message", [ ({"name": None}, 0, 5, ""), ({"name": ""}, 0, 5, ""), ({"name": "session_with_chat_assistant_1"}, 0, 1, ""), ({"name": "unknown"}, 0, 0, ""), ], ) def test_name(self, HttpApiAuth, add_sessions_with_chat_assistant, params, expected_code, expected_num, expected_message): chat_assistant_id, _ = add_sessions_with_chat_assistant res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, params=params) assert res["code"] == expected_code if expected_code == 0: if params["name"] == "session_with_chat_assistant_1": assert res["data"][0]["name"] == params["name"] else: assert len(res["data"]) == expected_num else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "session_id, expected_code, expected_num, expected_message", [ (None, 0, 5, ""), ("", 0, 5, ""), (lambda r: r[0], 0, 1, ""), ("unknown", 0, 0, "The chat doesn't exist"), ], ) def test_id(self, HttpApiAuth, add_sessions_with_chat_assistant, session_id, expected_code, expected_num, expected_message): chat_assistant_id, session_ids = add_sessions_with_chat_assistant if callable(session_id): params = {"id": session_id(session_ids)} else: params = {"id": session_id} res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, params=params) assert res["code"] == expected_code if expected_code == 0: if params["id"] == session_ids[0]: assert res["data"][0]["id"] == params["id"] else: assert len(res["data"]) == expected_num else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "session_id, name, expected_code, expected_num, expected_message", [ (lambda r: r[0], "session_with_chat_assistant_0", 0, 1, ""), (lambda r: r[0], "session_with_chat_assistant_100", 0, 0, ""), (lambda r: r[0], "unknown", 0, 0, ""), ("id", "session_with_chat_assistant_0", 0, 0, ""), ], ) def test_name_and_id(self, HttpApiAuth, add_sessions_with_chat_assistant, session_id, name, expected_code, expected_num, expected_message): chat_assistant_id, session_ids = add_sessions_with_chat_assistant if callable(session_id): params = {"id": session_id(session_ids), "name": name} else: params = {"id": session_id, "name": name} res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]) == expected_num else: assert res["message"] == expected_message @pytest.mark.p3 def test_concurrent_list(self, HttpApiAuth, add_sessions_with_chat_assistant): count = 100 chat_assistant_id, _ = add_sessions_with_chat_assistant with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(list_session_with_chat_assistants, HttpApiAuth, chat_assistant_id) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.p3 def test_invalid_params(self, HttpApiAuth, add_sessions_with_chat_assistant): chat_assistant_id, _ = add_sessions_with_chat_assistant params = {"a": "b"} res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, params=params) assert res["code"] == 0 assert len(res["data"]) == 5 @pytest.mark.p3 def test_list_chats_after_deleting_associated_chat_assistant(self, HttpApiAuth, add_sessions_with_chat_assistant): chat_assistant_id, _ = add_sessions_with_chat_assistant res = delete_chat_assistants(HttpApiAuth, {"ids": [chat_assistant_id]}) assert res["code"] == 0 res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id) assert res["code"] == 102 assert "You don't own the assistant" in res["message"]
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_session_management/test_delete_sessions_with_chat_assistant.py
test/testcases/test_http_api/test_session_management/test_delete_sessions_with_chat_assistant.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import batch_add_sessions_with_chat_assistant, delete_session_with_chat_assistants, list_session_with_chat_assistants from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = delete_session_with_chat_assistants(invalid_auth, "chat_assistant_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestSessionWithChatAssistantDelete: @pytest.mark.p3 @pytest.mark.parametrize( "chat_assistant_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_chat_assistant_id", 102, "You don't own the chat", ), ], ) def test_invalid_chat_assistant_id(self, HttpApiAuth, add_sessions_with_chat_assistant_func, chat_assistant_id, expected_code, expected_message): _, session_ids = add_sessions_with_chat_assistant_func res = delete_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, {"ids": session_ids}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.parametrize( "payload", [ pytest.param(lambda r: {"ids": ["invalid_id"] + r}, marks=pytest.mark.p3), pytest.param(lambda r: {"ids": r[:1] + ["invalid_id"] + r[1:5]}, marks=pytest.mark.p1), pytest.param(lambda r: {"ids": r + ["invalid_id"]}, marks=pytest.mark.p3), ], ) def test_delete_partial_invalid_id(self, HttpApiAuth, add_sessions_with_chat_assistant_func, payload): chat_assistant_id, session_ids = add_sessions_with_chat_assistant_func if callable(payload): payload = payload(session_ids) res = delete_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, payload) assert res["code"] == 0 assert res["data"]["errors"][0] == "The chat doesn't own the session invalid_id" res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id) if res["code"] != 0: assert False, res assert len(res["data"]) == 0 @pytest.mark.p3 def test_repeated_deletion(self, HttpApiAuth, add_sessions_with_chat_assistant_func): chat_assistant_id, session_ids = add_sessions_with_chat_assistant_func payload = {"ids": session_ids} res = delete_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, payload) assert res["code"] == 0 res = delete_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, payload) assert res["code"] == 102 assert "The chat doesn't own the session" in res["message"] @pytest.mark.p3 def test_duplicate_deletion(self, HttpApiAuth, add_sessions_with_chat_assistant_func): chat_assistant_id, session_ids = add_sessions_with_chat_assistant_func res = delete_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, {"ids": session_ids * 2}) assert res["code"] == 0 assert "Duplicate session ids" in res["data"]["errors"][0] assert res["data"]["success_count"] == 5 res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id) if res["code"] != 0: assert False, res assert len(res["data"]) == 0 @pytest.mark.p3 def test_concurrent_deletion(self, HttpApiAuth, add_chat_assistants): count = 100 _, _, chat_assistant_ids = add_chat_assistants session_ids = batch_add_sessions_with_chat_assistant(HttpApiAuth, chat_assistant_ids[0], count) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( delete_session_with_chat_assistants, HttpApiAuth, chat_assistant_ids[0], {"ids": session_ids[i : i + 1]}, ) for i in range(count) ] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.p3 def test_delete_1k(self, HttpApiAuth, add_chat_assistants): sessions_num = 1_000 _, _, chat_assistant_ids = add_chat_assistants session_ids = batch_add_sessions_with_chat_assistant(HttpApiAuth, chat_assistant_ids[0], sessions_num) res = delete_session_with_chat_assistants(HttpApiAuth, chat_assistant_ids[0], {"ids": session_ids}) assert res["code"] == 0 res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_ids[0]) if res["code"] != 0: assert False, res assert len(res["data"]) == 0 @pytest.mark.parametrize( "payload, expected_code, expected_message, remaining", [ pytest.param(None, 0, """TypeError("argument of type \'NoneType\' is not iterable")""", 0, marks=pytest.mark.skip), pytest.param({"ids": ["invalid_id"]}, 102, "The chat doesn't own the session invalid_id", 5, marks=pytest.mark.p3), pytest.param("not json", 100, """AttributeError("\'str\' object has no attribute \'get\'")""", 5, marks=pytest.mark.skip), pytest.param(lambda r: {"ids": r[:1]}, 0, "", 4, marks=pytest.mark.p3), pytest.param(lambda r: {"ids": r}, 0, "", 0, marks=pytest.mark.p1), pytest.param({"ids": []}, 0, "", 0, marks=pytest.mark.p3), ], ) def test_basic_scenarios( self, HttpApiAuth, add_sessions_with_chat_assistant_func, payload, expected_code, expected_message, remaining, ): chat_assistant_id, session_ids = add_sessions_with_chat_assistant_func if callable(payload): payload = payload(session_ids) res = delete_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, payload) assert res["code"] == expected_code if res["code"] != 0: assert res["message"] == expected_message res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id) if res["code"] != 0: assert False, res assert len(res["data"]) == remaining
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_session_management/conftest.py
test/testcases/test_http_api/test_session_management/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import batch_add_sessions_with_chat_assistant, delete_session_with_chat_assistants @pytest.fixture(scope="class") def add_sessions_with_chat_assistant(request, HttpApiAuth, add_chat_assistants): def cleanup(): for chat_assistant_id in chat_assistant_ids: delete_session_with_chat_assistants(HttpApiAuth, chat_assistant_id) request.addfinalizer(cleanup) _, _, chat_assistant_ids = add_chat_assistants return chat_assistant_ids[0], batch_add_sessions_with_chat_assistant(HttpApiAuth, chat_assistant_ids[0], 5) @pytest.fixture(scope="function") def add_sessions_with_chat_assistant_func(request, HttpApiAuth, add_chat_assistants): def cleanup(): for chat_assistant_id in chat_assistant_ids: delete_session_with_chat_assistants(HttpApiAuth, chat_assistant_id) request.addfinalizer(cleanup) _, _, chat_assistant_ids = add_chat_assistants return chat_assistant_ids[0], batch_add_sessions_with_chat_assistant(HttpApiAuth, chat_assistant_ids[0], 5)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_session_management/test_create_session_with_chat_assistant.py
test/testcases/test_http_api/test_session_management/test_create_session_with_chat_assistant.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import create_session_with_chat_assistant, delete_chat_assistants, list_session_with_chat_assistants from configs import INVALID_API_TOKEN, SESSION_WITH_CHAT_NAME_LIMIT from libs.auth import RAGFlowHttpApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = create_session_with_chat_assistant(invalid_auth, "chat_assistant_id") assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.usefixtures("clear_session_with_chat_assistants") class TestSessionWithChatAssistantCreate: @pytest.mark.p1 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"name": "valid_name"}, 0, ""), pytest.param({"name": "a" * (SESSION_WITH_CHAT_NAME_LIMIT + 1)}, 102, "", marks=pytest.mark.skip(reason="issues/")), pytest.param({"name": 1}, 100, "", marks=pytest.mark.skip(reason="issues/")), ({"name": ""}, 102, "`name` can not be empty."), ({"name": "duplicated_name"}, 0, ""), ({"name": "case insensitive"}, 0, ""), ], ) def test_name(self, HttpApiAuth, add_chat_assistants, payload, expected_code, expected_message): _, _, chat_assistant_ids = add_chat_assistants if payload["name"] == "duplicated_name": create_session_with_chat_assistant(HttpApiAuth, chat_assistant_ids[0], payload) elif payload["name"] == "case insensitive": create_session_with_chat_assistant(HttpApiAuth, chat_assistant_ids[0], {"name": payload["name"].upper()}) res = create_session_with_chat_assistant(HttpApiAuth, chat_assistant_ids[0], payload) assert res["code"] == expected_code, res if expected_code == 0: assert res["data"]["name"] == payload["name"] assert res["data"]["chat_id"] == chat_assistant_ids[0] else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "chat_assistant_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ("invalid_chat_assistant_id", 102, "You do not own the assistant."), ], ) def test_invalid_chat_assistant_id(self, HttpApiAuth, chat_assistant_id, expected_code, expected_message): res = create_session_with_chat_assistant(HttpApiAuth, chat_assistant_id, {"name": "valid_name"}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 def test_concurrent_create_session(self, HttpApiAuth, add_chat_assistants): count = 1000 _, _, chat_assistant_ids = add_chat_assistants res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_ids[0]) if res["code"] != 0: assert False, res sessions_count = len(res["data"]) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( create_session_with_chat_assistant, HttpApiAuth, chat_assistant_ids[0], {"name": f"session with chat assistant test {i}"}, ) for i in range(count) ] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_ids[0], {"page_size": count * 2}) if res["code"] != 0: assert False, res assert len(res["data"]) == sessions_count + count @pytest.mark.p3 def test_add_session_to_deleted_chat_assistant(self, HttpApiAuth, add_chat_assistants): _, _, chat_assistant_ids = add_chat_assistants res = delete_chat_assistants(HttpApiAuth, {"ids": [chat_assistant_ids[0]]}) assert res["code"] == 0 res = create_session_with_chat_assistant(HttpApiAuth, chat_assistant_ids[0], {"name": "valid_name"}) assert res["code"] == 102 assert res["message"] == "You do not own the assistant."
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_session_management/test_update_session_with_chat_assistant.py
test/testcases/test_http_api/test_session_management/test_update_session_with_chat_assistant.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed from random import randint import pytest from common import delete_chat_assistants, list_session_with_chat_assistants, update_session_with_chat_assistant from configs import INVALID_API_TOKEN, SESSION_WITH_CHAT_NAME_LIMIT from libs.auth import RAGFlowHttpApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = update_session_with_chat_assistant(invalid_auth, "chat_assistant_id", "session_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestSessionWithChatAssistantUpdate: @pytest.mark.parametrize( "payload, expected_code, expected_message", [ pytest.param({"name": "valid_name"}, 0, "", marks=pytest.mark.p1), pytest.param({"name": "a" * (SESSION_WITH_CHAT_NAME_LIMIT + 1)}, 102, "", marks=pytest.mark.skip(reason="issues/")), pytest.param({"name": 1}, 100, "", marks=pytest.mark.skip(reason="issues/")), pytest.param({"name": ""}, 102, "`name` can not be empty.", marks=pytest.mark.p3), pytest.param({"name": "duplicated_name"}, 0, "", marks=pytest.mark.p3), pytest.param({"name": "case insensitive"}, 0, "", marks=pytest.mark.p3), ], ) def test_name(self, HttpApiAuth, add_sessions_with_chat_assistant_func, payload, expected_code, expected_message): chat_assistant_id, session_ids = add_sessions_with_chat_assistant_func if payload["name"] == "duplicated_name": update_session_with_chat_assistant(HttpApiAuth, chat_assistant_id, session_ids[0], payload) elif payload["name"] == "case insensitive": update_session_with_chat_assistant(HttpApiAuth, chat_assistant_id, session_ids[0], {"name": payload["name"].upper()}) res = update_session_with_chat_assistant(HttpApiAuth, chat_assistant_id, session_ids[0], payload) assert res["code"] == expected_code, res if expected_code == 0: res = list_session_with_chat_assistants(HttpApiAuth, chat_assistant_id, {"id": session_ids[0]}) assert res["data"][0]["name"] == payload["name"] else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "chat_assistant_id, expected_code, expected_message", [ ("", 100, "<NotFound '404: Not Found'>"), pytest.param("invalid_chat_assistant_id", 102, "Session does not exist", marks=pytest.mark.skip(reason="issues/")), ], ) def test_invalid_chat_assistant_id(self, HttpApiAuth, add_sessions_with_chat_assistant_func, chat_assistant_id, expected_code, expected_message): _, session_ids = add_sessions_with_chat_assistant_func res = update_session_with_chat_assistant(HttpApiAuth, chat_assistant_id, session_ids[0], {"name": "valid_name"}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "session_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ("invalid_session_id", 102, "Session does not exist"), ], ) def test_invalid_session_id(self, HttpApiAuth, add_sessions_with_chat_assistant_func, session_id, expected_code, expected_message): chat_assistant_id, _ = add_sessions_with_chat_assistant_func res = update_session_with_chat_assistant(HttpApiAuth, chat_assistant_id, session_id, {"name": "valid_name"}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 def test_repeated_update_session(self, HttpApiAuth, add_sessions_with_chat_assistant_func): chat_assistant_id, session_ids = add_sessions_with_chat_assistant_func res = update_session_with_chat_assistant(HttpApiAuth, chat_assistant_id, session_ids[0], {"name": "valid_name_1"}) assert res["code"] == 0 res = update_session_with_chat_assistant(HttpApiAuth, chat_assistant_id, session_ids[0], {"name": "valid_name_2"}) assert res["code"] == 0 @pytest.mark.p3 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ pytest.param({"unknown_key": "unknown_value"}, 100, "ValueError", marks=pytest.mark.skip), ({}, 0, ""), pytest.param(None, 100, "TypeError", marks=pytest.mark.skip), ], ) def test_invalid_params(self, HttpApiAuth, add_sessions_with_chat_assistant_func, payload, expected_code, expected_message): chat_assistant_id, session_ids = add_sessions_with_chat_assistant_func res = update_session_with_chat_assistant(HttpApiAuth, chat_assistant_id, session_ids[0], payload) assert res["code"] == expected_code if expected_code != 0: assert expected_message in res["message"] @pytest.mark.p3 def test_concurrent_update_session(self, HttpApiAuth, add_sessions_with_chat_assistant_func): count = 50 chat_assistant_id, session_ids = add_sessions_with_chat_assistant_func with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( update_session_with_chat_assistant, HttpApiAuth, chat_assistant_id, session_ids[randint(0, 4)], {"name": f"update session test {i}"}, ) for i in range(count) ] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.p3 def test_update_session_to_deleted_chat_assistant(self, HttpApiAuth, add_sessions_with_chat_assistant_func): chat_assistant_id, session_ids = add_sessions_with_chat_assistant_func delete_chat_assistants(HttpApiAuth, {"ids": [chat_assistant_id]}) res = update_session_with_chat_assistant(HttpApiAuth, chat_assistant_id, session_ids[0], {"name": "valid_name"}) assert res["code"] == 102 assert res["message"] == "You do not own the session"
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_dataset_management/test_list_datasets.py
test/testcases/test_http_api/test_dataset_management/test_list_datasets.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import uuid from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import list_datasets from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth from utils import is_sorted class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = list_datasets(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestCapability: @pytest.mark.p3 def test_concurrent_list(self, HttpApiAuth): count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(list_datasets, HttpApiAuth) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.usefixtures("add_datasets") class TestDatasetsList: @pytest.mark.p1 def test_params_unset(self, HttpApiAuth): res = list_datasets(HttpApiAuth, None) assert res["code"] == 0, res assert len(res["data"]) == 5, res @pytest.mark.p2 def test_params_empty(self, HttpApiAuth): res = list_datasets(HttpApiAuth, {}) assert res["code"] == 0, res assert len(res["data"]) == 5, res @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_page_size", [ ({"page": 2, "page_size": 2}, 2), ({"page": 3, "page_size": 2}, 1), ({"page": 4, "page_size": 2}, 0), ({"page": "2", "page_size": 2}, 2), ({"page": 1, "page_size": 10}, 5), ], ids=["normal_middle_page", "normal_last_partial_page", "beyond_max_page", "string_page_number", "full_data_single_page"], ) def test_page(self, HttpApiAuth, params, expected_page_size): res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res assert len(res["data"]) == expected_page_size, res @pytest.mark.p2 @pytest.mark.parametrize( "params, expected_code, expected_message", [ ({"page": 0}, 101, "Input should be greater than or equal to 1"), ({"page": "a"}, 101, "Input should be a valid integer, unable to parse string as an integer"), ], ids=["page_0", "page_a"], ) def test_page_invalid(self, HttpApiAuth, params, expected_code, expected_message): res = list_datasets(HttpApiAuth, params=params) assert res["code"] == expected_code, res assert expected_message in res["message"], res @pytest.mark.p2 def test_page_none(self, HttpApiAuth): params = {"page": None} res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res assert len(res["data"]) == 5, res @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_page_size", [ ({"page_size": 1}, 1), ({"page_size": 3}, 3), ({"page_size": 5}, 5), ({"page_size": 6}, 5), ({"page_size": "1"}, 1), ], ids=["min_valid_page_size", "medium_page_size", "page_size_equals_total", "page_size_exceeds_total", "string_type_page_size"], ) def test_page_size(self, HttpApiAuth, params, expected_page_size): res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res assert len(res["data"]) == expected_page_size, res @pytest.mark.p2 @pytest.mark.parametrize( "params, expected_code, expected_message", [ ({"page_size": 0}, 101, "Input should be greater than or equal to 1"), ({"page_size": "a"}, 101, "Input should be a valid integer, unable to parse string as an integer"), ], ) def test_page_size_invalid(self, HttpApiAuth, params, expected_code, expected_message): res = list_datasets(HttpApiAuth, params) assert res["code"] == expected_code, res assert expected_message in res["message"], res @pytest.mark.p2 def test_page_size_none(self, HttpApiAuth): params = {"page_size": None} res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res assert len(res["data"]) == 5, res @pytest.mark.p2 @pytest.mark.parametrize( "params, assertions", [ ({"orderby": "create_time"}, lambda r: (is_sorted(r["data"], "create_time", True))), ({"orderby": "update_time"}, lambda r: (is_sorted(r["data"], "update_time", True))), ], ids=["orderby_create_time", "orderby_update_time"], ) def test_orderby(self, HttpApiAuth, params, assertions): res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res if callable(assertions): assert assertions(res), res @pytest.mark.p3 @pytest.mark.parametrize( "params", [ {"orderby": ""}, {"orderby": "unknown"}, {"orderby": "CREATE_TIME"}, {"orderby": "UPDATE_TIME"}, {"orderby": " create_time "}, ], ids=["empty", "unknown", "orderby_create_time_upper", "orderby_update_time_upper", "whitespace"], ) def test_orderby_invalid(self, HttpApiAuth, params): res = list_datasets(HttpApiAuth, params) assert res["code"] == 101, res assert "Input should be 'create_time' or 'update_time'" in res["message"], res @pytest.mark.p3 def test_orderby_none(self, HttpApiAuth): params = {"orderby": None} res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res assert is_sorted(res["data"], "create_time", True), res @pytest.mark.p2 @pytest.mark.parametrize( "params, assertions", [ ({"desc": True}, lambda r: (is_sorted(r["data"], "create_time", True))), ({"desc": False}, lambda r: (is_sorted(r["data"], "create_time", False))), ({"desc": "true"}, lambda r: (is_sorted(r["data"], "create_time", True))), ({"desc": "false"}, lambda r: (is_sorted(r["data"], "create_time", False))), ({"desc": 1}, lambda r: (is_sorted(r["data"], "create_time", True))), ({"desc": 0}, lambda r: (is_sorted(r["data"], "create_time", False))), ({"desc": "yes"}, lambda r: (is_sorted(r["data"], "create_time", True))), ({"desc": "no"}, lambda r: (is_sorted(r["data"], "create_time", False))), ({"desc": "y"}, lambda r: (is_sorted(r["data"], "create_time", True))), ({"desc": "n"}, lambda r: (is_sorted(r["data"], "create_time", False))), ], ids=["desc=True", "desc=False", "desc=true", "desc=false", "desc=1", "desc=0", "desc=yes", "desc=no", "desc=y", "desc=n"], ) def test_desc(self, HttpApiAuth, params, assertions): res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res if callable(assertions): assert assertions(res), res @pytest.mark.p3 @pytest.mark.parametrize( "params", [ {"desc": 3.14}, {"desc": "unknown"}, ], ids=["empty", "unknown"], ) def test_desc_invalid(self, HttpApiAuth, params): res = list_datasets(HttpApiAuth, params) assert res["code"] == 101, res assert "Input should be a valid boolean, unable to interpret input" in res["message"], res @pytest.mark.p3 def test_desc_none(self, HttpApiAuth): params = {"desc": None} res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res assert is_sorted(res["data"], "create_time", True), res @pytest.mark.p1 def test_name(self, HttpApiAuth): params = {"name": "dataset_1"} res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res assert len(res["data"]) == 1, res assert res["data"][0]["name"] == "dataset_1", res @pytest.mark.p2 def test_name_wrong(self, HttpApiAuth): params = {"name": "wrong name"} res = list_datasets(HttpApiAuth, params) assert res["code"] == 108, res assert "lacks permission for dataset" in res["message"], res @pytest.mark.p2 def test_name_empty(self, HttpApiAuth): params = {"name": ""} res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res assert len(res["data"]) == 5, res @pytest.mark.p2 def test_name_none(self, HttpApiAuth): params = {"name": None} res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res assert len(res["data"]) == 5, res @pytest.mark.p1 def test_id(self, HttpApiAuth, add_datasets): dataset_ids = add_datasets params = {"id": dataset_ids[0]} res = list_datasets(HttpApiAuth, params) assert res["code"] == 0 assert len(res["data"]) == 1 assert res["data"][0]["id"] == dataset_ids[0] @pytest.mark.p2 def test_id_not_uuid(self, HttpApiAuth): params = {"id": "not_uuid"} res = list_datasets(HttpApiAuth, params) assert res["code"] == 101, res assert "Invalid UUID1 format" in res["message"], res @pytest.mark.p2 def test_id_not_uuid1(self, HttpApiAuth): params = {"id": uuid.uuid4().hex} res = list_datasets(HttpApiAuth, params) assert res["code"] == 101, res assert "Invalid UUID1 format" in res["message"], res @pytest.mark.p2 def test_id_wrong_uuid(self, HttpApiAuth): params = {"id": "d94a8dc02c9711f0930f7fbc369eab6d"} res = list_datasets(HttpApiAuth, params) assert res["code"] == 108, res assert "lacks permission for dataset" in res["message"], res @pytest.mark.p2 def test_id_empty(self, HttpApiAuth): params = {"id": ""} res = list_datasets(HttpApiAuth, params) assert res["code"] == 101, res assert "Invalid UUID1 format" in res["message"], res @pytest.mark.p2 def test_id_none(self, HttpApiAuth): params = {"id": None} res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res assert len(res["data"]) == 5, res @pytest.mark.p2 @pytest.mark.parametrize( "func, name, expected_num", [ (lambda r: r[0], "dataset_0", 1), (lambda r: r[0], "dataset_1", 0), ], ids=["name_and_id_match", "name_and_id_mismatch"], ) def test_name_and_id(self, HttpApiAuth, add_datasets, func, name, expected_num): dataset_ids = add_datasets if callable(func): params = {"id": func(dataset_ids), "name": name} res = list_datasets(HttpApiAuth, params) assert res["code"] == 0, res assert len(res["data"]) == expected_num, res @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, name", [ (lambda r: r[0], "wrong_name"), (uuid.uuid1().hex, "dataset_0"), ], ids=["name", "id"], ) def test_name_and_id_wrong(self, HttpApiAuth, add_datasets, dataset_id, name): dataset_ids = add_datasets if callable(dataset_id): params = {"id": dataset_id(dataset_ids), "name": name} else: params = {"id": dataset_id, "name": name} res = list_datasets(HttpApiAuth, params) assert res["code"] == 108, res assert "lacks permission for dataset" in res["message"], res @pytest.mark.p2 def test_field_unsupported(self, HttpApiAuth): params = {"unknown_field": "unknown_field"} res = list_datasets(HttpApiAuth, params) assert res["code"] == 101, res assert "Extra inputs are not permitted" in res["message"], res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_dataset_management/test_delete_datasets.py
test/testcases/test_http_api/test_dataset_management/test_delete_datasets.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import uuid from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import ( batch_create_datasets, delete_datasets, list_datasets, ) from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = delete_datasets(invalid_auth) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestRquest: @pytest.mark.p3 def test_content_type_bad(self, HttpApiAuth): BAD_CONTENT_TYPE = "text/xml" res = delete_datasets(HttpApiAuth, headers={"Content-Type": BAD_CONTENT_TYPE}) assert res["code"] == 101, res assert res["message"] == f"Unsupported content type: Expected application/json, got {BAD_CONTENT_TYPE}", res @pytest.mark.p3 @pytest.mark.parametrize( "payload, expected_message", [ ("a", "Malformed JSON syntax: Missing commas/brackets or invalid encoding"), ('"a"', "Invalid request payload: expected object, got str"), ], ids=["malformed_json_syntax", "invalid_request_payload_type"], ) def test_payload_bad(self, HttpApiAuth, payload, expected_message): res = delete_datasets(HttpApiAuth, data=payload) assert res["code"] == 101, res assert res["message"] == expected_message, res @pytest.mark.p3 def test_payload_unset(self, HttpApiAuth): res = delete_datasets(HttpApiAuth, None) assert res["code"] == 101, res assert res["message"] == "Malformed JSON syntax: Missing commas/brackets or invalid encoding", res class TestCapability: @pytest.mark.p3 def test_delete_dataset_1k(self, HttpApiAuth): ids = batch_create_datasets(HttpApiAuth, 1_000) res = delete_datasets(HttpApiAuth, {"ids": ids}) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert len(res["data"]) == 0, res @pytest.mark.p3 def test_concurrent_deletion(self, HttpApiAuth): count = 1_000 ids = batch_create_datasets(HttpApiAuth, count) with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(delete_datasets, HttpApiAuth, {"ids": ids[i : i + 1]}) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) class TestDatasetsDelete: @pytest.mark.p1 @pytest.mark.parametrize( "func, expected_code, remaining", [ (lambda r: {"ids": r[:1]}, 0, 2), (lambda r: {"ids": r}, 0, 0), ], ids=["single_dataset", "multiple_datasets"], ) def test_ids(self, HttpApiAuth, add_datasets_func, func, expected_code, remaining): dataset_ids = add_datasets_func if callable(func): payload = func(dataset_ids) res = delete_datasets(HttpApiAuth, payload) assert res["code"] == expected_code, res res = list_datasets(HttpApiAuth) assert len(res["data"]) == remaining, res @pytest.mark.p1 @pytest.mark.usefixtures("add_dataset_func") def test_ids_empty(self, HttpApiAuth): payload = {"ids": []} res = delete_datasets(HttpApiAuth, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert len(res["data"]) == 1, res @pytest.mark.p1 @pytest.mark.usefixtures("add_datasets_func") def test_ids_none(self, HttpApiAuth): payload = {"ids": None} res = delete_datasets(HttpApiAuth, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert len(res["data"]) == 0, res @pytest.mark.p2 @pytest.mark.usefixtures("add_dataset_func") def test_id_not_uuid(self, HttpApiAuth): payload = {"ids": ["not_uuid"]} res = delete_datasets(HttpApiAuth, payload) assert res["code"] == 101, res assert "Invalid UUID1 format" in res["message"], res res = list_datasets(HttpApiAuth) assert len(res["data"]) == 1, res @pytest.mark.p3 @pytest.mark.usefixtures("add_dataset_func") def test_id_not_uuid1(self, HttpApiAuth): payload = {"ids": [uuid.uuid4().hex]} res = delete_datasets(HttpApiAuth, payload) assert res["code"] == 101, res assert "Invalid UUID1 format" in res["message"], res @pytest.mark.p2 @pytest.mark.usefixtures("add_dataset_func") def test_id_wrong_uuid(self, HttpApiAuth): payload = {"ids": ["d94a8dc02c9711f0930f7fbc369eab6d"]} res = delete_datasets(HttpApiAuth, payload) assert res["code"] == 108, res assert "lacks permission for dataset" in res["message"], res res = list_datasets(HttpApiAuth) assert len(res["data"]) == 1, res @pytest.mark.p2 @pytest.mark.parametrize( "func", [ lambda r: {"ids": ["d94a8dc02c9711f0930f7fbc369eab6d"] + r}, lambda r: {"ids": r[:1] + ["d94a8dc02c9711f0930f7fbc369eab6d"] + r[1:3]}, lambda r: {"ids": r + ["d94a8dc02c9711f0930f7fbc369eab6d"]}, ], ) def test_ids_partial_invalid(self, HttpApiAuth, add_datasets_func, func): dataset_ids = add_datasets_func if callable(func): payload = func(dataset_ids) res = delete_datasets(HttpApiAuth, payload) assert res["code"] == 108, res assert "lacks permission for dataset" in res["message"], res res = list_datasets(HttpApiAuth) assert len(res["data"]) == 3, res @pytest.mark.p2 def test_ids_duplicate(self, HttpApiAuth, add_datasets_func): dataset_ids = add_datasets_func payload = {"ids": dataset_ids + dataset_ids} res = delete_datasets(HttpApiAuth, payload) assert res["code"] == 101, res assert "Duplicate ids:" in res["message"], res res = list_datasets(HttpApiAuth) assert len(res["data"]) == 3, res @pytest.mark.p2 def test_repeated_delete(self, HttpApiAuth, add_datasets_func): dataset_ids = add_datasets_func payload = {"ids": dataset_ids} res = delete_datasets(HttpApiAuth, payload) assert res["code"] == 0, res res = delete_datasets(HttpApiAuth, payload) assert res["code"] == 108, res assert "lacks permission for dataset" in res["message"], res @pytest.mark.p2 @pytest.mark.usefixtures("add_dataset_func") def test_field_unsupported(self, HttpApiAuth): payload = {"unknown_field": "unknown_field"} res = delete_datasets(HttpApiAuth, payload) assert res["code"] == 101, res assert "Extra inputs are not permitted" in res["message"], res res = list_datasets(HttpApiAuth) assert len(res["data"]) == 1, res
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_dataset_management/test_create_dataset.py
test/testcases/test_http_api/test_dataset_management/test_create_dataset.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from configs import DATASET_NAME_LIMIT, DEFAULT_PARSER_CONFIG, INVALID_API_TOKEN from hypothesis import example, given, settings from libs.auth import RAGFlowHttpApiAuth from utils import encode_avatar from utils.file_utils import create_image_file from utils.hypothesis_utils import valid_names from common import create_dataset @pytest.mark.usefixtures("clear_datasets") class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ids=["empty_auth", "invalid_api_token"], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = create_dataset(invalid_auth, {"name": "auth_test"}) assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestRquest: @pytest.mark.p3 def test_content_type_bad(self, HttpApiAuth): BAD_CONTENT_TYPE = "text/xml" res = create_dataset(HttpApiAuth, {"name": "bad_content_type"}, headers={"Content-Type": BAD_CONTENT_TYPE}) assert res["code"] == 101, res assert res["message"] == f"Unsupported content type: Expected application/json, got {BAD_CONTENT_TYPE}", res @pytest.mark.p3 @pytest.mark.parametrize( "payload, expected_message", [ ("a", "Malformed JSON syntax: Missing commas/brackets or invalid encoding"), ('"a"', "Invalid request payload: expected object, got str"), ], ids=["malformed_json_syntax", "invalid_request_payload_type"], ) def test_payload_bad(self, HttpApiAuth, payload, expected_message): res = create_dataset(HttpApiAuth, data=payload) assert res["code"] == 101, res assert res["message"] == expected_message, res @pytest.mark.usefixtures("clear_datasets") class TestCapability: @pytest.mark.p3 def test_create_dataset_1k(self, HttpApiAuth): for i in range(1_000): payload = {"name": f"dataset_{i}"} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, f"Failed to create dataset {i}" @pytest.mark.p3 def test_create_dataset_concurrent(self, HttpApiAuth): count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(create_dataset, HttpApiAuth, {"name": f"dataset_{i}"}) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.usefixtures("clear_datasets") class TestDatasetCreate: @pytest.mark.p1 @given(name=valid_names()) @example("a" * 128) @settings(max_examples=20) def test_name(self, HttpApiAuth, name): res = create_dataset(HttpApiAuth, {"name": name}) assert res["code"] == 0, res assert res["data"]["name"] == name, res @pytest.mark.p2 @pytest.mark.parametrize( "name, expected_message", [ ("", "String should have at least 1 character"), (" ", "String should have at least 1 character"), ("a" * (DATASET_NAME_LIMIT + 1), "String should have at most 128 characters"), (0, "Input should be a valid string"), (None, "Input should be a valid string"), ], ids=["empty_name", "space_name", "too_long_name", "invalid_name", "None_name"], ) def test_name_invalid(self, HttpApiAuth, name, expected_message): payload = {"name": name} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101, res assert expected_message in res["message"], res @pytest.mark.p3 def test_name_duplicated(self, HttpApiAuth): name = "duplicated_name" payload = {"name": name} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["name"] == name + "(1)", res @pytest.mark.p3 def test_name_case_insensitive(self, HttpApiAuth): name = "CaseInsensitive" payload = {"name": name.upper()} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res payload = {"name": name.lower()} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["name"] == name.lower() + "(1)", res @pytest.mark.p2 def test_avatar(self, HttpApiAuth, tmp_path): fn = create_image_file(tmp_path / "ragflow_test.png") payload = { "name": "avatar", "avatar": f"data:image/png;base64,{encode_avatar(fn)}", } res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res @pytest.mark.p2 def test_avatar_exceeds_limit_length(self, HttpApiAuth): payload = {"name": "avatar_exceeds_limit_length", "avatar": "a" * 65536} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101, res assert "String should have at most 65535 characters" in res["message"], res @pytest.mark.p3 @pytest.mark.parametrize( "name, prefix, expected_message", [ ("empty_prefix", "", "Missing MIME prefix. Expected format: data:<mime>;base64,<data>"), ("missing_comma", "data:image/png;base64", "Missing MIME prefix. Expected format: data:<mime>;base64,<data>"), ("unsupported_mine_type", "invalid_mine_prefix:image/png;base64,", "Invalid MIME prefix format. Must start with 'data:'"), ("invalid_mine_type", "data:unsupported_mine_type;base64,", "Unsupported MIME type. Allowed: ['image/jpeg', 'image/png']"), ], ids=["empty_prefix", "missing_comma", "unsupported_mine_type", "invalid_mine_type"], ) def test_avatar_invalid_prefix(self, HttpApiAuth, tmp_path, name, prefix, expected_message): fn = create_image_file(tmp_path / "ragflow_test.png") payload = { "name": name, "avatar": f"{prefix}{encode_avatar(fn)}", } res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101, res assert expected_message in res["message"], res @pytest.mark.p3 def test_avatar_unset(self, HttpApiAuth): payload = {"name": "avatar_unset"} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["avatar"] is None, res @pytest.mark.p3 def test_avatar_none(self, HttpApiAuth): payload = {"name": "avatar_none", "avatar": None} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["avatar"] is None, res @pytest.mark.p2 def test_description(self, HttpApiAuth): payload = {"name": "description", "description": "description"} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["description"] == "description", res @pytest.mark.p2 def test_description_exceeds_limit_length(self, HttpApiAuth): payload = {"name": "description_exceeds_limit_length", "description": "a" * 65536} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101, res assert "String should have at most 65535 characters" in res["message"], res @pytest.mark.p3 def test_description_unset(self, HttpApiAuth): payload = {"name": "description_unset"} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["description"] is None, res @pytest.mark.p3 def test_description_none(self, HttpApiAuth): payload = {"name": "description_none", "description": None} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["description"] is None, res @pytest.mark.p1 @pytest.mark.parametrize( "name, embedding_model", [ ("BAAI/bge-small-en-v1.5@Builtin", "BAAI/bge-small-en-v1.5@Builtin"), ("embedding-3@ZHIPU-AI", "embedding-3@ZHIPU-AI"), ], ids=["builtin_baai", "tenant_zhipu"], ) def test_embedding_model(self, HttpApiAuth, name, embedding_model): payload = {"name": name, "embedding_model": embedding_model} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["embedding_model"] == embedding_model, res @pytest.mark.p2 @pytest.mark.parametrize( "name, embedding_model", [ ("unknown_llm_name", "unknown@ZHIPU-AI"), ("unknown_llm_factory", "embedding-3@unknown"), ("tenant_no_auth_default_tenant_llm", "text-embedding-v3@Tongyi-Qianwen"), ("tenant_no_auth", "text-embedding-3-small@OpenAI"), ], ids=["unknown_llm_name", "unknown_llm_factory", "tenant_no_auth_default_tenant_llm", "tenant_no_auth"], ) def test_embedding_model_invalid(self, HttpApiAuth, name, embedding_model): payload = {"name": name, "embedding_model": embedding_model} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101, res if "tenant_no_auth" in name: assert res["message"] == f"Unauthorized model: <{embedding_model}>", res else: assert res["message"] == f"Unsupported model: <{embedding_model}>", res @pytest.mark.p2 @pytest.mark.parametrize( "name, embedding_model", [ ("empty", ""), ("space", " "), ("missing_at", "BAAI/bge-small-en-v1.5Builtin"), ("missing_model_name", "@Builtin"), ("missing_provider", "BAAI/bge-small-en-v1.5@"), ("whitespace_only_model_name", " @Builtin"), ("whitespace_only_provider", "BAAI/bge-small-env1.5@ "), ], ids=["empty", "space", "missing_at", "empty_model_name", "empty_provider", "whitespace_only_model_name", "whitespace_only_provider"], ) def test_embedding_model_format(self, HttpApiAuth, name, embedding_model): payload = {"name": name, "embedding_model": embedding_model} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101, res if name in ["empty", "space", "missing_at"]: assert "Embedding model identifier must follow <model_name>@<provider> format" in res["message"], res else: assert "Both model_name and provider must be non-empty strings" in res["message"], res @pytest.mark.p2 def test_embedding_model_unset(self, HttpApiAuth): payload = {"name": "embedding_model_unset"} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["embedding_model"] == "BAAI/bge-small-en-v1.5@Builtin", res @pytest.mark.p2 def test_embedding_model_none(self, HttpApiAuth): payload = {"name": "embedding_model_none", "embedding_model": None} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["embedding_model"] == "BAAI/bge-small-en-v1.5@Builtin", res @pytest.mark.p1 @pytest.mark.parametrize( "name, permission", [ ("me", "me"), ("team", "team"), ], ids=["me", "team"], ) def test_permission(self, HttpApiAuth, name, permission): payload = {"name": name, "permission": permission} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["permission"] == permission.lower().strip(), res @pytest.mark.p2 @pytest.mark.parametrize( "name, permission", [ ("empty", ""), ("unknown", "unknown"), ("type_error", list()), ("me_upercase", "ME"), ("team_upercase", "TEAM"), ("whitespace", " ME "), ], ids=["empty", "unknown", "type_error", "me_upercase", "team_upercase", "whitespace"], ) def test_permission_invalid(self, HttpApiAuth, name, permission): payload = {"name": name, "permission": permission} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101 assert "Input should be 'me' or 'team'" in res["message"] @pytest.mark.p2 def test_permission_unset(self, HttpApiAuth): payload = {"name": "permission_unset"} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["permission"] == "me", res @pytest.mark.p3 def test_permission_none(self, HttpApiAuth): payload = {"name": "permission_none", "permission": None} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101, res assert "Input should be 'me' or 'team'" in res["message"], res @pytest.mark.p1 @pytest.mark.parametrize( "name, chunk_method", [ ("naive", "naive"), ("book", "book"), ("email", "email"), ("laws", "laws"), ("manual", "manual"), ("one", "one"), ("paper", "paper"), ("picture", "picture"), ("presentation", "presentation"), ("qa", "qa"), ("table", "table"), ("tag", "tag"), ], ids=["naive", "book", "email", "laws", "manual", "one", "paper", "picture", "presentation", "qa", "table", "tag"], ) def test_chunk_method(self, HttpApiAuth, name, chunk_method): payload = {"name": name, "chunk_method": chunk_method} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["chunk_method"] == chunk_method, res @pytest.mark.p2 @pytest.mark.parametrize( "name, chunk_method", [ ("empty", ""), ("unknown", "unknown"), ("type_error", list()), ], ids=["empty", "unknown", "type_error"], ) def test_chunk_method_invalid(self, HttpApiAuth, name, chunk_method): payload = {"name": name, "chunk_method": chunk_method} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101, res assert "Input should be 'naive', 'book', 'email', 'laws', 'manual', 'one', 'paper', 'picture', 'presentation', 'qa', 'table' or 'tag'" in res["message"], res @pytest.mark.p2 def test_chunk_method_unset(self, HttpApiAuth): payload = {"name": "chunk_method_unset"} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["chunk_method"] == "naive", res @pytest.mark.p3 def test_chunk_method_none(self, HttpApiAuth): payload = {"name": "chunk_method_none", "chunk_method": None} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101, res assert "Input should be 'naive', 'book', 'email', 'laws', 'manual', 'one', 'paper', 'picture', 'presentation', 'qa', 'table' or 'tag'" in res["message"], res @pytest.mark.p1 @pytest.mark.parametrize( "name, parser_config", [ ("auto_keywords_min", {"auto_keywords": 0}), ("auto_keywords_mid", {"auto_keywords": 16}), ("auto_keywords_max", {"auto_keywords": 32}), ("auto_questions_min", {"auto_questions": 0}), ("auto_questions_mid", {"auto_questions": 5}), ("auto_questions_max", {"auto_questions": 10}), ("chunk_token_num_min", {"chunk_token_num": 1}), ("chunk_token_num_mid", {"chunk_token_num": 1024}), ("chunk_token_num_max", {"chunk_token_num": 2048}), ("delimiter", {"delimiter": "\n"}), ("delimiter_space", {"delimiter": " "}), ("html4excel_true", {"html4excel": True}), ("html4excel_false", {"html4excel": False}), ("layout_recognize_DeepDOC", {"layout_recognize": "DeepDOC"}), ("layout_recognize_navie", {"layout_recognize": "Plain Text"}), ("tag_kb_ids", {"tag_kb_ids": ["1", "2"]}), ("topn_tags_min", {"topn_tags": 1}), ("topn_tags_mid", {"topn_tags": 5}), ("topn_tags_max", {"topn_tags": 10}), ("filename_embd_weight_min", {"filename_embd_weight": 0.1}), ("filename_embd_weight_mid", {"filename_embd_weight": 0.5}), ("filename_embd_weight_max", {"filename_embd_weight": 1.0}), ("task_page_size_min", {"task_page_size": 1}), ("task_page_size_None", {"task_page_size": None}), ("pages", {"pages": [[1, 100]]}), ("pages_none", {"pages": None}), ("graphrag_true", {"graphrag": {"use_graphrag": True}}), ("graphrag_false", {"graphrag": {"use_graphrag": False}}), ("graphrag_entity_types", {"graphrag": {"entity_types": ["age", "sex", "height", "weight"]}}), ("graphrag_method_general", {"graphrag": {"method": "general"}}), ("graphrag_method_light", {"graphrag": {"method": "light"}}), ("graphrag_community_true", {"graphrag": {"community": True}}), ("graphrag_community_false", {"graphrag": {"community": False}}), ("graphrag_resolution_true", {"graphrag": {"resolution": True}}), ("graphrag_resolution_false", {"graphrag": {"resolution": False}}), ("raptor_true", {"raptor": {"use_raptor": True}}), ("raptor_false", {"raptor": {"use_raptor": False}}), ("raptor_prompt", {"raptor": {"prompt": "Who are you?"}}), ("raptor_max_token_min", {"raptor": {"max_token": 1}}), ("raptor_max_token_mid", {"raptor": {"max_token": 1024}}), ("raptor_max_token_max", {"raptor": {"max_token": 2048}}), ("raptor_threshold_min", {"raptor": {"threshold": 0.0}}), ("raptor_threshold_mid", {"raptor": {"threshold": 0.5}}), ("raptor_threshold_max", {"raptor": {"threshold": 1.0}}), ("raptor_max_cluster_min", {"raptor": {"max_cluster": 1}}), ("raptor_max_cluster_mid", {"raptor": {"max_cluster": 512}}), ("raptor_max_cluster_max", {"raptor": {"max_cluster": 1024}}), ("raptor_random_seed_min", {"raptor": {"random_seed": 0}}), ], ids=[ "auto_keywords_min", "auto_keywords_mid", "auto_keywords_max", "auto_questions_min", "auto_questions_mid", "auto_questions_max", "chunk_token_num_min", "chunk_token_num_mid", "chunk_token_num_max", "delimiter", "delimiter_space", "html4excel_true", "html4excel_false", "layout_recognize_DeepDOC", "layout_recognize_navie", "tag_kb_ids", "topn_tags_min", "topn_tags_mid", "topn_tags_max", "filename_embd_weight_min", "filename_embd_weight_mid", "filename_embd_weight_max", "task_page_size_min", "task_page_size_None", "pages", "pages_none", "graphrag_true", "graphrag_false", "graphrag_entity_types", "graphrag_method_general", "graphrag_method_light", "graphrag_community_true", "graphrag_community_false", "graphrag_resolution_true", "graphrag_resolution_false", "raptor_true", "raptor_false", "raptor_prompt", "raptor_max_token_min", "raptor_max_token_mid", "raptor_max_token_max", "raptor_threshold_min", "raptor_threshold_mid", "raptor_threshold_max", "raptor_max_cluster_min", "raptor_max_cluster_mid", "raptor_max_cluster_max", "raptor_random_seed_min", ], ) def test_parser_config(self, HttpApiAuth, name, parser_config): payload = {"name": name, "parser_config": parser_config} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res for k, v in parser_config.items(): if isinstance(v, dict): for kk, vv in v.items(): assert res["data"]["parser_config"][k][kk] == vv, res else: assert res["data"]["parser_config"][k] == v, res @pytest.mark.p2 @pytest.mark.parametrize( "name, parser_config, expected_message", [ ("auto_keywords_min_limit", {"auto_keywords": -1}, "Input should be greater than or equal to 0"), ("auto_keywords_max_limit", {"auto_keywords": 33}, "Input should be less than or equal to 32"), ("auto_keywords_float_not_allowed", {"auto_keywords": 3.14}, "Input should be a valid integer"), ("auto_keywords_type_invalid", {"auto_keywords": "string"}, "Input should be a valid integer"), ("auto_questions_min_limit", {"auto_questions": -1}, "Input should be greater than or equal to 0"), ("auto_questions_max_limit", {"auto_questions": 11}, "Input should be less than or equal to 10"), ("auto_questions_float_not_allowed", {"auto_questions": 3.14}, "Input should be a valid integer"), ("auto_questions_type_invalid", {"auto_questions": "string"}, "Input should be a valid integer"), ("chunk_token_num_min_limit", {"chunk_token_num": 0}, "Input should be greater than or equal to 1"), ("chunk_token_num_max_limit", {"chunk_token_num": 2049}, "Input should be less than or equal to 2048"), ("chunk_token_num_float_not_allowed", {"chunk_token_num": 3.14}, "Input should be a valid integer"), ("chunk_token_num_type_invalid", {"chunk_token_num": "string"}, "Input should be a valid integer"), ("delimiter_empty", {"delimiter": ""}, "String should have at least 1 character"), ("html4excel_type_invalid", {"html4excel": "string"}, "Input should be a valid boolean"), ("tag_kb_ids_not_list", {"tag_kb_ids": "1,2"}, "Input should be a valid list"), ("tag_kb_ids_int_in_list", {"tag_kb_ids": [1, 2]}, "Input should be a valid string"), ("topn_tags_min_limit", {"topn_tags": 0}, "Input should be greater than or equal to 1"), ("topn_tags_max_limit", {"topn_tags": 11}, "Input should be less than or equal to 10"), ("topn_tags_float_not_allowed", {"topn_tags": 3.14}, "Input should be a valid integer"), ("topn_tags_type_invalid", {"topn_tags": "string"}, "Input should be a valid integer"), ("filename_embd_weight_min_limit", {"filename_embd_weight": -1}, "Input should be greater than or equal to 0"), ("filename_embd_weight_max_limit", {"filename_embd_weight": 1.1}, "Input should be less than or equal to 1"), ("filename_embd_weight_type_invalid", {"filename_embd_weight": "string"}, "Input should be a valid number"), ("task_page_size_min_limit", {"task_page_size": 0}, "Input should be greater than or equal to 1"), ("task_page_size_float_not_allowed", {"task_page_size": 3.14}, "Input should be a valid integer"), ("task_page_size_type_invalid", {"task_page_size": "string"}, "Input should be a valid integer"), ("pages_not_list", {"pages": "1,2"}, "Input should be a valid list"), ("pages_not_list_in_list", {"pages": ["1,2"]}, "Input should be a valid list"), ("pages_not_int_list", {"pages": [["string1", "string2"]]}, "Input should be a valid integer"), ("graphrag_type_invalid", {"graphrag": {"use_graphrag": "string"}}, "Input should be a valid boolean"), ("graphrag_entity_types_not_list", {"graphrag": {"entity_types": "1,2"}}, "Input should be a valid list"), ("graphrag_entity_types_not_str_in_list", {"graphrag": {"entity_types": [1, 2]}}, "nput should be a valid string"), ("graphrag_method_unknown", {"graphrag": {"method": "unknown"}}, "Input should be 'light' or 'general'"), ("graphrag_method_none", {"graphrag": {"method": None}}, "Input should be 'light' or 'general'"), ("graphrag_community_type_invalid", {"graphrag": {"community": "string"}}, "Input should be a valid boolean"), ("graphrag_resolution_type_invalid", {"graphrag": {"resolution": "string"}}, "Input should be a valid boolean"), ("raptor_type_invalid", {"raptor": {"use_raptor": "string"}}, "Input should be a valid boolean"), ("raptor_prompt_empty", {"raptor": {"prompt": ""}}, "String should have at least 1 character"), ("raptor_prompt_space", {"raptor": {"prompt": " "}}, "String should have at least 1 character"), ("raptor_max_token_min_limit", {"raptor": {"max_token": 0}}, "Input should be greater than or equal to 1"), ("raptor_max_token_max_limit", {"raptor": {"max_token": 2049}}, "Input should be less than or equal to 2048"), ("raptor_max_token_float_not_allowed", {"raptor": {"max_token": 3.14}}, "Input should be a valid integer"), ("raptor_max_token_type_invalid", {"raptor": {"max_token": "string"}}, "Input should be a valid integer"), ("raptor_threshold_min_limit", {"raptor": {"threshold": -0.1}}, "Input should be greater than or equal to 0"), ("raptor_threshold_max_limit", {"raptor": {"threshold": 1.1}}, "Input should be less than or equal to 1"), ("raptor_threshold_type_invalid", {"raptor": {"threshold": "string"}}, "Input should be a valid number"), ("raptor_max_cluster_min_limit", {"raptor": {"max_cluster": 0}}, "Input should be greater than or equal to 1"), ("raptor_max_cluster_max_limit", {"raptor": {"max_cluster": 1025}}, "Input should be less than or equal to 1024"), ("raptor_max_cluster_float_not_allowed", {"raptor": {"max_cluster": 3.14}}, "Input should be a valid integer"), ("raptor_max_cluster_type_invalid", {"raptor": {"max_cluster": "string"}}, "Input should be a valid integer"), ("raptor_random_seed_min_limit", {"raptor": {"random_seed": -1}}, "Input should be greater than or equal to 0"), ("raptor_random_seed_float_not_allowed", {"raptor": {"random_seed": 3.14}}, "Input should be a valid integer"), ("raptor_random_seed_type_invalid", {"raptor": {"random_seed": "string"}}, "Input should be a valid integer"), ("parser_config_type_invalid", {"delimiter": "a" * 65536}, "Parser config exceeds size limit (max 65,535 characters)"), ], ids=[ "auto_keywords_min_limit", "auto_keywords_max_limit", "auto_keywords_float_not_allowed", "auto_keywords_type_invalid", "auto_questions_min_limit", "auto_questions_max_limit", "auto_questions_float_not_allowed", "auto_questions_type_invalid", "chunk_token_num_min_limit", "chunk_token_num_max_limit", "chunk_token_num_float_not_allowed", "chunk_token_num_type_invalid", "delimiter_empty", "html4excel_type_invalid", "tag_kb_ids_not_list", "tag_kb_ids_int_in_list", "topn_tags_min_limit", "topn_tags_max_limit", "topn_tags_float_not_allowed", "topn_tags_type_invalid", "filename_embd_weight_min_limit", "filename_embd_weight_max_limit", "filename_embd_weight_type_invalid", "task_page_size_min_limit", "task_page_size_float_not_allowed", "task_page_size_type_invalid", "pages_not_list", "pages_not_list_in_list", "pages_not_int_list", "graphrag_type_invalid", "graphrag_entity_types_not_list", "graphrag_entity_types_not_str_in_list", "graphrag_method_unknown", "graphrag_method_none", "graphrag_community_type_invalid", "graphrag_resolution_type_invalid", "raptor_type_invalid", "raptor_prompt_empty", "raptor_prompt_space", "raptor_max_token_min_limit", "raptor_max_token_max_limit", "raptor_max_token_float_not_allowed", "raptor_max_token_type_invalid", "raptor_threshold_min_limit", "raptor_threshold_max_limit", "raptor_threshold_type_invalid", "raptor_max_cluster_min_limit", "raptor_max_cluster_max_limit", "raptor_max_cluster_float_not_allowed", "raptor_max_cluster_type_invalid", "raptor_random_seed_min_limit", "raptor_random_seed_float_not_allowed", "raptor_random_seed_type_invalid", "parser_config_type_invalid", ], ) def test_parser_config_invalid(self, HttpApiAuth, name, parser_config, expected_message): payload = {"name": name, "parser_config": parser_config} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101, res assert expected_message in res["message"], res @pytest.mark.p2 def test_parser_config_empty(self, HttpApiAuth): payload = {"name": "parser_config_empty", "parser_config": {}} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["parser_config"] == DEFAULT_PARSER_CONFIG, res @pytest.mark.p2 def test_parser_config_unset(self, HttpApiAuth): payload = {"name": "parser_config_unset"} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["parser_config"] == DEFAULT_PARSER_CONFIG, res @pytest.mark.p3 def test_parser_config_none(self, HttpApiAuth): payload = {"name": "parser_config_none", "parser_config": None} res = create_dataset(HttpApiAuth, payload) assert res["code"] == 0, res assert res["data"]["parser_config"] == DEFAULT_PARSER_CONFIG, res @pytest.mark.p2 @pytest.mark.parametrize( "payload", [ {"name": "id", "id": "id"}, {"name": "tenant_id", "tenant_id": "e57c1966f99211efb41e9e45646e0111"}, {"name": "created_by", "created_by": "created_by"}, {"name": "create_date", "create_date": "Tue, 11 Mar 2025 13:37:23 GMT"}, {"name": "create_time", "create_time": 1741671443322}, {"name": "update_date", "update_date": "Tue, 11 Mar 2025 13:37:23 GMT"}, {"name": "update_time", "update_time": 1741671443339}, {"name": "document_count", "document_count": 1}, {"name": "chunk_count", "chunk_count": 1}, {"name": "token_num", "token_num": 1}, {"name": "status", "status": "1"}, {"name": "pagerank", "pagerank": 50}, {"name": "unknown_field", "unknown_field": "unknown_field"}, ], ) def test_unsupported_field(self, HttpApiAuth, payload): res = create_dataset(HttpApiAuth, payload) assert res["code"] == 101, res assert "Extra inputs are not permitted" in res["message"], res @pytest.mark.usefixtures("clear_datasets") class TestParserConfigBugFix: @pytest.mark.p1 def test_parser_config_missing_raptor_and_graphrag(self, HttpApiAuth):
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_dataset_management/conftest.py
test/testcases/test_http_api/test_dataset_management/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import batch_create_datasets, delete_datasets @pytest.fixture(scope="class") def add_datasets(HttpApiAuth, request): def cleanup(): delete_datasets(HttpApiAuth, {"ids": None}) request.addfinalizer(cleanup) return batch_create_datasets(HttpApiAuth, 5) @pytest.fixture(scope="function") def add_datasets_func(HttpApiAuth, request): def cleanup(): delete_datasets(HttpApiAuth, {"ids": None}) request.addfinalizer(cleanup) return batch_create_datasets(HttpApiAuth, 3)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_dataset_management/test_update_dataset.py
test/testcases/test_http_api/test_dataset_management/test_update_dataset.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import os import uuid from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import list_datasets, update_dataset from configs import DATASET_NAME_LIMIT, INVALID_API_TOKEN from hypothesis import HealthCheck, example, given, settings from libs.auth import RAGFlowHttpApiAuth from utils import encode_avatar from utils.file_utils import create_image_file from utils.hypothesis_utils import valid_names from configs import DEFAULT_PARSER_CONFIG # TODO: Missing scenario for updating embedding_model with chunk_count != 0 class TestAuthorization: @pytest.mark.p1 @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ids=["empty_auth", "invalid_api_token"], ) def test_auth_invalid(self, invalid_auth, expected_code, expected_message): res = update_dataset(invalid_auth, "dataset_id") assert res["code"] == expected_code, res assert res["message"] == expected_message, res class TestRquest: @pytest.mark.p3 def test_bad_content_type(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func BAD_CONTENT_TYPE = "text/xml" res = update_dataset(HttpApiAuth, dataset_id, {"name": "bad_content_type"}, headers={"Content-Type": BAD_CONTENT_TYPE}) assert res["code"] == 101, res assert res["message"] == f"Unsupported content type: Expected application/json, got {BAD_CONTENT_TYPE}", res @pytest.mark.p3 @pytest.mark.parametrize( "payload, expected_message", [ ("a", "Malformed JSON syntax: Missing commas/brackets or invalid encoding"), ('"a"', "Invalid request payload: expected object, got str"), ], ids=["malformed_json_syntax", "invalid_request_payload_type"], ) def test_payload_bad(self, HttpApiAuth, add_dataset_func, payload, expected_message): dataset_id = add_dataset_func res = update_dataset(HttpApiAuth, dataset_id, data=payload) assert res["code"] == 101, res assert res["message"] == expected_message, res @pytest.mark.p2 def test_payload_empty(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func res = update_dataset(HttpApiAuth, dataset_id, {}) assert res["code"] == 101, res assert res["message"] == "No properties were modified", res @pytest.mark.p3 def test_payload_unset(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func res = update_dataset(HttpApiAuth, dataset_id, None) assert res["code"] == 101, res assert res["message"] == "Malformed JSON syntax: Missing commas/brackets or invalid encoding", res class TestCapability: @pytest.mark.p3 def test_update_dateset_concurrent(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(update_dataset, HttpApiAuth, dataset_id, {"name": f"dataset_{i}"}) for i in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) class TestDatasetUpdate: @pytest.mark.p3 def test_dataset_id_not_uuid(self, HttpApiAuth): payload = {"name": "not uuid"} res = update_dataset(HttpApiAuth, "not_uuid", payload) assert res["code"] == 101, res assert "Invalid UUID1 format" in res["message"], res @pytest.mark.p3 def test_dataset_id_not_uuid1(self, HttpApiAuth): payload = {"name": "not uuid1"} res = update_dataset(HttpApiAuth, uuid.uuid4().hex, payload) assert res["code"] == 101, res assert "Invalid UUID1 format" in res["message"], res @pytest.mark.p3 def test_dataset_id_wrong_uuid(self, HttpApiAuth): payload = {"name": "wrong uuid"} res = update_dataset(HttpApiAuth, "d94a8dc02c9711f0930f7fbc369eab6d", payload) assert res["code"] == 108, res assert "lacks permission for dataset" in res["message"], res @pytest.mark.p1 @given(name=valid_names()) @example("a" * 128) @settings(max_examples=20, suppress_health_check=[HealthCheck.function_scoped_fixture]) def test_name(self, HttpApiAuth, add_dataset_func, name): dataset_id = add_dataset_func payload = {"name": name} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert res["code"] == 0, res assert res["data"][0]["name"] == name, res @pytest.mark.p2 @pytest.mark.parametrize( "name, expected_message", [ ("", "String should have at least 1 character"), (" ", "String should have at least 1 character"), ("a" * (DATASET_NAME_LIMIT + 1), "String should have at most 128 characters"), (0, "Input should be a valid string"), (None, "Input should be a valid string"), ], ids=["empty_name", "space_name", "too_long_name", "invalid_name", "None_name"], ) def test_name_invalid(self, HttpApiAuth, add_dataset_func, name, expected_message): dataset_id = add_dataset_func payload = {"name": name} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res assert expected_message in res["message"], res @pytest.mark.p3 def test_name_duplicated(self, HttpApiAuth, add_datasets_func): dataset_id = add_datasets_func[0] name = "dataset_1" payload = {"name": name} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 102, res assert res["message"] == f"Dataset name '{name}' already exists", res @pytest.mark.p3 def test_name_case_insensitive(self, HttpApiAuth, add_datasets_func): dataset_id = add_datasets_func[0] name = "DATASET_1" payload = {"name": name} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 102, res assert res["message"] == f"Dataset name '{name}' already exists", res @pytest.mark.p2 def test_avatar(self, HttpApiAuth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fn = create_image_file(tmp_path / "ragflow_test.png") payload = { "avatar": f"data:image/png;base64,{encode_avatar(fn)}", } res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert res["code"] == 0, res assert res["data"][0]["avatar"] == f"data:image/png;base64,{encode_avatar(fn)}", res @pytest.mark.p2 def test_avatar_exceeds_limit_length(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"avatar": "a" * 65536} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res assert "String should have at most 65535 characters" in res["message"], res @pytest.mark.p3 @pytest.mark.parametrize( "avatar_prefix, expected_message", [ ("", "Missing MIME prefix. Expected format: data:<mime>;base64,<data>"), ("data:image/png;base64", "Missing MIME prefix. Expected format: data:<mime>;base64,<data>"), ("invalid_mine_prefix:image/png;base64,", "Invalid MIME prefix format. Must start with 'data:'"), ("data:unsupported_mine_type;base64,", "Unsupported MIME type. Allowed: ['image/jpeg', 'image/png']"), ], ids=["empty_prefix", "missing_comma", "unsupported_mine_type", "invalid_mine_type"], ) def test_avatar_invalid_prefix(self, HttpApiAuth, add_dataset_func, tmp_path, avatar_prefix, expected_message): dataset_id = add_dataset_func fn = create_image_file(tmp_path / "ragflow_test.png") payload = {"avatar": f"{avatar_prefix}{encode_avatar(fn)}"} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res assert expected_message in res["message"], res @pytest.mark.p3 def test_avatar_none(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"avatar": None} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert res["code"] == 0, res assert res["data"][0]["avatar"] is None, res @pytest.mark.p2 def test_description(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"description": "description"} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0 res = list_datasets(HttpApiAuth, {"id": dataset_id}) assert res["code"] == 0, res assert res["data"][0]["description"] == "description" @pytest.mark.p2 def test_description_exceeds_limit_length(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"description": "a" * 65536} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res assert "String should have at most 65535 characters" in res["message"], res @pytest.mark.p3 def test_description_none(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"description": None} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth, {"id": dataset_id}) assert res["code"] == 0, res assert res["data"][0]["description"] is None @pytest.mark.p1 @pytest.mark.parametrize( "embedding_model", [ "BAAI/bge-small-en-v1.5@Builtin", "embedding-3@ZHIPU-AI", ], ids=["builtin_baai", "tenant_zhipu"], ) def test_embedding_model(self, HttpApiAuth, add_dataset_func, embedding_model): dataset_id = add_dataset_func payload = {"embedding_model": embedding_model} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert res["code"] == 0, res assert res["data"][0]["embedding_model"] == embedding_model, res @pytest.mark.p2 @pytest.mark.parametrize( "name, embedding_model", [ ("unknown_llm_name", "unknown@ZHIPU-AI"), ("unknown_llm_factory", "embedding-3@unknown"), ("tenant_no_auth_default_tenant_llm", "text-embedding-v3@Tongyi-Qianwen"), ("tenant_no_auth", "text-embedding-3-small@OpenAI"), ], ids=["unknown_llm_name", "unknown_llm_factory", "tenant_no_auth_default_tenant_llm", "tenant_no_auth"], ) def test_embedding_model_invalid(self, HttpApiAuth, add_dataset_func, name, embedding_model): dataset_id = add_dataset_func payload = {"name": name, "embedding_model": embedding_model} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res if "tenant_no_auth" in name: assert res["message"] == f"Unauthorized model: <{embedding_model}>", res else: assert res["message"] == f"Unsupported model: <{embedding_model}>", res @pytest.mark.p2 @pytest.mark.parametrize( "name, embedding_model", [ ("empty", ""), ("space", " "), ("missing_at", "BAAI/bge-small-en-v1.5Builtin"), ("missing_model_name", "@Builtin"), ("missing_provider", "BAAI/bge-small-en-v1.5@"), ("whitespace_only_model_name", " @Builtin"), ("whitespace_only_provider", "BAAI/bge-small-en-v1.5@ "), ], ids=["empty", "space", "missing_at", "empty_model_name", "empty_provider", "whitespace_only_model_name", "whitespace_only_provider"], ) def test_embedding_model_format(self, HttpApiAuth, add_dataset_func, name, embedding_model): dataset_id = add_dataset_func payload = {"name": name, "embedding_model": embedding_model} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res if name in ["empty", "space", "missing_at"]: assert "Embedding model identifier must follow <model_name>@<provider> format" in res["message"], res else: assert "Both model_name and provider must be non-empty strings" in res["message"], res @pytest.mark.p2 def test_embedding_model_none(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"embedding_model": None} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert res["code"] == 0, res assert res["data"][0]["embedding_model"] == "BAAI/bge-small-en-v1.5@Builtin", res @pytest.mark.p1 @pytest.mark.parametrize( "permission", [ "me", "team", ], ids=["me", "team"], ) def test_permission(self, HttpApiAuth, add_dataset_func, permission): dataset_id = add_dataset_func payload = {"permission": permission} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert res["code"] == 0, res assert res["data"][0]["permission"] == permission.lower().strip(), res @pytest.mark.p2 @pytest.mark.parametrize( "permission", [ "", "unknown", list(), "ME", "TEAM", " ME ", ], ids=["empty", "unknown", "type_error", "me_upercase", "team_upercase", "whitespace"], ) def test_permission_invalid(self, HttpApiAuth, add_dataset_func, permission): dataset_id = add_dataset_func payload = {"permission": permission} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101 assert "Input should be 'me' or 'team'" in res["message"] @pytest.mark.p3 def test_permission_none(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"permission": None} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res assert "Input should be 'me' or 'team'" in res["message"], res @pytest.mark.p1 @pytest.mark.parametrize( "chunk_method", [ "naive", "book", "email", "laws", "manual", "one", "paper", "picture", "presentation", "qa", "table", "tag", ], ids=["naive", "book", "email", "laws", "manual", "one", "paper", "picture", "presentation", "qa", "table", "tag"], ) def test_chunk_method(self, HttpApiAuth, add_dataset_func, chunk_method): dataset_id = add_dataset_func payload = {"chunk_method": chunk_method} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert res["code"] == 0, res assert res["data"][0]["chunk_method"] == chunk_method, res @pytest.mark.p2 @pytest.mark.parametrize( "chunk_method", [ "", "unknown", list(), ], ids=["empty", "unknown", "type_error"], ) def test_chunk_method_invalid(self, HttpApiAuth, add_dataset_func, chunk_method): dataset_id = add_dataset_func payload = {"chunk_method": chunk_method} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res assert "Input should be 'naive', 'book', 'email', 'laws', 'manual', 'one', 'paper', 'picture', 'presentation', 'qa', 'table' or 'tag'" in res["message"], res @pytest.mark.p3 def test_chunk_method_none(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"chunk_method": None} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res assert "Input should be 'naive', 'book', 'email', 'laws', 'manual', 'one', 'paper', 'picture', 'presentation', 'qa', 'table' or 'tag'" in res["message"], res @pytest.mark.skipif(os.getenv("DOC_ENGINE") == "infinity", reason="#8208") @pytest.mark.p2 @pytest.mark.parametrize("pagerank", [0, 50, 100], ids=["min", "mid", "max"]) def test_pagerank(self, HttpApiAuth, add_dataset_func, pagerank): dataset_id = add_dataset_func payload = {"pagerank": pagerank} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0 res = list_datasets(HttpApiAuth, {"id": dataset_id}) assert res["code"] == 0, res assert res["data"][0]["pagerank"] == pagerank @pytest.mark.skipif(os.getenv("DOC_ENGINE") == "infinity", reason="#8208") @pytest.mark.p2 def test_pagerank_set_to_0(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"pagerank": 50} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth, {"id": dataset_id}) assert res["code"] == 0, res assert res["data"][0]["pagerank"] == 50, res payload = {"pagerank": 0} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0 res = list_datasets(HttpApiAuth, {"id": dataset_id}) assert res["code"] == 0, res assert res["data"][0]["pagerank"] == 0, res @pytest.mark.skipif(os.getenv("DOC_ENGINE") != "infinity", reason="#8208") @pytest.mark.p2 def test_pagerank_infinity(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"pagerank": 50} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res assert res["message"] == "'pagerank' can only be set when doc_engine is elasticsearch", res @pytest.mark.p2 @pytest.mark.parametrize( "pagerank, expected_message", [ (-1, "Input should be greater than or equal to 0"), (101, "Input should be less than or equal to 100"), ], ids=["min_limit", "max_limit"], ) def test_pagerank_invalid(self, HttpApiAuth, add_dataset_func, pagerank, expected_message): dataset_id = add_dataset_func payload = {"pagerank": pagerank} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res assert expected_message in res["message"], res @pytest.mark.p3 def test_pagerank_none(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"pagerank": None} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res assert "Input should be a valid integer" in res["message"], res @pytest.mark.p1 @pytest.mark.parametrize( "parser_config", [ {"auto_keywords": 0}, {"auto_keywords": 16}, {"auto_keywords": 32}, {"auto_questions": 0}, {"auto_questions": 5}, {"auto_questions": 10}, {"chunk_token_num": 1}, {"chunk_token_num": 1024}, {"chunk_token_num": 2048}, {"delimiter": "\n"}, {"delimiter": " "}, {"html4excel": True}, {"html4excel": False}, {"layout_recognize": "DeepDOC"}, {"layout_recognize": "Plain Text"}, {"tag_kb_ids": ["1", "2"]}, {"topn_tags": 1}, {"topn_tags": 5}, {"topn_tags": 10}, {"filename_embd_weight": 0.1}, {"filename_embd_weight": 0.5}, {"filename_embd_weight": 1.0}, {"task_page_size": 1}, {"task_page_size": None}, {"pages": [[1, 100]]}, {"pages": None}, {"graphrag": {"use_graphrag": True}}, {"graphrag": {"use_graphrag": False}}, {"graphrag": {"entity_types": ["age", "sex", "height", "weight"]}}, {"graphrag": {"method": "general"}}, {"graphrag": {"method": "light"}}, {"graphrag": {"community": True}}, {"graphrag": {"community": False}}, {"graphrag": {"resolution": True}}, {"graphrag": {"resolution": False}}, {"raptor": {"use_raptor": True}}, {"raptor": {"use_raptor": False}}, {"raptor": {"prompt": "Who are you?"}}, {"raptor": {"max_token": 1}}, {"raptor": {"max_token": 1024}}, {"raptor": {"max_token": 2048}}, {"raptor": {"threshold": 0.0}}, {"raptor": {"threshold": 0.5}}, {"raptor": {"threshold": 1.0}}, {"raptor": {"max_cluster": 1}}, {"raptor": {"max_cluster": 512}}, {"raptor": {"max_cluster": 1024}}, {"raptor": {"random_seed": 0}}, ], ids=[ "auto_keywords_min", "auto_keywords_mid", "auto_keywords_max", "auto_questions_min", "auto_questions_mid", "auto_questions_max", "chunk_token_num_min", "chunk_token_num_mid", "chunk_token_num_max", "delimiter", "delimiter_space", "html4excel_true", "html4excel_false", "layout_recognize_DeepDOC", "layout_recognize_navie", "tag_kb_ids", "topn_tags_min", "topn_tags_mid", "topn_tags_max", "filename_embd_weight_min", "filename_embd_weight_mid", "filename_embd_weight_max", "task_page_size_min", "task_page_size_None", "pages", "pages_none", "graphrag_true", "graphrag_false", "graphrag_entity_types", "graphrag_method_general", "graphrag_method_light", "graphrag_community_true", "graphrag_community_false", "graphrag_resolution_true", "graphrag_resolution_false", "raptor_true", "raptor_false", "raptor_prompt", "raptor_max_token_min", "raptor_max_token_mid", "raptor_max_token_max", "raptor_threshold_min", "raptor_threshold_mid", "raptor_threshold_max", "raptor_max_cluster_min", "raptor_max_cluster_mid", "raptor_max_cluster_max", "raptor_random_seed_min", ], ) def test_parser_config(self, HttpApiAuth, add_dataset_func, parser_config): dataset_id = add_dataset_func payload = {"parser_config": parser_config} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert res["code"] == 0, res for k, v in parser_config.items(): if isinstance(v, dict): for kk, vv in v.items(): assert res["data"][0]["parser_config"][k][kk] == vv, res else: assert res["data"][0]["parser_config"][k] == v, res @pytest.mark.p2 @pytest.mark.parametrize( "parser_config, expected_message", [ ({"auto_keywords": -1}, "Input should be greater than or equal to 0"), ({"auto_keywords": 33}, "Input should be less than or equal to 32"), ({"auto_keywords": 3.14}, "Input should be a valid integer"), ({"auto_keywords": "string"}, "Input should be a valid integer"), ({"auto_questions": -1}, "Input should be greater than or equal to 0"), ({"auto_questions": 11}, "Input should be less than or equal to 10"), ({"auto_questions": 3.14}, "Input should be a valid integer"), ({"auto_questions": "string"}, "Input should be a valid integer"), ({"chunk_token_num": 0}, "Input should be greater than or equal to 1"), ({"chunk_token_num": 2049}, "Input should be less than or equal to 2048"), ({"chunk_token_num": 3.14}, "Input should be a valid integer"), ({"chunk_token_num": "string"}, "Input should be a valid integer"), ({"delimiter": ""}, "String should have at least 1 character"), ({"html4excel": "string"}, "Input should be a valid boolean"), ({"tag_kb_ids": "1,2"}, "Input should be a valid list"), ({"tag_kb_ids": [1, 2]}, "Input should be a valid string"), ({"topn_tags": 0}, "Input should be greater than or equal to 1"), ({"topn_tags": 11}, "Input should be less than or equal to 10"), ({"topn_tags": 3.14}, "Input should be a valid integer"), ({"topn_tags": "string"}, "Input should be a valid integer"), ({"filename_embd_weight": -1}, "Input should be greater than or equal to 0"), ({"filename_embd_weight": 1.1}, "Input should be less than or equal to 1"), ({"filename_embd_weight": "string"}, "Input should be a valid number"), ({"task_page_size": 0}, "Input should be greater than or equal to 1"), ({"task_page_size": 3.14}, "Input should be a valid integer"), ({"task_page_size": "string"}, "Input should be a valid integer"), ({"pages": "1,2"}, "Input should be a valid list"), ({"pages": ["1,2"]}, "Input should be a valid list"), ({"pages": [["string1", "string2"]]}, "Input should be a valid integer"), ({"graphrag": {"use_graphrag": "string"}}, "Input should be a valid boolean"), ({"graphrag": {"entity_types": "1,2"}}, "Input should be a valid list"), ({"graphrag": {"entity_types": [1, 2]}}, "nput should be a valid string"), ({"graphrag": {"method": "unknown"}}, "Input should be 'light' or 'general'"), ({"graphrag": {"method": None}}, "Input should be 'light' or 'general'"), ({"graphrag": {"community": "string"}}, "Input should be a valid boolean"), ({"graphrag": {"resolution": "string"}}, "Input should be a valid boolean"), ({"raptor": {"use_raptor": "string"}}, "Input should be a valid boolean"), ({"raptor": {"prompt": ""}}, "String should have at least 1 character"), ({"raptor": {"prompt": " "}}, "String should have at least 1 character"), ({"raptor": {"max_token": 0}}, "Input should be greater than or equal to 1"), ({"raptor": {"max_token": 2049}}, "Input should be less than or equal to 2048"), ({"raptor": {"max_token": 3.14}}, "Input should be a valid integer"), ({"raptor": {"max_token": "string"}}, "Input should be a valid integer"), ({"raptor": {"threshold": -0.1}}, "Input should be greater than or equal to 0"), ({"raptor": {"threshold": 1.1}}, "Input should be less than or equal to 1"), ({"raptor": {"threshold": "string"}}, "Input should be a valid number"), ({"raptor": {"max_cluster": 0}}, "Input should be greater than or equal to 1"), ({"raptor": {"max_cluster": 1025}}, "Input should be less than or equal to 1024"), ({"raptor": {"max_cluster": 3.14}}, "Input should be a valid integer"), ({"raptor": {"max_cluster": "string"}}, "Input should be a valid integer"), ({"raptor": {"random_seed": -1}}, "Input should be greater than or equal to 0"), ({"raptor": {"random_seed": 3.14}}, "Input should be a valid integer"), ({"raptor": {"random_seed": "string"}}, "Input should be a valid integer"), ({"delimiter": "a" * 65536}, "Parser config exceeds size limit (max 65,535 characters)"), ], ids=[ "auto_keywords_min_limit", "auto_keywords_max_limit", "auto_keywords_float_not_allowed", "auto_keywords_type_invalid", "auto_questions_min_limit", "auto_questions_max_limit", "auto_questions_float_not_allowed", "auto_questions_type_invalid", "chunk_token_num_min_limit", "chunk_token_num_max_limit", "chunk_token_num_float_not_allowed", "chunk_token_num_type_invalid", "delimiter_empty", "html4excel_type_invalid", "tag_kb_ids_not_list", "tag_kb_ids_int_in_list", "topn_tags_min_limit", "topn_tags_max_limit", "topn_tags_float_not_allowed", "topn_tags_type_invalid", "filename_embd_weight_min_limit", "filename_embd_weight_max_limit", "filename_embd_weight_type_invalid", "task_page_size_min_limit", "task_page_size_float_not_allowed", "task_page_size_type_invalid", "pages_not_list", "pages_not_list_in_list", "pages_not_int_list", "graphrag_type_invalid", "graphrag_entity_types_not_list", "graphrag_entity_types_not_str_in_list", "graphrag_method_unknown", "graphrag_method_none", "graphrag_community_type_invalid", "graphrag_resolution_type_invalid", "raptor_type_invalid", "raptor_prompt_empty", "raptor_prompt_space", "raptor_max_token_min_limit", "raptor_max_token_max_limit", "raptor_max_token_float_not_allowed", "raptor_max_token_type_invalid", "raptor_threshold_min_limit", "raptor_threshold_max_limit", "raptor_threshold_type_invalid", "raptor_max_cluster_min_limit", "raptor_max_cluster_max_limit", "raptor_max_cluster_float_not_allowed", "raptor_max_cluster_type_invalid", "raptor_random_seed_min_limit", "raptor_random_seed_float_not_allowed", "raptor_random_seed_type_invalid", "parser_config_type_invalid", ], ) def test_parser_config_invalid(self, HttpApiAuth, add_dataset_func, parser_config, expected_message): dataset_id = add_dataset_func payload = {"parser_config": parser_config} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 101, res assert expected_message in res["message"], res @pytest.mark.p2 def test_parser_config_empty(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"parser_config": {}} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth) assert res["code"] == 0, res assert res["data"][0]["parser_config"] == DEFAULT_PARSER_CONFIG, res @pytest.mark.p3 def test_parser_config_none(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func payload = {"parser_config": None} res = update_dataset(HttpApiAuth, dataset_id, payload) assert res["code"] == 0, res res = list_datasets(HttpApiAuth, {"id": dataset_id})
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_file_management_within_dataset/test_stop_parse_documents.py
test/testcases/test_http_api/test_file_management_within_dataset/test_stop_parse_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor from time import sleep import pytest from common import bulk_upload_documents, list_documents, parse_documents, stop_parse_documents from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth from utils import wait_for def validate_document_parse_done(auth, dataset_id, document_ids): for document_id in document_ids: res = list_documents(auth, dataset_id, params={"id": document_id}) doc = res["data"]["docs"][0] assert doc["run"] == "DONE" assert len(doc["process_begin_at"]) > 0 assert doc["process_duration"] > 0 assert doc["progress"] > 0 assert "Task done" in doc["progress_msg"] def validate_document_parse_cancel(auth, dataset_id, document_ids): for document_id in document_ids: res = list_documents(auth, dataset_id, params={"id": document_id}) doc = res["data"]["docs"][0] assert doc["run"] == "CANCEL" assert len(doc["process_begin_at"]) > 0 assert doc["progress"] == 0.0 @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = stop_parse_documents(invalid_auth, "dataset_id") assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.skip class TestDocumentsParseStop: @pytest.mark.parametrize( "payload, expected_code, expected_message", [ pytest.param(None, 102, """AttributeError("\'NoneType\' object has no attribute \'get\'")""", marks=pytest.mark.skip), pytest.param({"document_ids": []}, 102, "`document_ids` is required", marks=pytest.mark.p1), pytest.param({"document_ids": ["invalid_id"]}, 102, "You don't own the document invalid_id.", marks=pytest.mark.p3), pytest.param({"document_ids": ["\n!?。;!?\"'"]}, 102, """You don\'t own the document \n!?。;!?"\'.""", marks=pytest.mark.p3), pytest.param("not json", 102, "AttributeError(\"'str' object has no attribute 'get'\")", marks=pytest.mark.skip), pytest.param(lambda r: {"document_ids": r[:1]}, 0, "", marks=pytest.mark.p1), pytest.param(lambda r: {"document_ids": r}, 0, "", marks=pytest.mark.p1), ], ) def test_basic_scenarios(self, HttpApiAuth, add_documents_func, payload, expected_code, expected_message): @wait_for(10, 1, "Document parsing timeout") def condition(_auth, _dataset_id, _document_ids): for _document_id in _document_ids: res = list_documents(_auth, _dataset_id, {"id": _document_id}) if res["data"]["docs"][0]["run"] != "DONE": return False return True dataset_id, document_ids = add_documents_func parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) if callable(payload): payload = payload(document_ids) res = stop_parse_documents(HttpApiAuth, dataset_id, payload) assert res["code"] == expected_code if expected_code != 0: assert res["message"] == expected_message else: completed_document_ids = list(set(document_ids) - set(payload["document_ids"])) condition(HttpApiAuth, dataset_id, completed_document_ids) validate_document_parse_cancel(HttpApiAuth, dataset_id, payload["document_ids"]) validate_document_parse_done(HttpApiAuth, dataset_id, completed_document_ids) @pytest.mark.p3 @pytest.mark.parametrize( "invalid_dataset_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id.", ), ], ) def test_invalid_dataset_id( self, HttpApiAuth, add_documents_func, invalid_dataset_id, expected_code, expected_message, ): dataset_id, document_ids = add_documents_func parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) res = stop_parse_documents(HttpApiAuth, invalid_dataset_id, {"document_ids": document_ids}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.skip @pytest.mark.parametrize( "payload", [ lambda r: {"document_ids": ["invalid_id"] + r}, lambda r: {"document_ids": r[:1] + ["invalid_id"] + r[1:3]}, lambda r: {"document_ids": r + ["invalid_id"]}, ], ) def test_stop_parse_partial_invalid_document_id(self, HttpApiAuth, add_documents_func, payload): dataset_id, document_ids = add_documents_func parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) if callable(payload): payload = payload(document_ids) res = stop_parse_documents(HttpApiAuth, dataset_id, payload) assert res["code"] == 102 assert res["message"] == "You don't own the document invalid_id." validate_document_parse_cancel(HttpApiAuth, dataset_id, document_ids) @pytest.mark.p3 def test_repeated_stop_parse(self, HttpApiAuth, add_documents_func): dataset_id, document_ids = add_documents_func parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) res = stop_parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 0 res = stop_parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 102 assert res["message"] == "Can't stop parsing document with progress at 0 or 1" @pytest.mark.p3 def test_duplicate_stop_parse(self, HttpApiAuth, add_documents_func): dataset_id, document_ids = add_documents_func parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) res = stop_parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids + document_ids}) assert res["code"] == 0 assert res["data"]["success_count"] == 3 assert f"Duplicate document ids: {document_ids[0]}" in res["data"]["errors"] @pytest.mark.skip(reason="unstable") def test_stop_parse_100_files(HttpApiAuth, add_dataset_func, tmp_path): document_num = 100 dataset_id = add_dataset_func document_ids = bulk_upload_documents(HttpApiAuth, dataset_id, document_num, tmp_path) parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) sleep(1) res = stop_parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 0 validate_document_parse_cancel(HttpApiAuth, dataset_id, document_ids) @pytest.mark.skip(reason="unstable") def test_concurrent_parse(HttpApiAuth, add_dataset_func, tmp_path): document_num = 50 dataset_id = add_dataset_func document_ids = bulk_upload_documents(HttpApiAuth, dataset_id, document_num, tmp_path) parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( stop_parse_documents, HttpApiAuth, dataset_id, {"document_ids": document_ids[i : i + 1]}, ) for i in range(document_num) ] responses = [f.result() for f in futures] assert all(r["code"] == 0 for r in responses) validate_document_parse_cancel(HttpApiAuth, dataset_id, document_ids)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_file_management_within_dataset/test_download_document.py
test/testcases/test_http_api/test_file_management_within_dataset/test_download_document.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import bulk_upload_documents, download_document, upload_documents from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth from requests import codes from utils import compare_by_hash @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, tmp_path, expected_code, expected_message): res = download_document(invalid_auth, "dataset_id", "document_id", tmp_path / "ragflow_tes.txt") assert res.status_code == codes.ok with (tmp_path / "ragflow_tes.txt").open("r") as f: response_json = json.load(f) assert response_json["code"] == expected_code assert response_json["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "generate_test_files", [ "docx", "excel", "ppt", "image", "pdf", "txt", "md", "json", "eml", "html", ], indirect=True, ) def test_file_type_validation(HttpApiAuth, add_dataset, generate_test_files, request): dataset_id = add_dataset fp = generate_test_files[request.node.callspec.params["generate_test_files"]] res = upload_documents(HttpApiAuth, dataset_id, [fp]) document_id = res["data"][0]["id"] res = download_document( HttpApiAuth, dataset_id, document_id, fp.with_stem("ragflow_test_download"), ) assert res.status_code == codes.ok assert compare_by_hash( fp, fp.with_stem("ragflow_test_download"), ) class TestDocumentDownload: @pytest.mark.p3 @pytest.mark.parametrize( "document_id, expected_code, expected_message", [ ( "invalid_document_id", 102, "The dataset not own the document invalid_document_id.", ), ], ) def test_invalid_document_id(self, HttpApiAuth, add_documents, tmp_path, document_id, expected_code, expected_message): dataset_id, _ = add_documents res = download_document( HttpApiAuth, dataset_id, document_id, tmp_path / "ragflow_test_download_1.txt", ) assert res.status_code == codes.ok with (tmp_path / "ragflow_test_download_1.txt").open("r") as f: response_json = json.load(f) assert response_json["code"] == expected_code assert response_json["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<NotFound '404: Not Found'>"), ( "invalid_dataset_id", 102, "You do not own the dataset invalid_dataset_id.", ), ], ) def test_invalid_dataset_id(self, HttpApiAuth, add_documents, tmp_path, dataset_id, expected_code, expected_message): _, document_ids = add_documents res = download_document( HttpApiAuth, dataset_id, document_ids[0], tmp_path / "ragflow_test_download_1.txt", ) assert res.status_code == codes.ok with (tmp_path / "ragflow_test_download_1.txt").open("r") as f: response_json = json.load(f) assert response_json["code"] == expected_code assert response_json["message"] == expected_message @pytest.mark.p3 def test_same_file_repeat(self, HttpApiAuth, add_documents, tmp_path, ragflow_tmp_dir): num = 5 dataset_id, document_ids = add_documents for i in range(num): res = download_document( HttpApiAuth, dataset_id, document_ids[0], tmp_path / f"ragflow_test_download_{i}.txt", ) assert res.status_code == codes.ok assert compare_by_hash( ragflow_tmp_dir / "ragflow_test_upload_0.txt", tmp_path / f"ragflow_test_download_{i}.txt", ) @pytest.mark.p3 def test_concurrent_download(HttpApiAuth, add_dataset, tmp_path): count = 20 dataset_id = add_dataset document_ids = bulk_upload_documents(HttpApiAuth, dataset_id, count, tmp_path) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( download_document, HttpApiAuth, dataset_id, document_ids[i], tmp_path / f"ragflow_test_download_{i}.txt", ) for i in range(count) ] responses = list(as_completed(futures)) assert len(responses) == count, responses for i in range(count): assert compare_by_hash( tmp_path / f"ragflow_test_upload_{i}.txt", tmp_path / f"ragflow_test_download_{i}.txt", )
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_file_management_within_dataset/test_upload_documents.py
test/testcases/test_http_api/test_file_management_within_dataset/test_upload_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import string from concurrent.futures import ThreadPoolExecutor, as_completed import pytest import requests from common import FILE_API_URL, list_datasets, upload_documents from configs import DOCUMENT_NAME_LIMIT, HOST_ADDRESS, INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth from requests_toolbelt import MultipartEncoder from utils.file_utils import create_txt_file @pytest.mark.p1 @pytest.mark.usefixtures("clear_datasets") class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = upload_documents(invalid_auth, "dataset_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestDocumentsUpload: @pytest.mark.p1 def test_valid_single_upload(self, HttpApiAuth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fp = create_txt_file(tmp_path / "ragflow_test.txt") res = upload_documents(HttpApiAuth, dataset_id, [fp]) assert res["code"] == 0 assert res["data"][0]["dataset_id"] == dataset_id assert res["data"][0]["name"] == fp.name @pytest.mark.p1 @pytest.mark.parametrize( "generate_test_files", [ "docx", "excel", "ppt", "image", "pdf", "txt", "md", "json", "eml", "html", ], indirect=True, ) def test_file_type_validation(self, HttpApiAuth, add_dataset_func, generate_test_files, request): dataset_id = add_dataset_func fp = generate_test_files[request.node.callspec.params["generate_test_files"]] res = upload_documents(HttpApiAuth, dataset_id, [fp]) assert res["code"] == 0 assert res["data"][0]["dataset_id"] == dataset_id assert res["data"][0]["name"] == fp.name @pytest.mark.p2 @pytest.mark.parametrize( "file_type", ["exe", "unknown"], ) def test_unsupported_file_type(self, HttpApiAuth, add_dataset_func, tmp_path, file_type): dataset_id = add_dataset_func fp = tmp_path / f"ragflow_test.{file_type}" fp.touch() res = upload_documents(HttpApiAuth, dataset_id, [fp]) assert res["code"] == 500 assert res["message"] == f"ragflow_test.{file_type}: This type of file has not been supported yet!" @pytest.mark.p2 def test_missing_file(self, HttpApiAuth, add_dataset_func): dataset_id = add_dataset_func res = upload_documents(HttpApiAuth, dataset_id) assert res["code"] == 101 assert res["message"] == "No file part!" @pytest.mark.p3 def test_empty_file(self, HttpApiAuth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fp = tmp_path / "empty.txt" fp.touch() res = upload_documents(HttpApiAuth, dataset_id, [fp]) assert res["code"] == 0 assert res["data"][0]["size"] == 0 @pytest.mark.p3 def test_filename_empty(self, HttpApiAuth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fp = create_txt_file(tmp_path / "ragflow_test.txt") url = f"{HOST_ADDRESS}{FILE_API_URL}".format(dataset_id=dataset_id) fields = (("file", ("", fp.open("rb"))),) m = MultipartEncoder(fields=fields) res = requests.post( url=url, headers={"Content-Type": m.content_type}, auth=HttpApiAuth, data=m, ) assert res.json()["code"] == 101 assert res.json()["message"] == "No file selected!" @pytest.mark.p2 def test_filename_max_length(self, HttpApiAuth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fp = create_txt_file(tmp_path / f"{'a' * (DOCUMENT_NAME_LIMIT - 4)}.txt") res = upload_documents(HttpApiAuth, dataset_id, [fp]) assert res["code"] == 0 assert res["data"][0]["name"] == fp.name @pytest.mark.p2 def test_invalid_dataset_id(self, HttpApiAuth, tmp_path): fp = create_txt_file(tmp_path / "ragflow_test.txt") res = upload_documents(HttpApiAuth, "invalid_dataset_id", [fp]) assert res["code"] == 100 assert res["message"] == """LookupError("Can\'t find the dataset with ID invalid_dataset_id!")""" @pytest.mark.p2 def test_duplicate_files(self, HttpApiAuth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fp = create_txt_file(tmp_path / "ragflow_test.txt") res = upload_documents(HttpApiAuth, dataset_id, [fp, fp]) assert res["code"] == 0 assert len(res["data"]) == 2 for i in range(len(res["data"])): assert res["data"][i]["dataset_id"] == dataset_id expected_name = fp.name if i != 0: expected_name = f"{fp.stem}({i}){fp.suffix}" assert res["data"][i]["name"] == expected_name @pytest.mark.p2 def test_same_file_repeat(self, HttpApiAuth, add_dataset_func, tmp_path): dataset_id = add_dataset_func fp = create_txt_file(tmp_path / "ragflow_test.txt") for i in range(3): res = upload_documents(HttpApiAuth, dataset_id, [fp]) assert res["code"] == 0 assert len(res["data"]) == 1 assert res["data"][0]["dataset_id"] == dataset_id expected_name = fp.name if i != 0: expected_name = f"{fp.stem}({i}){fp.suffix}" assert res["data"][0]["name"] == expected_name @pytest.mark.p3 def test_filename_special_characters(self, HttpApiAuth, add_dataset_func, tmp_path): dataset_id = add_dataset_func illegal_chars = '<>:"/\\|?*' translation_table = str.maketrans({char: "_" for char in illegal_chars}) safe_filename = string.punctuation.translate(translation_table) fp = tmp_path / f"{safe_filename}.txt" fp.write_text("Sample text content") res = upload_documents(HttpApiAuth, dataset_id, [fp]) assert res["code"] == 0 assert len(res["data"]) == 1 assert res["data"][0]["dataset_id"] == dataset_id assert res["data"][0]["name"] == fp.name @pytest.mark.p1 def test_multiple_files(self, HttpApiAuth, add_dataset_func, tmp_path): dataset_id = add_dataset_func expected_document_count = 20 fps = [] for i in range(expected_document_count): fp = create_txt_file(tmp_path / f"ragflow_test_{i}.txt") fps.append(fp) res = upload_documents(HttpApiAuth, dataset_id, fps) assert res["code"] == 0 res = list_datasets(HttpApiAuth, {"id": dataset_id}) assert res["data"][0]["document_count"] == expected_document_count @pytest.mark.p3 def test_concurrent_upload(self, HttpApiAuth, add_dataset_func, tmp_path): dataset_id = add_dataset_func count = 20 fps = [] for i in range(count): fp = create_txt_file(tmp_path / f"ragflow_test_{i}.txt") fps.append(fp) with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(upload_documents, HttpApiAuth, dataset_id, [fp]) for fp in fps] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) res = list_datasets(HttpApiAuth, {"id": dataset_id}) assert res["data"][0]["document_count"] == count
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_file_management_within_dataset/test_list_documents.py
test/testcases/test_http_api/test_file_management_within_dataset/test_list_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import list_documents from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth from utils import is_sorted @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = list_documents(invalid_auth, "dataset_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestDocumentsList: @pytest.mark.p1 def test_default(self, HttpApiAuth, add_documents): dataset_id, _ = add_documents res = list_documents(HttpApiAuth, dataset_id) assert res["code"] == 0 assert len(res["data"]["docs"]) == 5 assert res["data"]["total"] == 5 @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id. ", ), ], ) def test_invalid_dataset_id(self, HttpApiAuth, dataset_id, expected_code, expected_message): res = list_documents(HttpApiAuth, dataset_id) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page": None, "page_size": 2}, 0, 2, ""), ({"page": 0, "page_size": 2}, 0, 2, ""), ({"page": 2, "page_size": 2}, 0, 2, ""), ({"page": 3, "page_size": 2}, 0, 1, ""), ({"page": "3", "page_size": 2}, 0, 1, ""), pytest.param( {"page": -1, "page_size": 2}, 100, 0, "1064", marks=pytest.mark.skip(reason="issues/5851"), ), pytest.param( {"page": "a", "page_size": 2}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip(reason="issues/5851"), ), ], ) def test_page( self, HttpApiAuth, add_documents, params, expected_code, expected_page_size, expected_message, ): dataset_id, _ = add_documents res = list_documents(HttpApiAuth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["docs"]) == expected_page_size assert res["data"]["total"] == 5 else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_page_size, expected_message", [ ({"page_size": None}, 0, 5, ""), ({"page_size": 0}, 0, 0, ""), ({"page_size": 1}, 0, 1, ""), ({"page_size": 6}, 0, 5, ""), ({"page_size": "1"}, 0, 1, ""), pytest.param( {"page_size": -1}, 100, 0, "1064", marks=pytest.mark.skip(reason="issues/5851"), ), pytest.param( {"page_size": "a"}, 100, 0, """ValueError("invalid literal for int() with base 10: \'a\'")""", marks=pytest.mark.skip(reason="issues/5851"), ), ], ) def test_page_size( self, HttpApiAuth, add_documents, params, expected_code, expected_page_size, expected_message, ): dataset_id, _ = add_documents res = list_documents(HttpApiAuth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: assert len(res["data"]["docs"]) == expected_page_size else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "params, expected_code, assertions, expected_message", [ ({"orderby": None}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"orderby": "create_time"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"orderby": "update_time"}, 0, lambda r: (is_sorted(r["data"]["docs"], "update_time", True)), ""), pytest.param({"orderby": "name", "desc": "False"}, 0, lambda r: (is_sorted(r["data"]["docs"], "name", False)), "", marks=pytest.mark.skip(reason="issues/5851")), pytest.param({"orderby": "unknown"}, 102, 0, "orderby should be create_time or update_time", marks=pytest.mark.skip(reason="issues/5851")), ], ) def test_orderby( self, HttpApiAuth, add_documents, params, expected_code, assertions, expected_message, ): dataset_id, _ = add_documents res = list_documents(HttpApiAuth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: if callable(assertions): assert assertions(res) else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "params, expected_code, assertions, expected_message", [ ({"desc": None}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"desc": "true"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"desc": "True"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), ({"desc": True}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", True)), ""), pytest.param({"desc": "false"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", False)), "", marks=pytest.mark.skip(reason="issues/5851")), ({"desc": "False"}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", False)), ""), ({"desc": False}, 0, lambda r: (is_sorted(r["data"]["docs"], "create_time", False)), ""), ({"desc": "False", "orderby": "update_time"}, 0, lambda r: (is_sorted(r["data"]["docs"], "update_time", False)), ""), pytest.param({"desc": "unknown"}, 102, 0, "desc should be true or false", marks=pytest.mark.skip(reason="issues/5851")), ], ) def test_desc( self, HttpApiAuth, add_documents, params, expected_code, assertions, expected_message, ): dataset_id, _ = add_documents res = list_documents(HttpApiAuth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: if callable(assertions): assert assertions(res) else: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "params, expected_num", [ ({"keywords": None}, 5), ({"keywords": ""}, 5), ({"keywords": "0"}, 1), ({"keywords": "ragflow_test_upload"}, 5), ({"keywords": "unknown"}, 0), ], ) def test_keywords(self, HttpApiAuth, add_documents, params, expected_num): dataset_id, _ = add_documents res = list_documents(HttpApiAuth, dataset_id, params=params) assert res["code"] == 0 assert len(res["data"]["docs"]) == expected_num assert res["data"]["total"] == expected_num @pytest.mark.p1 @pytest.mark.parametrize( "params, expected_code, expected_num, expected_message", [ ({"name": None}, 0, 5, ""), ({"name": ""}, 0, 5, ""), ({"name": "ragflow_test_upload_0.txt"}, 0, 1, ""), ( {"name": "unknown.txt"}, 102, 0, "You don't own the document unknown.txt.", ), ], ) def test_name( self, HttpApiAuth, add_documents, params, expected_code, expected_num, expected_message, ): dataset_id, _ = add_documents res = list_documents(HttpApiAuth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: if params["name"] in [None, ""]: assert len(res["data"]["docs"]) == expected_num else: assert res["data"]["docs"][0]["name"] == params["name"] else: assert res["message"] == expected_message @pytest.mark.p1 @pytest.mark.parametrize( "document_id, expected_code, expected_num, expected_message", [ (None, 0, 5, ""), ("", 0, 5, ""), (lambda r: r[0], 0, 1, ""), ("unknown.txt", 102, 0, "You don't own the document unknown.txt."), ], ) def test_id( self, HttpApiAuth, add_documents, document_id, expected_code, expected_num, expected_message, ): dataset_id, document_ids = add_documents if callable(document_id): params = {"id": document_id(document_ids)} else: params = {"id": document_id} res = list_documents(HttpApiAuth, dataset_id, params=params) assert res["code"] == expected_code if expected_code == 0: if params["id"] in [None, ""]: assert len(res["data"]["docs"]) == expected_num else: assert res["data"]["docs"][0]["id"] == params["id"] else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "document_id, name, expected_code, expected_num, expected_message", [ (lambda r: r[0], "ragflow_test_upload_0.txt", 0, 1, ""), (lambda r: r[0], "ragflow_test_upload_1.txt", 0, 0, ""), (lambda r: r[0], "unknown", 102, 0, "You don't own the document unknown."), ( "id", "ragflow_test_upload_0.txt", 102, 0, "You don't own the document id.", ), ], ) def test_name_and_id( self, HttpApiAuth, add_documents, document_id, name, expected_code, expected_num, expected_message, ): dataset_id, document_ids = add_documents if callable(document_id): params = {"id": document_id(document_ids), "name": name} else: params = {"id": document_id, "name": name} res = list_documents(HttpApiAuth, dataset_id, params=params) if expected_code == 0: assert len(res["data"]["docs"]) == expected_num else: assert res["message"] == expected_message @pytest.mark.p3 def test_concurrent_list(self, HttpApiAuth, add_documents): dataset_id, _ = add_documents count = 100 with ThreadPoolExecutor(max_workers=5) as executor: futures = [executor.submit(list_documents, HttpApiAuth, dataset_id) for _ in range(count)] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.p3 def test_invalid_params(self, HttpApiAuth, add_documents): dataset_id, _ = add_documents params = {"a": "b"} res = list_documents(HttpApiAuth, dataset_id, params=params) assert res["code"] == 0 assert len(res["data"]["docs"]) == 5
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_file_management_within_dataset/conftest.py
test/testcases/test_http_api/test_file_management_within_dataset/conftest.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import bulk_upload_documents, delete_documents @pytest.fixture(scope="function") def add_document_func(request, HttpApiAuth, add_dataset, ragflow_tmp_dir): def cleanup(): delete_documents(HttpApiAuth, dataset_id, {"ids": None}) request.addfinalizer(cleanup) dataset_id = add_dataset return dataset_id, bulk_upload_documents(HttpApiAuth, dataset_id, 1, ragflow_tmp_dir)[0] @pytest.fixture(scope="class") def add_documents(request, HttpApiAuth, add_dataset, ragflow_tmp_dir): def cleanup(): delete_documents(HttpApiAuth, dataset_id, {"ids": None}) request.addfinalizer(cleanup) dataset_id = add_dataset return dataset_id, bulk_upload_documents(HttpApiAuth, dataset_id, 5, ragflow_tmp_dir) @pytest.fixture(scope="function") def add_documents_func(request, HttpApiAuth, add_dataset_func, ragflow_tmp_dir): def cleanup(): delete_documents(HttpApiAuth, dataset_id, {"ids": None}) request.addfinalizer(cleanup) dataset_id = add_dataset_func return dataset_id, bulk_upload_documents(HttpApiAuth, dataset_id, 3, ragflow_tmp_dir)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_file_management_within_dataset/test_delete_documents.py
test/testcases/test_http_api/test_file_management_within_dataset/test_delete_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import bulk_upload_documents, delete_documents, list_documents from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = delete_documents(invalid_auth, "dataset_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestDocumentsDeletion: @pytest.mark.p1 @pytest.mark.parametrize( "payload, expected_code, expected_message, remaining", [ (None, 0, "", 0), ({"ids": []}, 0, "", 0), ({"ids": ["invalid_id"]}, 102, "Documents not found: ['invalid_id']", 3), ( {"ids": ["\n!?。;!?\"'"]}, 102, """Documents not found: [\'\\n!?。;!?"\\\'\']""", 3, ), ( "not json", 100, "AttributeError(\"'str' object has no attribute 'get'\")", 3, ), (lambda r: {"ids": r[:1]}, 0, "", 2), (lambda r: {"ids": r}, 0, "", 0), ], ) def test_basic_scenarios( self, HttpApiAuth, add_documents_func, payload, expected_code, expected_message, remaining, ): dataset_id, document_ids = add_documents_func if callable(payload): payload = payload(document_ids) res = delete_documents(HttpApiAuth, dataset_id, payload) assert res["code"] == expected_code if res["code"] != 0: assert res["message"] == expected_message res = list_documents(HttpApiAuth, dataset_id) assert len(res["data"]["docs"]) == remaining assert res["data"]["total"] == remaining @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id. ", ), ], ) def test_invalid_dataset_id(self, HttpApiAuth, add_documents_func, dataset_id, expected_code, expected_message): _, document_ids = add_documents_func res = delete_documents(HttpApiAuth, dataset_id, {"ids": document_ids[:1]}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "payload", [ lambda r: {"ids": ["invalid_id"] + r}, lambda r: {"ids": r[:1] + ["invalid_id"] + r[1:3]}, lambda r: {"ids": r + ["invalid_id"]}, ], ) def test_delete_partial_invalid_id(self, HttpApiAuth, add_documents_func, payload): dataset_id, document_ids = add_documents_func if callable(payload): payload = payload(document_ids) res = delete_documents(HttpApiAuth, dataset_id, payload) assert res["code"] == 102 assert res["message"] == "Documents not found: ['invalid_id']" res = list_documents(HttpApiAuth, dataset_id) assert len(res["data"]["docs"]) == 0 assert res["data"]["total"] == 0 @pytest.mark.p2 def test_repeated_deletion(self, HttpApiAuth, add_documents_func): dataset_id, document_ids = add_documents_func res = delete_documents(HttpApiAuth, dataset_id, {"ids": document_ids}) assert res["code"] == 0 res = delete_documents(HttpApiAuth, dataset_id, {"ids": document_ids}) assert res["code"] == 102 assert "Documents not found" in res["message"] @pytest.mark.p2 def test_duplicate_deletion(self, HttpApiAuth, add_documents_func): dataset_id, document_ids = add_documents_func res = delete_documents(HttpApiAuth, dataset_id, {"ids": document_ids + document_ids}) assert res["code"] == 0 assert "Duplicate document ids" in res["data"]["errors"][0] assert res["data"]["success_count"] == 3 res = list_documents(HttpApiAuth, dataset_id) assert len(res["data"]["docs"]) == 0 assert res["data"]["total"] == 0 @pytest.mark.p3 def test_concurrent_deletion(HttpApiAuth, add_dataset, tmp_path): count = 100 dataset_id = add_dataset document_ids = bulk_upload_documents(HttpApiAuth, dataset_id, count, tmp_path) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( delete_documents, HttpApiAuth, dataset_id, {"ids": document_ids[i : i + 1]}, ) for i in range(count) ] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) @pytest.mark.p3 def test_delete_1k(HttpApiAuth, add_dataset, tmp_path): documents_num = 1_000 dataset_id = add_dataset document_ids = bulk_upload_documents(HttpApiAuth, dataset_id, documents_num, tmp_path) res = list_documents(HttpApiAuth, dataset_id) assert res["data"]["total"] == documents_num res = delete_documents(HttpApiAuth, dataset_id, {"ids": document_ids}) assert res["code"] == 0 res = list_documents(HttpApiAuth, dataset_id) assert res["data"]["total"] == 0
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_file_management_within_dataset/test_update_document.py
test/testcases/test_http_api/test_file_management_within_dataset/test_update_document.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common import list_documents, update_document from configs import DOCUMENT_NAME_LIMIT, INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth from configs import DEFAULT_PARSER_CONFIG @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = update_document(invalid_auth, "dataset_id", "document_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestDocumentsUpdated: @pytest.mark.p1 @pytest.mark.parametrize( "name, expected_code, expected_message", [ ("new_name.txt", 0, ""), ( f"{'a' * (DOCUMENT_NAME_LIMIT - 4)}.txt", 0, "", ), ( 0, 100, """AttributeError("\'int\' object has no attribute \'encode\'")""", ), ( None, 100, """AttributeError("\'NoneType\' object has no attribute \'encode\'")""", ), ( "", 101, "The extension of file can't be changed", ), ( "ragflow_test_upload_0", 101, "The extension of file can't be changed", ), ( "ragflow_test_upload_1.txt", 102, "Duplicated document name in the same dataset.", ), ( "RAGFLOW_TEST_UPLOAD_1.TXT", 0, "", ), ], ) def test_name(self, HttpApiAuth, add_documents, name, expected_code, expected_message): dataset_id, document_ids = add_documents res = update_document(HttpApiAuth, dataset_id, document_ids[0], {"name": name}) assert res["code"] == expected_code if expected_code == 0: res = list_documents(HttpApiAuth, dataset_id, {"id": document_ids[0]}) assert res["data"]["docs"][0]["name"] == name else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "document_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_document_id", 102, "The dataset doesn't own the document.", ), ], ) def test_invalid_document_id(self, HttpApiAuth, add_documents, document_id, expected_code, expected_message): dataset_id, _ = add_documents res = update_document(HttpApiAuth, dataset_id, document_id, {"name": "new_name.txt"}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<NotFound '404: Not Found'>"), ( "invalid_dataset_id", 102, "You don't own the dataset.", ), ], ) def test_invalid_dataset_id(self, HttpApiAuth, add_documents, dataset_id, expected_code, expected_message): _, document_ids = add_documents res = update_document(HttpApiAuth, dataset_id, document_ids[0], {"name": "new_name.txt"}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "meta_fields, expected_code, expected_message", [({"test": "test"}, 0, ""), ("test", 102, "meta_fields must be a dictionary")], ) def test_meta_fields(self, HttpApiAuth, add_documents, meta_fields, expected_code, expected_message): dataset_id, document_ids = add_documents res = update_document(HttpApiAuth, dataset_id, document_ids[0], {"meta_fields": meta_fields}) if expected_code == 0: res = list_documents(HttpApiAuth, dataset_id, {"id": document_ids[0]}) assert res["data"]["docs"][0]["meta_fields"] == meta_fields else: assert res["message"] == expected_message @pytest.mark.p2 @pytest.mark.parametrize( "chunk_method, expected_code, expected_message", [ ("naive", 0, ""), ("manual", 0, ""), ("qa", 0, ""), ("table", 0, ""), ("paper", 0, ""), ("book", 0, ""), ("laws", 0, ""), ("presentation", 0, ""), ("picture", 0, ""), ("one", 0, ""), ("knowledge_graph", 0, ""), ("email", 0, ""), ("tag", 0, ""), ("", 102, "`chunk_method` doesn't exist"), ( "other_chunk_method", 102, "`chunk_method` other_chunk_method doesn't exist", ), ], ) def test_chunk_method(self, HttpApiAuth, add_documents, chunk_method, expected_code, expected_message): dataset_id, document_ids = add_documents res = update_document(HttpApiAuth, dataset_id, document_ids[0], {"chunk_method": chunk_method}) assert res["code"] == expected_code if expected_code == 0: res = list_documents(HttpApiAuth, dataset_id, {"id": document_ids[0]}) if chunk_method == "": assert res["data"]["docs"][0]["chunk_method"] == "naive" else: assert res["data"]["docs"][0]["chunk_method"] == chunk_method else: assert res["message"] == expected_message @pytest.mark.p3 @pytest.mark.parametrize( "payload, expected_code, expected_message", [ ({"chunk_count": 1}, 102, "Can't change `chunk_count`."), pytest.param( {"create_date": "Fri, 14 Mar 2025 16:53:42 GMT"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"create_time": 1}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"created_by": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"dataset_id": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"id": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"location": "ragflow_test.txt"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"process_begin_at": 1}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"process_duration": 1.0}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param({"progress": 1.0}, 102, "Can't change `progress`."), pytest.param( {"progress_msg": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"run": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"size": 1}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"source_type": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"thumbnail": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), ({"token_count": 1}, 102, "Can't change `token_count`."), pytest.param( {"type": "ragflow_test"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"update_date": "Fri, 14 Mar 2025 16:33:17 GMT"}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), pytest.param( {"update_time": 1}, 102, "The input parameters are invalid.", marks=pytest.mark.skip(reason="issues/6104"), ), ], ) def test_invalid_field( self, HttpApiAuth, add_documents, payload, expected_code, expected_message, ): dataset_id, document_ids = add_documents res = update_document(HttpApiAuth, dataset_id, document_ids[0], payload) assert res["code"] == expected_code assert res["message"] == expected_message class TestUpdateDocumentParserConfig: @pytest.mark.p2 @pytest.mark.parametrize( "chunk_method, parser_config, expected_code, expected_message", [ ("naive", {}, 0, ""), ( "naive", DEFAULT_PARSER_CONFIG, 0, "", ), pytest.param( "naive", {"chunk_token_num": -1}, 100, "AssertionError('chunk_token_num should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"chunk_token_num": 0}, 100, "AssertionError('chunk_token_num should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"chunk_token_num": 100000000}, 100, "AssertionError('chunk_token_num should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"chunk_token_num": 3.14}, 102, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"chunk_token_num": "1024"}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), ( "naive", {"layout_recognize": "DeepDOC"}, 0, "", ), ( "naive", {"layout_recognize": "Naive"}, 0, "", ), ("naive", {"html4excel": True}, 0, ""), ("naive", {"html4excel": False}, 0, ""), pytest.param( "naive", {"html4excel": 1}, 100, "AssertionError('html4excel should be True or False')", marks=pytest.mark.skip(reason="issues/6098"), ), ("naive", {"delimiter": ""}, 0, ""), ("naive", {"delimiter": "`##`"}, 0, ""), pytest.param( "naive", {"delimiter": 1}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"task_page_size": -1}, 100, "AssertionError('task_page_size should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"task_page_size": 0}, 100, "AssertionError('task_page_size should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"task_page_size": 100000000}, 100, "AssertionError('task_page_size should be in range from 1 to 100000000')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"task_page_size": 3.14}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"task_page_size": "1024"}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), ("naive", {"raptor": {"use_raptor": { "use_raptor": True, "prompt": "Please summarize the following paragraphs. Be careful with the numbers, do not make things up. Paragraphs as following:\n {cluster_content}\nThe above is the content you need to summarize.", "max_token": 256, "threshold": 0.1, "max_cluster": 64, "random_seed": 0, },}}, 0, ""), ("naive", {"raptor": {"use_raptor": False}}, 0, ""), pytest.param( "naive", {"invalid_key": "invalid_value"}, 100, """AssertionError("Abnormal \'parser_config\'. Invalid key: invalid_key")""", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_keywords": -1}, 100, "AssertionError('auto_keywords should be in range from 0 to 32')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_keywords": 32}, 100, "AssertionError('auto_keywords should be in range from 0 to 32')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_questions": 3.14}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_keywords": "1024"}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_questions": -1}, 100, "AssertionError('auto_questions should be in range from 0 to 10')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_questions": 10}, 100, "AssertionError('auto_questions should be in range from 0 to 10')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_questions": 3.14}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"auto_questions": "1024"}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"topn_tags": -1}, 100, "AssertionError('topn_tags should be in range from 0 to 10')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"topn_tags": 10}, 100, "AssertionError('topn_tags should be in range from 0 to 10')", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"topn_tags": 3.14}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), pytest.param( "naive", {"topn_tags": "1024"}, 100, "", marks=pytest.mark.skip(reason="issues/6098"), ), ], ) def test_parser_config( self, HttpApiAuth, add_documents, chunk_method, parser_config, expected_code, expected_message, ): dataset_id, document_ids = add_documents res = update_document( HttpApiAuth, dataset_id, document_ids[0], {"chunk_method": chunk_method, "parser_config": parser_config}, ) assert res["code"] == expected_code if expected_code == 0: res = list_documents(HttpApiAuth, dataset_id, {"id": document_ids[0]}) if parser_config == {}: assert res["data"]["docs"][0]["parser_config"] == DEFAULT_PARSER_CONFIG else: for k, v in parser_config.items(): assert res["data"]["docs"][0]["parser_config"][k] == v if expected_code != 0 or expected_message: assert res["message"] == expected_message
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/test_http_api/test_file_management_within_dataset/test_parse_documents.py
test/testcases/test_http_api/test_file_management_within_dataset/test_parse_documents.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from concurrent.futures import ThreadPoolExecutor, as_completed import pytest from common import bulk_upload_documents, list_documents, parse_documents from configs import INVALID_API_TOKEN from libs.auth import RAGFlowHttpApiAuth from utils import wait_for @wait_for(30, 1, "Document parsing timeout") def condition(_auth, _dataset_id, _document_ids=None): res = list_documents(_auth, _dataset_id) target_docs = res["data"]["docs"] if _document_ids is None: for doc in target_docs: if doc["run"] != "DONE": return False return True target_ids = set(_document_ids) for doc in target_docs: if doc["id"] in target_ids: if doc.get("run") != "DONE": return False return True def validate_document_details(auth, dataset_id, document_ids): for document_id in document_ids: res = list_documents(auth, dataset_id, params={"id": document_id}) doc = res["data"]["docs"][0] assert doc["run"] == "DONE" assert len(doc["process_begin_at"]) > 0 assert doc["process_duration"] > 0 assert doc["progress"] > 0 assert "Task done" in doc["progress_msg"] @pytest.mark.p1 class TestAuthorization: @pytest.mark.parametrize( "invalid_auth, expected_code, expected_message", [ (None, 0, "`Authorization` can't be empty"), ( RAGFlowHttpApiAuth(INVALID_API_TOKEN), 109, "Authentication error: API key is invalid!", ), ], ) def test_invalid_auth(self, invalid_auth, expected_code, expected_message): res = parse_documents(invalid_auth, "dataset_id") assert res["code"] == expected_code assert res["message"] == expected_message class TestDocumentsParse: @pytest.mark.parametrize( "payload, expected_code, expected_message", [ pytest.param(None, 102, """AttributeError("\'NoneType\' object has no attribute \'get\'")""", marks=pytest.mark.skip), pytest.param({"document_ids": []}, 102, "`document_ids` is required", marks=pytest.mark.p1), pytest.param({"document_ids": ["invalid_id"]}, 102, "Documents not found: ['invalid_id']", marks=pytest.mark.p3), pytest.param({"document_ids": ["\n!?。;!?\"'"]}, 102, """Documents not found: [\'\\n!?。;!?"\\\'\']""", marks=pytest.mark.p3), pytest.param("not json", 102, "AttributeError(\"'str' object has no attribute 'get'\")", marks=pytest.mark.skip), pytest.param(lambda r: {"document_ids": r[:1]}, 0, "", marks=pytest.mark.p1), pytest.param(lambda r: {"document_ids": r}, 0, "", marks=pytest.mark.p1), ], ) def test_basic_scenarios(self, HttpApiAuth, add_documents_func, payload, expected_code, expected_message): dataset_id, document_ids = add_documents_func if callable(payload): payload = payload(document_ids) res = parse_documents(HttpApiAuth, dataset_id, payload) assert res["code"] == expected_code if expected_code != 0: assert res["message"] == expected_message if expected_code == 0: condition(HttpApiAuth, dataset_id, payload["document_ids"]) validate_document_details(HttpApiAuth, dataset_id, payload["document_ids"]) @pytest.mark.p3 @pytest.mark.parametrize( "dataset_id, expected_code, expected_message", [ ("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"), ( "invalid_dataset_id", 102, "You don't own the dataset invalid_dataset_id.", ), ], ) def test_invalid_dataset_id( self, HttpApiAuth, add_documents_func, dataset_id, expected_code, expected_message, ): _, document_ids = add_documents_func res = parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) assert res["code"] == expected_code assert res["message"] == expected_message @pytest.mark.parametrize( "payload", [ pytest.param(lambda r: {"document_ids": ["invalid_id"] + r}, marks=pytest.mark.p3), pytest.param(lambda r: {"document_ids": r[:1] + ["invalid_id"] + r[1:3]}, marks=pytest.mark.p1), pytest.param(lambda r: {"document_ids": r + ["invalid_id"]}, marks=pytest.mark.p3), ], ) def test_parse_partial_invalid_document_id(self, HttpApiAuth, add_documents_func, payload): dataset_id, document_ids = add_documents_func if callable(payload): payload = payload(document_ids) res = parse_documents(HttpApiAuth, dataset_id, payload) assert res["code"] == 102 assert res["message"] == "Documents not found: ['invalid_id']" condition(HttpApiAuth, dataset_id) validate_document_details(HttpApiAuth, dataset_id, document_ids) @pytest.mark.p3 def test_repeated_parse(self, HttpApiAuth, add_documents_func): dataset_id, document_ids = add_documents_func res = parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 0 condition(HttpApiAuth, dataset_id) res = parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 0 @pytest.mark.p3 def test_duplicate_parse(self, HttpApiAuth, add_documents_func): dataset_id, document_ids = add_documents_func res = parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids + document_ids}) assert res["code"] == 0 assert "Duplicate document ids" in res["data"]["errors"][0] assert res["data"]["success_count"] == 3 condition(HttpApiAuth, dataset_id) validate_document_details(HttpApiAuth, dataset_id, document_ids) @pytest.mark.p3 def test_parse_100_files(HttpApiAuth, add_dataset_func, tmp_path): @wait_for(200, 1, "Document parsing timeout") def condition(_auth, _dataset_id, _document_num): res = list_documents(_auth, _dataset_id, {"page_size": _document_num}) for doc in res["data"]["docs"]: if doc["run"] != "DONE": return False return True document_num = 100 dataset_id = add_dataset_func document_ids = bulk_upload_documents(HttpApiAuth, dataset_id, document_num, tmp_path) res = parse_documents(HttpApiAuth, dataset_id, {"document_ids": document_ids}) assert res["code"] == 0 condition(HttpApiAuth, dataset_id, document_num) validate_document_details(HttpApiAuth, dataset_id, document_ids) @pytest.mark.p3 def test_concurrent_parse(HttpApiAuth, add_dataset_func, tmp_path): @wait_for(200, 1, "Document parsing timeout") def condition(_auth, _dataset_id, _document_num): res = list_documents(_auth, _dataset_id, {"page_size": _document_num}) for doc in res["data"]["docs"]: if doc["run"] != "DONE": return False return True count = 100 dataset_id = add_dataset_func document_ids = bulk_upload_documents(HttpApiAuth, dataset_id, count, tmp_path) with ThreadPoolExecutor(max_workers=5) as executor: futures = [ executor.submit( parse_documents, HttpApiAuth, dataset_id, {"document_ids": document_ids[i : i + 1]}, ) for i in range(count) ] responses = list(as_completed(futures)) assert len(responses) == count, responses assert all(future.result()["code"] == 0 for future in futures) condition(HttpApiAuth, dataset_id, count) validate_document_details(HttpApiAuth, dataset_id, document_ids)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/libs/__init__.py
test/testcases/libs/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/testcases/libs/auth.py
test/testcases/libs/auth.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from requests.auth import AuthBase class RAGFlowHttpApiAuth(AuthBase): def __init__(self, token): self._token = token def __call__(self, r): r.headers["Authorization"] = f"Bearer {self._token}" return r class RAGFlowWebApiAuth(AuthBase): def __init__(self, token): self._token = token def __call__(self, r): r.headers["Authorization"] = self._token return r
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/unit_test/utils/test_raptor_utils.py
test/unit_test/utils/test_raptor_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ Unit tests for Raptor utility functions. """ import pytest from rag.utils.raptor_utils import ( is_structured_file_type, is_tabular_pdf, should_skip_raptor, get_skip_reason, EXCEL_EXTENSIONS, CSV_EXTENSIONS, STRUCTURED_EXTENSIONS ) class TestIsStructuredFileType: """Test file type detection for structured data""" @pytest.mark.parametrize("file_type,expected", [ (".xlsx", True), (".xls", True), (".xlsm", True), (".xlsb", True), (".csv", True), (".tsv", True), ("xlsx", True), # Without leading dot ("XLSX", True), # Uppercase (".pdf", False), (".docx", False), (".txt", False), ("", False), (None, False), ]) def test_file_type_detection(self, file_type, expected): """Test detection of various file types""" assert is_structured_file_type(file_type) == expected def test_excel_extensions_defined(self): """Test that Excel extensions are properly defined""" assert ".xlsx" in EXCEL_EXTENSIONS assert ".xls" in EXCEL_EXTENSIONS assert len(EXCEL_EXTENSIONS) >= 4 def test_csv_extensions_defined(self): """Test that CSV extensions are properly defined""" assert ".csv" in CSV_EXTENSIONS assert ".tsv" in CSV_EXTENSIONS def test_structured_extensions_combined(self): """Test that structured extensions include both Excel and CSV""" assert EXCEL_EXTENSIONS.issubset(STRUCTURED_EXTENSIONS) assert CSV_EXTENSIONS.issubset(STRUCTURED_EXTENSIONS) class TestIsTabularPDF: """Test tabular PDF detection""" def test_table_parser_detected(self): """Test that table parser is detected as tabular""" assert is_tabular_pdf("table", {}) is True assert is_tabular_pdf("TABLE", {}) is True def test_html4excel_detected(self): """Test that html4excel config is detected as tabular""" assert is_tabular_pdf("naive", {"html4excel": True}) is True assert is_tabular_pdf("", {"html4excel": True}) is True def test_non_tabular_pdf(self): """Test that non-tabular PDFs are not detected""" assert is_tabular_pdf("naive", {}) is False assert is_tabular_pdf("naive", {"html4excel": False}) is False assert is_tabular_pdf("", {}) is False def test_combined_conditions(self): """Test combined table parser and html4excel""" assert is_tabular_pdf("table", {"html4excel": True}) is True assert is_tabular_pdf("table", {"html4excel": False}) is True class TestShouldSkipRaptor: """Test Raptor skip logic""" def test_skip_excel_files(self): """Test that Excel files skip Raptor""" assert should_skip_raptor(".xlsx") is True assert should_skip_raptor(".xls") is True assert should_skip_raptor(".xlsm") is True def test_skip_csv_files(self): """Test that CSV files skip Raptor""" assert should_skip_raptor(".csv") is True assert should_skip_raptor(".tsv") is True def test_skip_tabular_pdf_with_table_parser(self): """Test that tabular PDFs skip Raptor""" assert should_skip_raptor(".pdf", parser_id="table") is True assert should_skip_raptor("pdf", parser_id="TABLE") is True def test_skip_tabular_pdf_with_html4excel(self): """Test that PDFs with html4excel skip Raptor""" assert should_skip_raptor(".pdf", parser_config={"html4excel": True}) is True def test_dont_skip_regular_pdf(self): """Test that regular PDFs don't skip Raptor""" assert should_skip_raptor(".pdf", parser_id="naive") is False assert should_skip_raptor(".pdf", parser_config={}) is False def test_dont_skip_text_files(self): """Test that text files don't skip Raptor""" assert should_skip_raptor(".txt") is False assert should_skip_raptor(".docx") is False assert should_skip_raptor(".md") is False def test_override_with_config(self): """Test that auto-disable can be overridden""" raptor_config = {"auto_disable_for_structured_data": False} # Should not skip even for Excel files assert should_skip_raptor(".xlsx", raptor_config=raptor_config) is False assert should_skip_raptor(".csv", raptor_config=raptor_config) is False assert should_skip_raptor(".pdf", parser_id="table", raptor_config=raptor_config) is False def test_default_auto_disable_enabled(self): """Test that auto-disable is enabled by default""" # Empty raptor_config should default to auto_disable=True assert should_skip_raptor(".xlsx", raptor_config={}) is True assert should_skip_raptor(".xlsx", raptor_config=None) is True def test_explicit_auto_disable_enabled(self): """Test explicit auto-disable enabled""" raptor_config = {"auto_disable_for_structured_data": True} assert should_skip_raptor(".xlsx", raptor_config=raptor_config) is True class TestGetSkipReason: """Test skip reason generation""" def test_excel_skip_reason(self): """Test skip reason for Excel files""" reason = get_skip_reason(".xlsx") assert "Structured data file" in reason assert ".xlsx" in reason assert "auto-disabled" in reason.lower() def test_csv_skip_reason(self): """Test skip reason for CSV files""" reason = get_skip_reason(".csv") assert "Structured data file" in reason assert ".csv" in reason def test_tabular_pdf_skip_reason(self): """Test skip reason for tabular PDFs""" reason = get_skip_reason(".pdf", parser_id="table") assert "Tabular PDF" in reason assert "table" in reason.lower() assert "auto-disabled" in reason.lower() def test_html4excel_skip_reason(self): """Test skip reason for html4excel PDFs""" reason = get_skip_reason(".pdf", parser_config={"html4excel": True}) assert "Tabular PDF" in reason def test_no_skip_reason_for_regular_files(self): """Test that regular files have no skip reason""" assert get_skip_reason(".txt") == "" assert get_skip_reason(".docx") == "" assert get_skip_reason(".pdf", parser_id="naive") == "" class TestEdgeCases: """Test edge cases and error handling""" def test_none_values(self): """Test handling of None values""" assert should_skip_raptor(None) is False assert should_skip_raptor("") is False assert get_skip_reason(None) == "" def test_empty_strings(self): """Test handling of empty strings""" assert should_skip_raptor("") is False assert get_skip_reason("") == "" def test_case_insensitivity(self): """Test case insensitive handling""" assert is_structured_file_type("XLSX") is True assert is_structured_file_type("XlSx") is True assert is_tabular_pdf("TABLE", {}) is True assert is_tabular_pdf("TaBlE", {}) is True def test_with_and_without_dot(self): """Test file extensions with and without leading dot""" assert should_skip_raptor(".xlsx") is True assert should_skip_raptor("xlsx") is True assert should_skip_raptor(".CSV") is True assert should_skip_raptor("csv") is True class TestIntegrationScenarios: """Test real-world integration scenarios""" def test_financial_excel_report(self): """Test scenario: Financial quarterly Excel report""" file_type = ".xlsx" parser_id = "naive" parser_config = {} raptor_config = {"use_raptor": True} # Should skip Raptor assert should_skip_raptor(file_type, parser_id, parser_config, raptor_config) is True reason = get_skip_reason(file_type, parser_id, parser_config) assert "Structured data file" in reason def test_scientific_csv_data(self): """Test scenario: Scientific experimental CSV results""" file_type = ".csv" # Should skip Raptor assert should_skip_raptor(file_type) is True reason = get_skip_reason(file_type) assert ".csv" in reason def test_legal_contract_with_tables(self): """Test scenario: Legal contract PDF with tables""" file_type = ".pdf" parser_id = "table" parser_config = {} # Should skip Raptor assert should_skip_raptor(file_type, parser_id, parser_config) is True reason = get_skip_reason(file_type, parser_id, parser_config) assert "Tabular PDF" in reason def test_text_heavy_pdf_document(self): """Test scenario: Text-heavy PDF document""" file_type = ".pdf" parser_id = "naive" parser_config = {} # Should NOT skip Raptor assert should_skip_raptor(file_type, parser_id, parser_config) is False reason = get_skip_reason(file_type, parser_id, parser_config) assert reason == "" def test_mixed_dataset_processing(self): """Test scenario: Mixed dataset with various file types""" files = [ (".xlsx", "naive", {}, True), # Excel - skip (".csv", "naive", {}, True), # CSV - skip (".pdf", "table", {}, True), # Tabular PDF - skip (".pdf", "naive", {}, False), # Regular PDF - don't skip (".docx", "naive", {}, False), # Word doc - don't skip (".txt", "naive", {}, False), # Text file - don't skip ] for file_type, parser_id, parser_config, expected_skip in files: result = should_skip_raptor(file_type, parser_id, parser_config) assert result == expected_skip, f"Failed for {file_type}" def test_override_for_special_excel(self): """Test scenario: Override auto-disable for special Excel processing""" file_type = ".xlsx" raptor_config = {"auto_disable_for_structured_data": False} # Should NOT skip when explicitly disabled assert should_skip_raptor(file_type, raptor_config=raptor_config) is False if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/unit_test/common/test_token_utils.py
test/unit_test/common/test_token_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from common.token_utils import num_tokens_from_string, total_token_count_from_response, truncate, encoder import pytest class TestNumTokensFromString: """Test cases for num_tokens_from_string function""" def test_empty_string(self): """Test that empty string returns zero tokens""" result = num_tokens_from_string("") assert result == 0 def test_single_word(self): """Test token count for a single word""" # "hello" should be 1 token with cl100k_base encoding result = num_tokens_from_string("hello") assert result == 1 def test_multiple_words(self): """Test token count for multiple words""" # "hello world" typically becomes 2 tokens result = num_tokens_from_string("hello world") assert result == 2 def test_special_characters(self): """Test token count with special characters""" result = num_tokens_from_string("hello, world!") # Special characters may be separate tokens assert result == 4 def test_hanzi_characters(self): """Test token count with special characters""" result = num_tokens_from_string("世界") # Special characters may be separate tokens assert result > 0 def test_unicode_characters(self): """Test token count with unicode characters""" result = num_tokens_from_string("Hello 世界 🌍") # Unicode characters typically require multiple tokens assert result > 0 def test_long_text(self): """Test token count for longer text""" long_text = "This is a longer piece of text that should contain multiple sentences. " \ "It will help verify that the token counting works correctly for substantial input." result = num_tokens_from_string(long_text) assert result > 10 def test_whitespace_only(self): """Test token count for whitespace-only strings""" result = num_tokens_from_string(" \n\t ") # Whitespace may or may not be tokens depending on the encoding assert result >= 0 def test_numbers(self): """Test token count with numerical values""" result = num_tokens_from_string("12345 678.90") assert result > 0 def test_mixed_content(self): """Test token count with mixed content types""" mixed_text = "Hello! 123 Main St. Price: $19.99 🎉" result = num_tokens_from_string(mixed_text) assert result > 0 def test_encoding_error_handling(self): """Test that function handles encoding errors gracefully""" # This test verifies the exception handling in the function. # The function should return 0 when encoding fails # Note: We can't easily simulate encoding errors without mocking pass # Additional parameterized tests for efficiency @pytest.mark.parametrize("input_string,expected_min_tokens", [ ("a", 1), # Single character ("test", 1), # Single word ("hello world", 2), # Two words ("This is a sentence.", 4), # Short sentence # ("A" * 100, 100), # Repeated characters ]) def test_token_count_ranges(input_string, expected_min_tokens): """Parameterized test for various input strings""" result = num_tokens_from_string(input_string) assert result >= expected_min_tokens def test_consistency(): """Test that the same input produces consistent results""" test_string = "Consistent token counting" first_result = num_tokens_from_string(test_string) second_result = num_tokens_from_string(test_string) assert first_result == second_result assert first_result > 0 class TestTotalTokenCountFromResponse: """Test cases for total_token_count_from_response function""" def test_dict_with_usage_total_tokens(self): """Test dictionary response with usage['total_tokens']""" resp_dict = { 'usage': { 'total_tokens': 175 } } result = total_token_count_from_response(resp_dict) assert result == 175 def test_dict_with_usage_input_output_tokens(self): """Test dictionary response with input_tokens and output_tokens in usage""" resp_dict = { 'usage': { 'input_tokens': 100, 'output_tokens': 50 } } result = total_token_count_from_response(resp_dict) assert result == 150 def test_dict_with_meta_tokens_input_output(self): """Test dictionary response with meta.tokens.input_tokens and output_tokens""" resp_dict = { 'meta': { 'tokens': { 'input_tokens': 80, 'output_tokens': 40 } } } result = total_token_count_from_response(resp_dict) assert result == 120 def test_priority_order_dict_usage_total_tokens_third(self): """Test that dict['usage']['total_tokens'] is third in priority""" resp_dict = { 'usage': { 'total_tokens': 180, 'input_tokens': 100, 'output_tokens': 80 }, 'meta': { 'tokens': { 'input_tokens': 200, 'output_tokens': 100 } } } result = total_token_count_from_response(resp_dict) assert result == 180 # Should use total_tokens from usage def test_priority_order_dict_usage_input_output_fourth(self): """Test that dict['usage']['input_tokens'] + output_tokens is fourth in priority""" resp_dict = { 'usage': { 'input_tokens': 120, 'output_tokens': 60 }, 'meta': { 'tokens': { 'input_tokens': 200, 'output_tokens': 100 } } } result = total_token_count_from_response(resp_dict) assert result == 180 # Should sum input_tokens + output_tokens from usage def test_priority_order_meta_tokens_last(self): """Test that meta.tokens is the last option in priority""" resp_dict = { 'meta': { 'tokens': { 'input_tokens': 90, 'output_tokens': 30 } } } result = total_token_count_from_response(resp_dict) assert result == 120 def test_no_token_info_returns_zero(self): """Test that function returns 0 when no token information is found""" empty_resp = {} result = total_token_count_from_response(empty_resp) assert result == 0 def test_partial_dict_usage_missing_output_tokens(self): """Test dictionary with usage but missing output_tokens""" resp_dict = { 'usage': { 'input_tokens': 100 # Missing output_tokens } } result = total_token_count_from_response(resp_dict) assert result == 0 # Should not match the condition and return 0 def test_partial_meta_tokens_missing_input_tokens(self): """Test dictionary with meta.tokens but missing input_tokens""" resp_dict = { 'meta': { 'tokens': { 'output_tokens': 50 # Missing input_tokens } } } result = total_token_count_from_response(resp_dict) assert result == 0 # Should not match the condition and return 0 def test_none_response(self): """Test that function handles None response gracefully""" result = total_token_count_from_response(None) assert result == 0 def test_invalid_response_type(self): """Test that function handles invalid response types gracefully""" result = total_token_count_from_response("invalid response") assert result == 0 # result = total_token_count_from_response(123) # assert result == 0 class TestTruncate: """Test cases for truncate function""" def test_empty_string(self): """Test truncation of empty string""" result = truncate("", 5) assert result == "" assert isinstance(result, str) def test_string_shorter_than_max_len(self): """Test string that is shorter than max_len""" original_string = "hello" result = truncate(original_string, 10) assert result == original_string assert len(encoder.encode(result)) <= 10 def test_string_equal_to_max_len(self): """Test string that exactly equals max_len in tokens""" # Create a string that encodes to exactly 5 tokens test_string = "hello world test" encoded = encoder.encode(test_string) exact_length = len(encoded) result = truncate(test_string, exact_length) assert result == test_string assert len(encoder.encode(result)) == exact_length def test_string_longer_than_max_len(self): """Test string that is longer than max_len""" long_string = "This is a longer string that will be truncated" max_len = 5 result = truncate(long_string, max_len) assert len(encoder.encode(result)) == max_len assert result != long_string def test_truncation_preserves_beginning(self): """Test that truncation preserves the beginning of the string""" test_string = "The quick brown fox jumps over the lazy dog" max_len = 3 result = truncate(test_string, max_len) encoded_result = encoder.encode(result) # The truncated result should match the beginning of the original encoding original_encoded = encoder.encode(test_string) assert encoded_result == original_encoded[:max_len] def test_unicode_characters(self): """Test truncation with unicode characters""" unicode_string = "Hello 世界 🌍 测试" max_len = 4 result = truncate(unicode_string, max_len) assert len(encoder.encode(result)) == max_len # Should be a valid string assert isinstance(result, str) def test_special_characters(self): """Test truncation with special characters""" special_string = "Hello, world! @#$%^&*()" max_len = 3 result = truncate(special_string, max_len) assert len(encoder.encode(result)) == max_len def test_whitespace_string(self): """Test truncation of whitespace-only string""" whitespace_string = " \n\t " max_len = 2 result = truncate(whitespace_string, max_len) assert len(encoder.encode(result)) <= max_len assert isinstance(result, str) def test_max_len_zero(self): """Test truncation with max_len = 0""" test_string = "hello world" result = truncate(test_string, 0) assert result == "" assert len(encoder.encode(result)) == 0 def test_max_len_one(self): """Test truncation with max_len = 1""" test_string = "hello world" result = truncate(test_string, 1) assert len(encoder.encode(result)) == 1 def test_preserves_decoding_encoding_consistency(self): """Test that truncation preserves encoding-decoding consistency""" test_string = "This is a test string for encoding consistency" max_len = 6 result = truncate(test_string, max_len) # Re-encoding the result should give the same token count re_encoded = encoder.encode(result) assert len(re_encoded) == max_len def test_multibyte_characters_truncation(self): """Test truncation with multibyte characters that span multiple tokens""" # Some unicode characters may require multiple tokens multibyte_string = "🚀🌟🎉✨🔥💫" max_len = 3 result = truncate(multibyte_string, max_len) assert len(encoder.encode(result)) == max_len def test_mixed_english_chinese_text(self): """Test truncation with mixed English and Chinese text""" mixed_string = "Hello 世界, this is a test 测试" max_len = 5 result = truncate(mixed_string, max_len) assert len(encoder.encode(result)) == max_len def test_numbers_and_symbols(self): """Test truncation with numbers and symbols""" number_string = "12345 678.90 $100.00 @username #tag" max_len = 4 result = truncate(number_string, max_len) assert len(encoder.encode(result)) == max_len
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/unit_test/common/test_string_utils.py
test/unit_test/common/test_string_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import pytest from common.string_utils import remove_redundant_spaces, clean_markdown_block class TestRemoveRedundantSpaces: # Basic punctuation tests @pytest.mark.skip(reason="Failed") def test_remove_spaces_before_commas(self): """Test removing spaces before commas""" input_text = "Hello , world" expected = "Hello, world" assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_remove_spaces_before_periods(self): """Test removing spaces before periods""" input_text = "This is a test ." expected = "This is a test." assert remove_redundant_spaces(input_text) == expected def test_remove_spaces_before_exclamation(self): """Test removing spaces before exclamation marks""" input_text = "Amazing !" expected = "Amazing!" assert remove_redundant_spaces(input_text) == expected def test_remove_spaces_after_opening_parenthesis(self): """Test removing spaces after opening parenthesis""" input_text = "This is ( test)" expected = "This is (test)" assert remove_redundant_spaces(input_text) == expected def test_remove_spaces_before_closing_parenthesis(self): """Test removing spaces before closing parenthesis""" input_text = "This is (test )" expected = "This is (test)" assert remove_redundant_spaces(input_text) == expected def test_keep_spaces_between_words(self): """Test preserving normal spaces between words""" input_text = "This should remain unchanged" expected = "This should remain unchanged" assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_mixed_punctuation(self): """Test mixed punctuation scenarios""" input_text = "Hello , world ! This is ( test ) ." expected = "Hello, world! This is (test)." assert remove_redundant_spaces(input_text) == expected # Numbers and special formats @pytest.mark.skip(reason="Failed") def test_with_numbers(self): """Test handling of numbers""" input_text = "I have 100 , 000 dollars ." expected = "I have 100, 000 dollars." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_decimal_numbers(self): """Test decimal numbers""" input_text = "The value is 3 . 14 ." expected = "The value is 3.14." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_time_format(self): """Test time format handling""" input_text = "Time is 12 : 30 PM ." expected = "Time is 12:30 PM." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_currency_symbols(self): """Test currency symbols""" input_text = "Price : € 100 , £ 50 , ¥ 1000 ." expected = "Price: €100, £50, ¥1000." assert remove_redundant_spaces(input_text) == expected # Edge cases and special characters def test_empty_string(self): """Test empty string input""" assert remove_redundant_spaces("") == "" def test_only_spaces(self): """Test input with only spaces""" input_text = " " expected = " " assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_no_redundant_spaces(self): """Test text without redundant spaces""" input_text = "Hello, world! This is (test)." expected = "Hello, world! This is (test)." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_multiple_spaces(self): """Test multiple consecutive spaces""" input_text = "Hello , world !" expected = "Hello, world!" assert remove_redundant_spaces(input_text) == expected def test_angle_brackets(self): """Test angle brackets handling""" input_text = "This is < test >" expected = "This is <test>" assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_case_insensitive(self): """Test case insensitivity""" input_text = "HELLO , World !" expected = "HELLO, World!" assert remove_redundant_spaces(input_text) == expected # Additional punctuation marks @pytest.mark.skip(reason="Failed") def test_semicolon_and_colon(self): """Test semicolon and colon handling""" input_text = "Items : apple ; banana ; orange ." expected = "Items: apple; banana; orange." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_quotation_marks(self): """Test quotation marks handling""" input_text = 'He said , " Hello " .' expected = 'He said, "Hello".' assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_abbreviations(self): """Test abbreviations""" input_text = "Dr . Smith and Mr . Jones ." expected = "Dr. Smith and Mr. Jones." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_multiple_punctuation(self): """Test multiple consecutive punctuation marks""" input_text = "Wow !! ... Really ??" expected = "Wow!! ... Really??" assert remove_redundant_spaces(input_text) == expected # Special text formats @pytest.mark.skip(reason="Failed") def test_email_addresses(self): """Test email addresses (should not be modified ideally)""" input_text = "Contact me at test @ example . com ." expected = "Contact me at test@example.com." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_urls(self): """Test URLs (might be modified by current function)""" input_text = "Visit https : //example.com / path ." expected = "Visit https://example.com/path." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_hashtags_and_mentions(self): """Test hashtags and mentions""" input_text = "Check out # topic and @ user ." expected = "Check out #topic and @user." assert remove_redundant_spaces(input_text) == expected # Complex structures @pytest.mark.skip(reason="Failed") def test_nested_parentheses(self): """Test nested parentheses""" input_text = "Outer ( inner ( deep ) ) ." expected = "Outer (inner (deep))." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_math_expressions(self): """Test mathematical expressions""" input_text = "Calculate 2 + 2 = 4 ." expected = "Calculate 2 + 2 = 4." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_html_tags(self): """Test HTML tags""" input_text = "< p > This is a paragraph . < / p >" expected = "<p> This is a paragraph. </p>" assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_programming_code(self): """Test programming code snippets""" input_text = "Code : if ( x > 0 ) { print ( 'hello' ) ; }" expected = "Code: if (x > 0) {print ('hello');}" assert remove_redundant_spaces(input_text) == expected # Unicode and special symbols @pytest.mark.skip(reason="Failed") def test_unicode_and_special_symbols(self): """Test Unicode characters and special symbols""" input_text = "Copyright © 2023 , All rights reserved ." expected = "Copyright © 2023, All rights reserved." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_mixed_chinese_english(self): """Test mixed Chinese and English text""" input_text = "你好 , world ! 这是 ( 测试 ) ." expected = "你好, world! 这是 (测试)." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_special_characters_in_pattern(self): """Test special characters in the pattern""" input_text = "Price is $ 100 . 00 , tax included ." expected = "Price is $100.00, tax included." assert remove_redundant_spaces(input_text) == expected @pytest.mark.skip(reason="Failed") def test_tabs_and_newlines(self): """Test tabs and newlines handling""" input_text = "Hello ,\tworld !\nThis is ( test ) ." expected = "Hello,\tworld!\nThis is (test)." assert remove_redundant_spaces(input_text) == expected class TestCleanMarkdownBlock: def test_standard_markdown_block(self): """Test standard Markdown code block syntax""" input_text = "```markdown\nHello world\n```" expected = "Hello world" assert clean_markdown_block(input_text) == expected def test_with_whitespace_variations(self): """Test markdown blocks with various whitespace patterns""" input_text = " ```markdown \n Content here \n ``` " expected = "Content here" assert clean_markdown_block(input_text) == expected def test_multiline_content(self): """Test markdown blocks with multiple lines of content""" input_text = "```markdown\nLine 1\nLine 2\nLine 3\n```" expected = "Line 1\nLine 2\nLine 3" assert clean_markdown_block(input_text) == expected def test_no_opening_newline(self): """Test markdown block without newline after opening tag""" input_text = "```markdownHello world\n```" expected = "Hello world" assert clean_markdown_block(input_text) == expected def test_no_closing_newline(self): """Test markdown block without newline before closing tag""" input_text = "```markdown\nHello world```" expected = "Hello world" assert clean_markdown_block(input_text) == expected def test_empty_markdown_block(self): """Test empty Markdown code block""" input_text = "```markdown\n```" expected = "" assert clean_markdown_block(input_text) == expected def test_only_whitespace_content(self): """Test markdown block containing only whitespace""" input_text = "```markdown\n \n\t\n\n```" expected = "" assert clean_markdown_block(input_text) == expected def test_plain_text_without_markdown(self): """Test text that doesn't contain markdown block syntax""" input_text = "This is plain text without any code blocks" expected = "This is plain text without any code blocks" assert clean_markdown_block(input_text) == expected def test_partial_markdown_syntax(self): """Test text with only opening or closing tags""" input_text = "```markdown\nUnclosed block" expected = "Unclosed block" assert clean_markdown_block(input_text) == expected input_text = "Unopened block\n```" expected = "Unopened block" assert clean_markdown_block(input_text) == expected def test_mixed_whitespace_characters(self): """Test with tabs, spaces, and mixed whitespace""" input_text = "\t```markdown\t\n\tContent with tabs\n\t```\t" expected = "Content with tabs" assert clean_markdown_block(input_text) == expected def test_preserves_internal_whitespace(self): """Test that internal whitespace is preserved""" input_text = "```markdown\n Preserve internal \n whitespace \n```" expected = "Preserve internal \n whitespace" assert clean_markdown_block(input_text) == expected def test_special_characters_content(self): """Test markdown block with special characters""" input_text = "```markdown\n# Header\n**Bold** and *italic*\n```" expected = "# Header\n**Bold** and *italic*" assert clean_markdown_block(input_text) == expected def test_empty_string(self): """Test empty string input""" input_text = "" expected = "" assert clean_markdown_block(input_text) == expected def test_only_markdown_tags(self): """Test input containing only Markdown tags""" input_text = "```markdown```" expected = "" assert clean_markdown_block(input_text) == expected def test_windows_line_endings(self): """Test markdown block with Windows line endings""" input_text = "```markdown\r\nHello world\r\n```" expected = "Hello world" assert clean_markdown_block(input_text) == expected def test_unix_line_endings(self): """Test markdown block with Unix line endings""" input_text = "```markdown\nHello world\n```" expected = "Hello world" assert clean_markdown_block(input_text) == expected def test_nested_code_blocks_preserved(self): """Test that nested code blocks within content are preserved""" input_text = "```markdown\nText with ```nested``` blocks\n```" expected = "Text with ```nested``` blocks" assert clean_markdown_block(input_text) == expected def test_multiple_markdown_blocks(self): """Test behavior with multiple markdown blocks (takes first and last)""" input_text = "```markdown\nFirst line\n```\n```markdown\nSecond line\n```" expected = "First line\n```\n```markdown\nSecond line" assert clean_markdown_block(input_text) == expected
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/unit_test/common/test_decorator.py
test/unit_test/common/test_decorator.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from common.decorator import singleton # Test class for demonstration @singleton class TestClass: def __init__(self): self.counter = 0 def increment(self): self.counter += 1 return self.counter # Test cases class TestSingleton: def test_state_persistence(self): """Test that instance state persists across multiple calls""" instance1 = TestClass() instance1.increment() instance1.increment() instance2 = TestClass() assert instance2.counter == 2 # State should persist def test_multiple_calls_consistency(self): """Test consistency across multiple calls""" instances = [TestClass() for _ in range(5)] # All references should point to the same object first_instance = instances[0] for instance in instances: assert instance is first_instance def test_instance_methods_work(self): """Test that instance methods work correctly""" instance = TestClass() # Test method calls result1 = instance.increment() result2 = instance.increment() assert result1 == 3 assert result2 == 4 assert instance.counter == 4 # Test decorator itself def test_singleton_decorator_returns_callable(): """Test that the decorator returns a callable""" class PlainClass: pass decorated_class = singleton(PlainClass) # Should return a function assert callable(decorated_class) # Calling should return an instance of PlainClass instance = decorated_class() assert isinstance(instance, PlainClass)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/unit_test/common/test_file_utils.py
test/unit_test/common/test_file_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import os import pytest from unittest.mock import patch from common import file_utils from common.file_utils import get_project_base_directory class TestGetProjectBaseDirectory: """Test cases for get_project_base_directory function""" def test_returns_project_base_when_no_args(self): """Test that function returns project base directory when no arguments provided""" result = get_project_base_directory() assert result is not None assert isinstance(result, str) assert os.path.isabs(result) # Should return absolute path def test_returns_path_with_single_argument(self): """Test that function joins project base with single additional path component""" result = get_project_base_directory("subfolder") assert result is not None assert "subfolder" in result assert result.endswith("subfolder") def test_returns_path_with_multiple_arguments(self): """Test that function joins project base with multiple path components""" result = get_project_base_directory("folder1", "folder2", "file.txt") assert result is not None assert "folder1" in result assert "folder2" in result assert "file.txt" in result assert os.path.basename(result) == "file.txt" def test_uses_environment_variable_when_available(self): """Test that function uses RAG_PROJECT_BASE environment variable when set""" test_path = "/custom/project/path" file_utils.PROJECT_BASE = test_path result = get_project_base_directory() assert result == test_path def test_calculates_default_path_when_no_env_vars(self): """Test that function calculates default path when no environment variables are set""" with patch.dict(os.environ, {}, clear=True): # Clear all environment variables # Reset the global variable to force re-initialization result = get_project_base_directory() # Should return a valid absolute path assert result is not None assert os.path.isabs(result) assert os.path.basename(result) != "" # Should not be root directory def test_caches_project_base_value(self): """Test that PROJECT_BASE is cached after first calculation""" # Reset the global variable # First call should calculate the value first_result = get_project_base_directory() # Store the current value cached_value = file_utils.PROJECT_BASE # Second call should use cached value second_result = get_project_base_directory() assert first_result == second_result assert file_utils.PROJECT_BASE == cached_value def test_path_components_joined_correctly(self): """Test that path components are properly joined with the base directory""" base_path = get_project_base_directory() expected_path = os.path.join(base_path, "data", "files", "document.txt") result = get_project_base_directory("data", "files", "document.txt") assert result == expected_path def test_handles_empty_string_arguments(self): """Test that function handles empty string arguments correctly""" result = get_project_base_directory("") # Should still return a valid path (base directory) assert result is not None assert os.path.isabs(result) # Parameterized tests for different path combinations @pytest.mark.parametrize("path_args,expected_suffix", [ ((), ""), # No additional arguments (("src",), "src"), (("data", "models"), os.path.join("data", "models")), (("config", "app", "settings.json"), os.path.join("config", "app", "settings.json")), ]) def test_various_path_combinations(path_args, expected_suffix): """Test various combinations of path arguments""" base_path = get_project_base_directory() result = get_project_base_directory(*path_args) if expected_suffix: assert result.endswith(expected_suffix) else: assert result == base_path
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/unit_test/common/test_time_utils.py
test/unit_test/common/test_time_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import time import datetime import pytest from common.time_utils import current_timestamp, timestamp_to_date, date_string_to_timestamp, datetime_format, delta_seconds class TestCurrentTimestamp: """Test cases for current_timestamp function""" def test_returns_integer(self): """Test that function returns an integer""" result = current_timestamp() assert isinstance(result, int) def test_returns_13_digits(self): """Test that returned timestamp has 13 digits (milliseconds)""" result = current_timestamp() assert len(str(result)) == 13 def test_approximately_correct_value(self): """Test that returned value is approximately correct compared to current time""" # Get timestamps before and after function call for comparison before = int(time.time() * 1000) result = current_timestamp() after = int(time.time() * 1000) assert before <= result <= after def test_consistent_with_time_module(self): """Test that result matches time.time() * 1000 calculation""" expected = int(time.time() * 1000) result = current_timestamp() # Allow small difference due to execution time (typically 1-2ms) assert abs(result - expected) <= 10 def test_multiple_calls_increase(self): """Test that multiple calls return increasing timestamps""" results = [current_timestamp() for _ in range(5)] # Check if timestamps are monotonically increasing # (allow equal values as they might be in the same millisecond) for i in range(1, len(results)): assert results[i] >= results[i - 1] class TestTimestampToDate: """Test cases for timestamp_to_date function""" def test_basic_timestamp_conversion(self): """Test basic timestamp to date conversion with default format""" # Test with a specific timestamp timestamp = 1704067200000 # 2024-01-01 00:00:00 UTC result = timestamp_to_date(timestamp) expected = "2024-01-01 08:00:00" assert result == expected def test_custom_format_string(self): """Test conversion with custom format string""" timestamp = 1704067200000 # 2024-01-01 00:00:00 UTC # Test different format strings result1 = timestamp_to_date(timestamp, "%Y-%m-%d") assert result1 == "2024-01-01" result2 = timestamp_to_date(timestamp, "%H:%M:%S") assert result2 == "08:00:00" result3 = timestamp_to_date(timestamp, "%Y/%m/%d %H:%M") assert result3 == "2024/01/01 08:00" def test_zero_timestamp(self): """Test conversion with zero timestamp (epoch)""" timestamp = 0 # 1970-01-01 00:00:00 UTC result = timestamp_to_date(timestamp) # Note: Actual result depends on local timezone assert isinstance(result, str) assert len(result) > 0 def test_negative_timestamp(self): """Test conversion with negative timestamp (pre-epoch)""" timestamp = -1000000 # Some time before 1970 result = timestamp_to_date(timestamp) assert isinstance(result, str) assert len(result) > 0 def test_string_timestamp_input(self): """Test that string timestamp input is handled correctly""" timestamp_str = "1704067200000" result = timestamp_to_date(timestamp_str) expected = "2024-01-01 08:00:00" assert result == expected def test_float_timestamp_input(self): """Test that float timestamp input is handled correctly""" timestamp_float = 1704067200000.0 result = timestamp_to_date(timestamp_float) expected = "2024-01-01 08:00:00" assert result == expected def test_different_timezones_handled(self): """Test that function handles timezone conversion properly""" timestamp = 1704067200000 # 2024-01-01 00:00:00 UTC # The actual result will depend on the system's local timezone result = timestamp_to_date(timestamp) assert isinstance(result, str) # Should contain date components assert "2024" in result or "08:00:00" in result def test_millisecond_precision(self): """Test that milliseconds are properly handled (truncated)""" # Test timestamp with milliseconds component timestamp = 1704067200123 # 2024-01-01 00:00:00.123 UTC result = timestamp_to_date(timestamp) # Should still return "08:00:00" since milliseconds are truncated assert "08:00:00" in result def test_various_timestamps(self): """Test conversion with various timestamp values""" test_cases = [ (1609459200000, "2021-01-01 08:00:00"), # 2020-12-31 16:00:00 UTC (4102444800000, "2100-01-01"), # Future date ] for timestamp, expected_prefix in test_cases: result = timestamp_to_date(timestamp) assert expected_prefix in result def test_return_type_always_string(self): """Test that return type is always string regardless of input""" test_inputs = [1704067200000, None, "", 0, -1000, "1704067200000"] for timestamp in test_inputs: result = timestamp_to_date(timestamp) assert isinstance(result, str) def test_edge_case_format_strings(self): """Test edge cases with unusual format strings""" timestamp = 1704067200000 # Empty format string result = timestamp_to_date(timestamp, "") assert result == "" # Single character format result = timestamp_to_date(timestamp, "Y") assert isinstance(result, str) # Format with only separators result = timestamp_to_date(timestamp, "---") assert result == "---" class TestDateStringToTimestamp: """Test cases for date_string_to_timestamp function""" def test_basic_date_string_conversion(self): """Test basic date string to timestamp conversion with default format""" date_string = "2024-01-01 08:00:00" result = date_string_to_timestamp(date_string) expected = 1704067200000 assert result == expected def test_custom_format_string(self): """Test conversion with custom format strings""" # Test different date formats test_cases = [ ("2024-01-01", "%Y-%m-%d", 1704038400000), ("2024/01/01 12:30:45", "%Y/%m/%d %H:%M:%S", 1704083445000), ("01-01-2024", "%m-%d-%Y", 1704038400000), ("20240101", "%Y%m%d", 1704038400000), ] for date_string, format_string, expected in test_cases: result = date_string_to_timestamp(date_string, format_string) assert result == expected def test_return_type_integer(self): """Test that function always returns integer""" date_string = "2024-01-01 00:00:00" result = date_string_to_timestamp(date_string) assert isinstance(result, int) def test_timestamp_in_milliseconds(self): """Test that returned timestamp is in milliseconds (13 digits)""" date_string = "2024-01-01 00:00:00" result = date_string_to_timestamp(date_string) assert len(str(result)) == 13 # Verify it's milliseconds by checking it's 1000x larger than seconds timestamp seconds_timestamp = time.mktime(time.strptime(date_string, "%Y-%m-%d %H:%M:%S")) expected_milliseconds = int(seconds_timestamp * 1000) assert result == expected_milliseconds def test_different_dates(self): """Test conversion with various date strings""" test_cases = [ ("2024-01-01 00:00:00", 1704038400000), ("2020-12-31 16:00:00", 1609401600000), ("2023-06-15 14:30:00", 1686810600000), ("2025-12-25 23:59:59", 1766678399000), ] for date_string, expected in test_cases: result = date_string_to_timestamp(date_string) assert result == expected def test_epoch_date(self): """Test conversion with epoch date (1970-01-01)""" # Note: The actual value depends on the local timezone date_string = "1970-01-01 00:00:00" result = date_string_to_timestamp(date_string) assert isinstance(result, int) # Should be a small positive or negative number depending on timezone assert abs(result) < 86400000 # Within 24 hours in milliseconds def test_leap_year_date(self): """Test conversion with leap year date""" date_string = "2024-02-29 12:00:00" # Valid leap year date result = date_string_to_timestamp(date_string) expected = 1709179200000 # 2024-02-29 12:00:00 in milliseconds assert result == expected def test_date_only_string(self): """Test conversion with date-only format (assumes 00:00:00 time)""" date_string = "2024-01-01" result = date_string_to_timestamp(date_string, "%Y-%m-%d") # Should be equivalent to "2024-01-01 00:00:00" expected = 1704038400000 assert result == expected def test_with_whitespace(self): """Test that function handles whitespace properly""" test_cases = [ " 2024-01-01 00:00:00 ", "\t2024-01-01 00:00:00\n", ] for date_string in test_cases: # These should raise ValueError due to extra whitespace with pytest.raises(ValueError): date_string_to_timestamp(date_string) def test_invalid_date_string(self): """Test that invalid date string raises ValueError""" invalid_cases = [ "invalid-date", "2024-13-01 00:00:00", # Invalid month "2024-01-32 00:00:00", # Invalid day "2024-01-01 25:00:00", # Invalid hour "2024-01-01 00:60:00", # Invalid minute "2024-02-30 00:00:00", # Invalid date (Feb 30) ] for invalid_date in invalid_cases: with pytest.raises(ValueError): date_string_to_timestamp(invalid_date) def test_mismatched_format_string(self): """Test that mismatched format string raises ValueError""" test_cases = [ ("2024-01-01 00:00:00", "%Y-%m-%d"), # Missing time in format ("2024-01-01", "%Y-%m-%d %H:%M:%S"), # Missing time in date string ("01/01/2024", "%Y-%m-%d"), # Wrong separator ] for date_string, format_string in test_cases: with pytest.raises(ValueError): date_string_to_timestamp(date_string, format_string) def test_empty_string_input(self): """Test that empty string input raises ValueError""" with pytest.raises(ValueError): date_string_to_timestamp("") def test_none_input(self): """Test that None input raises TypeError""" with pytest.raises(TypeError): date_string_to_timestamp(None) class TestDatetimeFormat: """Test cases for datetime_format function""" def test_remove_microseconds(self): """Test that microseconds are removed from datetime object""" original_dt = datetime.datetime(2024, 1, 1, 12, 30, 45, 123456) result = datetime_format(original_dt) # Verify microseconds are 0 assert result.microsecond == 0 # Verify other components remain the same assert result.year == 2024 assert result.month == 1 assert result.day == 1 assert result.hour == 12 assert result.minute == 30 assert result.second == 45 def test_datetime_with_zero_microseconds(self): """Test datetime that already has zero microseconds""" original_dt = datetime.datetime(2024, 1, 1, 12, 30, 45, 0) result = datetime_format(original_dt) # Should remain the same assert result == original_dt assert result.microsecond == 0 def test_datetime_with_max_microseconds(self): """Test datetime with maximum microseconds value""" original_dt = datetime.datetime(2024, 1, 1, 12, 30, 45, 999999) result = datetime_format(original_dt) # Microseconds should be removed assert result.microsecond == 0 # Other components should remain assert result.year == 2024 assert result.month == 1 assert result.day == 1 assert result.hour == 12 assert result.minute == 30 assert result.second == 45 def test_datetime_with_only_date_components(self): """Test datetime with only date components (time defaults to 00:00:00)""" original_dt = datetime.datetime(2024, 1, 1) result = datetime_format(original_dt) # Should have zero time components and zero microseconds assert result.year == 2024 assert result.month == 1 assert result.day == 1 assert result.hour == 0 assert result.minute == 0 assert result.second == 0 assert result.microsecond == 0 def test_datetime_with_midnight(self): """Test datetime at midnight""" original_dt = datetime.datetime(2024, 1, 1, 0, 0, 0, 123456) result = datetime_format(original_dt) assert result.hour == 0 assert result.minute == 0 assert result.second == 0 assert result.microsecond == 0 def test_datetime_with_end_of_day(self): """Test datetime at end of day (23:59:59)""" original_dt = datetime.datetime(2024, 1, 1, 23, 59, 59, 999999) result = datetime_format(original_dt) assert result.hour == 23 assert result.minute == 59 assert result.second == 59 assert result.microsecond == 0 def test_leap_year_datetime(self): """Test datetime on leap day""" original_dt = datetime.datetime(2024, 2, 29, 14, 30, 15, 500000) result = datetime_format(original_dt) assert result.year == 2024 assert result.month == 2 assert result.day == 29 assert result.hour == 14 assert result.minute == 30 assert result.second == 15 assert result.microsecond == 0 def test_returns_new_object(self): """Test that function returns a new datetime object, not the original""" original_dt = datetime.datetime(2024, 1, 1, 12, 30, 45, 123456) result = datetime_format(original_dt) # Verify it's a different object assert result is not original_dt # Verify original is unchanged assert original_dt.microsecond == 123456 def test_datetime_with_only_seconds(self): """Test datetime with only seconds specified""" original_dt = datetime.datetime(2024, 1, 1, 12, 30, 45) result = datetime_format(original_dt) # Should have zero microseconds assert result.microsecond == 0 # Other components should match assert result == original_dt.replace(microsecond=0) def test_immutability_of_original(self): """Test that original datetime object is not modified""" original_dt = datetime.datetime(2024, 1, 1, 12, 30, 45, 123456) original_microsecond = original_dt.microsecond # Original should remain unchanged assert original_dt.microsecond == original_microsecond assert original_dt.microsecond == 123456 def test_minimum_datetime_value(self): """Test with minimum datetime value""" original_dt = datetime.datetime.min result = datetime_format(original_dt) # Should have zero microseconds assert result.microsecond == 0 # Other components should match assert result.year == original_dt.year assert result.month == original_dt.month assert result.day == original_dt.day def test_maximum_datetime_value(self): """Test with maximum datetime value""" original_dt = datetime.datetime.max result = datetime_format(original_dt) # Should have zero microseconds assert result.microsecond == 0 # Other components should match assert result.year == original_dt.year assert result.month == original_dt.month assert result.day == original_dt.day def test_timezone_naive_datetime(self): """Test with timezone-naive datetime (should remain naive)""" original_dt = datetime.datetime(2024, 1, 1, 12, 30, 45, 123456) result = datetime_format(original_dt) # Should remain timezone-naive assert result.tzinfo is None def test_equality_with_replaced_datetime(self): """Test that result equals datetime.replace(microsecond=0)""" original_dt = datetime.datetime(2024, 1, 1, 12, 30, 45, 123456) result = datetime_format(original_dt) expected = original_dt.replace(microsecond=0) assert result == expected @pytest.mark.parametrize("year,month,day,hour,minute,second,microsecond", [ (2024, 1, 1, 0, 0, 0, 0), # Start of day (2024, 12, 31, 23, 59, 59, 999999), # End of year (2000, 6, 15, 12, 30, 45, 500000), # Random date (1970, 1, 1, 0, 0, 0, 123456), # Epoch equivalent (2030, 3, 20, 6, 15, 30, 750000), # Future date ]) def test_parametrized_datetimes(self, year, month, day, hour, minute, second, microsecond): """Test multiple datetime scenarios using parametrization""" original_dt = datetime.datetime(year, month, day, hour, minute, second, microsecond) result = datetime_format(original_dt) # Verify microseconds are removed assert result.microsecond == 0 # Verify other components remain the same assert result.year == year assert result.month == month assert result.day == day assert result.hour == hour assert result.minute == minute assert result.second == second def test_consistency_across_multiple_calls(self): """Test that multiple calls with same input produce same output""" original_dt = datetime.datetime(2024, 1, 1, 12, 30, 45, 123456) result1 = datetime_format(original_dt) result2 = datetime_format(original_dt) result3 = datetime_format(original_dt) # All results should be equal assert result1 == result2 == result3 # All should have zero microseconds assert result1.microsecond == result2.microsecond == result3.microsecond == 0 def test_type_return(self): """Test that return type is datetime.datetime""" original_dt = datetime.datetime(2024, 1, 1, 12, 30, 45, 123456) result = datetime_format(original_dt) assert isinstance(result, datetime.datetime) class TestDeltaSeconds: """Test cases for delta_seconds function""" def test_zero_seconds_difference(self): """Test when given time equals current time""" # Use a time very close to now to minimize test flakiness now = datetime.datetime.now() date_string = now.strftime("%Y-%m-%d %H:%M:%S") result = delta_seconds(date_string) # Should be very close to 0 assert abs(result) < 1.0 def test_positive_seconds_difference(self): """Test positive time difference (past date)""" now = datetime.datetime.now() past_time = now - datetime.timedelta(hours=1) date_string = past_time.strftime("%Y-%m-%d %H:%M:%S") result = delta_seconds(date_string) # Should be approximately 3600 seconds (1 hour) assert abs(result - 3600.0) < 1.0 def test_negative_seconds_difference(self): """Test negative time difference (future date)""" now = datetime.datetime.now() future_time = now + datetime.timedelta(hours=1) date_string = future_time.strftime("%Y-%m-%d %H:%M:%S") result = delta_seconds(date_string) # Should be approximately -3600 seconds (1 hour) assert abs(result + 3600.0) < 1.0 def test_minutes_difference(self): """Test difference in minutes""" now = datetime.datetime.now() past_time = now - datetime.timedelta(minutes=5) date_string = past_time.strftime("%Y-%m-%d %H:%M:%S") result = delta_seconds(date_string) # Should be approximately 300 seconds (5 minutes) assert abs(result - 300.0) < 1.0 def test_return_type_float(self): """Test that function returns float""" now = datetime.datetime.now() date_string = now.strftime("%Y-%m-%d %H:%M:%S") result = delta_seconds(date_string) assert isinstance(result, float) def test_days_difference(self): """Test difference across multiple days""" now = datetime.datetime.now() past_time = now - datetime.timedelta(days=1) date_string = past_time.strftime("%Y-%m-%d %H:%M:%S") result = delta_seconds(date_string) # Should be approximately 86400 seconds (24 hours) assert abs(result - 86400.0) < 1.0 def test_complex_time_difference(self): """Test complex time difference with all components""" now = datetime.datetime.now() past_time = now - datetime.timedelta(hours=2, minutes=30, seconds=15) date_string = past_time.strftime("%Y-%m-%d %H:%M:%S") result = delta_seconds(date_string) expected = 2 * 3600 + 30 * 60 + 15 # 2 hours + 30 minutes + 15 seconds assert abs(result - expected) < 1.0 def test_invalid_date_format(self): """Test that invalid date format raises ValueError""" invalid_cases = [ "2024-01-01", # Missing time "2024-01-01 12:00", # Missing seconds "2024/01/01 12:00:00", # Wrong date separator "01-01-2024 12:00:00", # Wrong date format "2024-13-01 12:00:00", # Invalid month "2024-01-32 12:00:00", # Invalid day "2024-01-01 25:00:00", # Invalid hour "2024-01-01 12:60:00", # Invalid minute "2024-01-01 12:00:60", # Invalid second "invalid datetime string", # Completely invalid ] for invalid_date in invalid_cases: with pytest.raises(ValueError): delta_seconds(invalid_date) def test_empty_string(self): """Test that empty string raises ValueError""" with pytest.raises(ValueError): delta_seconds("") def test_none_input(self): """Test that None input raises TypeError""" with pytest.raises(TypeError): delta_seconds(None) def test_whitespace_string(self): """Test that whitespace-only string raises ValueError""" with pytest.raises(ValueError): delta_seconds(" ") def test_very_old_date(self): """Test with very old date""" date_string = "2000-01-01 12:00:00" result = delta_seconds(date_string) # Should be a large positive number (many years in seconds) assert result > 0 assert isinstance(result, float) def test_very_future_date(self): """Test with very future date""" date_string = "2030-01-01 12:00:00" result = delta_seconds(date_string) # Should be a large negative number assert result < 0 assert isinstance(result, float) def test_consistency_across_calls(self): """Test that same input produces consistent results""" now = datetime.datetime.now() past_time = now - datetime.timedelta(minutes=10) date_string = past_time.strftime("%Y-%m-%d %H:%M:%S") result1 = delta_seconds(date_string) result2 = delta_seconds(date_string) result3 = delta_seconds(date_string) # All results should be very close (within 0.1 seconds) assert abs(result1 - result2) < 0.1 assert abs(result2 - result3) < 0.1 def test_leap_year_date(self): """Test with leap year date (basic functionality)""" # This test verifies the function can handle leap year dates # without checking specific time differences date_string = "2024-02-29 12:00:00" result = delta_seconds(date_string) assert isinstance(result, float) def test_month_boundary(self): """Test crossing month boundary""" now = datetime.datetime.now() # Use first day of current month at a specific time first_day = datetime.datetime(now.year, now.month, 1, 12, 0, 0) if first_day < now: date_string = first_day.strftime("%Y-%m-%d %H:%M:%S") result = delta_seconds(date_string) assert result > 0 # Should be positive if first_day is in past else: # If we're testing on the first day of month date_string = "2024-01-31 12:00:00" # Use a known past date result = delta_seconds(date_string) assert result > 0
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/unit_test/common/test_misc_utils.py
test/unit_test/common/test_misc_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import uuid import hashlib from common.misc_utils import get_uuid, download_img, hash_str2int, convert_bytes class TestGetUuid: """Test cases for get_uuid function""" def test_returns_string(self): """Test that function returns a string""" result = get_uuid() assert isinstance(result, str) def test_hex_format(self): """Test that returned string is in hex format""" result = get_uuid() # UUID v1 hex should be 32 characters (without dashes) assert len(result) == 32 # Should only contain hexadecimal characters assert all(c in '0123456789abcdef' for c in result) def test_no_dashes_in_result(self): """Test that result contains no dashes""" result = get_uuid() assert '-' not in result def test_unique_results(self): """Test that multiple calls return different UUIDs""" results = [get_uuid() for _ in range(10)] # All results should be unique assert len(results) == len(set(results)) # All should be valid hex strings of correct length for result in results: assert len(result) == 32 assert all(c in '0123456789abcdef' for c in result) def test_valid_uuid_structure(self): """Test that the hex string can be converted back to UUID""" result = get_uuid() # Should be able to create UUID from the hex string reconstructed_uuid = uuid.UUID(hex=result) assert isinstance(reconstructed_uuid, uuid.UUID) # The hex representation should match the original assert reconstructed_uuid.hex == result def test_uuid1_specific_characteristics(self): """Test that UUID v1 characteristics are present""" result = get_uuid() uuid_obj = uuid.UUID(hex=result) # UUID v1 should have version 1 assert uuid_obj.version == 1 # Variant should be RFC 4122 assert uuid_obj.variant == 'specified in RFC 4122' def test_result_length_consistency(self): """Test that all generated UUIDs have consistent length""" for _ in range(100): result = get_uuid() assert len(result) == 32 def test_hex_characters_only(self): """Test that only valid hex characters are used""" for _ in range(100): result = get_uuid() # Should only contain lowercase hex characters (UUID hex is lowercase) assert result.islower() assert all(c in '0123456789abcdef' for c in result) class TestDownloadImg: """Test cases for download_img function""" def test_empty_url_returns_empty_string(self): """Test that empty URL returns empty string""" result = download_img("") assert result == "" def test_none_url_returns_empty_string(self): """Test that None URL returns empty string""" result = download_img(None) assert result == "" class TestHashStr2Int: """Test cases for hash_str2int function""" def test_basic_hashing(self): """Test basic string hashing functionality""" result = hash_str2int("hello") assert isinstance(result, int) assert 0 <= result < 10 ** 8 def test_default_mod_value(self): """Test that default mod value is 10^8""" result = hash_str2int("test") assert 0 <= result < 10 ** 8 def test_custom_mod_value(self): """Test with custom mod value""" result = hash_str2int("test", mod=1000) assert isinstance(result, int) assert 0 <= result < 1000 def test_same_input_same_output(self): """Test that same input produces same output""" result1 = hash_str2int("consistent") result2 = hash_str2int("consistent") result3 = hash_str2int("consistent") assert result1 == result2 == result3 def test_different_input_different_output(self): """Test that different inputs produce different outputs (usually)""" result1 = hash_str2int("hello") result2 = hash_str2int("world") result3 = hash_str2int("hello world") # While hash collisions are possible, they're very unlikely for these inputs results = [result1, result2, result3] assert len(set(results)) == len(results) def test_empty_string(self): """Test hashing empty string""" result = hash_str2int("") assert isinstance(result, int) assert 0 <= result < 10 ** 8 def test_unicode_string(self): """Test hashing unicode strings""" test_strings = [ "中文", "🚀火箭", "café", "🎉", "Hello 世界" ] for test_str in test_strings: result = hash_str2int(test_str) assert isinstance(result, int) assert 0 <= result < 10 ** 8 def test_special_characters(self): """Test hashing strings with special characters""" test_strings = [ "hello@world.com", "test#123", "line\nwith\nnewlines", "tab\tcharacter", "space in string" ] for test_str in test_strings: result = hash_str2int(test_str) assert isinstance(result, int) assert 0 <= result < 10 ** 8 def test_large_string(self): """Test hashing large string""" large_string = "x" * 10000 result = hash_str2int(large_string) assert isinstance(result, int) assert 0 <= result < 10 ** 8 def test_mod_value_1(self): """Test with mod value 1 (should always return 0)""" result = hash_str2int("any string", mod=1) assert result == 0 def test_mod_value_2(self): """Test with mod value 2 (should return 0 or 1)""" result = hash_str2int("test", mod=2) assert result in [0, 1] def test_very_large_mod(self): """Test with very large mod value""" result = hash_str2int("test", mod=10 ** 12) assert isinstance(result, int) assert 0 <= result < 10 ** 12 def test_hash_algorithm_sha1(self): """Test that SHA1 algorithm is used""" test_string = "hello" expected_hash = hashlib.sha1(test_string.encode("utf-8")).hexdigest() expected_int = int(expected_hash, 16) % (10 ** 8) result = hash_str2int(test_string) assert result == expected_int def test_utf8_encoding(self): """Test that UTF-8 encoding is used""" # This should work without encoding errors result = hash_str2int("café 🎉") assert isinstance(result, int) def test_range_with_different_mods(self): """Test that result is always in correct range for different mod values""" test_cases = [ ("test1", 100), ("test2", 1000), ("test3", 10000), ("test4", 999999), ] for test_str, mod_val in test_cases: result = hash_str2int(test_str, mod=mod_val) assert 0 <= result < mod_val def test_hexdigest_conversion(self): """Test the hexdigest to integer conversion""" test_string = "hello" hash_obj = hashlib.sha1(test_string.encode("utf-8")) hex_digest = hash_obj.hexdigest() expected_int = int(hex_digest, 16) % (10 ** 8) result = hash_str2int(test_string) assert result == expected_int def test_consistent_with_direct_calculation(self): """Test that function matches direct hashlib usage""" test_strings = ["a", "b", "abc", "hello world", "12345"] for test_str in test_strings: direct_result = int(hashlib.sha1(test_str.encode("utf-8")).hexdigest(), 16) % (10 ** 8) function_result = hash_str2int(test_str) assert function_result == direct_result def test_numeric_strings(self): """Test hashing numeric strings""" test_strings = ["123", "0", "999999", "3.14159", "-42"] for test_str in test_strings: result = hash_str2int(test_str) assert isinstance(result, int) assert 0 <= result < 10 ** 8 def test_whitespace_strings(self): """Test hashing strings with various whitespace""" test_strings = [ " leading", "trailing ", " both ", "\ttab", "new\nline", "\r\nwindows" ] for test_str in test_strings: result = hash_str2int(test_str) assert isinstance(result, int) assert 0 <= result < 10 ** 8 class TestConvertBytes: """Test suite for convert_bytes function""" def test_zero_bytes(self): """Test that 0 bytes returns '0 B'""" assert convert_bytes(0) == "0 B" def test_single_byte(self): """Test single byte values""" assert convert_bytes(1) == "1 B" assert convert_bytes(999) == "999 B" def test_kilobyte_range(self): """Test values in kilobyte range with different precisions""" # Exactly 1 KB assert convert_bytes(1024) == "1.00 KB" # Values that should show 1 decimal place (10-99.9 range) assert convert_bytes(15360) == "15.0 KB" # 15 KB exactly assert convert_bytes(10752) == "10.5 KB" # 10.5 KB # Values that should show 2 decimal places (1-9.99 range) assert convert_bytes(2048) == "2.00 KB" # 2 KB exactly assert convert_bytes(3072) == "3.00 KB" # 3 KB exactly assert convert_bytes(5120) == "5.00 KB" # 5 KB exactly def test_megabyte_range(self): """Test values in megabyte range""" # Exactly 1 MB assert convert_bytes(1048576) == "1.00 MB" # Values with different precision requirements assert convert_bytes(15728640) == "15.0 MB" # 15.0 MB assert convert_bytes(11010048) == "10.5 MB" # 10.5 MB def test_gigabyte_range(self): """Test values in gigabyte range""" # Exactly 1 GB assert convert_bytes(1073741824) == "1.00 GB" # Large value that should show 0 decimal places assert convert_bytes(3221225472) == "3.00 GB" # 3 GB exactly def test_terabyte_range(self): """Test values in terabyte range""" assert convert_bytes(1099511627776) == "1.00 TB" # 1 TB def test_petabyte_range(self): """Test values in petabyte range""" assert convert_bytes(1125899906842624) == "1.00 PB" # 1 PB def test_boundary_values(self): """Test values at unit boundaries""" # Just below 1 KB assert convert_bytes(1023) == "1023 B" # Just above 1 KB assert convert_bytes(1025) == "1.00 KB" # At 100 KB boundary (should switch to 0 decimal places) assert convert_bytes(102400) == "100 KB" assert convert_bytes(102300) == "99.9 KB" def test_precision_transitions(self): """Test the precision formatting transitions""" # Test transition from 2 decimal places to 1 decimal place assert convert_bytes(9216) == "9.00 KB" # 9.00 KB (2 decimal places) assert convert_bytes(10240) == "10.0 KB" # 10.0 KB (1 decimal place) # Test transition from 1 decimal place to 0 decimal places assert convert_bytes(102400) == "100 KB" # 100 KB (0 decimal places) def test_large_values_no_overflow(self): """Test that very large values don't cause issues""" # Very large value that should use PB large_value = 10 * 1125899906842624 # 10 PB assert "PB" in convert_bytes(large_value) # Ensure we don't exceed available units huge_value = 100 * 1125899906842624 # 100 PB (still within PB range) assert "PB" in convert_bytes(huge_value)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/test/unit_test/common/test_float_utils.py
test/unit_test/common/test_float_utils.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import math from common.float_utils import get_float class TestGetFloat: def test_valid_float_string(self): """Test conversion of valid float strings""" assert get_float("3.14") == 3.14 assert get_float("-2.5") == -2.5 assert get_float("0.0") == 0.0 assert get_float("123.456") == 123.456 def test_valid_integer_string(self): """Test conversion of valid integer strings""" assert get_float("42") == 42.0 assert get_float("-100") == -100.0 assert get_float("0") == 0.0 def test_valid_numbers(self): """Test conversion of actual number types""" assert get_float(3.14) == 3.14 assert get_float(-2.5) == -2.5 assert get_float(42) == 42.0 assert get_float(0) == 0.0 def test_none_input(self): """Test handling of None input""" result = get_float(None) assert math.isinf(result) assert result < 0 # Should be negative infinity def test_invalid_strings(self): """Test handling of invalid string inputs""" result = get_float("invalid") assert math.isinf(result) assert result < 0 result = get_float("12.34.56") assert math.isinf(result) assert result < 0 result = get_float("") assert math.isinf(result) assert result < 0 def test_boolean_input(self): """Test conversion of boolean values""" assert get_float(True) == 1.0 assert get_float(False) == 0.0 def test_special_float_strings(self): """Test handling of special float strings""" assert get_float("inf") == float('inf') assert get_float("-inf") == float('-inf') # NaN should return -inf according to our function's design result = get_float("nan") assert math.isnan(result) def test_very_large_numbers(self): """Test very large number strings""" assert get_float("1e308") == 1e308 # This will become inf in Python, but let's test it large_result = get_float("1e500") assert math.isinf(large_result) def test_whitespace_strings(self): """Test strings with whitespace""" assert get_float(" 3.14 ") == 3.14 result = get_float(" invalid ") assert math.isinf(result) assert result < 0
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/memory/__init__.py
memory/__init__.py
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/memory/services/query.py
memory/services/query.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import re import logging import json import numpy as np from common.query_base import QueryBase from common.doc_store.doc_store_base import MatchDenseExpr, MatchTextExpr from common.float_utils import get_float from rag.nlp import rag_tokenizer, term_weight, synonym def get_vector(txt, emb_mdl, topk=10, similarity=0.1): if isinstance(similarity, str) and len(similarity) > 0: try: similarity = float(similarity) except Exception as e: logging.warning(f"Convert similarity '{similarity}' to float failed: {e}. Using default 0.1") similarity = 0.1 qv, _ = emb_mdl.encode_queries(txt) shape = np.array(qv).shape if len(shape) > 1: raise Exception( f"Dealer.get_vector returned array's shape {shape} doesn't match expectation(exact one dimension).") embedding_data = [get_float(v) for v in qv] vector_column_name = f"q_{len(embedding_data)}_vec" return MatchDenseExpr(vector_column_name, embedding_data, 'float', 'cosine', topk, {"similarity": similarity}) class MsgTextQuery(QueryBase): def __init__(self): self.tw = term_weight.Dealer() self.syn = synonym.Dealer() self.query_fields = [ "content" ] def question(self, txt, tbl="messages", min_match: float=0.6): original_query = txt txt = MsgTextQuery.add_space_between_eng_zh(txt) txt = re.sub( r"[ :|\r\n\t,,。??/`!!&^%%()\[\]{}<>]+", " ", rag_tokenizer.tradi2simp(rag_tokenizer.strQ2B(txt.lower())), ).strip() otxt = txt txt = MsgTextQuery.rmWWW(txt) if not self.is_chinese(txt): txt = self.rmWWW(txt) tks = rag_tokenizer.tokenize(txt).split() keywords = [t for t in tks if t] tks_w = self.tw.weights(tks, preprocess=False) tks_w = [(re.sub(r"[ \\\"'^]", "", tk), w) for tk, w in tks_w] tks_w = [(re.sub(r"^[a-z0-9]$", "", tk), w) for tk, w in tks_w if tk] tks_w = [(re.sub(r"^[\+-]", "", tk), w) for tk, w in tks_w if tk] tks_w = [(tk.strip(), w) for tk, w in tks_w if tk.strip()] syns = [] for tk, w in tks_w[:256]: syn = self.syn.lookup(tk) syn = rag_tokenizer.tokenize(" ".join(syn)).split() keywords.extend(syn) syn = ["\"{}\"^{:.4f}".format(s, w / 4.) for s in syn if s.strip()] syns.append(" ".join(syn)) q = ["({}^{:.4f}".format(tk, w) + " {})".format(syn) for (tk, w), syn in zip(tks_w, syns) if tk and not re.match(r"[.^+\(\)-]", tk)] for i in range(1, len(tks_w)): left, right = tks_w[i - 1][0].strip(), tks_w[i][0].strip() if not left or not right: continue q.append( '"%s %s"^%.4f' % ( tks_w[i - 1][0], tks_w[i][0], max(tks_w[i - 1][1], tks_w[i][1]) * 2, ) ) if not q: q.append(txt) query = " ".join(q) return MatchTextExpr( self.query_fields, query, 100, {"original_query": original_query} ), keywords def need_fine_grained_tokenize(tk): if len(tk) < 3: return False if re.match(r"[0-9a-z\.\+#_\*-]+$", tk): return False return True txt = self.rmWWW(txt) qs, keywords = [], [] for tt in self.tw.split(txt)[:256]: # .split(): if not tt: continue keywords.append(tt) twts = self.tw.weights([tt]) syns = self.syn.lookup(tt) if syns and len(keywords) < 32: keywords.extend(syns) logging.debug(json.dumps(twts, ensure_ascii=False)) tms = [] for tk, w in sorted(twts, key=lambda x: x[1] * -1): sm = ( rag_tokenizer.fine_grained_tokenize(tk).split() if need_fine_grained_tokenize(tk) else [] ) sm = [ re.sub( r"[ ,\./;'\[\]\\`~!@#$%\^&\*\(\)=\+_<>\?:\"\{\}\|,。;‘’【】、!¥……()——《》?:“”-]+", "", m, ) for m in sm ] sm = [self.sub_special_char(m) for m in sm if len(m) > 1] sm = [m for m in sm if len(m) > 1] if len(keywords) < 32: keywords.append(re.sub(r"[ \\\"']+", "", tk)) keywords.extend(sm) tk_syns = self.syn.lookup(tk) tk_syns = [self.sub_special_char(s) for s in tk_syns] if len(keywords) < 32: keywords.extend([s for s in tk_syns if s]) tk_syns = [rag_tokenizer.fine_grained_tokenize(s) for s in tk_syns if s] tk_syns = [f"\"{s}\"" if s.find(" ") > 0 else s for s in tk_syns] if len(keywords) >= 32: break tk = self.sub_special_char(tk) if tk.find(" ") > 0: tk = '"%s"' % tk if tk_syns: tk = f"({tk} OR (%s)^0.2)" % " ".join(tk_syns) if sm: tk = f'{tk} OR "%s" OR ("%s"~2)^0.5' % (" ".join(sm), " ".join(sm)) if tk.strip(): tms.append((tk, w)) tms = " ".join([f"({t})^{w}" for t, w in tms]) if len(twts) > 1: tms += ' ("%s"~2)^1.5' % rag_tokenizer.tokenize(tt) syns = " OR ".join( [ '"%s"' % rag_tokenizer.tokenize(self.sub_special_char(s)) for s in syns ] ) if syns and tms: tms = f"({tms})^5 OR ({syns})^0.7" qs.append(tms) if qs: query = " OR ".join([f"({t})" for t in qs if t]) if not query: query = otxt return MatchTextExpr( self.query_fields, query, 100, {"minimum_should_match": min_match, "original_query": original_query} ), keywords return None, keywords
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/memory/services/__init__.py
memory/services/__init__.py
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/memory/services/messages.py
memory/services/messages.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import sys from typing import List from common import settings from common.doc_store.doc_store_base import OrderByExpr, MatchExpr def index_name(uid: str): return f"memory_{uid}" class MessageService: @classmethod def has_index(cls, uid: str, memory_id: str): index = index_name(uid) return settings.msgStoreConn.index_exist(index, memory_id) @classmethod def create_index(cls, uid: str, memory_id: str, vector_size: int): index = index_name(uid) return settings.msgStoreConn.create_idx(index, memory_id, vector_size) @classmethod def delete_index(cls, uid: str, memory_id: str): index = index_name(uid) return settings.msgStoreConn.delete_idx(index, memory_id) @classmethod def insert_message(cls, messages: List[dict], uid: str, memory_id: str): index = index_name(uid) [m.update({ "id": f'{memory_id}_{m["message_id"]}', "status": 1 if m["status"] else 0 }) for m in messages] return settings.msgStoreConn.insert(messages, index, memory_id) @classmethod def update_message(cls, condition: dict, update_dict: dict, uid: str, memory_id: str): index = index_name(uid) if "status" in update_dict: update_dict["status"] = 1 if update_dict["status"] else 0 return settings.msgStoreConn.update(condition, update_dict, index, memory_id) @classmethod def delete_message(cls, condition: dict, uid: str, memory_id: str): index = index_name(uid) return settings.msgStoreConn.delete(condition, index, memory_id) @classmethod def list_message(cls, uid: str, memory_id: str, agent_ids: List[str]=None, keywords: str=None, page: int=1, page_size: int=50): index = index_name(uid) filter_dict = {} if agent_ids: filter_dict["agent_id"] = agent_ids if keywords: filter_dict["session_id"] = keywords order_by = OrderByExpr() order_by.desc("valid_at") res, total_count = settings.msgStoreConn.search( select_fields=[ "message_id", "message_type", "source_id", "memory_id", "user_id", "agent_id", "session_id", "valid_at", "invalid_at", "forget_at", "status" ], highlight_fields=[], condition=filter_dict, match_expressions=[], order_by=order_by, offset=(page-1)*page_size, limit=page_size, index_names=index, memory_ids=[memory_id], agg_fields=[], hide_forgotten=False ) if not total_count: return { "message_list": [], "total_count": 0 } doc_mapping = settings.msgStoreConn.get_fields(res, [ "message_id", "message_type", "source_id", "memory_id", "user_id", "agent_id", "session_id", "valid_at", "invalid_at", "forget_at", "status" ]) return { "message_list": list(doc_mapping.values()), "total_count": total_count } @classmethod def get_recent_messages(cls, uid_list: List[str], memory_ids: List[str], agent_id: str, session_id: str, limit: int): index_names = [index_name(uid) for uid in uid_list] condition_dict = { "agent_id": agent_id, "session_id": session_id } order_by = OrderByExpr() order_by.desc("valid_at") res, total_count = settings.msgStoreConn.search( select_fields=[ "message_id", "message_type", "source_id", "memory_id", "user_id", "agent_id", "session_id", "valid_at", "invalid_at", "forget_at", "status", "content" ], highlight_fields=[], condition=condition_dict, match_expressions=[], order_by=order_by, offset=0, limit=limit, index_names=index_names, memory_ids=memory_ids, agg_fields=[] ) if not total_count: return [] doc_mapping = settings.msgStoreConn.get_fields(res, [ "message_id", "message_type", "source_id", "memory_id","user_id", "agent_id", "session_id", "valid_at", "invalid_at", "forget_at", "status", "content" ]) return list(doc_mapping.values()) @classmethod def search_message(cls, memory_ids: List[str], condition_dict: dict, uid_list: List[str], match_expressions:list[MatchExpr], top_n: int): index_names = [index_name(uid) for uid in uid_list] # filter only valid messages by default if "status" not in condition_dict: condition_dict["status"] = 1 order_by = OrderByExpr() order_by.desc("valid_at") res, total_count = settings.msgStoreConn.search( select_fields=[ "message_id", "message_type", "source_id", "memory_id", "user_id", "agent_id", "session_id", "valid_at", "invalid_at", "forget_at", "status", "content" ], highlight_fields=[], condition=condition_dict, match_expressions=match_expressions, order_by=order_by, offset=0, limit=top_n, index_names=index_names, memory_ids=memory_ids, agg_fields=[] ) if not total_count: return [] docs = settings.msgStoreConn.get_fields(res, [ "message_id", "message_type", "source_id", "memory_id", "user_id", "agent_id", "session_id", "valid_at", "invalid_at", "forget_at", "status", "content" ]) return list(docs.values()) @staticmethod def calculate_message_size(message: dict): return sys.getsizeof(message["content"]) + sys.getsizeof(message["content_embed"][0]) * len(message["content_embed"]) @classmethod def calculate_memory_size(cls, memory_ids: List[str], uid_list: List[str]): index_names = [index_name(uid) for uid in uid_list] order_by = OrderByExpr() order_by.desc("valid_at") res, count = settings.msgStoreConn.search( select_fields=["memory_id", "content", "content_embed"], highlight_fields=[], condition={}, match_expressions=[], order_by=order_by, offset=0, limit=2048*len(memory_ids), index_names=index_names, memory_ids=memory_ids, agg_fields=[], hide_forgotten=False ) if count == 0: return {} docs = settings.msgStoreConn.get_fields(res, ["memory_id", "content", "content_embed"]) size_dict = {} for doc in docs.values(): if size_dict.get(doc["memory_id"]): size_dict[doc["memory_id"]] += cls.calculate_message_size(doc) else: size_dict[doc["memory_id"]] = cls.calculate_message_size(doc) return size_dict @classmethod def pick_messages_to_delete_by_fifo(cls, memory_id: str, uid: str, size_to_delete: int): select_fields = ["message_id", "content", "content_embed"] _index_name = index_name(uid) res = settings.msgStoreConn.get_forgotten_messages(select_fields, _index_name, memory_id) current_size = 0 ids_to_remove = [] if res: message_list = settings.msgStoreConn.get_fields(res, select_fields) for message in message_list.values(): if current_size < size_to_delete: current_size += cls.calculate_message_size(message) ids_to_remove.append(message["message_id"]) else: return ids_to_remove, current_size if current_size >= size_to_delete: return ids_to_remove, current_size order_by = OrderByExpr() order_by.asc("valid_at") res, total_count = settings.msgStoreConn.search( select_fields=select_fields, highlight_fields=[], condition={}, match_expressions=[], order_by=order_by, offset=0, limit=512, index_names=[_index_name], memory_ids=[memory_id], agg_fields=[] ) docs = settings.msgStoreConn.get_fields(res, select_fields) for doc in docs.values(): if current_size < size_to_delete: current_size += cls.calculate_message_size(doc) ids_to_remove.append(doc["message_id"]) else: return ids_to_remove, current_size return ids_to_remove, current_size @classmethod def get_by_message_id(cls, memory_id: str, message_id: int, uid: str): index = index_name(uid) doc_id = f'{memory_id}_{message_id}' return settings.msgStoreConn.get(doc_id, index, [memory_id]) @classmethod def get_max_message_id(cls, uid_list: List[str], memory_ids: List[str]): order_by = OrderByExpr() order_by.desc("message_id") index_names = [index_name(uid) for uid in uid_list] res, total_count = settings.msgStoreConn.search( select_fields=["message_id"], highlight_fields=[], condition={}, match_expressions=[], order_by=order_by, offset=0, limit=1, index_names=index_names, memory_ids=memory_ids, agg_fields=[], hide_forgotten=False ) if not total_count: return 1 docs = settings.msgStoreConn.get_fields(res, ["message_id"]) if not docs: return 1 else: latest_msg = list(docs.values())[0] return int(latest_msg["message_id"])
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/memory/utils/infinity_conn.py
memory/utils/infinity_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import re import json import copy from infinity.common import InfinityException, SortType from infinity.errors import ErrorCode from common.decorator import singleton import pandas as pd from common.doc_store.doc_store_base import MatchExpr, MatchTextExpr, MatchDenseExpr, FusionExpr, OrderByExpr from common.doc_store.infinity_conn_base import InfinityConnectionBase from common.time_utils import date_string_to_timestamp @singleton class InfinityConnection(InfinityConnectionBase): def __init__(self): super().__init__(mapping_file_name="message_infinity_mapping.json") """ Dataframe and fields convert """ @staticmethod def field_keyword(field_name: str): # no keywords right now return False @staticmethod def convert_message_field_to_infinity(field_name: str, table_fields: list[str]=None): match field_name: case "message_type": return "message_type_kwd" case "status": return "status_int" case "content_embed": if not table_fields: raise Exception("Can't convert 'content_embed' to vector field name with empty table fields.") vector_field = [tf for tf in table_fields if re.match(r"q_\d+_vec", tf)] if not vector_field: raise Exception("Can't convert 'content_embed' to vector field name. No match field name found.") return vector_field[0] case _: return field_name @staticmethod def convert_infinity_field_to_message(field_name: str): if field_name.startswith("message_type"): return "message_type" if field_name.startswith("status"): return "status" if re.match(r"q_\d+_vec", field_name): return "content_embed" return field_name def convert_select_fields(self, output_fields: list[str], table_fields: list[str]=None) -> list[str]: return list({self.convert_message_field_to_infinity(f, table_fields) for f in output_fields}) @staticmethod def convert_matching_field(field_weight_str: str) -> str: tokens = field_weight_str.split("^") field = tokens[0] if field == "content": field = "content@ft_content_rag_fine" tokens[0] = field return "^".join(tokens) @staticmethod def convert_condition_and_order_field(field_name: str): match field_name: case "message_type": return "message_type_kwd" case "status": return "status_int" case "valid_at": return "valid_at_flt" case "invalid_at": return "invalid_at_flt" case "forget_at": return "forget_at_flt" case _: return field_name """ CRUD operations """ def search( self, select_fields: list[str], highlight_fields: list[str], condition: dict, match_expressions: list[MatchExpr], order_by: OrderByExpr, offset: int, limit: int, index_names: str | list[str], memory_ids: list[str], agg_fields: list[str] | None = None, rank_feature: dict | None = None, hide_forgotten: bool = True, ) -> tuple[pd.DataFrame, int]: """ BUG: Infinity returns empty for a highlight field if the query string doesn't use that field. """ if isinstance(index_names, str): index_names = index_names.split(",") assert isinstance(index_names, list) and len(index_names) > 0 inf_conn = self.connPool.get_conn() db_instance = inf_conn.get_database(self.dbName) df_list = list() table_list = list() if hide_forgotten: condition.update({"must_not": {"exists": "forget_at_flt"}}) output = select_fields.copy() if agg_fields is None: agg_fields = [] for essential_field in ["id"] + agg_fields: if essential_field not in output: output.append(essential_field) score_func = "" score_column = "" for matchExpr in match_expressions: if isinstance(matchExpr, MatchTextExpr): score_func = "score()" score_column = "SCORE" break if not score_func: for matchExpr in match_expressions: if isinstance(matchExpr, MatchDenseExpr): score_func = "similarity()" score_column = "SIMILARITY" break if match_expressions: if score_func not in output: output.append(score_func) output = [f for f in output if f != "_score"] if limit <= 0: # ElasticSearch default limit is 10000 limit = 10000 # Prepare expressions common to all tables filter_cond = None filter_fulltext = "" if condition: condition_dict = {self.convert_condition_and_order_field(k): v for k, v in condition.items()} table_found = False for indexName in index_names: for mem_id in memory_ids: table_name = f"{indexName}_{mem_id}" try: filter_cond = self.equivalent_condition_to_str(condition_dict, db_instance.get_table(table_name)) table_found = True break except Exception: pass if table_found: break if not table_found: self.logger.error(f"No valid tables found for indexNames {index_names} and memoryIds {memory_ids}") return pd.DataFrame(), 0 for matchExpr in match_expressions: if isinstance(matchExpr, MatchTextExpr): if filter_cond and "filter" not in matchExpr.extra_options: matchExpr.extra_options.update({"filter": filter_cond}) matchExpr.fields = [self.convert_matching_field(field) for field in matchExpr.fields] fields = ",".join(matchExpr.fields) filter_fulltext = f"filter_fulltext('{fields}', '{matchExpr.matching_text}')" if filter_cond: filter_fulltext = f"({filter_cond}) AND {filter_fulltext}" minimum_should_match = matchExpr.extra_options.get("minimum_should_match", 0.0) if isinstance(minimum_should_match, float): str_minimum_should_match = str(int(minimum_should_match * 100)) + "%" matchExpr.extra_options["minimum_should_match"] = str_minimum_should_match for k, v in matchExpr.extra_options.items(): if not isinstance(v, str): matchExpr.extra_options[k] = str(v) self.logger.debug(f"INFINITY search MatchTextExpr: {json.dumps(matchExpr.__dict__)}") elif isinstance(matchExpr, MatchDenseExpr): if filter_fulltext and "filter" not in matchExpr.extra_options: matchExpr.extra_options.update({"filter": filter_fulltext}) for k, v in matchExpr.extra_options.items(): if not isinstance(v, str): matchExpr.extra_options[k] = str(v) similarity = matchExpr.extra_options.get("similarity") if similarity: matchExpr.extra_options["threshold"] = similarity del matchExpr.extra_options["similarity"] self.logger.debug(f"INFINITY search MatchDenseExpr: {json.dumps(matchExpr.__dict__)}") elif isinstance(matchExpr, FusionExpr): if matchExpr.method == "weighted_sum": # The default is "minmax" which gives a zero score for the last doc. matchExpr.fusion_params["normalize"] = "atan" self.logger.debug(f"INFINITY search FusionExpr: {json.dumps(matchExpr.__dict__)}") order_by_expr_list = list() if order_by.fields: for order_field in order_by.fields: order_field_name = self.convert_condition_and_order_field(order_field[0]) if order_field[1] == 0: order_by_expr_list.append((order_field_name, SortType.Asc)) else: order_by_expr_list.append((order_field_name, SortType.Desc)) total_hits_count = 0 # Scatter search tables and gather the results column_name_list = [] for indexName in index_names: for memory_id in memory_ids: table_name = f"{indexName}_{memory_id}" try: table_instance = db_instance.get_table(table_name) except Exception: continue table_list.append(table_name) if not column_name_list: column_name_list = [r[0] for r in table_instance.show_columns().rows()] output = self.convert_select_fields(output, column_name_list) builder = table_instance.output(output) if len(match_expressions) > 0: for matchExpr in match_expressions: if isinstance(matchExpr, MatchTextExpr): fields = ",".join(matchExpr.fields) builder = builder.match_text( fields, matchExpr.matching_text, matchExpr.topn, matchExpr.extra_options.copy(), ) elif isinstance(matchExpr, MatchDenseExpr): builder = builder.match_dense( matchExpr.vector_column_name, matchExpr.embedding_data, matchExpr.embedding_data_type, matchExpr.distance_type, matchExpr.topn, matchExpr.extra_options.copy(), ) elif isinstance(matchExpr, FusionExpr): builder = builder.fusion(matchExpr.method, matchExpr.topn, matchExpr.fusion_params) else: if filter_cond and len(filter_cond) > 0: builder.filter(filter_cond) if order_by.fields: builder.sort(order_by_expr_list) builder.offset(offset).limit(limit) mem_res, extra_result = builder.option({"total_hits_count": True}).to_df() if extra_result: total_hits_count += int(extra_result["total_hits_count"]) self.logger.debug(f"INFINITY search table: {str(table_name)}, result: {str(mem_res)}") df_list.append(mem_res) self.connPool.release_conn(inf_conn) res = self.concat_dataframes(df_list, output) if match_expressions: res["_score"] = res[score_column] res = res.sort_values(by="_score", ascending=False).reset_index(drop=True) res = res.head(limit) self.logger.debug(f"INFINITY search final result: {str(res)}") return res, total_hits_count def get_forgotten_messages(self, select_fields: list[str], index_name: str, memory_id: str, limit: int=512): condition = {"memory_id": memory_id, "exists": "forget_at_flt"} order_by = OrderByExpr() order_by.asc("forget_at_flt") # query inf_conn = self.connPool.get_conn() db_instance = inf_conn.get_database(self.dbName) table_name = f"{index_name}_{memory_id}" table_instance = db_instance.get_table(table_name) column_name_list = [r[0] for r in table_instance.show_columns().rows()] output_fields = [self.convert_message_field_to_infinity(f, column_name_list) for f in select_fields] builder = table_instance.output(output_fields) filter_cond = self.equivalent_condition_to_str(condition, db_instance.get_table(table_name)) builder.filter(filter_cond) order_by_expr_list = list() if order_by.fields: for order_field in order_by.fields: order_field_name = self.convert_condition_and_order_field(order_field[0]) if order_field[1] == 0: order_by_expr_list.append((order_field_name, SortType.Asc)) else: order_by_expr_list.append((order_field_name, SortType.Desc)) builder.sort(order_by_expr_list) builder.offset(0).limit(limit) mem_res, _ = builder.option({"total_hits_count": True}).to_df() res = self.concat_dataframes(mem_res, output_fields) res.head(limit) self.connPool.release_conn(inf_conn) return res def get(self, message_id: str, index_name: str, memory_ids: list[str]) -> dict | None: inf_conn = self.connPool.get_conn() db_instance = inf_conn.get_database(self.dbName) df_list = list() assert isinstance(memory_ids, list) table_list = list() for memoryId in memory_ids: table_name = f"{index_name}_{memoryId}" table_list.append(table_name) try: table_instance = db_instance.get_table(table_name) except Exception: self.logger.warning(f"Table not found: {table_name}, this memory isn't created in Infinity. Maybe it is created in other document engine.") continue mem_res, _ = table_instance.output(["*"]).filter(f"id = '{message_id}'").to_df() self.logger.debug(f"INFINITY get table: {str(table_list)}, result: {str(mem_res)}") df_list.append(mem_res) self.connPool.release_conn(inf_conn) res = self.concat_dataframes(df_list, ["id"]) fields = set(res.columns.tolist()) res_fields = self.get_fields(res, list(fields)) return {self.convert_infinity_field_to_message(k): v for k, v in res_fields[message_id].items()} if res_fields.get(message_id) else {} def insert(self, documents: list[dict], index_name: str, memory_id: str = None) -> list[str]: if not documents: return [] inf_conn = self.connPool.get_conn() db_instance = inf_conn.get_database(self.dbName) table_name = f"{index_name}_{memory_id}" vector_size = int(len(documents[0]["content_embed"])) try: table_instance = db_instance.get_table(table_name) except InfinityException as e: # src/common/status.cppm, kTableNotExist = 3022 if e.error_code != ErrorCode.TABLE_NOT_EXIST: raise if vector_size == 0: raise ValueError("Cannot infer vector size from documents") self.create_idx(index_name, memory_id, vector_size) table_instance = db_instance.get_table(table_name) # embedding fields can't have a default value.... embedding_columns = [] table_columns = table_instance.show_columns().rows() for n, ty, _, _ in table_columns: r = re.search(r"Embedding\([a-z]+,([0-9]+)\)", ty) if not r: continue embedding_columns.append((n, int(r.group(1)))) docs = copy.deepcopy(documents) for d in docs: assert "_id" not in d assert "id" in d for k, v in list(d.items()): if k == "content_embed": d[f"q_{vector_size}_vec"] = d["content_embed"] d.pop("content_embed") continue field_name = self.convert_message_field_to_infinity(k) if field_name in ["valid_at", "invalid_at", "forget_at"]: d[f"{field_name}_flt"] = date_string_to_timestamp(v) if v else 0 if v is None: d[field_name] = "" elif self.field_keyword(k): if isinstance(v, list): d[k] = "###".join(v) else: d[k] = v elif k == "memory_id": if isinstance(d[k], list): d[k] = d[k][0] # since d[k] is a list, but we need a str else: d[field_name] = v if k != field_name: d.pop(k) for n, vs in embedding_columns: if n in d: continue d[n] = [0] * vs ids = ["'{}'".format(d["id"]) for d in docs] str_ids = ", ".join(ids) str_filter = f"id IN ({str_ids})" table_instance.delete(str_filter) table_instance.insert(docs) self.connPool.release_conn(inf_conn) self.logger.debug(f"INFINITY inserted into {table_name} {str_ids}.") return [] def update(self, condition: dict, new_value: dict, index_name: str, memory_id: str) -> bool: inf_conn = self.connPool.get_conn() db_instance = inf_conn.get_database(self.dbName) table_name = f"{index_name}_{memory_id}" table_instance = db_instance.get_table(table_name) columns = {} if table_instance: for n, ty, de, _ in table_instance.show_columns().rows(): columns[n] = (ty, de) condition_dict = {self.convert_condition_and_order_field(k): v for k, v in condition.items()} filter = self.equivalent_condition_to_str(condition_dict, table_instance) update_dict = {self.convert_message_field_to_infinity(k): v for k, v in new_value.items()} date_floats = {} for k, v in update_dict.items(): if k in ["valid_at", "invalid_at", "forget_at"]: date_floats[f"{k}_flt"] = date_string_to_timestamp(v) if v else 0 elif self.field_keyword(k): if isinstance(v, list): update_dict[k] = "###".join(v) else: update_dict[k] = v elif k == "memory_id": if isinstance(update_dict[k], list): update_dict[k] = update_dict[k][0] # since d[k] is a list, but we need a str else: update_dict[k] = v if date_floats: update_dict.update(date_floats) self.logger.debug(f"INFINITY update table {table_name}, filter {filter}, newValue {new_value}.") table_instance.update(filter, update_dict) self.connPool.release_conn(inf_conn) return True """ Helper functions for search result """ def get_fields(self, res: tuple[pd.DataFrame, int] | pd.DataFrame, fields: list[str]) -> dict[str, dict]: if isinstance(res, tuple): res_df = res[0] else: res_df = res if not fields: return {} fields_all = fields.copy() fields_all.append("id") fields_all = self.convert_select_fields(fields_all, res_df.columns.tolist()) column_map = {col.lower(): col for col in res_df.columns} matched_columns = {column_map[col.lower()]: col for col in fields_all if col.lower() in column_map} none_columns = [col for col in fields_all if col.lower() not in column_map] selected_res = res_df[matched_columns.keys()] selected_res = selected_res.rename(columns=matched_columns) selected_res.drop_duplicates(subset=["id"], inplace=True) for column in list(selected_res.columns): k = column.lower() if self.field_keyword(k): selected_res[column] = selected_res[column].apply(lambda v: [kwd for kwd in v.split("###") if kwd]) else: pass for column in none_columns: selected_res[column] = None res_dict = selected_res.set_index("id").to_dict(orient="index") return {_id: {self.convert_infinity_field_to_message(k): v for k, v in doc.items()} for _id, doc in res_dict.items()}
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/memory/utils/prompt_util.py
memory/utils/prompt_util.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from typing import Optional, List from common.constants import MemoryType from common.time_utils import current_timestamp class PromptAssembler: SYSTEM_BASE_TEMPLATE = """**Memory Extraction Specialist** You are an expert at analyzing conversations to extract structured memory. {type_specific_instructions} **OUTPUT REQUIREMENTS:** 1. Output MUST be valid JSON 2. Follow the specified output format exactly 3. Each extracted item MUST have: content, valid_at, invalid_at 4. Timestamps in {timestamp_format} format 5. Only extract memory types specified above 6. Maximum {max_items} items per type """ TYPE_INSTRUCTIONS = { MemoryType.SEMANTIC.name.lower(): """ **EXTRACT SEMANTIC KNOWLEDGE:** - Universal facts, definitions, concepts, relationships - Time-invariant, generally true information - Examples: "The capital of France is Paris", "Water boils at 100°C" **Timestamp Rules for Semantic Knowledge:** - valid_at: When the fact became true (e.g., law enactment, discovery) - invalid_at: When it becomes false (e.g., repeal, disproven) or empty if still true - Default: valid_at = conversation time, invalid_at = "" for timeless facts """, MemoryType.EPISODIC.name.lower(): """ **EXTRACT EPISODIC KNOWLEDGE:** - Specific experiences, events, personal stories - Time-bound, person-specific, contextual - Examples: "Yesterday I fixed the bug", "User reported issue last week" **Timestamp Rules for Episodic Knowledge:** - valid_at: Event start/occurrence time - invalid_at: Event end time or empty if instantaneous - Extract explicit times: "at 3 PM", "last Monday", "from X to Y" """, MemoryType.PROCEDURAL.name.lower(): """ **EXTRACT PROCEDURAL KNOWLEDGE:** - Processes, methods, step-by-step instructions - Goal-oriented, actionable, often includes conditions - Examples: "To reset password, click...", "Debugging steps: 1)..." **Timestamp Rules for Procedural Knowledge:** - valid_at: When procedure becomes valid/effective - invalid_at: When it expires/becomes obsolete or empty if current - For version-specific: use release dates - For best practices: invalid_at = "" """ } OUTPUT_TEMPLATES = { MemoryType.SEMANTIC.name.lower(): """ "semantic": [ { "content": "Clear factual statement", "valid_at": "timestamp or empty", "invalid_at": "timestamp or empty" } ] """, MemoryType.EPISODIC.name.lower(): """ "episodic": [ { "content": "Narrative event description", "valid_at": "event start timestamp", "invalid_at": "event end timestamp or empty" } ] """, MemoryType.PROCEDURAL.name.lower(): """ "procedural": [ { "content": "Actionable instructions", "valid_at": "procedure effective timestamp", "invalid_at": "procedure expiration timestamp or empty" } ] """ } BASE_USER_PROMPT = """ **CONVERSATION:** {conversation} **CONVERSATION TIME:** {conversation_time} **CURRENT TIME:** {current_time} """ @classmethod def assemble_system_prompt(cls, config: dict) -> str: types_to_extract = cls._get_types_to_extract(config["memory_type"]) type_instructions = cls._generate_type_instructions(types_to_extract) output_format = cls._generate_output_format(types_to_extract) full_prompt = cls.SYSTEM_BASE_TEMPLATE.format( type_specific_instructions=type_instructions, timestamp_format=config.get("timestamp_format", "ISO 8601"), max_items=config.get("max_items_per_type", 5) ) full_prompt += f"\n**REQUIRED OUTPUT FORMAT (JSON):**\n```json\n{{\n{output_format}\n}}\n```\n" examples = cls._generate_examples(types_to_extract) if examples: full_prompt += f"\n**EXAMPLES:**\n{examples}\n" return full_prompt @staticmethod def _get_types_to_extract(requested_types: List[str]) -> List[str]: types = set() for rt in requested_types: if rt in [e.name.lower() for e in MemoryType] and rt != MemoryType.RAW.name.lower(): types.add(rt) return list(types) @classmethod def _generate_type_instructions(cls, types_to_extract: List[str]) -> str: target_types = set(types_to_extract) instructions = [cls.TYPE_INSTRUCTIONS[mt] for mt in target_types] return "\n".join(instructions) @classmethod def _generate_output_format(cls, types_to_extract: List[str]) -> str: target_types = set(types_to_extract) output_parts = [cls.OUTPUT_TEMPLATES[mt] for mt in target_types] return ",\n".join(output_parts) @staticmethod def _generate_examples(types_to_extract: list[str]) -> str: examples = [] if MemoryType.SEMANTIC.name.lower() in types_to_extract: examples.append(""" **Semantic Example:** Input: "Python lists are mutable and support various operations." Output: {"semantic": [{"content": "Python lists are mutable data structures", "valid_at": "2024-01-15T10:00:00", "invalid_at": ""}]} """) if MemoryType.EPISODIC.name.lower() in types_to_extract: examples.append(""" **Episodic Example:** Input: "I deployed the new feature yesterday afternoon." Output: {"episodic": [{"content": "User deployed new feature", "valid_at": "2024-01-14T14:00:00", "invalid_at": "2024-01-14T18:00:00"}]} """) if MemoryType.PROCEDURAL.name.lower() in types_to_extract: examples.append(""" **Procedural Example:** Input: "To debug API errors: 1) Check logs 2) Verify endpoints 3) Test connectivity." Output: {"procedural": [{"content": "API error debugging: 1. Check logs 2. Verify endpoints 3. Test connectivity", "valid_at": "2024-01-15T10:00:00", "invalid_at": ""}]} """) return "\n".join(examples) @classmethod def assemble_user_prompt( cls, conversation: str, conversation_time: Optional[str] = None, current_time: Optional[str] = None ) -> str: return cls.BASE_USER_PROMPT.format( conversation=conversation, conversation_time=conversation_time or "Not specified", current_time=current_time or current_timestamp(), ) @classmethod def get_raw_user_prompt(cls): return cls.BASE_USER_PROMPT
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/memory/utils/msg_util.py
memory/utils/msg_util.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json def get_json_result_from_llm_response(response_str: str) -> dict: """ Parse the LLM response string to extract JSON content. The function looks for the first and last curly braces to identify the JSON part. If parsing fails, it returns an empty dictionary. :param response_str: The response string from the LLM. :return: A dictionary parsed from the JSON content in the response. """ try: clean_str = response_str.strip() if clean_str.startswith('```json'): clean_str = clean_str[7:] # Remove the starting ```json if clean_str.endswith('```'): clean_str = clean_str[:-3] # Remove the ending ``` return json.loads(clean_str.strip()) except (ValueError, json.JSONDecodeError): return {}
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/memory/utils/es_conn.py
memory/utils/es_conn.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import re import json import time import copy from elasticsearch import NotFoundError from elasticsearch_dsl import UpdateByQuery, Q, Search from elastic_transport import ConnectionTimeout from common.decorator import singleton from common.doc_store.doc_store_base import MatchExpr, OrderByExpr, MatchTextExpr, MatchDenseExpr, FusionExpr from common.doc_store.es_conn_base import ESConnectionBase from common.float_utils import get_float from common.constants import PAGERANK_FLD, TAG_FLD ATTEMPT_TIME = 2 @singleton class ESConnection(ESConnectionBase): @staticmethod def convert_field_name(field_name: str) -> str: match field_name: case "message_type": return "message_type_kwd" case "status": return "status_int" case "content": return "content_ltks" case _: return field_name @staticmethod def map_message_to_es_fields(message: dict) -> dict: """ Map message dictionary fields to Elasticsearch document/Infinity fields. :param message: A dictionary containing message details. :return: A dictionary formatted for Elasticsearch/Infinity indexing. """ storage_doc = { "id": message.get("id"), "message_id": message["message_id"], "message_type_kwd": message["message_type"], "source_id": message["source_id"], "memory_id": message["memory_id"], "user_id": message["user_id"], "agent_id": message["agent_id"], "session_id": message["session_id"], "valid_at": message["valid_at"], "invalid_at": message["invalid_at"], "forget_at": message["forget_at"], "status_int": 1 if message["status"] else 0, "zone_id": message.get("zone_id", 0), "content_ltks": message["content"], f"q_{len(message['content_embed'])}_vec": message["content_embed"], } return storage_doc @staticmethod def get_message_from_es_doc(doc: dict) -> dict: """ Convert an Elasticsearch/Infinity document back to a message dictionary. :param doc: A dictionary representing the Elasticsearch/Infinity document. :return: A dictionary formatted as a message. """ embd_field_name = next((key for key in doc.keys() if re.match(r"q_\d+_vec", key)), None) message = { "message_id": doc["message_id"], "message_type": doc["message_type_kwd"], "source_id": doc["source_id"] if doc["source_id"] else None, "memory_id": doc["memory_id"], "user_id": doc.get("user_id", ""), "agent_id": doc["agent_id"], "session_id": doc["session_id"], "zone_id": doc.get("zone_id", 0), "valid_at": doc["valid_at"], "invalid_at": doc.get("invalid_at", "-"), "forget_at": doc.get("forget_at", "-"), "status": bool(int(doc["status_int"])), "content": doc.get("content_ltks", ""), "content_embed": doc.get(embd_field_name, []) if embd_field_name else [], } if doc.get("id"): message["id"] = doc["id"] return message """ CRUD operations """ def search( self, select_fields: list[str], highlight_fields: list[str], condition: dict, match_expressions: list[MatchExpr], order_by: OrderByExpr, offset: int, limit: int, index_names: str | list[str], memory_ids: list[str], agg_fields: list[str] | None = None, rank_feature: dict | None = None, hide_forgotten: bool = True ): """ Refers to https://www.elastic.co/guide/en/elasticsearch/reference/current/query-dsl.html """ if isinstance(index_names, str): index_names = index_names.split(",") assert isinstance(index_names, list) and len(index_names) > 0 assert "_id" not in condition exist_index_list = [idx for idx in index_names if self.index_exist(idx)] if not exist_index_list: return None, 0 bool_query = Q("bool", must=[], must_not=[]) if hide_forgotten: # filter not forget bool_query.must_not.append(Q("exists", field="forget_at")) condition["memory_id"] = memory_ids for k, v in condition.items(): field_name = self.convert_field_name(k) if field_name == "session_id" and v: bool_query.filter.append(Q("query_string", **{"query": f"*{v}*", "fields": ["session_id"], "analyze_wildcard": True})) continue if not v: continue if isinstance(v, list): bool_query.filter.append(Q("terms", **{field_name: v})) elif isinstance(v, str) or isinstance(v, int): bool_query.filter.append(Q("term", **{field_name: v})) else: raise Exception( f"Condition `{str(k)}={str(v)}` value type is {str(type(v))}, expected to be int, str or list.") s = Search() vector_similarity_weight = 0.5 for m in match_expressions: if isinstance(m, FusionExpr) and m.method == "weighted_sum" and "weights" in m.fusion_params: assert len(match_expressions) == 3 and isinstance(match_expressions[0], MatchTextExpr) and isinstance(match_expressions[1], MatchDenseExpr) and isinstance( match_expressions[2], FusionExpr) weights = m.fusion_params["weights"] vector_similarity_weight = get_float(weights.split(",")[1]) for m in match_expressions: if isinstance(m, MatchTextExpr): minimum_should_match = m.extra_options.get("minimum_should_match", 0.0) if isinstance(minimum_should_match, float): minimum_should_match = str(int(minimum_should_match * 100)) + "%" bool_query.must.append(Q("query_string", fields=[self.convert_field_name(f) for f in m.fields], type="best_fields", query=m.matching_text, minimum_should_match=minimum_should_match, boost=1)) bool_query.boost = 1.0 - vector_similarity_weight elif isinstance(m, MatchDenseExpr): assert (bool_query is not None) similarity = 0.0 if "similarity" in m.extra_options: similarity = m.extra_options["similarity"] s = s.knn(self.convert_field_name(m.vector_column_name), m.topn, m.topn * 2, query_vector=list(m.embedding_data), filter=bool_query.to_dict(), similarity=similarity, ) if bool_query and rank_feature: for fld, sc in rank_feature.items(): if fld != PAGERANK_FLD: fld = f"{TAG_FLD}.{fld}" bool_query.should.append(Q("rank_feature", field=fld, linear={}, boost=sc)) if bool_query: s = s.query(bool_query) for field in highlight_fields: s = s.highlight(field) if order_by: orders = list() for field, order in order_by.fields: order = "asc" if order == 0 else "desc" if field.endswith("_int") or field.endswith("_flt"): order_info = {"order": order, "unmapped_type": "float"} else: order_info = {"order": order, "unmapped_type": "text"} orders.append({field: order_info}) s = s.sort(*orders) if agg_fields: for fld in agg_fields: s.aggs.bucket(f'aggs_{fld}', 'terms', field=fld, size=1000000) if limit > 0: s = s[offset:offset + limit] q = s.to_dict() self.logger.debug(f"ESConnection.search {str(index_names)} query: " + json.dumps(q)) for i in range(ATTEMPT_TIME): try: #print(json.dumps(q, ensure_ascii=False)) res = self.es.search(index=exist_index_list, body=q, timeout="600s", # search_type="dfs_query_then_fetch", track_total_hits=True, _source=True) if str(res.get("timed_out", "")).lower() == "true": raise Exception("Es Timeout.") self.logger.debug(f"ESConnection.search {str(index_names)} res: " + str(res)) return res, self.get_total(res) except ConnectionTimeout: self.logger.exception("ES request timeout") self._connect() continue except NotFoundError as e: self.logger.debug(f"ESConnection.search {str(index_names)} query: " + str(q) + str(e)) return None, 0 except Exception as e: self.logger.exception(f"ESConnection.search {str(index_names)} query: " + str(q) + str(e)) raise e self.logger.error(f"ESConnection.search timeout for {ATTEMPT_TIME} times!") raise Exception("ESConnection.search timeout.") def get_forgotten_messages(self, select_fields: list[str], index_name: str, memory_id: str, limit: int=512): bool_query = Q("bool", must=[]) bool_query.must.append(Q("exists", field="forget_at")) bool_query.filter.append(Q("term", memory_id=memory_id)) # from old to new order_by = OrderByExpr() order_by.asc("forget_at") # build search s = Search() s = s.query(bool_query) orders = list() for field, order in order_by.fields: order = "asc" if order == 0 else "desc" if field.endswith("_int") or field.endswith("_flt"): order_info = {"order": order, "unmapped_type": "float"} else: order_info = {"order": order, "unmapped_type": "text"} orders.append({field: order_info}) s = s.sort(*orders) s = s[:limit] q = s.to_dict() # search for i in range(ATTEMPT_TIME): try: res = self.es.search(index=index_name, body=q, timeout="600s", track_total_hits=True, _source=True) if str(res.get("timed_out", "")).lower() == "true": raise Exception("Es Timeout.") self.logger.debug(f"ESConnection.search {str(index_name)} res: " + str(res)) return res except ConnectionTimeout: self.logger.exception("ES request timeout") self._connect() continue except NotFoundError as e: self.logger.debug(f"ESConnection.search {str(index_name)} query: " + str(q) + str(e)) return None except Exception as e: self.logger.exception(f"ESConnection.search {str(index_name)} query: " + str(q) + str(e)) raise e self.logger.error(f"ESConnection.search timeout for {ATTEMPT_TIME} times!") raise Exception("ESConnection.search timeout.") def get(self, doc_id: str, index_name: str, memory_ids: list[str]) -> dict | None: for i in range(ATTEMPT_TIME): try: res = self.es.get(index=index_name, id=doc_id, source=True, ) if str(res.get("timed_out", "")).lower() == "true": raise Exception("Es Timeout.") message = res["_source"] message["id"] = doc_id return self.get_message_from_es_doc(message) except NotFoundError: return None except Exception as e: self.logger.exception(f"ESConnection.get({doc_id}) got exception") raise e self.logger.error(f"ESConnection.get timeout for {ATTEMPT_TIME} times!") raise Exception("ESConnection.get timeout.") def insert(self, documents: list[dict], index_name: str, memory_id: str = None) -> list[str]: # Refers to https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-bulk.html operations = [] for d in documents: assert "_id" not in d assert "id" in d d_copy_raw = copy.deepcopy(d) d_copy = self.map_message_to_es_fields(d_copy_raw) d_copy["memory_id"] = memory_id meta_id = d_copy.pop("id", "") operations.append( {"index": {"_index": index_name, "_id": meta_id}}) operations.append(d_copy) res = [] for _ in range(ATTEMPT_TIME): try: res = [] r = self.es.bulk(index=index_name, operations=operations, refresh=False, timeout="60s") if re.search(r"False", str(r["errors"]), re.IGNORECASE): return res for item in r["items"]: for action in ["create", "delete", "index", "update"]: if action in item and "error" in item[action]: res.append(str(item[action]["_id"]) + ":" + str(item[action]["error"])) return res except ConnectionTimeout: self.logger.exception("ES request timeout") time.sleep(3) self._connect() continue except Exception as e: res.append(str(e)) self.logger.warning("ESConnection.insert got exception: " + str(e)) return res def update(self, condition: dict, new_value: dict, index_name: str, memory_id: str) -> bool: doc = copy.deepcopy(new_value) update_dict = {self.convert_field_name(k): v for k, v in doc.items()} update_dict.pop("id", None) condition_dict = {self.convert_field_name(k): v for k, v in condition.items()} condition_dict["memory_id"] = memory_id if "id" in condition_dict and isinstance(condition_dict["id"], str): # update specific single document message_id = condition_dict["id"] for i in range(ATTEMPT_TIME): for k in update_dict.keys(): if "feas" != k.split("_")[-1]: continue try: self.es.update(index=index_name, id=message_id, script=f"ctx._source.remove(\"{k}\");") except Exception: self.logger.exception(f"ESConnection.update(index={index_name}, id={message_id}, doc={json.dumps(condition, ensure_ascii=False)}) got exception") try: self.es.update(index=index_name, id=message_id, doc=update_dict) return True except Exception as e: self.logger.exception( f"ESConnection.update(index={index_name}, id={message_id}, doc={json.dumps(condition, ensure_ascii=False)}) got exception: " + str(e)) break return False # update unspecific maybe-multiple documents bool_query = Q("bool") for k, v in condition_dict.items(): if not isinstance(k, str) or not v: continue if k == "exists": bool_query.filter.append(Q("exists", field=v)) continue if isinstance(v, list): bool_query.filter.append(Q("terms", **{k: v})) elif isinstance(v, str) or isinstance(v, int): bool_query.filter.append(Q("term", **{k: v})) else: raise Exception( f"Condition `{str(k)}={str(v)}` value type is {str(type(v))}, expected to be int, str or list.") scripts = [] params = {} for k, v in update_dict.items(): if k == "remove": if isinstance(v, str): scripts.append(f"ctx._source.remove('{v}');") if isinstance(v, dict): for kk, vv in v.items(): scripts.append(f"int i=ctx._source.{kk}.indexOf(params.p_{kk});ctx._source.{kk}.remove(i);") params[f"p_{kk}"] = vv continue if k == "add": if isinstance(v, dict): for kk, vv in v.items(): scripts.append(f"ctx._source.{kk}.add(params.pp_{kk});") params[f"pp_{kk}"] = vv.strip() continue if (not isinstance(k, str) or not v) and k != "status_int": continue if isinstance(v, str): v = re.sub(r"(['\n\r]|\\.)", " ", v) params[f"pp_{k}"] = v scripts.append(f"ctx._source.{k}=params.pp_{k};") elif isinstance(v, int) or isinstance(v, float): scripts.append(f"ctx._source.{k}={v};") elif isinstance(v, list): scripts.append(f"ctx._source.{k}=params.pp_{k};") params[f"pp_{k}"] = json.dumps(v, ensure_ascii=False) else: raise Exception( f"newValue `{str(k)}={str(v)}` value type is {str(type(v))}, expected to be int, str.") ubq = UpdateByQuery( index=index_name).using( self.es).query(bool_query) ubq = ubq.script(source="".join(scripts), params=params) ubq = ubq.params(refresh=True) ubq = ubq.params(slices=5) ubq = ubq.params(conflicts="proceed") for _ in range(ATTEMPT_TIME): try: _ = ubq.execute() return True except ConnectionTimeout: self.logger.exception("ES request timeout") time.sleep(3) self._connect() continue except Exception as e: self.logger.error("ESConnection.update got exception: " + str(e) + "\n".join(scripts)) break return False def delete(self, condition: dict, index_name: str, memory_id: str) -> int: assert "_id" not in condition condition_dict = {self.convert_field_name(k): v for k, v in condition.items()} condition_dict["memory_id"] = memory_id if "id" in condition_dict: message_ids = condition_dict["id"] if not isinstance(message_ids, list): message_ids = [message_ids] if not message_ids: # when message_ids is empty, delete all qry = Q("match_all") else: qry = Q("ids", values=message_ids) else: qry = Q("bool") for k, v in condition_dict.items(): if k == "exists": qry.filter.append(Q("exists", field=v)) elif k == "must_not": if isinstance(v, dict): for kk, vv in v.items(): if kk == "exists": qry.must_not.append(Q("exists", field=vv)) elif isinstance(v, list): qry.must.append(Q("terms", **{k: v})) elif isinstance(v, str) or isinstance(v, int): qry.must.append(Q("term", **{k: v})) else: raise Exception("Condition value must be int, str or list.") self.logger.debug("ESConnection.delete query: " + json.dumps(qry.to_dict())) for _ in range(ATTEMPT_TIME): try: res = self.es.delete_by_query( index=index_name, body=Search().query(qry).to_dict(), refresh=True) return res["deleted"] except ConnectionTimeout: self.logger.exception("ES request timeout") time.sleep(3) self._connect() continue except Exception as e: self.logger.warning("ESConnection.delete got exception: " + str(e)) if re.search(r"(not_found)", str(e), re.IGNORECASE): return 0 return 0 """ Helper functions for search result """ def get_fields(self, res, fields: list[str]) -> dict[str, dict]: res_fields = {} if not fields: return {} for doc in self._get_source(res): message = self.get_message_from_es_doc(doc) m = {} for n, v in message.items(): if n not in fields: continue if isinstance(v, list): m[n] = v continue if n in ["message_id", "source_id", "valid_at", "invalid_at", "forget_at", "status"] and isinstance(v, (int, float, bool)): m[n] = v continue if not isinstance(v, str): m[n] = str(v) else: m[n] = v if m: res_fields[doc["id"]] = m return res_fields
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/memory/utils/__init__.py
memory/utils/__init__.py
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/ragflow_server.py
api/ragflow_server.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # from beartype import BeartypeConf # from beartype.claw import beartype_all # <-- you didn't sign up for this # beartype_all(conf=BeartypeConf(violation_type=UserWarning)) # <-- emit warnings from all code from common.log_utils import init_root_logger from plugin import GlobalPluginManager import logging import os import signal import sys import traceback import threading import uuid import faulthandler from api.apps import app from api.db.runtime_config import RuntimeConfig from api.db.services.document_service import DocumentService from common.file_utils import get_project_base_directory from common import settings from api.db.db_models import init_database_tables as init_web_db from api.db.init_data import init_web_data, init_superuser from common.versions import get_ragflow_version from common.config_utils import show_configs from common.mcp_tool_call_conn import shutdown_all_mcp_sessions from rag.utils.redis_conn import RedisDistributedLock stop_event = threading.Event() RAGFLOW_DEBUGPY_LISTEN = int(os.environ.get('RAGFLOW_DEBUGPY_LISTEN', "0")) def update_progress(): lock_value = str(uuid.uuid4()) redis_lock = RedisDistributedLock("update_progress", lock_value=lock_value, timeout=60) logging.info(f"update_progress lock_value: {lock_value}") while not stop_event.is_set(): try: if redis_lock.acquire(): DocumentService.update_progress() redis_lock.release() except Exception: logging.exception("update_progress exception") finally: try: redis_lock.release() except Exception: logging.exception("update_progress exception") stop_event.wait(6) def signal_handler(sig, frame): logging.info("Received interrupt signal, shutting down...") shutdown_all_mcp_sessions() stop_event.set() stop_event.wait(1) sys.exit(0) if __name__ == '__main__': faulthandler.enable() init_root_logger("ragflow_server") logging.info(r""" ____ ___ ______ ______ __ / __ \ / | / ____// ____// /____ _ __ / /_/ // /| | / / __ / /_ / // __ \| | /| / / / _, _// ___ |/ /_/ // __/ / // /_/ /| |/ |/ / /_/ |_|/_/ |_|\____//_/ /_/ \____/ |__/|__/ """) logging.info( f'RAGFlow version: {get_ragflow_version()}' ) logging.info( f'project base: {get_project_base_directory()}' ) show_configs() settings.init_settings() settings.print_rag_settings() if RAGFLOW_DEBUGPY_LISTEN > 0: logging.info(f"debugpy listen on {RAGFLOW_DEBUGPY_LISTEN}") import debugpy debugpy.listen(("0.0.0.0", RAGFLOW_DEBUGPY_LISTEN)) # init db init_web_db() init_web_data() # init runtime config import argparse parser = argparse.ArgumentParser() parser.add_argument( "--version", default=False, help="RAGFlow version", action="store_true" ) parser.add_argument( "--debug", default=False, help="debug mode", action="store_true" ) parser.add_argument( "--init-superuser", default=False, help="init superuser", action="store_true" ) args = parser.parse_args() if args.version: print(get_ragflow_version()) sys.exit(0) if args.init_superuser: init_superuser() RuntimeConfig.DEBUG = args.debug if RuntimeConfig.DEBUG: logging.info("run on debug mode") RuntimeConfig.init_env() RuntimeConfig.init_config(JOB_SERVER_HOST=settings.HOST_IP, HTTP_PORT=settings.HOST_PORT) GlobalPluginManager.load_plugins() signal.signal(signal.SIGINT, signal_handler) signal.signal(signal.SIGTERM, signal_handler) def delayed_start_update_progress(): logging.info("Starting update_progress thread (delayed)") t = threading.Thread(target=update_progress, daemon=True) t.start() if RuntimeConfig.DEBUG: if os.environ.get("WERKZEUG_RUN_MAIN") == "true": threading.Timer(1.0, delayed_start_update_progress).start() else: threading.Timer(1.0, delayed_start_update_progress).start() # start http server try: logging.info("RAGFlow HTTP server start...") app.run(host=settings.HOST_IP, port=settings.HOST_PORT) except Exception: traceback.print_exc() stop_event.set() stop_event.wait(1) os.kill(os.getpid(), signal.SIGKILL)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/constants.py
api/constants.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. NAME_LENGTH_LIMIT = 2**10 IMG_BASE64_PREFIX = "data:image/png;base64," API_VERSION = "v1" RAG_FLOW_SERVICE_NAME = "ragflow" REQUEST_WAIT_SEC = 2 REQUEST_MAX_WAIT_SEC = 300 DATASET_NAME_LIMIT = 128 FILE_NAME_LEN_LIMIT = 255 MEMORY_NAME_LIMIT = 128 MEMORY_SIZE_LIMIT = 10*1024*1024 # Byte
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/validation.py
api/validation.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import sys def python_version_validation(): # Check python version required_python_version = (3, 10) if sys.version_info < required_python_version: logging.info( f"Required Python: >= {required_python_version[0]}.{required_python_version[1]}. Current Python version: {sys.version_info[0]}.{sys.version_info[1]}." ) sys.exit(1) else: logging.info(f"Python version: {sys.version_info[0]}.{sys.version_info[1]}") python_version_validation() # Download nltk data def download_nltk_data(): import nltk nltk.download('wordnet', halt_on_error=False, quiet=True) nltk.download('punkt_tab', halt_on_error=False, quiet=True) try: from multiprocessing import Pool pool = Pool(processes=1) thread = pool.apply_async(download_nltk_data) binary = thread.get(timeout=60) except Exception: print('\x1b[6;37;41m WARNING \x1b[0m' + "Downloading NLTK data failure.", flush=True)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/settings.py
api/settings.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/__init__.py
api/__init__.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # from beartype.claw import beartype_this_package # beartype_this_package()
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/file2document_app.py
api/apps/file2document_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License # from pathlib import Path from api.db.services.file2document_service import File2DocumentService from api.db.services.file_service import FileService from api.apps import login_required, current_user from api.db.services.knowledgebase_service import KnowledgebaseService from api.utils.api_utils import get_data_error_result, get_json_result, get_request_json, server_error_response, validate_request from common.misc_utils import get_uuid from common.constants import RetCode from api.db import FileType from api.db.services.document_service import DocumentService @manager.route('/convert', methods=['POST']) # noqa: F821 @login_required @validate_request("file_ids", "kb_ids") async def convert(): req = await get_request_json() kb_ids = req["kb_ids"] file_ids = req["file_ids"] file2documents = [] try: files = FileService.get_by_ids(file_ids) files_set = dict({file.id: file for file in files}) for file_id in file_ids: file = files_set[file_id] if not file: return get_data_error_result(message="File not found!") file_ids_list = [file_id] if file.type == FileType.FOLDER.value: file_ids_list = FileService.get_all_innermost_file_ids(file_id, []) for id in file_ids_list: informs = File2DocumentService.get_by_file_id(id) # delete for inform in informs: doc_id = inform.document_id e, doc = DocumentService.get_by_id(doc_id) if not e: return get_data_error_result(message="Document not found!") tenant_id = DocumentService.get_tenant_id(doc_id) if not tenant_id: return get_data_error_result(message="Tenant not found!") if not DocumentService.remove_document(doc, tenant_id): return get_data_error_result( message="Database error (Document removal)!") File2DocumentService.delete_by_file_id(id) # insert for kb_id in kb_ids: e, kb = KnowledgebaseService.get_by_id(kb_id) if not e: return get_data_error_result( message="Can't find this dataset!") e, file = FileService.get_by_id(id) if not e: return get_data_error_result( message="Can't find this file!") doc = DocumentService.insert({ "id": get_uuid(), "kb_id": kb.id, "parser_id": kb.parser_id, "pipeline_id": kb.pipeline_id, "parser_config": kb.parser_config, "created_by": current_user.id, "type": file.type, "name": file.name, "suffix": Path(file.name).suffix.lstrip("."), "location": file.location, "size": file.size }) file2document = File2DocumentService.insert({ "id": get_uuid(), "file_id": id, "document_id": doc.id, }) file2documents.append(file2document.to_json()) return get_json_result(data=file2documents) except Exception as e: return server_error_response(e) @manager.route('/rm', methods=['POST']) # noqa: F821 @login_required @validate_request("file_ids") async def rm(): req = await get_request_json() file_ids = req["file_ids"] if not file_ids: return get_json_result( data=False, message='Lack of "Files ID"', code=RetCode.ARGUMENT_ERROR) try: for file_id in file_ids: informs = File2DocumentService.get_by_file_id(file_id) if not informs: return get_data_error_result(message="Inform not found!") for inform in informs: if not inform: return get_data_error_result(message="Inform not found!") File2DocumentService.delete_by_file_id(file_id) doc_id = inform.document_id e, doc = DocumentService.get_by_id(doc_id) if not e: return get_data_error_result(message="Document not found!") tenant_id = DocumentService.get_tenant_id(doc_id) if not tenant_id: return get_data_error_result(message="Tenant not found!") if not DocumentService.remove_document(doc, tenant_id): return get_data_error_result( message="Database error (Document removal)!") return get_json_result(data=True) except Exception as e: return server_error_response(e)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/canvas_app.py
api/apps/canvas_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import asyncio import inspect import json import logging from functools import partial from quart import request, Response, make_response from agent.component import LLM from api.db import CanvasCategory from api.db.services.canvas_service import CanvasTemplateService, UserCanvasService, API4ConversationService from api.db.services.document_service import DocumentService from api.db.services.file_service import FileService from api.db.services.pipeline_operation_log_service import PipelineOperationLogService from api.db.services.task_service import queue_dataflow, CANVAS_DEBUG_DOC_ID, TaskService from api.db.services.user_service import TenantService from api.db.services.user_canvas_version import UserCanvasVersionService from common.constants import RetCode from common.misc_utils import get_uuid from api.utils.api_utils import get_json_result, server_error_response, validate_request, get_data_error_result, \ get_request_json from agent.canvas import Canvas from peewee import MySQLDatabase, PostgresqlDatabase from api.db.db_models import APIToken, Task import time from rag.flow.pipeline import Pipeline from rag.nlp import search from rag.utils.redis_conn import REDIS_CONN from common import settings from api.apps import login_required, current_user @manager.route('/templates', methods=['GET']) # noqa: F821 @login_required def templates(): return get_json_result(data=[c.to_dict() for c in CanvasTemplateService.get_all()]) @manager.route('/rm', methods=['POST']) # noqa: F821 @validate_request("canvas_ids") @login_required async def rm(): req = await get_request_json() for i in req["canvas_ids"]: if not UserCanvasService.accessible(i, current_user.id): return get_json_result( data=False, message='Only owner of canvas authorized for this operation.', code=RetCode.OPERATING_ERROR) UserCanvasService.delete_by_id(i) return get_json_result(data=True) @manager.route('/set', methods=['POST']) # noqa: F821 @validate_request("dsl", "title") @login_required async def save(): req = await get_request_json() if not isinstance(req["dsl"], str): req["dsl"] = json.dumps(req["dsl"], ensure_ascii=False) req["dsl"] = json.loads(req["dsl"]) cate = req.get("canvas_category", CanvasCategory.Agent) if "id" not in req: req["user_id"] = current_user.id if UserCanvasService.query(user_id=current_user.id, title=req["title"].strip(), canvas_category=cate): return get_data_error_result(message=f"{req['title'].strip()} already exists.") req["id"] = get_uuid() if not UserCanvasService.save(**req): return get_data_error_result(message="Fail to save canvas.") else: if not UserCanvasService.accessible(req["id"], current_user.id): return get_json_result( data=False, message='Only owner of canvas authorized for this operation.', code=RetCode.OPERATING_ERROR) UserCanvasService.update_by_id(req["id"], req) # save version UserCanvasVersionService.insert(user_canvas_id=req["id"], dsl=req["dsl"], title="{0}_{1}".format(req["title"], time.strftime("%Y_%m_%d_%H_%M_%S"))) UserCanvasVersionService.delete_all_versions(req["id"]) return get_json_result(data=req) @manager.route('/get/<canvas_id>', methods=['GET']) # noqa: F821 @login_required def get(canvas_id): if not UserCanvasService.accessible(canvas_id, current_user.id): return get_data_error_result(message="canvas not found.") e, c = UserCanvasService.get_by_canvas_id(canvas_id) return get_json_result(data=c) @manager.route('/getsse/<canvas_id>', methods=['GET']) # type: ignore # noqa: F821 def getsse(canvas_id): token = request.headers.get('Authorization').split() if len(token) != 2: return get_data_error_result(message='Authorization is not valid!"') token = token[1] objs = APIToken.query(beta=token) if not objs: return get_data_error_result(message='Authentication error: API key is invalid!"') tenant_id = objs[0].tenant_id if not UserCanvasService.query(user_id=tenant_id, id=canvas_id): return get_json_result( data=False, message='Only owner of canvas authorized for this operation.', code=RetCode.OPERATING_ERROR ) e, c = UserCanvasService.get_by_id(canvas_id) if not e or c.user_id != tenant_id: return get_data_error_result(message="canvas not found.") return get_json_result(data=c.to_dict()) @manager.route('/completion', methods=['POST']) # noqa: F821 @validate_request("id") @login_required async def run(): req = await get_request_json() query = req.get("query", "") files = req.get("files", []) inputs = req.get("inputs", {}) user_id = req.get("user_id", current_user.id) if not await asyncio.to_thread(UserCanvasService.accessible, req["id"], current_user.id): return get_json_result( data=False, message='Only owner of canvas authorized for this operation.', code=RetCode.OPERATING_ERROR) e, cvs = await asyncio.to_thread(UserCanvasService.get_by_id, req["id"]) if not e: return get_data_error_result(message="canvas not found.") if not isinstance(cvs.dsl, str): cvs.dsl = json.dumps(cvs.dsl, ensure_ascii=False) if cvs.canvas_category == CanvasCategory.DataFlow: task_id = get_uuid() Pipeline(cvs.dsl, tenant_id=current_user.id, doc_id=CANVAS_DEBUG_DOC_ID, task_id=task_id, flow_id=req["id"]) ok, error_message = await asyncio.to_thread(queue_dataflow, user_id, req["id"], task_id, CANVAS_DEBUG_DOC_ID, files[0], 0) if not ok: return get_data_error_result(message=error_message) return get_json_result(data={"message_id": task_id}) try: canvas = Canvas(cvs.dsl, current_user.id, canvas_id=cvs.id) except Exception as e: return server_error_response(e) async def sse(): nonlocal canvas, user_id try: async for ans in canvas.run(query=query, files=files, user_id=user_id, inputs=inputs): yield "data:" + json.dumps(ans, ensure_ascii=False) + "\n\n" cvs.dsl = json.loads(str(canvas)) UserCanvasService.update_by_id(req["id"], cvs.to_dict()) except Exception as e: logging.exception(e) canvas.cancel_task() yield "data:" + json.dumps({"code": 500, "message": str(e), "data": False}, ensure_ascii=False) + "\n\n" resp = Response(sse(), mimetype="text/event-stream") resp.headers.add_header("Cache-control", "no-cache") resp.headers.add_header("Connection", "keep-alive") resp.headers.add_header("X-Accel-Buffering", "no") resp.headers.add_header("Content-Type", "text/event-stream; charset=utf-8") #resp.call_on_close(lambda: canvas.cancel_task()) return resp @manager.route('/rerun', methods=['POST']) # noqa: F821 @validate_request("id", "dsl", "component_id") @login_required async def rerun(): req = await get_request_json() doc = PipelineOperationLogService.get_documents_info(req["id"]) if not doc: return get_data_error_result(message="Document not found.") doc = doc[0] if 0 < doc["progress"] < 1: return get_data_error_result(message=f"`{doc['name']}` is processing...") if settings.docStoreConn.index_exist(search.index_name(current_user.id), doc["kb_id"]): settings.docStoreConn.delete({"doc_id": doc["id"]}, search.index_name(current_user.id), doc["kb_id"]) doc["progress_msg"] = "" doc["chunk_num"] = 0 doc["token_num"] = 0 DocumentService.clear_chunk_num_when_rerun(doc["id"]) DocumentService.update_by_id(id, doc) TaskService.filter_delete([Task.doc_id == id]) dsl = req["dsl"] dsl["path"] = [req["component_id"]] PipelineOperationLogService.update_by_id(req["id"], {"dsl": dsl}) queue_dataflow(tenant_id=current_user.id, flow_id=req["id"], task_id=get_uuid(), doc_id=doc["id"], priority=0, rerun=True) return get_json_result(data=True) @manager.route('/cancel/<task_id>', methods=['PUT']) # noqa: F821 @login_required def cancel(task_id): try: REDIS_CONN.set(f"{task_id}-cancel", "x") except Exception as e: logging.exception(e) return get_json_result(data=True) @manager.route('/reset', methods=['POST']) # noqa: F821 @validate_request("id") @login_required async def reset(): req = await get_request_json() if not UserCanvasService.accessible(req["id"], current_user.id): return get_json_result( data=False, message='Only owner of canvas authorized for this operation.', code=RetCode.OPERATING_ERROR) try: e, user_canvas = UserCanvasService.get_by_id(req["id"]) if not e: return get_data_error_result(message="canvas not found.") canvas = Canvas(json.dumps(user_canvas.dsl), current_user.id, canvas_id=user_canvas.id) canvas.reset() req["dsl"] = json.loads(str(canvas)) UserCanvasService.update_by_id(req["id"], {"dsl": req["dsl"]}) return get_json_result(data=req["dsl"]) except Exception as e: return server_error_response(e) @manager.route("/upload/<canvas_id>", methods=["POST"]) # noqa: F821 async def upload(canvas_id): e, cvs = UserCanvasService.get_by_canvas_id(canvas_id) if not e: return get_data_error_result(message="canvas not found.") user_id = cvs["user_id"] files = await request.files file = files['file'] if files and files.get("file") else None try: return get_json_result(data=FileService.upload_info(user_id, file, request.args.get("url"))) except Exception as e: return server_error_response(e) @manager.route('/input_form', methods=['GET']) # noqa: F821 @login_required def input_form(): cvs_id = request.args.get("id") cpn_id = request.args.get("component_id") try: e, user_canvas = UserCanvasService.get_by_id(cvs_id) if not e: return get_data_error_result(message="canvas not found.") if not UserCanvasService.query(user_id=current_user.id, id=cvs_id): return get_json_result( data=False, message='Only owner of canvas authorized for this operation.', code=RetCode.OPERATING_ERROR) canvas = Canvas(json.dumps(user_canvas.dsl), current_user.id, canvas_id=user_canvas.id) return get_json_result(data=canvas.get_component_input_form(cpn_id)) except Exception as e: return server_error_response(e) @manager.route('/debug', methods=['POST']) # noqa: F821 @validate_request("id", "component_id", "params") @login_required async def debug(): req = await get_request_json() if not UserCanvasService.accessible(req["id"], current_user.id): return get_json_result( data=False, message='Only owner of canvas authorized for this operation.', code=RetCode.OPERATING_ERROR) try: e, user_canvas = UserCanvasService.get_by_id(req["id"]) canvas = Canvas(json.dumps(user_canvas.dsl), current_user.id, canvas_id=user_canvas.id) canvas.reset() canvas.message_id = get_uuid() component = canvas.get_component(req["component_id"])["obj"] component.reset() if isinstance(component, LLM): component.set_debug_inputs(req["params"]) component.invoke(**{k: o["value"] for k,o in req["params"].items()}) outputs = component.output() for k in outputs.keys(): if isinstance(outputs[k], partial): txt = "" iter_obj = outputs[k]() if inspect.isasyncgen(iter_obj): async for c in iter_obj: txt += c else: for c in iter_obj: txt += c outputs[k] = txt return get_json_result(data=outputs) except Exception as e: return server_error_response(e) @manager.route('/test_db_connect', methods=['POST']) # noqa: F821 @validate_request("db_type", "database", "username", "host", "port", "password") @login_required async def test_db_connect(): req = await get_request_json() try: if req["db_type"] in ["mysql", "mariadb"]: db = MySQLDatabase(req["database"], user=req["username"], host=req["host"], port=req["port"], password=req["password"]) elif req["db_type"] == 'postgres': db = PostgresqlDatabase(req["database"], user=req["username"], host=req["host"], port=req["port"], password=req["password"]) elif req["db_type"] == 'mssql': import pyodbc connection_string = ( f"DRIVER={{ODBC Driver 17 for SQL Server}};" f"SERVER={req['host']},{req['port']};" f"DATABASE={req['database']};" f"UID={req['username']};" f"PWD={req['password']};" ) db = pyodbc.connect(connection_string) cursor = db.cursor() cursor.execute("SELECT 1") cursor.close() elif req["db_type"] == 'IBM DB2': import ibm_db conn_str = ( f"DATABASE={req['database']};" f"HOSTNAME={req['host']};" f"PORT={req['port']};" f"PROTOCOL=TCPIP;" f"UID={req['username']};" f"PWD={req['password']};" ) redacted_conn_str = ( f"DATABASE={req['database']};" f"HOSTNAME={req['host']};" f"PORT={req['port']};" f"PROTOCOL=TCPIP;" f"UID={req['username']};" f"PWD=****;" ) logging.info(redacted_conn_str) conn = ibm_db.connect(conn_str, "", "") stmt = ibm_db.exec_immediate(conn, "SELECT 1 FROM sysibm.sysdummy1") ibm_db.fetch_assoc(stmt) ibm_db.close(conn) return get_json_result(data="Database Connection Successful!") elif req["db_type"] == 'trino': def _parse_catalog_schema(db_name: str): if not db_name: return None, None if "." in db_name: catalog_name, schema_name = db_name.split(".", 1) elif "/" in db_name: catalog_name, schema_name = db_name.split("/", 1) else: catalog_name, schema_name = db_name, "default" return catalog_name, schema_name try: import trino import os except Exception as e: return server_error_response(f"Missing dependency 'trino'. Please install: pip install trino, detail: {e}") catalog, schema = _parse_catalog_schema(req["database"]) if not catalog: return server_error_response("For Trino, 'database' must be 'catalog.schema' or at least 'catalog'.") http_scheme = "https" if os.environ.get("TRINO_USE_TLS", "0") == "1" else "http" auth = None if http_scheme == "https" and req.get("password"): auth = trino.BasicAuthentication(req.get("username") or "ragflow", req["password"]) conn = trino.dbapi.connect( host=req["host"], port=int(req["port"] or 8080), user=req["username"] or "ragflow", catalog=catalog, schema=schema or "default", http_scheme=http_scheme, auth=auth ) cur = conn.cursor() cur.execute("SELECT 1") cur.fetchall() cur.close() conn.close() return get_json_result(data="Database Connection Successful!") else: return server_error_response("Unsupported database type.") if req["db_type"] != 'mssql': db.connect() db.close() return get_json_result(data="Database Connection Successful!") except Exception as e: return server_error_response(e) #api get list version dsl of canvas @manager.route('/getlistversion/<canvas_id>', methods=['GET']) # noqa: F821 @login_required def getlistversion(canvas_id): try: versions =sorted([c.to_dict() for c in UserCanvasVersionService.list_by_canvas_id(canvas_id)], key=lambda x: x["update_time"]*-1) return get_json_result(data=versions) except Exception as e: return get_data_error_result(message=f"Error getting history files: {e}") #api get version dsl of canvas @manager.route('/getversion/<version_id>', methods=['GET']) # noqa: F821 @login_required def getversion( version_id): try: e, version = UserCanvasVersionService.get_by_id(version_id) if version: return get_json_result(data=version.to_dict()) except Exception as e: return get_json_result(data=f"Error getting history file: {e}") @manager.route('/list', methods=['GET']) # noqa: F821 @login_required def list_canvas(): keywords = request.args.get("keywords", "") page_number = int(request.args.get("page", 0)) items_per_page = int(request.args.get("page_size", 0)) orderby = request.args.get("orderby", "create_time") canvas_category = request.args.get("canvas_category") if request.args.get("desc", "true").lower() == "false": desc = False else: desc = True owner_ids = [id for id in request.args.get("owner_ids", "").strip().split(",") if id] if not owner_ids: tenants = TenantService.get_joined_tenants_by_user_id(current_user.id) tenants = [m["tenant_id"] for m in tenants] tenants.append(current_user.id) canvas, total = UserCanvasService.get_by_tenant_ids( tenants, current_user.id, page_number, items_per_page, orderby, desc, keywords, canvas_category) else: tenants = owner_ids canvas, total = UserCanvasService.get_by_tenant_ids( tenants, current_user.id, 0, 0, orderby, desc, keywords, canvas_category) return get_json_result(data={"canvas": canvas, "total": total}) @manager.route('/setting', methods=['POST']) # noqa: F821 @validate_request("id", "title", "permission") @login_required async def setting(): req = await get_request_json() req["user_id"] = current_user.id if not UserCanvasService.accessible(req["id"], current_user.id): return get_json_result( data=False, message='Only owner of canvas authorized for this operation.', code=RetCode.OPERATING_ERROR) e,flow = UserCanvasService.get_by_id(req["id"]) if not e: return get_data_error_result(message="canvas not found.") flow = flow.to_dict() flow["title"] = req["title"] for key in ["description", "permission", "avatar"]: if value := req.get(key): flow[key] = value num= UserCanvasService.update_by_id(req["id"], flow) return get_json_result(data=num) @manager.route('/trace', methods=['GET']) # noqa: F821 def trace(): cvs_id = request.args.get("canvas_id") msg_id = request.args.get("message_id") try: binary = REDIS_CONN.get(f"{cvs_id}-{msg_id}-logs") if not binary: return get_json_result(data={}) return get_json_result(data=json.loads(binary.encode("utf-8"))) except Exception as e: logging.exception(e) @manager.route('/<canvas_id>/sessions', methods=['GET']) # noqa: F821 @login_required def sessions(canvas_id): tenant_id = current_user.id if not UserCanvasService.accessible(canvas_id, tenant_id): return get_json_result( data=False, message='Only owner of canvas authorized for this operation.', code=RetCode.OPERATING_ERROR) user_id = request.args.get("user_id") page_number = int(request.args.get("page", 1)) items_per_page = int(request.args.get("page_size", 30)) keywords = request.args.get("keywords") from_date = request.args.get("from_date") to_date = request.args.get("to_date") orderby = request.args.get("orderby", "update_time") if request.args.get("desc") == "False" or request.args.get("desc") == "false": desc = False else: desc = True # dsl defaults to True in all cases except for False and false include_dsl = request.args.get("dsl") != "False" and request.args.get("dsl") != "false" total, sess = API4ConversationService.get_list(canvas_id, tenant_id, page_number, items_per_page, orderby, desc, None, user_id, include_dsl, keywords, from_date, to_date) try: return get_json_result(data={"total": total, "sessions": sess}) except Exception as e: return server_error_response(e) @manager.route('/prompts', methods=['GET']) # noqa: F821 @login_required def prompts(): from rag.prompts.generator import ANALYZE_TASK_SYSTEM, ANALYZE_TASK_USER, NEXT_STEP, REFLECT, CITATION_PROMPT_TEMPLATE return get_json_result(data={ "task_analysis": ANALYZE_TASK_SYSTEM +"\n\n"+ ANALYZE_TASK_USER, "plan_generation": NEXT_STEP, "reflection": REFLECT, #"context_summary": SUMMARY4MEMORY, #"context_ranking": RANK_MEMORY, "citation_guidelines": CITATION_PROMPT_TEMPLATE }) @manager.route('/download', methods=['GET']) # noqa: F821 async def download(): id = request.args.get("id") created_by = request.args.get("created_by") blob = FileService.get_blob(created_by, id) return await make_response(blob)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/api_app.py
api/apps/api_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from datetime import datetime, timedelta from quart import request from api.db.db_models import APIToken from api.db.services.api_service import APITokenService, API4ConversationService from api.db.services.user_service import UserTenantService from api.utils.api_utils import generate_confirmation_token, get_data_error_result, get_json_result, get_request_json, server_error_response, validate_request from common.time_utils import current_timestamp, datetime_format from api.apps import login_required, current_user @manager.route('/new_token', methods=['POST']) # noqa: F821 @login_required async def new_token(): req = await get_request_json() try: tenants = UserTenantService.query(user_id=current_user.id) if not tenants: return get_data_error_result(message="Tenant not found!") tenant_id = tenants[0].tenant_id obj = {"tenant_id": tenant_id, "token": generate_confirmation_token(), "create_time": current_timestamp(), "create_date": datetime_format(datetime.now()), "update_time": None, "update_date": None } if req.get("canvas_id"): obj["dialog_id"] = req["canvas_id"] obj["source"] = "agent" else: obj["dialog_id"] = req["dialog_id"] if not APITokenService.save(**obj): return get_data_error_result(message="Fail to new a dialog!") return get_json_result(data=obj) except Exception as e: return server_error_response(e) @manager.route('/token_list', methods=['GET']) # noqa: F821 @login_required def token_list(): try: tenants = UserTenantService.query(user_id=current_user.id) if not tenants: return get_data_error_result(message="Tenant not found!") id = request.args["dialog_id"] if "dialog_id" in request.args else request.args["canvas_id"] objs = APITokenService.query(tenant_id=tenants[0].tenant_id, dialog_id=id) return get_json_result(data=[o.to_dict() for o in objs]) except Exception as e: return server_error_response(e) @manager.route('/rm', methods=['POST']) # noqa: F821 @validate_request("tokens", "tenant_id") @login_required async def rm(): req = await get_request_json() try: for token in req["tokens"]: APITokenService.filter_delete( [APIToken.tenant_id == req["tenant_id"], APIToken.token == token]) return get_json_result(data=True) except Exception as e: return server_error_response(e) @manager.route('/stats', methods=['GET']) # noqa: F821 @login_required def stats(): try: tenants = UserTenantService.query(user_id=current_user.id) if not tenants: return get_data_error_result(message="Tenant not found!") objs = API4ConversationService.stats( tenants[0].tenant_id, request.args.get( "from_date", (datetime.now() - timedelta( days=7)).strftime("%Y-%m-%d 00:00:00")), request.args.get( "to_date", datetime.now().strftime("%Y-%m-%d %H:%M:%S")), "agent" if "canvas_id" in request.args else None) res = {"pv": [], "uv": [], "speed": [], "tokens": [], "round": [], "thumb_up": []} for obj in objs: dt = obj["dt"] res["pv"].append((dt, obj["pv"])) res["uv"].append((dt, obj["uv"])) res["speed"].append((dt, float(obj["tokens"]) / (float(obj["duration"]) + 0.1))) # +0.1 to avoid division by zero res["tokens"].append((dt, float(obj["tokens"]) / 1000.0)) # convert to thousands res["round"].append((dt, obj["round"])) res["thumb_up"].append((dt, obj["thumb_up"])) return get_json_result(data=res) except Exception as e: return server_error_response(e)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/system_app.py
api/apps/system_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License # import logging from datetime import datetime import json from api.apps import login_required, current_user from api.db.db_models import APIToken from api.db.services.api_service import APITokenService from api.db.services.knowledgebase_service import KnowledgebaseService from api.db.services.user_service import UserTenantService from api.utils.api_utils import ( get_json_result, get_data_error_result, server_error_response, generate_confirmation_token, ) from common.versions import get_ragflow_version from common.time_utils import current_timestamp, datetime_format from timeit import default_timer as timer from rag.utils.redis_conn import REDIS_CONN from quart import jsonify from api.utils.health_utils import run_health_checks from common import settings @manager.route("/version", methods=["GET"]) # noqa: F821 @login_required def version(): """ Get the current version of the application. --- tags: - System security: - ApiKeyAuth: [] responses: 200: description: Version retrieved successfully. schema: type: object properties: version: type: string description: Version number. """ return get_json_result(data=get_ragflow_version()) @manager.route("/status", methods=["GET"]) # noqa: F821 @login_required def status(): """ Get the system status. --- tags: - System security: - ApiKeyAuth: [] responses: 200: description: System is operational. schema: type: object properties: es: type: object description: Elasticsearch status. storage: type: object description: Storage status. database: type: object description: Database status. 503: description: Service unavailable. schema: type: object properties: error: type: string description: Error message. """ res = {} st = timer() try: res["doc_engine"] = settings.docStoreConn.health() res["doc_engine"]["elapsed"] = "{:.1f}".format((timer() - st) * 1000.0) except Exception as e: res["doc_engine"] = { "type": "unknown", "status": "red", "elapsed": "{:.1f}".format((timer() - st) * 1000.0), "error": str(e), } st = timer() try: settings.STORAGE_IMPL.health() res["storage"] = { "storage": settings.STORAGE_IMPL_TYPE.lower(), "status": "green", "elapsed": "{:.1f}".format((timer() - st) * 1000.0), } except Exception as e: res["storage"] = { "storage": settings.STORAGE_IMPL_TYPE.lower(), "status": "red", "elapsed": "{:.1f}".format((timer() - st) * 1000.0), "error": str(e), } st = timer() try: KnowledgebaseService.get_by_id("x") res["database"] = { "database": settings.DATABASE_TYPE.lower(), "status": "green", "elapsed": "{:.1f}".format((timer() - st) * 1000.0), } except Exception as e: res["database"] = { "database": settings.DATABASE_TYPE.lower(), "status": "red", "elapsed": "{:.1f}".format((timer() - st) * 1000.0), "error": str(e), } st = timer() try: if not REDIS_CONN.health(): raise Exception("Lost connection!") res["redis"] = { "status": "green", "elapsed": "{:.1f}".format((timer() - st) * 1000.0), } except Exception as e: res["redis"] = { "status": "red", "elapsed": "{:.1f}".format((timer() - st) * 1000.0), "error": str(e), } task_executor_heartbeats = {} try: task_executors = REDIS_CONN.smembers("TASKEXE") now = datetime.now().timestamp() for task_executor_id in task_executors: heartbeats = REDIS_CONN.zrangebyscore(task_executor_id, now - 60 * 30, now) heartbeats = [json.loads(heartbeat) for heartbeat in heartbeats] task_executor_heartbeats[task_executor_id] = heartbeats except Exception: logging.exception("get task executor heartbeats failed!") res["task_executor_heartbeats"] = task_executor_heartbeats return get_json_result(data=res) @manager.route("/healthz", methods=["GET"]) # noqa: F821 def healthz(): result, all_ok = run_health_checks() return jsonify(result), (200 if all_ok else 500) @manager.route("/ping", methods=["GET"]) # noqa: F821 def ping(): return "pong", 200 @manager.route("/new_token", methods=["POST"]) # noqa: F821 @login_required def new_token(): """ Generate a new API token. --- tags: - API Tokens security: - ApiKeyAuth: [] parameters: - in: query name: name type: string required: false description: Name of the token. responses: 200: description: Token generated successfully. schema: type: object properties: token: type: string description: The generated API token. """ try: tenants = UserTenantService.query(user_id=current_user.id) if not tenants: return get_data_error_result(message="Tenant not found!") tenant_id = [tenant for tenant in tenants if tenant.role == "owner"][0].tenant_id obj = { "tenant_id": tenant_id, "token": generate_confirmation_token(), "beta": generate_confirmation_token().replace("ragflow-", "")[:32], "create_time": current_timestamp(), "create_date": datetime_format(datetime.now()), "update_time": None, "update_date": None, } if not APITokenService.save(**obj): return get_data_error_result(message="Fail to new a dialog!") return get_json_result(data=obj) except Exception as e: return server_error_response(e) @manager.route("/token_list", methods=["GET"]) # noqa: F821 @login_required def token_list(): """ List all API tokens for the current user. --- tags: - API Tokens security: - ApiKeyAuth: [] responses: 200: description: List of API tokens. schema: type: object properties: tokens: type: array items: type: object properties: token: type: string description: The API token. name: type: string description: Name of the token. create_time: type: string description: Token creation time. """ try: tenants = UserTenantService.query(user_id=current_user.id) if not tenants: return get_data_error_result(message="Tenant not found!") tenant_id = [tenant for tenant in tenants if tenant.role == "owner"][0].tenant_id objs = APITokenService.query(tenant_id=tenant_id) objs = [o.to_dict() for o in objs] for o in objs: if not o["beta"]: o["beta"] = generate_confirmation_token().replace("ragflow-", "")[:32] APITokenService.filter_update([APIToken.tenant_id == tenant_id, APIToken.token == o["token"]], o) return get_json_result(data=objs) except Exception as e: return server_error_response(e) @manager.route("/token/<token>", methods=["DELETE"]) # noqa: F821 @login_required def rm(token): """ Remove an API token. --- tags: - API Tokens security: - ApiKeyAuth: [] parameters: - in: path name: token type: string required: true description: The API token to remove. responses: 200: description: Token removed successfully. schema: type: object properties: success: type: boolean description: Deletion status. """ try: tenants = UserTenantService.query(user_id=current_user.id) if not tenants: return get_data_error_result(message="Tenant not found!") tenant_id = tenants[0].tenant_id APITokenService.filter_delete([APIToken.tenant_id == tenant_id, APIToken.token == token]) return get_json_result(data=True) except Exception as e: return server_error_response(e) @manager.route("/config", methods=["GET"]) # noqa: F821 def get_config(): """ Get system configuration. --- tags: - System responses: 200: description: Return system configuration schema: type: object properties: registerEnable: type: integer 0 means disabled, 1 means enabled description: Whether user registration is enabled """ return get_json_result(data={"registerEnabled": settings.REGISTER_ENABLED})
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/evaluation_app.py
api/apps/evaluation_app.py
# # Copyright 2025 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ RAG Evaluation API Endpoints Provides REST API for RAG evaluation functionality including: - Dataset management - Test case management - Evaluation execution - Results retrieval - Configuration recommendations """ from quart import request from api.apps import login_required, current_user from api.db.services.evaluation_service import EvaluationService from api.utils.api_utils import ( get_data_error_result, get_json_result, get_request_json, server_error_response, validate_request ) from common.constants import RetCode # ==================== Dataset Management ==================== @manager.route('/dataset/create', methods=['POST']) # noqa: F821 @login_required @validate_request("name", "kb_ids") async def create_dataset(): """ Create a new evaluation dataset. Request body: { "name": "Dataset name", "description": "Optional description", "kb_ids": ["kb_id1", "kb_id2"] } """ try: req = await get_request_json() name = req.get("name", "").strip() description = req.get("description", "") kb_ids = req.get("kb_ids", []) if not name: return get_data_error_result(message="Dataset name cannot be empty") if not kb_ids or not isinstance(kb_ids, list): return get_data_error_result(message="kb_ids must be a non-empty list") success, result = EvaluationService.create_dataset( name=name, description=description, kb_ids=kb_ids, tenant_id=current_user.id, user_id=current_user.id ) if not success: return get_data_error_result(message=result) return get_json_result(data={"dataset_id": result}) except Exception as e: return server_error_response(e) @manager.route('/dataset/list', methods=['GET']) # noqa: F821 @login_required async def list_datasets(): """ List evaluation datasets for current tenant. Query params: - page: Page number (default: 1) - page_size: Items per page (default: 20) """ try: page = int(request.args.get("page", 1)) page_size = int(request.args.get("page_size", 20)) result = EvaluationService.list_datasets( tenant_id=current_user.id, user_id=current_user.id, page=page, page_size=page_size ) return get_json_result(data=result) except Exception as e: return server_error_response(e) @manager.route('/dataset/<dataset_id>', methods=['GET']) # noqa: F821 @login_required async def get_dataset(dataset_id): """Get dataset details by ID""" try: dataset = EvaluationService.get_dataset(dataset_id) if not dataset: return get_data_error_result( message="Dataset not found", code=RetCode.DATA_ERROR ) return get_json_result(data=dataset) except Exception as e: return server_error_response(e) @manager.route('/dataset/<dataset_id>', methods=['PUT']) # noqa: F821 @login_required async def update_dataset(dataset_id): """ Update dataset. Request body: { "name": "New name", "description": "New description", "kb_ids": ["kb_id1", "kb_id2"] } """ try: req = await get_request_json() # Remove fields that shouldn't be updated req.pop("id", None) req.pop("tenant_id", None) req.pop("created_by", None) req.pop("create_time", None) success = EvaluationService.update_dataset(dataset_id, **req) if not success: return get_data_error_result(message="Failed to update dataset") return get_json_result(data={"dataset_id": dataset_id}) except Exception as e: return server_error_response(e) @manager.route('/dataset/<dataset_id>', methods=['DELETE']) # noqa: F821 @login_required async def delete_dataset(dataset_id): """Delete dataset (soft delete)""" try: success = EvaluationService.delete_dataset(dataset_id) if not success: return get_data_error_result(message="Failed to delete dataset") return get_json_result(data={"dataset_id": dataset_id}) except Exception as e: return server_error_response(e) # ==================== Test Case Management ==================== @manager.route('/dataset/<dataset_id>/case/add', methods=['POST']) # noqa: F821 @login_required @validate_request("question") async def add_test_case(dataset_id): """ Add a test case to a dataset. Request body: { "question": "Test question", "reference_answer": "Optional ground truth answer", "relevant_doc_ids": ["doc_id1", "doc_id2"], "relevant_chunk_ids": ["chunk_id1", "chunk_id2"], "metadata": {"key": "value"} } """ try: req = await get_request_json() question = req.get("question", "").strip() if not question: return get_data_error_result(message="Question cannot be empty") success, result = EvaluationService.add_test_case( dataset_id=dataset_id, question=question, reference_answer=req.get("reference_answer"), relevant_doc_ids=req.get("relevant_doc_ids"), relevant_chunk_ids=req.get("relevant_chunk_ids"), metadata=req.get("metadata") ) if not success: return get_data_error_result(message=result) return get_json_result(data={"case_id": result}) except Exception as e: return server_error_response(e) @manager.route('/dataset/<dataset_id>/case/import', methods=['POST']) # noqa: F821 @login_required @validate_request("cases") async def import_test_cases(dataset_id): """ Bulk import test cases. Request body: { "cases": [ { "question": "Question 1", "reference_answer": "Answer 1", ... }, { "question": "Question 2", ... } ] } """ try: req = await get_request_json() cases = req.get("cases", []) if not cases or not isinstance(cases, list): return get_data_error_result(message="cases must be a non-empty list") success_count, failure_count = EvaluationService.import_test_cases( dataset_id=dataset_id, cases=cases ) return get_json_result(data={ "success_count": success_count, "failure_count": failure_count, "total": len(cases) }) except Exception as e: return server_error_response(e) @manager.route('/dataset/<dataset_id>/cases', methods=['GET']) # noqa: F821 @login_required async def get_test_cases(dataset_id): """Get all test cases for a dataset""" try: cases = EvaluationService.get_test_cases(dataset_id) return get_json_result(data={"cases": cases, "total": len(cases)}) except Exception as e: return server_error_response(e) @manager.route('/case/<case_id>', methods=['DELETE']) # noqa: F821 @login_required async def delete_test_case(case_id): """Delete a test case""" try: success = EvaluationService.delete_test_case(case_id) if not success: return get_data_error_result(message="Failed to delete test case") return get_json_result(data={"case_id": case_id}) except Exception as e: return server_error_response(e) # ==================== Evaluation Execution ==================== @manager.route('/run/start', methods=['POST']) # noqa: F821 @login_required @validate_request("dataset_id", "dialog_id") async def start_evaluation(): """ Start an evaluation run. Request body: { "dataset_id": "dataset_id", "dialog_id": "dialog_id", "name": "Optional run name" } """ try: req = await get_request_json() dataset_id = req.get("dataset_id") dialog_id = req.get("dialog_id") name = req.get("name") success, result = EvaluationService.start_evaluation( dataset_id=dataset_id, dialog_id=dialog_id, user_id=current_user.id, name=name ) if not success: return get_data_error_result(message=result) return get_json_result(data={"run_id": result}) except Exception as e: return server_error_response(e) @manager.route('/run/<run_id>', methods=['GET']) # noqa: F821 @login_required async def get_evaluation_run(run_id): """Get evaluation run details""" try: result = EvaluationService.get_run_results(run_id) if not result: return get_data_error_result( message="Evaluation run not found", code=RetCode.DATA_ERROR ) return get_json_result(data=result) except Exception as e: return server_error_response(e) @manager.route('/run/<run_id>/results', methods=['GET']) # noqa: F821 @login_required async def get_run_results(run_id): """Get detailed results for an evaluation run""" try: result = EvaluationService.get_run_results(run_id) if not result: return get_data_error_result( message="Evaluation run not found", code=RetCode.DATA_ERROR ) return get_json_result(data=result) except Exception as e: return server_error_response(e) @manager.route('/run/list', methods=['GET']) # noqa: F821 @login_required async def list_evaluation_runs(): """ List evaluation runs. Query params: - dataset_id: Filter by dataset (optional) - dialog_id: Filter by dialog (optional) - page: Page number (default: 1) - page_size: Items per page (default: 20) """ try: # TODO: Implement list_runs in EvaluationService return get_json_result(data={"runs": [], "total": 0}) except Exception as e: return server_error_response(e) @manager.route('/run/<run_id>', methods=['DELETE']) # noqa: F821 @login_required async def delete_evaluation_run(run_id): """Delete an evaluation run""" try: # TODO: Implement delete_run in EvaluationService return get_json_result(data={"run_id": run_id}) except Exception as e: return server_error_response(e) # ==================== Analysis & Recommendations ==================== @manager.route('/run/<run_id>/recommendations', methods=['GET']) # noqa: F821 @login_required async def get_recommendations(run_id): """Get configuration recommendations based on evaluation results""" try: recommendations = EvaluationService.get_recommendations(run_id) return get_json_result(data={"recommendations": recommendations}) except Exception as e: return server_error_response(e) @manager.route('/compare', methods=['POST']) # noqa: F821 @login_required @validate_request("run_ids") async def compare_runs(): """ Compare multiple evaluation runs. Request body: { "run_ids": ["run_id1", "run_id2", "run_id3"] } """ try: req = await get_request_json() run_ids = req.get("run_ids", []) if not run_ids or not isinstance(run_ids, list) or len(run_ids) < 2: return get_data_error_result( message="run_ids must be a list with at least 2 run IDs" ) # TODO: Implement compare_runs in EvaluationService return get_json_result(data={"comparison": {}}) except Exception as e: return server_error_response(e) @manager.route('/run/<run_id>/export', methods=['GET']) # noqa: F821 @login_required async def export_results(run_id): """Export evaluation results as JSON/CSV""" try: # format_type = request.args.get("format", "json") # TODO: Use for CSV export result = EvaluationService.get_run_results(run_id) if not result: return get_data_error_result( message="Evaluation run not found", code=RetCode.DATA_ERROR ) # TODO: Implement CSV export return get_json_result(data=result) except Exception as e: return server_error_response(e) # ==================== Real-time Evaluation ==================== @manager.route('/evaluate_single', methods=['POST']) # noqa: F821 @login_required @validate_request("question", "dialog_id") async def evaluate_single(): """ Evaluate a single question-answer pair in real-time. Request body: { "question": "Test question", "dialog_id": "dialog_id", "reference_answer": "Optional ground truth", "relevant_chunk_ids": ["chunk_id1", "chunk_id2"] } """ try: # req = await get_request_json() # TODO: Use for single evaluation implementation # TODO: Implement single evaluation # This would execute the RAG pipeline and return metrics immediately return get_json_result(data={ "answer": "", "metrics": {}, "retrieved_chunks": [] }) except Exception as e: return server_error_response(e)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/document_app.py
api/apps/document_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License # import asyncio import json import os.path import pathlib import re from pathlib import Path from quart import request, make_response from api.apps import current_user, login_required from api.common.check_team_permission import check_kb_team_permission from api.constants import FILE_NAME_LEN_LIMIT, IMG_BASE64_PREFIX from api.db import VALID_FILE_TYPES, FileType from api.db.db_models import Task from api.db.services import duplicate_name from api.db.services.document_service import DocumentService, doc_upload_and_parse from common.metadata_utils import meta_filter, convert_conditions from api.db.services.file2document_service import File2DocumentService from api.db.services.file_service import FileService from api.db.services.knowledgebase_service import KnowledgebaseService from api.db.services.task_service import TaskService, cancel_all_task_of from api.db.services.user_service import UserTenantService from common.misc_utils import get_uuid from api.utils.api_utils import ( get_data_error_result, get_json_result, server_error_response, validate_request, get_request_json, ) from api.utils.file_utils import filename_type, thumbnail from common.file_utils import get_project_base_directory from common.constants import RetCode, VALID_TASK_STATUS, ParserType, TaskStatus from api.utils.web_utils import CONTENT_TYPE_MAP, html2pdf, is_valid_url from deepdoc.parser.html_parser import RAGFlowHtmlParser from rag.nlp import search, rag_tokenizer from common import settings @manager.route("/upload", methods=["POST"]) # noqa: F821 @login_required @validate_request("kb_id") async def upload(): form = await request.form kb_id = form.get("kb_id") if not kb_id: return get_json_result(data=False, message='Lack of "KB ID"', code=RetCode.ARGUMENT_ERROR) files = await request.files if "file" not in files: return get_json_result(data=False, message="No file part!", code=RetCode.ARGUMENT_ERROR) file_objs = files.getlist("file") for file_obj in file_objs: if file_obj.filename == "": return get_json_result(data=False, message="No file selected!", code=RetCode.ARGUMENT_ERROR) if len(file_obj.filename.encode("utf-8")) > FILE_NAME_LEN_LIMIT: return get_json_result(data=False, message=f"File name must be {FILE_NAME_LEN_LIMIT} bytes or less.", code=RetCode.ARGUMENT_ERROR) e, kb = KnowledgebaseService.get_by_id(kb_id) if not e: raise LookupError("Can't find this dataset!") if not check_kb_team_permission(kb, current_user.id): return get_json_result(data=False, message="No authorization.", code=RetCode.AUTHENTICATION_ERROR) err, files = await asyncio.to_thread(FileService.upload_document, kb, file_objs, current_user.id) if err: return get_json_result(data=files, message="\n".join(err), code=RetCode.SERVER_ERROR) if not files: return get_json_result(data=files, message="There seems to be an issue with your file format. Please verify it is correct and not corrupted.", code=RetCode.DATA_ERROR) files = [f[0] for f in files] # remove the blob return get_json_result(data=files) @manager.route("/web_crawl", methods=["POST"]) # noqa: F821 @login_required @validate_request("kb_id", "name", "url") async def web_crawl(): form = await request.form kb_id = form.get("kb_id") if not kb_id: return get_json_result(data=False, message='Lack of "KB ID"', code=RetCode.ARGUMENT_ERROR) name = form.get("name") url = form.get("url") if not is_valid_url(url): return get_json_result(data=False, message="The URL format is invalid", code=RetCode.ARGUMENT_ERROR) e, kb = KnowledgebaseService.get_by_id(kb_id) if not e: raise LookupError("Can't find this dataset!") if check_kb_team_permission(kb, current_user.id): return get_json_result(data=False, message="No authorization.", code=RetCode.AUTHENTICATION_ERROR) blob = html2pdf(url) if not blob: return server_error_response(ValueError("Download failure.")) root_folder = FileService.get_root_folder(current_user.id) pf_id = root_folder["id"] FileService.init_knowledgebase_docs(pf_id, current_user.id) kb_root_folder = FileService.get_kb_folder(current_user.id) kb_folder = FileService.new_a_file_from_kb(kb.tenant_id, kb.name, kb_root_folder["id"]) try: filename = duplicate_name(DocumentService.query, name=name + ".pdf", kb_id=kb.id) filetype = filename_type(filename) if filetype == FileType.OTHER.value: raise RuntimeError("This type of file has not been supported yet!") location = filename while settings.STORAGE_IMPL.obj_exist(kb_id, location): location += "_" settings.STORAGE_IMPL.put(kb_id, location, blob) doc = { "id": get_uuid(), "kb_id": kb.id, "parser_id": kb.parser_id, "parser_config": kb.parser_config, "created_by": current_user.id, "type": filetype, "name": filename, "location": location, "size": len(blob), "thumbnail": thumbnail(filename, blob), "suffix": Path(filename).suffix.lstrip("."), } if doc["type"] == FileType.VISUAL: doc["parser_id"] = ParserType.PICTURE.value if doc["type"] == FileType.AURAL: doc["parser_id"] = ParserType.AUDIO.value if re.search(r"\.(ppt|pptx|pages)$", filename): doc["parser_id"] = ParserType.PRESENTATION.value if re.search(r"\.(eml)$", filename): doc["parser_id"] = ParserType.EMAIL.value DocumentService.insert(doc) FileService.add_file_from_kb(doc, kb_folder["id"], kb.tenant_id) except Exception as e: return server_error_response(e) return get_json_result(data=True) @manager.route("/create", methods=["POST"]) # noqa: F821 @login_required @validate_request("name", "kb_id") async def create(): req = await get_request_json() kb_id = req["kb_id"] if not kb_id: return get_json_result(data=False, message='Lack of "KB ID"', code=RetCode.ARGUMENT_ERROR) if len(req["name"].encode("utf-8")) > FILE_NAME_LEN_LIMIT: return get_json_result(data=False, message=f"File name must be {FILE_NAME_LEN_LIMIT} bytes or less.", code=RetCode.ARGUMENT_ERROR) if req["name"].strip() == "": return get_json_result(data=False, message="File name can't be empty.", code=RetCode.ARGUMENT_ERROR) req["name"] = req["name"].strip() try: e, kb = KnowledgebaseService.get_by_id(kb_id) if not e: return get_data_error_result(message="Can't find this dataset!") if DocumentService.query(name=req["name"], kb_id=kb_id): return get_data_error_result(message="Duplicated document name in the same dataset.") kb_root_folder = FileService.get_kb_folder(kb.tenant_id) if not kb_root_folder: return get_data_error_result(message="Cannot find the root folder.") kb_folder = FileService.new_a_file_from_kb( kb.tenant_id, kb.name, kb_root_folder["id"], ) if not kb_folder: return get_data_error_result(message="Cannot find the kb folder for this file.") doc = DocumentService.insert( { "id": get_uuid(), "kb_id": kb.id, "parser_id": kb.parser_id, "pipeline_id": kb.pipeline_id, "parser_config": kb.parser_config, "created_by": current_user.id, "type": FileType.VIRTUAL, "name": req["name"], "suffix": Path(req["name"]).suffix.lstrip("."), "location": "", "size": 0, } ) FileService.add_file_from_kb(doc.to_dict(), kb_folder["id"], kb.tenant_id) return get_json_result(data=doc.to_json()) except Exception as e: return server_error_response(e) @manager.route("/list", methods=["POST"]) # noqa: F821 @login_required async def list_docs(): kb_id = request.args.get("kb_id") if not kb_id: return get_json_result(data=False, message='Lack of "KB ID"', code=RetCode.ARGUMENT_ERROR) tenants = UserTenantService.query(user_id=current_user.id) for tenant in tenants: if KnowledgebaseService.query(tenant_id=tenant.tenant_id, id=kb_id): break else: return get_json_result(data=False, message="Only owner of dataset authorized for this operation.", code=RetCode.OPERATING_ERROR) keywords = request.args.get("keywords", "") page_number = int(request.args.get("page", 0)) items_per_page = int(request.args.get("page_size", 0)) orderby = request.args.get("orderby", "create_time") if request.args.get("desc", "true").lower() == "false": desc = False else: desc = True create_time_from = int(request.args.get("create_time_from", 0)) create_time_to = int(request.args.get("create_time_to", 0)) req = await get_request_json() return_empty_metadata = req.get("return_empty_metadata", False) if isinstance(return_empty_metadata, str): return_empty_metadata = return_empty_metadata.lower() == "true" run_status = req.get("run_status", []) if run_status: invalid_status = {s for s in run_status if s not in VALID_TASK_STATUS} if invalid_status: return get_data_error_result(message=f"Invalid filter run status conditions: {', '.join(invalid_status)}") types = req.get("types", []) if types: invalid_types = {t for t in types if t not in VALID_FILE_TYPES} if invalid_types: return get_data_error_result(message=f"Invalid filter conditions: {', '.join(invalid_types)} type{'s' if len(invalid_types) > 1 else ''}") suffix = req.get("suffix", []) metadata_condition = req.get("metadata_condition", {}) or {} metadata = req.get("metadata", {}) or {} if isinstance(metadata, dict) and metadata.get("empty_metadata"): return_empty_metadata = True metadata = {k: v for k, v in metadata.items() if k != "empty_metadata"} if return_empty_metadata: metadata_condition = {} metadata = {} else: if metadata_condition and not isinstance(metadata_condition, dict): return get_data_error_result(message="metadata_condition must be an object.") if metadata and not isinstance(metadata, dict): return get_data_error_result(message="metadata must be an object.") doc_ids_filter = None metas = None if metadata_condition or metadata: metas = DocumentService.get_flatted_meta_by_kbs([kb_id]) if metadata_condition: doc_ids_filter = set(meta_filter(metas, convert_conditions(metadata_condition), metadata_condition.get("logic", "and"))) if metadata_condition.get("conditions") and not doc_ids_filter: return get_json_result(data={"total": 0, "docs": []}) if metadata: metadata_doc_ids = None for key, values in metadata.items(): if not values: continue if not isinstance(values, list): values = [values] values = [str(v) for v in values if v is not None and str(v).strip()] if not values: continue key_doc_ids = set() for value in values: key_doc_ids.update(metas.get(key, {}).get(value, [])) if metadata_doc_ids is None: metadata_doc_ids = key_doc_ids else: metadata_doc_ids &= key_doc_ids if not metadata_doc_ids: return get_json_result(data={"total": 0, "docs": []}) if metadata_doc_ids is not None: if doc_ids_filter is None: doc_ids_filter = metadata_doc_ids else: doc_ids_filter &= metadata_doc_ids if not doc_ids_filter: return get_json_result(data={"total": 0, "docs": []}) if doc_ids_filter is not None: doc_ids_filter = list(doc_ids_filter) try: docs, tol = DocumentService.get_by_kb_id( kb_id, page_number, items_per_page, orderby, desc, keywords, run_status, types, suffix, doc_ids_filter, return_empty_metadata=return_empty_metadata, ) if create_time_from or create_time_to: filtered_docs = [] for doc in docs: doc_create_time = doc.get("create_time", 0) if (create_time_from == 0 or doc_create_time >= create_time_from) and (create_time_to == 0 or doc_create_time <= create_time_to): filtered_docs.append(doc) docs = filtered_docs for doc_item in docs: if doc_item["thumbnail"] and not doc_item["thumbnail"].startswith(IMG_BASE64_PREFIX): doc_item["thumbnail"] = f"/v1/document/image/{kb_id}-{doc_item['thumbnail']}" if doc_item.get("source_type"): doc_item["source_type"] = doc_item["source_type"].split("/")[0] return get_json_result(data={"total": tol, "docs": docs}) except Exception as e: return server_error_response(e) @manager.route("/filter", methods=["POST"]) # noqa: F821 @login_required async def get_filter(): req = await get_request_json() kb_id = req.get("kb_id") if not kb_id: return get_json_result(data=False, message='Lack of "KB ID"', code=RetCode.ARGUMENT_ERROR) tenants = UserTenantService.query(user_id=current_user.id) for tenant in tenants: if KnowledgebaseService.query(tenant_id=tenant.tenant_id, id=kb_id): break else: return get_json_result(data=False, message="Only owner of dataset authorized for this operation.", code=RetCode.OPERATING_ERROR) keywords = req.get("keywords", "") suffix = req.get("suffix", []) run_status = req.get("run_status", []) if run_status: invalid_status = {s for s in run_status if s not in VALID_TASK_STATUS} if invalid_status: return get_data_error_result(message=f"Invalid filter run status conditions: {', '.join(invalid_status)}") types = req.get("types", []) if types: invalid_types = {t for t in types if t not in VALID_FILE_TYPES} if invalid_types: return get_data_error_result(message=f"Invalid filter conditions: {', '.join(invalid_types)} type{'s' if len(invalid_types) > 1 else ''}") try: filter, total = DocumentService.get_filter_by_kb_id(kb_id, keywords, run_status, types, suffix) return get_json_result(data={"total": total, "filter": filter}) except Exception as e: return server_error_response(e) @manager.route("/infos", methods=["POST"]) # noqa: F821 @login_required async def doc_infos(): req = await get_request_json() doc_ids = req["doc_ids"] for doc_id in doc_ids: if not DocumentService.accessible(doc_id, current_user.id): return get_json_result(data=False, message="No authorization.", code=RetCode.AUTHENTICATION_ERROR) docs = DocumentService.get_by_ids(doc_ids) return get_json_result(data=list(docs.dicts())) @manager.route("/metadata/summary", methods=["POST"]) # noqa: F821 @login_required async def metadata_summary(): req = await get_request_json() kb_id = req.get("kb_id") if not kb_id: return get_json_result(data=False, message='Lack of "KB ID"', code=RetCode.ARGUMENT_ERROR) tenants = UserTenantService.query(user_id=current_user.id) for tenant in tenants: if KnowledgebaseService.query(tenant_id=tenant.tenant_id, id=kb_id): break else: return get_json_result(data=False, message="Only owner of dataset authorized for this operation.", code=RetCode.OPERATING_ERROR) try: summary = DocumentService.get_metadata_summary(kb_id) return get_json_result(data={"summary": summary}) except Exception as e: return server_error_response(e) @manager.route("/metadata/update", methods=["POST"]) # noqa: F821 @login_required async def metadata_update(): req = await get_request_json() kb_id = req.get("kb_id") if not kb_id: return get_json_result(data=False, message='Lack of "KB ID"', code=RetCode.ARGUMENT_ERROR) tenants = UserTenantService.query(user_id=current_user.id) for tenant in tenants: if KnowledgebaseService.query(tenant_id=tenant.tenant_id, id=kb_id): break else: return get_json_result(data=False, message="Only owner of dataset authorized for this operation.", code=RetCode.OPERATING_ERROR) selector = req.get("selector", {}) or {} updates = req.get("updates", []) or [] deletes = req.get("deletes", []) or [] if not isinstance(selector, dict): return get_json_result(data=False, message="selector must be an object.", code=RetCode.ARGUMENT_ERROR) if not isinstance(updates, list) or not isinstance(deletes, list): return get_json_result(data=False, message="updates and deletes must be lists.", code=RetCode.ARGUMENT_ERROR) metadata_condition = selector.get("metadata_condition", {}) or {} if metadata_condition and not isinstance(metadata_condition, dict): return get_json_result(data=False, message="metadata_condition must be an object.", code=RetCode.ARGUMENT_ERROR) document_ids = selector.get("document_ids", []) or [] if document_ids and not isinstance(document_ids, list): return get_json_result(data=False, message="document_ids must be a list.", code=RetCode.ARGUMENT_ERROR) for upd in updates: if not isinstance(upd, dict) or not upd.get("key") or "value" not in upd: return get_json_result(data=False, message="Each update requires key and value.", code=RetCode.ARGUMENT_ERROR) for d in deletes: if not isinstance(d, dict) or not d.get("key"): return get_json_result(data=False, message="Each delete requires key.", code=RetCode.ARGUMENT_ERROR) kb_doc_ids = KnowledgebaseService.list_documents_by_ids([kb_id]) target_doc_ids = set(kb_doc_ids) if document_ids: invalid_ids = set(document_ids) - set(kb_doc_ids) if invalid_ids: return get_json_result(data=False, message=f"These documents do not belong to dataset {kb_id}: {', '.join(invalid_ids)}", code=RetCode.ARGUMENT_ERROR) target_doc_ids = set(document_ids) if metadata_condition: metas = DocumentService.get_flatted_meta_by_kbs([kb_id]) filtered_ids = set(meta_filter(metas, convert_conditions(metadata_condition), metadata_condition.get("logic", "and"))) target_doc_ids = target_doc_ids & filtered_ids if metadata_condition.get("conditions") and not target_doc_ids: return get_json_result(data={"updated": 0, "matched_docs": 0}) target_doc_ids = list(target_doc_ids) updated = DocumentService.batch_update_metadata(kb_id, target_doc_ids, updates, deletes) return get_json_result(data={"updated": updated, "matched_docs": len(target_doc_ids)}) @manager.route("/update_metadata_setting", methods=["POST"]) # noqa: F821 @login_required @validate_request("doc_id", "metadata") async def update_metadata_setting(): req = await get_request_json() if not DocumentService.accessible(req["doc_id"], current_user.id): return get_json_result(data=False, message="No authorization.", code=RetCode.AUTHENTICATION_ERROR) e, doc = DocumentService.get_by_id(req["doc_id"]) if not e: return get_data_error_result(message="Document not found!") DocumentService.update_parser_config(doc.id, {"metadata": req["metadata"]}) e, doc = DocumentService.get_by_id(doc.id) if not e: return get_data_error_result(message="Document not found!") return get_json_result(data=doc.to_dict()) @manager.route("/thumbnails", methods=["GET"]) # noqa: F821 # @login_required def thumbnails(): doc_ids = request.args.getlist("doc_ids") if not doc_ids: return get_json_result(data=False, message='Lack of "Document ID"', code=RetCode.ARGUMENT_ERROR) try: docs = DocumentService.get_thumbnails(doc_ids) for doc_item in docs: if doc_item["thumbnail"] and not doc_item["thumbnail"].startswith(IMG_BASE64_PREFIX): doc_item["thumbnail"] = f"/v1/document/image/{doc_item['kb_id']}-{doc_item['thumbnail']}" return get_json_result(data={d["id"]: d["thumbnail"] for d in docs}) except Exception as e: return server_error_response(e) @manager.route("/change_status", methods=["POST"]) # noqa: F821 @login_required @validate_request("doc_ids", "status") async def change_status(): req = await get_request_json() doc_ids = req.get("doc_ids", []) status = str(req.get("status", "")) if status not in ["0", "1"]: return get_json_result(data=False, message='"Status" must be either 0 or 1!', code=RetCode.ARGUMENT_ERROR) result = {} for doc_id in doc_ids: if not DocumentService.accessible(doc_id, current_user.id): result[doc_id] = {"error": "No authorization."} continue try: e, doc = DocumentService.get_by_id(doc_id) if not e: result[doc_id] = {"error": "No authorization."} continue e, kb = KnowledgebaseService.get_by_id(doc.kb_id) if not e: result[doc_id] = {"error": "Can't find this dataset!"} continue if not DocumentService.update_by_id(doc_id, {"status": str(status)}): result[doc_id] = {"error": "Database error (Document update)!"} continue status_int = int(status) if not settings.docStoreConn.update({"doc_id": doc_id}, {"available_int": status_int}, search.index_name(kb.tenant_id), doc.kb_id): result[doc_id] = {"error": "Database error (docStore update)!"} result[doc_id] = {"status": status} except Exception as e: result[doc_id] = {"error": f"Internal server error: {str(e)}"} return get_json_result(data=result) @manager.route("/rm", methods=["POST"]) # noqa: F821 @login_required @validate_request("doc_id") async def rm(): req = await get_request_json() doc_ids = req["doc_id"] if isinstance(doc_ids, str): doc_ids = [doc_ids] for doc_id in doc_ids: if not DocumentService.accessible4deletion(doc_id, current_user.id): return get_json_result(data=False, message="No authorization.", code=RetCode.AUTHENTICATION_ERROR) errors = await asyncio.to_thread(FileService.delete_docs, doc_ids, current_user.id) if errors: return get_json_result(data=False, message=errors, code=RetCode.SERVER_ERROR) return get_json_result(data=True) @manager.route("/run", methods=["POST"]) # noqa: F821 @login_required @validate_request("doc_ids", "run") async def run(): req = await get_request_json() try: def _run_sync(): for doc_id in req["doc_ids"]: if not DocumentService.accessible(doc_id, current_user.id): return get_json_result(data=False, message="No authorization.", code=RetCode.AUTHENTICATION_ERROR) kb_table_num_map = {} for id in req["doc_ids"]: info = {"run": str(req["run"]), "progress": 0} if str(req["run"]) == TaskStatus.RUNNING.value and req.get("delete", False): info["progress_msg"] = "" info["chunk_num"] = 0 info["token_num"] = 0 tenant_id = DocumentService.get_tenant_id(id) if not tenant_id: return get_data_error_result(message="Tenant not found!") e, doc = DocumentService.get_by_id(id) if not e: return get_data_error_result(message="Document not found!") if str(req["run"]) == TaskStatus.CANCEL.value: if str(doc.run) == TaskStatus.RUNNING.value: cancel_all_task_of(id) else: return get_data_error_result(message="Cannot cancel a task that is not in RUNNING status") if all([("delete" not in req or req["delete"]), str(req["run"]) == TaskStatus.RUNNING.value, str(doc.run) == TaskStatus.DONE.value]): DocumentService.clear_chunk_num_when_rerun(doc.id) DocumentService.update_by_id(id, info) if req.get("delete", False): TaskService.filter_delete([Task.doc_id == id]) if settings.docStoreConn.index_exist(search.index_name(tenant_id), doc.kb_id): settings.docStoreConn.delete({"doc_id": id}, search.index_name(tenant_id), doc.kb_id) if str(req["run"]) == TaskStatus.RUNNING.value: if req.get("apply_kb"): e, kb = KnowledgebaseService.get_by_id(doc.kb_id) if not e: raise LookupError("Can't find this dataset!") doc.parser_config["enable_metadata"] = kb.parser_config.get("enable_metadata", False) doc.parser_config["metadata"] = kb.parser_config.get("metadata", {}) DocumentService.update_parser_config(doc.id, doc.parser_config) doc_dict = doc.to_dict() DocumentService.run(tenant_id, doc_dict, kb_table_num_map) return get_json_result(data=True) return await asyncio.to_thread(_run_sync) except Exception as e: return server_error_response(e) @manager.route("/rename", methods=["POST"]) # noqa: F821 @login_required @validate_request("doc_id", "name") async def rename(): req = await get_request_json() try: def _rename_sync(): if not DocumentService.accessible(req["doc_id"], current_user.id): return get_json_result(data=False, message="No authorization.", code=RetCode.AUTHENTICATION_ERROR) e, doc = DocumentService.get_by_id(req["doc_id"]) if not e: return get_data_error_result(message="Document not found!") if pathlib.Path(req["name"].lower()).suffix != pathlib.Path(doc.name.lower()).suffix: return get_json_result(data=False, message="The extension of file can't be changed", code=RetCode.ARGUMENT_ERROR) if len(req["name"].encode("utf-8")) > FILE_NAME_LEN_LIMIT: return get_json_result(data=False, message=f"File name must be {FILE_NAME_LEN_LIMIT} bytes or less.", code=RetCode.ARGUMENT_ERROR) for d in DocumentService.query(name=req["name"], kb_id=doc.kb_id): if d.name == req["name"]: return get_data_error_result(message="Duplicated document name in the same dataset.") if not DocumentService.update_by_id(req["doc_id"], {"name": req["name"]}): return get_data_error_result(message="Database error (Document rename)!") informs = File2DocumentService.get_by_document_id(req["doc_id"]) if informs: e, file = FileService.get_by_id(informs[0].file_id) FileService.update_by_id(file.id, {"name": req["name"]}) tenant_id = DocumentService.get_tenant_id(req["doc_id"]) title_tks = rag_tokenizer.tokenize(req["name"]) es_body = { "docnm_kwd": req["name"], "title_tks": title_tks, "title_sm_tks": rag_tokenizer.fine_grained_tokenize(title_tks), } if settings.docStoreConn.index_exist(search.index_name(tenant_id), doc.kb_id): settings.docStoreConn.update( {"doc_id": req["doc_id"]}, es_body, search.index_name(tenant_id), doc.kb_id, ) return get_json_result(data=True) return await asyncio.to_thread(_rename_sync) except Exception as e: return server_error_response(e) @manager.route("/get/<doc_id>", methods=["GET"]) # noqa: F821 # @login_required async def get(doc_id): try: e, doc = DocumentService.get_by_id(doc_id) if not e: return get_data_error_result(message="Document not found!") b, n = File2DocumentService.get_storage_address(doc_id=doc_id) data = await asyncio.to_thread(settings.STORAGE_IMPL.get, b, n) response = await make_response(data) ext = re.search(r"\.([^.]+)$", doc.name.lower()) ext = ext.group(1) if ext else None if ext: if doc.type == FileType.VISUAL.value: content_type = CONTENT_TYPE_MAP.get(ext, f"image/{ext}") else: content_type = CONTENT_TYPE_MAP.get(ext, f"application/{ext}") response.headers.set("Content-Type", content_type) return response except Exception as e: return server_error_response(e) @manager.route("/download/<attachment_id>", methods=["GET"]) # noqa: F821 @login_required async def download_attachment(attachment_id): try: ext = request.args.get("ext", "markdown") data = await asyncio.to_thread(settings.STORAGE_IMPL.get, current_user.id, attachment_id) response = await make_response(data) response.headers.set("Content-Type", CONTENT_TYPE_MAP.get(ext, f"application/{ext}")) return response except Exception as e: return server_error_response(e) @manager.route("/change_parser", methods=["POST"]) # noqa: F821 @login_required @validate_request("doc_id") async def change_parser(): req = await get_request_json() if not DocumentService.accessible(req["doc_id"], current_user.id): return get_json_result(data=False, message="No authorization.", code=RetCode.AUTHENTICATION_ERROR) e, doc = DocumentService.get_by_id(req["doc_id"]) if not e: return get_data_error_result(message="Document not found!") def reset_doc(): nonlocal doc e = DocumentService.update_by_id(doc.id, {"pipeline_id": req["pipeline_id"], "parser_id": req["parser_id"], "progress": 0, "progress_msg": "", "run": TaskStatus.UNSTART.value}) if not e: return get_data_error_result(message="Document not found!") if doc.token_num > 0: e = DocumentService.increment_chunk_num(doc.id, doc.kb_id, doc.token_num * -1, doc.chunk_num * -1, doc.process_duration * -1) if not e: return get_data_error_result(message="Document not found!") tenant_id = DocumentService.get_tenant_id(req["doc_id"]) if not tenant_id: return get_data_error_result(message="Tenant not found!") DocumentService.delete_chunk_images(doc, tenant_id) if settings.docStoreConn.index_exist(search.index_name(tenant_id), doc.kb_id): settings.docStoreConn.delete({"doc_id": doc.id}, search.index_name(tenant_id), doc.kb_id) return None try: if "pipeline_id" in req and req["pipeline_id"] != "": if doc.pipeline_id == req["pipeline_id"]: return get_json_result(data=True) DocumentService.update_by_id(doc.id, {"pipeline_id": req["pipeline_id"]}) reset_doc() return get_json_result(data=True) if doc.parser_id.lower() == req["parser_id"].lower(): if "parser_config" in req: if req["parser_config"] == doc.parser_config: return get_json_result(data=True) else:
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/user_app.py
api/apps/user_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json import logging import string import os import re import secrets import time from datetime import datetime import base64 from quart import make_response, redirect, request, session from werkzeug.security import check_password_hash, generate_password_hash from api.apps.auth import get_auth_client from api.db import FileType, UserTenantRole from api.db.db_models import TenantLLM from api.db.services.file_service import FileService from api.db.services.llm_service import get_init_tenant_llm from api.db.services.tenant_llm_service import TenantLLMService from api.db.services.user_service import TenantService, UserService, UserTenantService from common.time_utils import current_timestamp, datetime_format, get_format_time from common.misc_utils import download_img, get_uuid from common.constants import RetCode from common.connection_utils import construct_response from api.utils.api_utils import ( get_data_error_result, get_json_result, get_request_json, server_error_response, validate_request, ) from api.utils.crypt import decrypt from rag.utils.redis_conn import REDIS_CONN from api.apps import login_required, current_user, login_user, logout_user from api.utils.web_utils import ( send_email_html, OTP_LENGTH, OTP_TTL_SECONDS, ATTEMPT_LIMIT, ATTEMPT_LOCK_SECONDS, RESEND_COOLDOWN_SECONDS, otp_keys, hash_code, captcha_key, ) from common import settings from common.http_client import async_request @manager.route("/login", methods=["POST", "GET"]) # noqa: F821 async def login(): """ User login endpoint. --- tags: - User parameters: - in: body name: body description: Login credentials. required: true schema: type: object properties: email: type: string description: User email. password: type: string description: User password. responses: 200: description: Login successful. schema: type: object 401: description: Authentication failed. schema: type: object """ json_body = await get_request_json() if not json_body: return get_json_result(data=False, code=RetCode.AUTHENTICATION_ERROR, message="Unauthorized!") email = json_body.get("email", "") if email == "admin@ragflow.io": return get_json_result(data=False, code=RetCode.AUTHENTICATION_ERROR, message="Default admin account cannot be used to login normal services!") users = UserService.query(email=email) if not users: return get_json_result( data=False, code=RetCode.AUTHENTICATION_ERROR, message=f"Email: {email} is not registered!", ) password = json_body.get("password") try: password = decrypt(password) except BaseException: return get_json_result(data=False, code=RetCode.SERVER_ERROR, message="Fail to crypt password") user = UserService.query_user(email, password) if user and hasattr(user, 'is_active') and user.is_active == "0": return get_json_result( data=False, code=RetCode.FORBIDDEN, message="This account has been disabled, please contact the administrator!", ) elif user: response_data = user.to_json() user.access_token = get_uuid() login_user(user) user.update_time = current_timestamp() user.update_date = datetime_format(datetime.now()) user.save() msg = "Welcome back!" return await construct_response(data=response_data, auth=user.get_id(), message=msg) else: return get_json_result( data=False, code=RetCode.AUTHENTICATION_ERROR, message="Email and password do not match!", ) @manager.route("/login/channels", methods=["GET"]) # noqa: F821 async def get_login_channels(): """ Get all supported authentication channels. """ try: channels = [] for channel, config in settings.OAUTH_CONFIG.items(): channels.append( { "channel": channel, "display_name": config.get("display_name", channel.title()), "icon": config.get("icon", "sso"), } ) return get_json_result(data=channels) except Exception as e: logging.exception(e) return get_json_result(data=[], message=f"Load channels failure, error: {str(e)}", code=RetCode.EXCEPTION_ERROR) @manager.route("/login/<channel>", methods=["GET"]) # noqa: F821 async def oauth_login(channel): channel_config = settings.OAUTH_CONFIG.get(channel) if not channel_config: raise ValueError(f"Invalid channel name: {channel}") auth_cli = get_auth_client(channel_config) state = get_uuid() session["oauth_state"] = state auth_url = auth_cli.get_authorization_url(state) return redirect(auth_url) @manager.route("/oauth/callback/<channel>", methods=["GET"]) # noqa: F821 async def oauth_callback(channel): """ Handle the OAuth/OIDC callback for various channels dynamically. """ try: channel_config = settings.OAUTH_CONFIG.get(channel) if not channel_config: raise ValueError(f"Invalid channel name: {channel}") auth_cli = get_auth_client(channel_config) # Check the state state = request.args.get("state") if not state or state != session.get("oauth_state"): return redirect("/?error=invalid_state") session.pop("oauth_state", None) # Obtain the authorization code code = request.args.get("code") if not code: return redirect("/?error=missing_code") # Exchange authorization code for access token if hasattr(auth_cli, "async_exchange_code_for_token"): token_info = await auth_cli.async_exchange_code_for_token(code) else: token_info = auth_cli.exchange_code_for_token(code) access_token = token_info.get("access_token") if not access_token: return redirect("/?error=token_failed") id_token = token_info.get("id_token") # Fetch user info if hasattr(auth_cli, "async_fetch_user_info"): user_info = await auth_cli.async_fetch_user_info(access_token, id_token=id_token) else: user_info = auth_cli.fetch_user_info(access_token, id_token=id_token) if not user_info.email: return redirect("/?error=email_missing") # Login or register users = UserService.query(email=user_info.email) user_id = get_uuid() if not users: try: try: avatar = download_img(user_info.avatar_url) except Exception as e: logging.exception(e) avatar = "" users = user_register( user_id, { "access_token": get_uuid(), "email": user_info.email, "avatar": avatar, "nickname": user_info.nickname, "login_channel": channel, "last_login_time": get_format_time(), "is_superuser": False, }, ) if not users: raise Exception(f"Failed to register {user_info.email}") if len(users) > 1: raise Exception(f"Same email: {user_info.email} exists!") # Try to log in user = users[0] login_user(user) return redirect(f"/?auth={user.get_id()}") except Exception as e: rollback_user_registration(user_id) logging.exception(e) return redirect(f"/?error={str(e)}") # User exists, try to log in user = users[0] user.access_token = get_uuid() if user and hasattr(user, 'is_active') and user.is_active == "0": return redirect("/?error=user_inactive") login_user(user) user.save() return redirect(f"/?auth={user.get_id()}") except Exception as e: logging.exception(e) return redirect(f"/?error={str(e)}") @manager.route("/github_callback", methods=["GET"]) # noqa: F821 async def github_callback(): """ **Deprecated**, Use `/oauth/callback/<channel>` instead. GitHub OAuth callback endpoint. --- tags: - OAuth parameters: - in: query name: code type: string required: true description: Authorization code from GitHub. responses: 200: description: Authentication successful. schema: type: object """ res = await async_request( "POST", settings.GITHUB_OAUTH.get("url"), data={ "client_id": settings.GITHUB_OAUTH.get("client_id"), "client_secret": settings.GITHUB_OAUTH.get("secret_key"), "code": request.args.get("code"), }, headers={"Accept": "application/json"}, ) res = res.json() if "error" in res: return redirect("/?error=%s" % res["error_description"]) if "user:email" not in res["scope"].split(","): return redirect("/?error=user:email not in scope") session["access_token"] = res["access_token"] session["access_token_from"] = "github" user_info = await user_info_from_github(session["access_token"]) email_address = user_info["email"] users = UserService.query(email=email_address) user_id = get_uuid() if not users: # User isn't try to register try: try: avatar = download_img(user_info["avatar_url"]) except Exception as e: logging.exception(e) avatar = "" users = user_register( user_id, { "access_token": session["access_token"], "email": email_address, "avatar": avatar, "nickname": user_info["login"], "login_channel": "github", "last_login_time": get_format_time(), "is_superuser": False, }, ) if not users: raise Exception(f"Fail to register {email_address}.") if len(users) > 1: raise Exception(f"Same email: {email_address} exists!") # Try to log in user = users[0] login_user(user) return redirect("/?auth=%s" % user.get_id()) except Exception as e: rollback_user_registration(user_id) logging.exception(e) return redirect("/?error=%s" % str(e)) # User has already registered, try to log in user = users[0] user.access_token = get_uuid() if user and hasattr(user, 'is_active') and user.is_active == "0": return redirect("/?error=user_inactive") login_user(user) user.save() return redirect("/?auth=%s" % user.get_id()) @manager.route("/feishu_callback", methods=["GET"]) # noqa: F821 async def feishu_callback(): """ Feishu OAuth callback endpoint. --- tags: - OAuth parameters: - in: query name: code type: string required: true description: Authorization code from Feishu. responses: 200: description: Authentication successful. schema: type: object """ app_access_token_res = await async_request( "POST", settings.FEISHU_OAUTH.get("app_access_token_url"), data=json.dumps( { "app_id": settings.FEISHU_OAUTH.get("app_id"), "app_secret": settings.FEISHU_OAUTH.get("app_secret"), } ), headers={"Content-Type": "application/json; charset=utf-8"}, ) app_access_token_res = app_access_token_res.json() if app_access_token_res["code"] != 0: return redirect("/?error=%s" % app_access_token_res) res = await async_request( "POST", settings.FEISHU_OAUTH.get("user_access_token_url"), data=json.dumps( { "grant_type": settings.FEISHU_OAUTH.get("grant_type"), "code": request.args.get("code"), } ), headers={ "Content-Type": "application/json; charset=utf-8", "Authorization": f"Bearer {app_access_token_res['app_access_token']}", }, ) res = res.json() if res["code"] != 0: return redirect("/?error=%s" % res["message"]) if "contact:user.email:readonly" not in res["data"]["scope"].split(): return redirect("/?error=contact:user.email:readonly not in scope") session["access_token"] = res["data"]["access_token"] session["access_token_from"] = "feishu" user_info = await user_info_from_feishu(session["access_token"]) email_address = user_info["email"] users = UserService.query(email=email_address) user_id = get_uuid() if not users: # User isn't try to register try: try: avatar = download_img(user_info["avatar_url"]) except Exception as e: logging.exception(e) avatar = "" users = user_register( user_id, { "access_token": session["access_token"], "email": email_address, "avatar": avatar, "nickname": user_info["en_name"], "login_channel": "feishu", "last_login_time": get_format_time(), "is_superuser": False, }, ) if not users: raise Exception(f"Fail to register {email_address}.") if len(users) > 1: raise Exception(f"Same email: {email_address} exists!") # Try to log in user = users[0] login_user(user) return redirect("/?auth=%s" % user.get_id()) except Exception as e: rollback_user_registration(user_id) logging.exception(e) return redirect("/?error=%s" % str(e)) # User has already registered, try to log in user = users[0] if user and hasattr(user, 'is_active') and user.is_active == "0": return redirect("/?error=user_inactive") user.access_token = get_uuid() login_user(user) user.save() return redirect("/?auth=%s" % user.get_id()) async def user_info_from_feishu(access_token): headers = { "Content-Type": "application/json; charset=utf-8", "Authorization": f"Bearer {access_token}", } res = await async_request("GET", "https://open.feishu.cn/open-apis/authen/v1/user_info", headers=headers) user_info = res.json()["data"] user_info["email"] = None if user_info.get("email") == "" else user_info["email"] return user_info async def user_info_from_github(access_token): headers = {"Accept": "application/json", "Authorization": f"token {access_token}"} res = await async_request("GET", f"https://api.github.com/user?access_token={access_token}", headers=headers) user_info = res.json() email_info_response = await async_request( "GET", f"https://api.github.com/user/emails?access_token={access_token}", headers=headers, ) email_info = email_info_response.json() user_info["email"] = next((email for email in email_info if email["primary"]), None)["email"] return user_info @manager.route("/logout", methods=["GET"]) # noqa: F821 @login_required async def log_out(): """ User logout endpoint. --- tags: - User security: - ApiKeyAuth: [] responses: 200: description: Logout successful. schema: type: object """ current_user.access_token = f"INVALID_{secrets.token_hex(16)}" current_user.save() logout_user() return get_json_result(data=True) @manager.route("/setting", methods=["POST"]) # noqa: F821 @login_required async def setting_user(): """ Update user settings. --- tags: - User security: - ApiKeyAuth: [] parameters: - in: body name: body description: User settings to update. required: true schema: type: object properties: nickname: type: string description: New nickname. email: type: string description: New email. responses: 200: description: Settings updated successfully. schema: type: object """ update_dict = {} request_data = await get_request_json() if request_data.get("password"): new_password = request_data.get("new_password") if not check_password_hash(current_user.password, decrypt(request_data["password"])): return get_json_result( data=False, code=RetCode.AUTHENTICATION_ERROR, message="Password error!", ) if new_password: update_dict["password"] = generate_password_hash(decrypt(new_password)) for k in request_data.keys(): if k in [ "password", "new_password", "email", "status", "is_superuser", "login_channel", "is_anonymous", "is_active", "is_authenticated", "last_login_time", ]: continue update_dict[k] = request_data[k] try: UserService.update_by_id(current_user.id, update_dict) return get_json_result(data=True) except Exception as e: logging.exception(e) return get_json_result(data=False, message="Update failure!", code=RetCode.EXCEPTION_ERROR) @manager.route("/info", methods=["GET"]) # noqa: F821 @login_required async def user_profile(): """ Get user profile information. --- tags: - User security: - ApiKeyAuth: [] responses: 200: description: User profile retrieved successfully. schema: type: object properties: id: type: string description: User ID. nickname: type: string description: User nickname. email: type: string description: User email. """ return get_json_result(data=current_user.to_dict()) def rollback_user_registration(user_id): try: UserService.delete_by_id(user_id) except Exception: pass try: TenantService.delete_by_id(user_id) except Exception: pass try: u = UserTenantService.query(tenant_id=user_id) if u: UserTenantService.delete_by_id(u[0].id) except Exception: pass try: TenantLLM.delete().where(TenantLLM.tenant_id == user_id).execute() except Exception: pass def user_register(user_id, user): user["id"] = user_id tenant = { "id": user_id, "name": user["nickname"] + "‘s Kingdom", "llm_id": settings.CHAT_MDL, "embd_id": settings.EMBEDDING_MDL, "asr_id": settings.ASR_MDL, "parser_ids": settings.PARSERS, "img2txt_id": settings.IMAGE2TEXT_MDL, "rerank_id": settings.RERANK_MDL, } usr_tenant = { "tenant_id": user_id, "user_id": user_id, "invited_by": user_id, "role": UserTenantRole.OWNER, } file_id = get_uuid() file = { "id": file_id, "parent_id": file_id, "tenant_id": user_id, "created_by": user_id, "name": "/", "type": FileType.FOLDER.value, "size": 0, "location": "", } tenant_llm = get_init_tenant_llm(user_id) if not UserService.save(**user): return None TenantService.insert(**tenant) UserTenantService.insert(**usr_tenant) TenantLLMService.insert_many(tenant_llm) FileService.insert(file) return UserService.query(email=user["email"]) @manager.route("/register", methods=["POST"]) # noqa: F821 @validate_request("nickname", "email", "password") async def user_add(): """ Register a new user. --- tags: - User parameters: - in: body name: body description: Registration details. required: true schema: type: object properties: nickname: type: string description: User nickname. email: type: string description: User email. password: type: string description: User password. responses: 200: description: Registration successful. schema: type: object """ if not settings.REGISTER_ENABLED: return get_json_result( data=False, message="User registration is disabled!", code=RetCode.OPERATING_ERROR, ) req = await get_request_json() email_address = req["email"] # Validate the email address if not re.match(r"^[\w\._-]+@([\w_-]+\.)+[\w-]{2,}$", email_address): return get_json_result( data=False, message=f"Invalid email address: {email_address}!", code=RetCode.OPERATING_ERROR, ) # Check if the email address is already used if UserService.query(email=email_address): return get_json_result( data=False, message=f"Email: {email_address} has already registered!", code=RetCode.OPERATING_ERROR, ) # Construct user info data nickname = req["nickname"] user_dict = { "access_token": get_uuid(), "email": email_address, "nickname": nickname, "password": decrypt(req["password"]), "login_channel": "password", "last_login_time": get_format_time(), "is_superuser": False, } user_id = get_uuid() try: users = user_register(user_id, user_dict) if not users: raise Exception(f"Fail to register {email_address}.") if len(users) > 1: raise Exception(f"Same email: {email_address} exists!") user = users[0] login_user(user) return await construct_response( data=user.to_json(), auth=user.get_id(), message=f"{nickname}, welcome aboard!", ) except Exception as e: rollback_user_registration(user_id) logging.exception(e) return get_json_result( data=False, message=f"User registration failure, error: {str(e)}", code=RetCode.EXCEPTION_ERROR, ) @manager.route("/tenant_info", methods=["GET"]) # noqa: F821 @login_required async def tenant_info(): """ Get tenant information. --- tags: - Tenant security: - ApiKeyAuth: [] responses: 200: description: Tenant information retrieved successfully. schema: type: object properties: tenant_id: type: string description: Tenant ID. name: type: string description: Tenant name. llm_id: type: string description: LLM ID. embd_id: type: string description: Embedding model ID. """ try: tenants = TenantService.get_info_by(current_user.id) if not tenants: return get_data_error_result(message="Tenant not found!") return get_json_result(data=tenants[0]) except Exception as e: return server_error_response(e) @manager.route("/set_tenant_info", methods=["POST"]) # noqa: F821 @login_required @validate_request("tenant_id", "asr_id", "embd_id", "img2txt_id", "llm_id") async def set_tenant_info(): """ Update tenant information. --- tags: - Tenant security: - ApiKeyAuth: [] parameters: - in: body name: body description: Tenant information to update. required: true schema: type: object properties: tenant_id: type: string description: Tenant ID. llm_id: type: string description: LLM ID. embd_id: type: string description: Embedding model ID. asr_id: type: string description: ASR model ID. img2txt_id: type: string description: Image to Text model ID. responses: 200: description: Tenant information updated successfully. schema: type: object """ req = await get_request_json() try: tid = req.pop("tenant_id") TenantService.update_by_id(tid, req) return get_json_result(data=True) except Exception as e: return server_error_response(e) @manager.route("/forget/captcha", methods=["GET"]) # noqa: F821 async def forget_get_captcha(): """ GET /forget/captcha?email=<email> - Generate an image captcha and cache it in Redis under key captcha:{email} with TTL = OTP_TTL_SECONDS. - Returns the captcha as a PNG image. """ email = (request.args.get("email") or "") if not email: return get_json_result(data=False, code=RetCode.ARGUMENT_ERROR, message="email is required") users = UserService.query(email=email) if not users: return get_json_result(data=False, code=RetCode.DATA_ERROR, message="invalid email") # Generate captcha text allowed = string.ascii_uppercase + string.digits captcha_text = "".join(secrets.choice(allowed) for _ in range(OTP_LENGTH)) REDIS_CONN.set(captcha_key(email), captcha_text, 60) # Valid for 60 seconds from captcha.image import ImageCaptcha image = ImageCaptcha(width=300, height=120, font_sizes=[50, 60, 70]) img_bytes = image.generate(captcha_text).read() response = await make_response(img_bytes) response.headers.set("Content-Type", "image/JPEG") return response @manager.route("/forget/otp", methods=["POST"]) # noqa: F821 async def forget_send_otp(): """ POST /forget/otp - Verify the image captcha stored at captcha:{email} (case-insensitive). - On success, generate an email OTP (A–Z with length = OTP_LENGTH), store hash + salt (and timestamp) in Redis with TTL, reset attempts and cooldown, and send the OTP via email. """ req = await get_request_json() email = req.get("email") or "" captcha = (req.get("captcha") or "").strip() if not email or not captcha: return get_json_result(data=False, code=RetCode.ARGUMENT_ERROR, message="email and captcha required") users = UserService.query(email=email) if not users: return get_json_result(data=False, code=RetCode.DATA_ERROR, message="invalid email") stored_captcha = REDIS_CONN.get(captcha_key(email)) if not stored_captcha: return get_json_result(data=False, code=RetCode.NOT_EFFECTIVE, message="invalid or expired captcha") if (stored_captcha or "").strip().lower() != captcha.lower(): return get_json_result(data=False, code=RetCode.AUTHENTICATION_ERROR, message="invalid or expired captcha") # Delete captcha to prevent reuse REDIS_CONN.delete(captcha_key(email)) k_code, k_attempts, k_last, k_lock = otp_keys(email) now = int(time.time()) last_ts = REDIS_CONN.get(k_last) if last_ts: try: elapsed = now - int(last_ts) except Exception: elapsed = RESEND_COOLDOWN_SECONDS remaining = RESEND_COOLDOWN_SECONDS - elapsed if remaining > 0: return get_json_result(data=False, code=RetCode.NOT_EFFECTIVE, message=f"you still have to wait {remaining} seconds") # Generate OTP (uppercase letters only) and store hashed otp = "".join(secrets.choice(string.ascii_uppercase) for _ in range(OTP_LENGTH)) salt = os.urandom(16) code_hash = hash_code(otp, salt) REDIS_CONN.set(k_code, f"{code_hash}:{salt.hex()}", OTP_TTL_SECONDS) REDIS_CONN.set(k_attempts, 0, OTP_TTL_SECONDS) REDIS_CONN.set(k_last, now, OTP_TTL_SECONDS) REDIS_CONN.delete(k_lock) ttl_min = OTP_TTL_SECONDS // 60 try: await send_email_html( subject="Your Password Reset Code", to_email=email, template_key="reset_code", code=otp, ttl_min=ttl_min, ) except Exception as e: logging.exception(e) return get_json_result(data=False, code=RetCode.SERVER_ERROR, message="failed to send email") return get_json_result(data=True, code=RetCode.SUCCESS, message="verification passed, email sent") def _verified_key(email: str) -> str: return f"otp:verified:{email}" @manager.route("/forget/verify-otp", methods=["POST"]) # noqa: F821 async def forget_verify_otp(): """ Verify email + OTP only. On success: - consume the OTP and attempt counters - set a short-lived verified flag in Redis for the email Request JSON: { email, otp } """ req = await get_request_json() email = req.get("email") or "" otp = (req.get("otp") or "").strip() if not all([email, otp]): return get_json_result(data=False, code=RetCode.ARGUMENT_ERROR, message="email and otp are required") users = UserService.query(email=email) if not users: return get_json_result(data=False, code=RetCode.DATA_ERROR, message="invalid email") # Verify OTP from Redis k_code, k_attempts, k_last, k_lock = otp_keys(email) if REDIS_CONN.get(k_lock): return get_json_result(data=False, code=RetCode.NOT_EFFECTIVE, message="too many attempts, try later") stored = REDIS_CONN.get(k_code) if not stored: return get_json_result(data=False, code=RetCode.NOT_EFFECTIVE, message="expired otp") try: stored_hash, salt_hex = str(stored).split(":", 1) salt = bytes.fromhex(salt_hex) except Exception: return get_json_result(data=False, code=RetCode.EXCEPTION_ERROR, message="otp storage corrupted") calc = hash_code(otp.upper(), salt) if calc != stored_hash: # bump attempts try: attempts = int(REDIS_CONN.get(k_attempts) or 0) + 1 except Exception: attempts = 1 REDIS_CONN.set(k_attempts, attempts, OTP_TTL_SECONDS) if attempts >= ATTEMPT_LIMIT: REDIS_CONN.set(k_lock, int(time.time()), ATTEMPT_LOCK_SECONDS) return get_json_result(data=False, code=RetCode.AUTHENTICATION_ERROR, message="expired otp") # Success: consume OTP and attempts; mark verified REDIS_CONN.delete(k_code) REDIS_CONN.delete(k_attempts) REDIS_CONN.delete(k_last) REDIS_CONN.delete(k_lock) # set verified flag with limited TTL, reuse OTP_TTL_SECONDS or smaller window try: REDIS_CONN.set(_verified_key(email), "1", OTP_TTL_SECONDS) except Exception: return get_json_result(data=False, code=RetCode.SERVER_ERROR, message="failed to set verification state") return get_json_result(data=True, code=RetCode.SUCCESS, message="otp verified") @manager.route("/forget/reset-password", methods=["POST"]) # noqa: F821 async def forget_reset_password(): """ Reset password after successful OTP verification. Requires: { email, new_password, confirm_new_password } Steps: - check verified flag in Redis
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
true
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/conversation_app.py
api/apps/conversation_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import json import os import re import logging from copy import deepcopy import tempfile from quart import Response, request from api.apps import current_user, login_required from api.db.db_models import APIToken from api.db.services.conversation_service import ConversationService, structure_answer from api.db.services.dialog_service import DialogService, async_ask, async_chat, gen_mindmap from api.db.services.llm_service import LLMBundle from api.db.services.search_service import SearchService from api.db.services.tenant_llm_service import TenantLLMService from api.db.services.user_service import TenantService, UserTenantService from api.utils.api_utils import get_data_error_result, get_json_result, get_request_json, server_error_response, validate_request from rag.prompts.template import load_prompt from rag.prompts.generator import chunks_format from common.constants import RetCode, LLMType @manager.route("/set", methods=["POST"]) # noqa: F821 @login_required async def set_conversation(): req = await get_request_json() conv_id = req.get("conversation_id") is_new = req.get("is_new") name = req.get("name", "New conversation") req["user_id"] = current_user.id if len(name) > 255: name = name[0:255] del req["is_new"] if not is_new: del req["conversation_id"] try: if not ConversationService.update_by_id(conv_id, req): return get_data_error_result(message="Conversation not found!") e, conv = ConversationService.get_by_id(conv_id) if not e: return get_data_error_result(message="Fail to update a conversation!") conv = conv.to_dict() return get_json_result(data=conv) except Exception as e: return server_error_response(e) try: e, dia = DialogService.get_by_id(req["dialog_id"]) if not e: return get_data_error_result(message="Dialog not found") conv = { "id": conv_id, "dialog_id": req["dialog_id"], "name": name, "message": [{"role": "assistant", "content": dia.prompt_config["prologue"]}], "user_id": current_user.id, "reference": [], } ConversationService.save(**conv) return get_json_result(data=conv) except Exception as e: return server_error_response(e) @manager.route("/get", methods=["GET"]) # noqa: F821 @login_required async def get(): conv_id = request.args["conversation_id"] try: e, conv = ConversationService.get_by_id(conv_id) if not e: return get_data_error_result(message="Conversation not found!") tenants = UserTenantService.query(user_id=current_user.id) for tenant in tenants: dialog = DialogService.query(tenant_id=tenant.tenant_id, id=conv.dialog_id) if dialog and len(dialog) > 0: avatar = dialog[0].icon break else: return get_json_result(data=False, message="Only owner of conversation authorized for this operation.", code=RetCode.OPERATING_ERROR) for ref in conv.reference: if isinstance(ref, list): continue ref["chunks"] = chunks_format(ref) conv = conv.to_dict() conv["avatar"] = avatar return get_json_result(data=conv) except Exception as e: return server_error_response(e) @manager.route("/getsse/<dialog_id>", methods=["GET"]) # type: ignore # noqa: F821 def getsse(dialog_id): token = request.headers.get("Authorization").split() if len(token) != 2: return get_data_error_result(message='Authorization is not valid!"') token = token[1] objs = APIToken.query(beta=token) if not objs: return get_data_error_result(message='Authentication error: API key is invalid!"') try: e, conv = DialogService.get_by_id(dialog_id) if not e: return get_data_error_result(message="Dialog not found!") conv = conv.to_dict() conv["avatar"] = conv["icon"] del conv["icon"] return get_json_result(data=conv) except Exception as e: return server_error_response(e) @manager.route("/rm", methods=["POST"]) # noqa: F821 @login_required async def rm(): req = await get_request_json() conv_ids = req["conversation_ids"] try: for cid in conv_ids: exist, conv = ConversationService.get_by_id(cid) if not exist: return get_data_error_result(message="Conversation not found!") tenants = UserTenantService.query(user_id=current_user.id) for tenant in tenants: if DialogService.query(tenant_id=tenant.tenant_id, id=conv.dialog_id): break else: return get_json_result(data=False, message="Only owner of conversation authorized for this operation.", code=RetCode.OPERATING_ERROR) ConversationService.delete_by_id(cid) return get_json_result(data=True) except Exception as e: return server_error_response(e) @manager.route("/list", methods=["GET"]) # noqa: F821 @login_required async def list_conversation(): dialog_id = request.args["dialog_id"] try: if not DialogService.query(tenant_id=current_user.id, id=dialog_id): return get_json_result(data=False, message="Only owner of dialog authorized for this operation.", code=RetCode.OPERATING_ERROR) convs = ConversationService.query(dialog_id=dialog_id, order_by=ConversationService.model.create_time, reverse=True) convs = [d.to_dict() for d in convs] return get_json_result(data=convs) except Exception as e: return server_error_response(e) @manager.route("/completion", methods=["POST"]) # noqa: F821 @login_required @validate_request("conversation_id", "messages") async def completion(): req = await get_request_json() msg = [] for m in req["messages"]: if m["role"] == "system": continue if m["role"] == "assistant" and not msg: continue msg.append(m) message_id = msg[-1].get("id") chat_model_id = req.get("llm_id", "") req.pop("llm_id", None) chat_model_config = {} for model_config in [ "temperature", "top_p", "frequency_penalty", "presence_penalty", "max_tokens", ]: config = req.get(model_config) if config: chat_model_config[model_config] = config try: e, conv = ConversationService.get_by_id(req["conversation_id"]) if not e: return get_data_error_result(message="Conversation not found!") conv.message = deepcopy(req["messages"]) e, dia = DialogService.get_by_id(conv.dialog_id) if not e: return get_data_error_result(message="Dialog not found!") del req["conversation_id"] del req["messages"] if not conv.reference: conv.reference = [] conv.reference = [r for r in conv.reference if r] conv.reference.append({"chunks": [], "doc_aggs": []}) if chat_model_id: if not TenantLLMService.get_api_key(tenant_id=dia.tenant_id, model_name=chat_model_id): req.pop("chat_model_id", None) req.pop("chat_model_config", None) return get_data_error_result(message=f"Cannot use specified model {chat_model_id}.") dia.llm_id = chat_model_id dia.llm_setting = chat_model_config is_embedded = bool(chat_model_id) async def stream(): nonlocal dia, msg, req, conv try: async for ans in async_chat(dia, msg, True, **req): ans = structure_answer(conv, ans, message_id, conv.id) yield "data:" + json.dumps({"code": 0, "message": "", "data": ans}, ensure_ascii=False) + "\n\n" if not is_embedded: ConversationService.update_by_id(conv.id, conv.to_dict()) except Exception as e: logging.exception(e) yield "data:" + json.dumps({"code": 500, "message": str(e), "data": {"answer": "**ERROR**: " + str(e), "reference": []}}, ensure_ascii=False) + "\n\n" yield "data:" + json.dumps({"code": 0, "message": "", "data": True}, ensure_ascii=False) + "\n\n" if req.get("stream", True): resp = Response(stream(), mimetype="text/event-stream") resp.headers.add_header("Cache-control", "no-cache") resp.headers.add_header("Connection", "keep-alive") resp.headers.add_header("X-Accel-Buffering", "no") resp.headers.add_header("Content-Type", "text/event-stream; charset=utf-8") return resp else: answer = None async for ans in async_chat(dia, msg, **req): answer = structure_answer(conv, ans, message_id, conv.id) if not is_embedded: ConversationService.update_by_id(conv.id, conv.to_dict()) break return get_json_result(data=answer) except Exception as e: return server_error_response(e) @manager.route("/sequence2txt", methods=["POST"]) # noqa: F821 @login_required async def sequence2txt(): req = await request.form stream_mode = req.get("stream", "false").lower() == "true" files = await request.files if "file" not in files: return get_data_error_result(message="Missing 'file' in multipart form-data") uploaded = files["file"] ALLOWED_EXTS = { ".wav", ".mp3", ".m4a", ".aac", ".flac", ".ogg", ".webm", ".opus", ".wma" } filename = uploaded.filename or "" suffix = os.path.splitext(filename)[-1].lower() if suffix not in ALLOWED_EXTS: return get_data_error_result(message= f"Unsupported audio format: {suffix}. " f"Allowed: {', '.join(sorted(ALLOWED_EXTS))}" ) fd, temp_audio_path = tempfile.mkstemp(suffix=suffix) os.close(fd) await uploaded.save(temp_audio_path) tenants = TenantService.get_info_by(current_user.id) if not tenants: return get_data_error_result(message="Tenant not found!") asr_id = tenants[0]["asr_id"] if not asr_id: return get_data_error_result(message="No default ASR model is set") asr_mdl=LLMBundle(tenants[0]["tenant_id"], LLMType.SPEECH2TEXT, asr_id) if not stream_mode: text = asr_mdl.transcription(temp_audio_path) try: os.remove(temp_audio_path) except Exception as e: logging.error(f"Failed to remove temp audio file: {str(e)}") return get_json_result(data={"text": text}) async def event_stream(): try: for evt in asr_mdl.stream_transcription(temp_audio_path): yield f"data: {json.dumps(evt, ensure_ascii=False)}\n\n" except Exception as e: err = {"event": "error", "text": str(e)} yield f"data: {json.dumps(err, ensure_ascii=False)}\n\n" finally: try: os.remove(temp_audio_path) except Exception as e: logging.error(f"Failed to remove temp audio file: {str(e)}") return Response(event_stream(), content_type="text/event-stream") @manager.route("/tts", methods=["POST"]) # noqa: F821 @login_required async def tts(): req = await get_request_json() text = req["text"] tenants = TenantService.get_info_by(current_user.id) if not tenants: return get_data_error_result(message="Tenant not found!") tts_id = tenants[0]["tts_id"] if not tts_id: return get_data_error_result(message="No default TTS model is set") tts_mdl = LLMBundle(tenants[0]["tenant_id"], LLMType.TTS, tts_id) def stream_audio(): try: for txt in re.split(r"[,。/《》?;:!\n\r:;]+", text): for chunk in tts_mdl.tts(txt): yield chunk except Exception as e: yield ("data:" + json.dumps({"code": 500, "message": str(e), "data": {"answer": "**ERROR**: " + str(e)}}, ensure_ascii=False)).encode("utf-8") resp = Response(stream_audio(), mimetype="audio/mpeg") resp.headers.add_header("Cache-Control", "no-cache") resp.headers.add_header("Connection", "keep-alive") resp.headers.add_header("X-Accel-Buffering", "no") return resp @manager.route("/delete_msg", methods=["POST"]) # noqa: F821 @login_required @validate_request("conversation_id", "message_id") async def delete_msg(): req = await get_request_json() e, conv = ConversationService.get_by_id(req["conversation_id"]) if not e: return get_data_error_result(message="Conversation not found!") conv = conv.to_dict() for i, msg in enumerate(conv["message"]): if req["message_id"] != msg.get("id", ""): continue assert conv["message"][i + 1]["id"] == req["message_id"] conv["message"].pop(i) conv["message"].pop(i) conv["reference"].pop(max(0, i // 2 - 1)) break ConversationService.update_by_id(conv["id"], conv) return get_json_result(data=conv) @manager.route("/thumbup", methods=["POST"]) # noqa: F821 @login_required @validate_request("conversation_id", "message_id") async def thumbup(): req = await get_request_json() e, conv = ConversationService.get_by_id(req["conversation_id"]) if not e: return get_data_error_result(message="Conversation not found!") up_down = req.get("thumbup") feedback = req.get("feedback", "") conv = conv.to_dict() for i, msg in enumerate(conv["message"]): if req["message_id"] == msg.get("id", "") and msg.get("role", "") == "assistant": if up_down: msg["thumbup"] = True if "feedback" in msg: del msg["feedback"] else: msg["thumbup"] = False if feedback: msg["feedback"] = feedback break ConversationService.update_by_id(conv["id"], conv) return get_json_result(data=conv) @manager.route("/ask", methods=["POST"]) # noqa: F821 @login_required @validate_request("question", "kb_ids") async def ask_about(): req = await get_request_json() uid = current_user.id search_id = req.get("search_id", "") search_app = None search_config = {} if search_id: search_app = SearchService.get_detail(search_id) if search_app: search_config = search_app.get("search_config", {}) async def stream(): nonlocal req, uid try: async for ans in async_ask(req["question"], req["kb_ids"], uid, search_config=search_config): yield "data:" + json.dumps({"code": 0, "message": "", "data": ans}, ensure_ascii=False) + "\n\n" except Exception as e: yield "data:" + json.dumps({"code": 500, "message": str(e), "data": {"answer": "**ERROR**: " + str(e), "reference": []}}, ensure_ascii=False) + "\n\n" yield "data:" + json.dumps({"code": 0, "message": "", "data": True}, ensure_ascii=False) + "\n\n" resp = Response(stream(), mimetype="text/event-stream") resp.headers.add_header("Cache-control", "no-cache") resp.headers.add_header("Connection", "keep-alive") resp.headers.add_header("X-Accel-Buffering", "no") resp.headers.add_header("Content-Type", "text/event-stream; charset=utf-8") return resp @manager.route("/mindmap", methods=["POST"]) # noqa: F821 @login_required @validate_request("question", "kb_ids") async def mindmap(): req = await get_request_json() search_id = req.get("search_id", "") search_app = SearchService.get_detail(search_id) if search_id else {} search_config = search_app.get("search_config", {}) if search_app else {} kb_ids = search_config.get("kb_ids", []) kb_ids.extend(req["kb_ids"]) kb_ids = list(set(kb_ids)) mind_map = await gen_mindmap(req["question"], kb_ids, search_app.get("tenant_id", current_user.id), search_config) if "error" in mind_map: return server_error_response(Exception(mind_map["error"])) return get_json_result(data=mind_map) @manager.route("/related_questions", methods=["POST"]) # noqa: F821 @login_required @validate_request("question") async def related_questions(): req = await get_request_json() search_id = req.get("search_id", "") search_config = {} if search_id: if search_app := SearchService.get_detail(search_id): search_config = search_app.get("search_config", {}) question = req["question"] chat_id = search_config.get("chat_id", "") chat_mdl = LLMBundle(current_user.id, LLMType.CHAT, chat_id) gen_conf = search_config.get("llm_setting", {"temperature": 0.9}) if "parameter" in gen_conf: del gen_conf["parameter"] prompt = load_prompt("related_question") ans = await chat_mdl.async_chat( prompt, [ { "role": "user", "content": f""" Keywords: {question} Related search terms: """, } ], gen_conf, ) return get_json_result(data=[re.sub(r"^[0-9]\. ", "", a) for a in ans.split("\n") if re.match(r"^[0-9]\. ", a)])
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/file_app.py
api/apps/file_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License # import logging import asyncio import os import pathlib import re from quart import request, make_response from api.apps import login_required, current_user from api.common.check_team_permission import check_file_team_permission from api.db.services.document_service import DocumentService from api.db.services.file2document_service import File2DocumentService from api.utils.api_utils import server_error_response, get_data_error_result, validate_request from common.misc_utils import get_uuid from common.constants import RetCode, FileSource from api.db import FileType from api.db.services import duplicate_name from api.db.services.file_service import FileService from api.utils.api_utils import get_json_result, get_request_json from api.utils.file_utils import filename_type from api.utils.web_utils import CONTENT_TYPE_MAP from common import settings @manager.route('/upload', methods=['POST']) # noqa: F821 @login_required # @validate_request("parent_id") async def upload(): form = await request.form pf_id = form.get("parent_id") if not pf_id: root_folder = FileService.get_root_folder(current_user.id) pf_id = root_folder["id"] files = await request.files if 'file' not in files: return get_json_result( data=False, message='No file part!', code=RetCode.ARGUMENT_ERROR) file_objs = files.getlist('file') for file_obj in file_objs: if file_obj.filename == '': return get_json_result( data=False, message='No file selected!', code=RetCode.ARGUMENT_ERROR) file_res = [] try: e, pf_folder = FileService.get_by_id(pf_id) if not e: return get_data_error_result( message="Can't find this folder!") async def _handle_single_file(file_obj): MAX_FILE_NUM_PER_USER: int = int(os.environ.get('MAX_FILE_NUM_PER_USER', 0)) if 0 < MAX_FILE_NUM_PER_USER <= await asyncio.to_thread(DocumentService.get_doc_count, current_user.id): return get_data_error_result( message="Exceed the maximum file number of a free user!") # split file name path if not file_obj.filename: file_obj_names = [pf_folder.name, file_obj.filename] else: full_path = '/' + file_obj.filename file_obj_names = full_path.split('/') file_len = len(file_obj_names) # get folder file_id_list = await asyncio.to_thread(FileService.get_id_list_by_id, pf_id, file_obj_names, 1, [pf_id]) len_id_list = len(file_id_list) # create folder if file_len != len_id_list: e, file = await asyncio.to_thread(FileService.get_by_id, file_id_list[len_id_list - 1]) if not e: return get_data_error_result(message="Folder not found!") last_folder = await asyncio.to_thread(FileService.create_folder, file, file_id_list[len_id_list - 1], file_obj_names, len_id_list) else: e, file = await asyncio.to_thread(FileService.get_by_id, file_id_list[len_id_list - 2]) if not e: return get_data_error_result(message="Folder not found!") last_folder = await asyncio.to_thread(FileService.create_folder, file, file_id_list[len_id_list - 2], file_obj_names, len_id_list) # file type filetype = filename_type(file_obj_names[file_len - 1]) location = file_obj_names[file_len - 1] while await asyncio.to_thread(settings.STORAGE_IMPL.obj_exist, last_folder.id, location): location += "_" blob = await asyncio.to_thread(file_obj.read) filename = await asyncio.to_thread( duplicate_name, FileService.query, name=file_obj_names[file_len - 1], parent_id=last_folder.id) await asyncio.to_thread(settings.STORAGE_IMPL.put, last_folder.id, location, blob) file_data = { "id": get_uuid(), "parent_id": last_folder.id, "tenant_id": current_user.id, "created_by": current_user.id, "type": filetype, "name": filename, "location": location, "size": len(blob), } inserted = await asyncio.to_thread(FileService.insert, file_data) return inserted.to_json() for file_obj in file_objs: res = await _handle_single_file(file_obj) file_res.append(res) return get_json_result(data=file_res) except Exception as e: return server_error_response(e) @manager.route('/create', methods=['POST']) # noqa: F821 @login_required @validate_request("name") async def create(): req = await get_request_json() pf_id = req.get("parent_id") input_file_type = req.get("type") if not pf_id: root_folder = FileService.get_root_folder(current_user.id) pf_id = root_folder["id"] try: if not FileService.is_parent_folder_exist(pf_id): return get_json_result( data=False, message="Parent Folder Doesn't Exist!", code=RetCode.OPERATING_ERROR) if FileService.query(name=req["name"], parent_id=pf_id): return get_data_error_result( message="Duplicated folder name in the same folder.") if input_file_type == FileType.FOLDER.value: file_type = FileType.FOLDER.value else: file_type = FileType.VIRTUAL.value file = FileService.insert({ "id": get_uuid(), "parent_id": pf_id, "tenant_id": current_user.id, "created_by": current_user.id, "name": req["name"], "location": "", "size": 0, "type": file_type }) return get_json_result(data=file.to_json()) except Exception as e: return server_error_response(e) @manager.route('/list', methods=['GET']) # noqa: F821 @login_required def list_files(): pf_id = request.args.get("parent_id") keywords = request.args.get("keywords", "") page_number = int(request.args.get("page", 1)) items_per_page = int(request.args.get("page_size", 15)) orderby = request.args.get("orderby", "create_time") desc = request.args.get("desc", True) if not pf_id: root_folder = FileService.get_root_folder(current_user.id) pf_id = root_folder["id"] FileService.init_knowledgebase_docs(pf_id, current_user.id) try: e, file = FileService.get_by_id(pf_id) if not e: return get_data_error_result(message="Folder not found!") files, total = FileService.get_by_pf_id( current_user.id, pf_id, page_number, items_per_page, orderby, desc, keywords) parent_folder = FileService.get_parent_folder(pf_id) if not parent_folder: return get_json_result(message="File not found!") return get_json_result(data={"total": total, "files": files, "parent_folder": parent_folder.to_json()}) except Exception as e: return server_error_response(e) @manager.route('/root_folder', methods=['GET']) # noqa: F821 @login_required def get_root_folder(): try: root_folder = FileService.get_root_folder(current_user.id) return get_json_result(data={"root_folder": root_folder}) except Exception as e: return server_error_response(e) @manager.route('/parent_folder', methods=['GET']) # noqa: F821 @login_required def get_parent_folder(): file_id = request.args.get("file_id") try: e, file = FileService.get_by_id(file_id) if not e: return get_data_error_result(message="Folder not found!") parent_folder = FileService.get_parent_folder(file_id) return get_json_result(data={"parent_folder": parent_folder.to_json()}) except Exception as e: return server_error_response(e) @manager.route('/all_parent_folder', methods=['GET']) # noqa: F821 @login_required def get_all_parent_folders(): file_id = request.args.get("file_id") try: e, file = FileService.get_by_id(file_id) if not e: return get_data_error_result(message="Folder not found!") parent_folders = FileService.get_all_parent_folders(file_id) parent_folders_res = [] for parent_folder in parent_folders: parent_folders_res.append(parent_folder.to_json()) return get_json_result(data={"parent_folders": parent_folders_res}) except Exception as e: return server_error_response(e) @manager.route("/rm", methods=["POST"]) # noqa: F821 @login_required @validate_request("file_ids") async def rm(): req = await get_request_json() file_ids = req["file_ids"] try: def _delete_single_file(file): try: if file.location: settings.STORAGE_IMPL.rm(file.parent_id, file.location) except Exception as e: logging.exception(f"Fail to remove object: {file.parent_id}/{file.location}, error: {e}") informs = File2DocumentService.get_by_file_id(file.id) for inform in informs: doc_id = inform.document_id e, doc = DocumentService.get_by_id(doc_id) if e and doc: tenant_id = DocumentService.get_tenant_id(doc_id) if tenant_id: DocumentService.remove_document(doc, tenant_id) File2DocumentService.delete_by_file_id(file.id) FileService.delete(file) def _delete_folder_recursive(folder, tenant_id): sub_files = FileService.list_all_files_by_parent_id(folder.id) for sub_file in sub_files: if sub_file.type == FileType.FOLDER.value: _delete_folder_recursive(sub_file, tenant_id) else: _delete_single_file(sub_file) FileService.delete(folder) def _rm_sync(): for file_id in file_ids: e, file = FileService.get_by_id(file_id) if not e or not file: return get_data_error_result(message="File or Folder not found!") if not file.tenant_id: return get_data_error_result(message="Tenant not found!") if not check_file_team_permission(file, current_user.id): return get_json_result(data=False, message="No authorization.", code=RetCode.AUTHENTICATION_ERROR) if file.source_type == FileSource.KNOWLEDGEBASE: continue if file.type == FileType.FOLDER.value: _delete_folder_recursive(file, current_user.id) continue _delete_single_file(file) return get_json_result(data=True) return await asyncio.to_thread(_rm_sync) except Exception as e: return server_error_response(e) @manager.route('/rename', methods=['POST']) # noqa: F821 @login_required @validate_request("file_id", "name") async def rename(): req = await get_request_json() try: e, file = FileService.get_by_id(req["file_id"]) if not e: return get_data_error_result(message="File not found!") if not check_file_team_permission(file, current_user.id): return get_json_result(data=False, message='No authorization.', code=RetCode.AUTHENTICATION_ERROR) if file.type != FileType.FOLDER.value \ and pathlib.Path(req["name"].lower()).suffix != pathlib.Path( file.name.lower()).suffix: return get_json_result( data=False, message="The extension of file can't be changed", code=RetCode.ARGUMENT_ERROR) for file in FileService.query(name=req["name"], pf_id=file.parent_id): if file.name == req["name"]: return get_data_error_result( message="Duplicated file name in the same folder.") if not FileService.update_by_id( req["file_id"], {"name": req["name"]}): return get_data_error_result( message="Database error (File rename)!") informs = File2DocumentService.get_by_file_id(req["file_id"]) if informs: if not DocumentService.update_by_id( informs[0].document_id, {"name": req["name"]}): return get_data_error_result( message="Database error (Document rename)!") return get_json_result(data=True) except Exception as e: return server_error_response(e) @manager.route('/get/<file_id>', methods=['GET']) # noqa: F821 @login_required async def get(file_id): try: e, file = FileService.get_by_id(file_id) if not e: return get_data_error_result(message="Document not found!") if not check_file_team_permission(file, current_user.id): return get_json_result(data=False, message='No authorization.', code=RetCode.AUTHENTICATION_ERROR) blob = await asyncio.to_thread(settings.STORAGE_IMPL.get, file.parent_id, file.location) if not blob: b, n = File2DocumentService.get_storage_address(file_id=file_id) blob = await asyncio.to_thread(settings.STORAGE_IMPL.get, b, n) response = await make_response(blob) ext = re.search(r"\.([^.]+)$", file.name.lower()) ext = ext.group(1) if ext else None if ext: if file.type == FileType.VISUAL.value: content_type = CONTENT_TYPE_MAP.get(ext, f"image/{ext}") else: content_type = CONTENT_TYPE_MAP.get(ext, f"application/{ext}") response.headers.set("Content-Type", content_type) return response except Exception as e: return server_error_response(e) @manager.route("/mv", methods=["POST"]) # noqa: F821 @login_required @validate_request("src_file_ids", "dest_file_id") async def move(): req = await get_request_json() try: file_ids = req["src_file_ids"] dest_parent_id = req["dest_file_id"] ok, dest_folder = FileService.get_by_id(dest_parent_id) if not ok or not dest_folder: return get_data_error_result(message="Parent folder not found!") files = FileService.get_by_ids(file_ids) if not files: return get_data_error_result(message="Source files not found!") files_dict = {f.id: f for f in files} for file_id in file_ids: file = files_dict.get(file_id) if not file: return get_data_error_result(message="File or folder not found!") if not file.tenant_id: return get_data_error_result(message="Tenant not found!") if not check_file_team_permission(file, current_user.id): return get_json_result( data=False, message="No authorization.", code=RetCode.AUTHENTICATION_ERROR, ) def _move_entry_recursive(source_file_entry, dest_folder): if source_file_entry.type == FileType.FOLDER.value: existing_folder = FileService.query(name=source_file_entry.name, parent_id=dest_folder.id) if existing_folder: new_folder = existing_folder[0] else: new_folder = FileService.insert( { "id": get_uuid(), "parent_id": dest_folder.id, "tenant_id": source_file_entry.tenant_id, "created_by": current_user.id, "name": source_file_entry.name, "location": "", "size": 0, "type": FileType.FOLDER.value, } ) sub_files = FileService.list_all_files_by_parent_id(source_file_entry.id) for sub_file in sub_files: _move_entry_recursive(sub_file, new_folder) FileService.delete_by_id(source_file_entry.id) return old_parent_id = source_file_entry.parent_id old_location = source_file_entry.location filename = source_file_entry.name new_location = filename while settings.STORAGE_IMPL.obj_exist(dest_folder.id, new_location): new_location += "_" try: settings.STORAGE_IMPL.move(old_parent_id, old_location, dest_folder.id, new_location) except Exception as storage_err: raise RuntimeError(f"Move file failed at storage layer: {str(storage_err)}") FileService.update_by_id( source_file_entry.id, { "parent_id": dest_folder.id, "location": new_location, }, ) def _move_sync(): for file in files: _move_entry_recursive(file, dest_folder) return get_json_result(data=True) return await asyncio.to_thread(_move_sync) except Exception as e: return server_error_response(e)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/chunk_app.py
api/apps/chunk_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import asyncio import datetime import json import re import base64 import xxhash from quart import request from api.db.services.document_service import DocumentService from api.db.services.knowledgebase_service import KnowledgebaseService from api.db.services.llm_service import LLMBundle from common.metadata_utils import apply_meta_data_filter from api.db.services.search_service import SearchService from api.db.services.user_service import UserTenantService from api.utils.api_utils import get_data_error_result, get_json_result, server_error_response, validate_request, \ get_request_json from rag.app.qa import beAdoc, rmPrefix from rag.app.tag import label_question from rag.nlp import rag_tokenizer, search from rag.prompts.generator import cross_languages, keyword_extraction from common.string_utils import remove_redundant_spaces from common.constants import RetCode, LLMType, ParserType, PAGERANK_FLD from common import settings from api.apps import login_required, current_user @manager.route('/list', methods=['POST']) # noqa: F821 @login_required @validate_request("doc_id") async def list_chunk(): req = await get_request_json() doc_id = req["doc_id"] page = int(req.get("page", 1)) size = int(req.get("size", 30)) question = req.get("keywords", "") try: tenant_id = DocumentService.get_tenant_id(req["doc_id"]) if not tenant_id: return get_data_error_result(message="Tenant not found!") e, doc = DocumentService.get_by_id(doc_id) if not e: return get_data_error_result(message="Document not found!") kb_ids = KnowledgebaseService.get_kb_ids(tenant_id) query = { "doc_ids": [doc_id], "page": page, "size": size, "question": question, "sort": True } if "available_int" in req: query["available_int"] = int(req["available_int"]) sres = settings.retriever.search(query, search.index_name(tenant_id), kb_ids, highlight=["content_ltks"]) res = {"total": sres.total, "chunks": [], "doc": doc.to_dict()} for id in sres.ids: d = { "chunk_id": id, "content_with_weight": remove_redundant_spaces(sres.highlight[id]) if question and id in sres.highlight else sres.field[ id].get( "content_with_weight", ""), "doc_id": sres.field[id]["doc_id"], "docnm_kwd": sres.field[id]["docnm_kwd"], "important_kwd": sres.field[id].get("important_kwd", []), "question_kwd": sres.field[id].get("question_kwd", []), "image_id": sres.field[id].get("img_id", ""), "available_int": int(sres.field[id].get("available_int", 1)), "positions": sres.field[id].get("position_int", []), "doc_type_kwd": sres.field[id].get("doc_type_kwd") } assert isinstance(d["positions"], list) assert len(d["positions"]) == 0 or (isinstance(d["positions"][0], list) and len(d["positions"][0]) == 5) res["chunks"].append(d) return get_json_result(data=res) except Exception as e: if str(e).find("not_found") > 0: return get_json_result(data=False, message='No chunk found!', code=RetCode.DATA_ERROR) return server_error_response(e) @manager.route('/get', methods=['GET']) # noqa: F821 @login_required def get(): chunk_id = request.args["chunk_id"] try: chunk = None tenants = UserTenantService.query(user_id=current_user.id) if not tenants: return get_data_error_result(message="Tenant not found!") for tenant in tenants: kb_ids = KnowledgebaseService.get_kb_ids(tenant.tenant_id) chunk = settings.docStoreConn.get(chunk_id, search.index_name(tenant.tenant_id), kb_ids) if chunk: break if chunk is None: return server_error_response(Exception("Chunk not found")) k = [] for n in chunk.keys(): if re.search(r"(_vec$|_sm_|_tks|_ltks)", n): k.append(n) for n in k: del chunk[n] return get_json_result(data=chunk) except Exception as e: if str(e).find("NotFoundError") >= 0: return get_json_result(data=False, message='Chunk not found!', code=RetCode.DATA_ERROR) return server_error_response(e) @manager.route('/set', methods=['POST']) # noqa: F821 @login_required @validate_request("doc_id", "chunk_id", "content_with_weight") async def set(): req = await get_request_json() d = { "id": req["chunk_id"], "content_with_weight": req["content_with_weight"]} d["content_ltks"] = rag_tokenizer.tokenize(req["content_with_weight"]) d["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(d["content_ltks"]) if "important_kwd" in req: if not isinstance(req["important_kwd"], list): return get_data_error_result(message="`important_kwd` should be a list") d["important_kwd"] = req["important_kwd"] d["important_tks"] = rag_tokenizer.tokenize(" ".join(req["important_kwd"])) if "question_kwd" in req: if not isinstance(req["question_kwd"], list): return get_data_error_result(message="`question_kwd` should be a list") d["question_kwd"] = req["question_kwd"] d["question_tks"] = rag_tokenizer.tokenize("\n".join(req["question_kwd"])) if "tag_kwd" in req: d["tag_kwd"] = req["tag_kwd"] if "tag_feas" in req: d["tag_feas"] = req["tag_feas"] if "available_int" in req: d["available_int"] = req["available_int"] try: def _set_sync(): tenant_id = DocumentService.get_tenant_id(req["doc_id"]) if not tenant_id: return get_data_error_result(message="Tenant not found!") embd_id = DocumentService.get_embd_id(req["doc_id"]) embd_mdl = LLMBundle(tenant_id, LLMType.EMBEDDING, embd_id) e, doc = DocumentService.get_by_id(req["doc_id"]) if not e: return get_data_error_result(message="Document not found!") _d = d if doc.parser_id == ParserType.QA: arr = [ t for t in re.split( r"[\n\t]", req["content_with_weight"]) if len(t) > 1] q, a = rmPrefix(arr[0]), rmPrefix("\n".join(arr[1:])) _d = beAdoc(d, q, a, not any( [rag_tokenizer.is_chinese(t) for t in q + a])) v, c = embd_mdl.encode([doc.name, req["content_with_weight"] if not _d.get("question_kwd") else "\n".join(_d["question_kwd"])]) v = 0.1 * v[0] + 0.9 * v[1] if doc.parser_id != ParserType.QA else v[1] _d["q_%d_vec" % len(v)] = v.tolist() settings.docStoreConn.update({"id": req["chunk_id"]}, _d, search.index_name(tenant_id), doc.kb_id) # update image image_base64 = req.get("image_base64", None) if image_base64: bkt, name = req.get("img_id", "-").split("-") image_binary = base64.b64decode(image_base64) settings.STORAGE_IMPL.put(bkt, name, image_binary) return get_json_result(data=True) return await asyncio.to_thread(_set_sync) except Exception as e: return server_error_response(e) @manager.route('/switch', methods=['POST']) # noqa: F821 @login_required @validate_request("chunk_ids", "available_int", "doc_id") async def switch(): req = await get_request_json() try: def _switch_sync(): e, doc = DocumentService.get_by_id(req["doc_id"]) if not e: return get_data_error_result(message="Document not found!") for cid in req["chunk_ids"]: if not settings.docStoreConn.update({"id": cid}, {"available_int": int(req["available_int"])}, search.index_name(DocumentService.get_tenant_id(req["doc_id"])), doc.kb_id): return get_data_error_result(message="Index updating failure") return get_json_result(data=True) return await asyncio.to_thread(_switch_sync) except Exception as e: return server_error_response(e) @manager.route('/rm', methods=['POST']) # noqa: F821 @login_required @validate_request("chunk_ids", "doc_id") async def rm(): req = await get_request_json() try: def _rm_sync(): e, doc = DocumentService.get_by_id(req["doc_id"]) if not e: return get_data_error_result(message="Document not found!") if not settings.docStoreConn.delete({"id": req["chunk_ids"]}, search.index_name(DocumentService.get_tenant_id(req["doc_id"])), doc.kb_id): return get_data_error_result(message="Chunk deleting failure") deleted_chunk_ids = req["chunk_ids"] chunk_number = len(deleted_chunk_ids) DocumentService.decrement_chunk_num(doc.id, doc.kb_id, 1, chunk_number, 0) for cid in deleted_chunk_ids: if settings.STORAGE_IMPL.obj_exist(doc.kb_id, cid): settings.STORAGE_IMPL.rm(doc.kb_id, cid) return get_json_result(data=True) return await asyncio.to_thread(_rm_sync) except Exception as e: return server_error_response(e) @manager.route('/create', methods=['POST']) # noqa: F821 @login_required @validate_request("doc_id", "content_with_weight") async def create(): req = await get_request_json() chunck_id = xxhash.xxh64((req["content_with_weight"] + req["doc_id"]).encode("utf-8")).hexdigest() d = {"id": chunck_id, "content_ltks": rag_tokenizer.tokenize(req["content_with_weight"]), "content_with_weight": req["content_with_weight"]} d["content_sm_ltks"] = rag_tokenizer.fine_grained_tokenize(d["content_ltks"]) d["important_kwd"] = req.get("important_kwd", []) if not isinstance(d["important_kwd"], list): return get_data_error_result(message="`important_kwd` is required to be a list") d["important_tks"] = rag_tokenizer.tokenize(" ".join(d["important_kwd"])) d["question_kwd"] = req.get("question_kwd", []) if not isinstance(d["question_kwd"], list): return get_data_error_result(message="`question_kwd` is required to be a list") d["question_tks"] = rag_tokenizer.tokenize("\n".join(d["question_kwd"])) d["create_time"] = str(datetime.datetime.now()).replace("T", " ")[:19] d["create_timestamp_flt"] = datetime.datetime.now().timestamp() if "tag_feas" in req: d["tag_feas"] = req["tag_feas"] try: def _create_sync(): e, doc = DocumentService.get_by_id(req["doc_id"]) if not e: return get_data_error_result(message="Document not found!") d["kb_id"] = [doc.kb_id] d["docnm_kwd"] = doc.name d["title_tks"] = rag_tokenizer.tokenize(doc.name) d["doc_id"] = doc.id tenant_id = DocumentService.get_tenant_id(req["doc_id"]) if not tenant_id: return get_data_error_result(message="Tenant not found!") e, kb = KnowledgebaseService.get_by_id(doc.kb_id) if not e: return get_data_error_result(message="Knowledgebase not found!") if kb.pagerank: d[PAGERANK_FLD] = kb.pagerank embd_id = DocumentService.get_embd_id(req["doc_id"]) embd_mdl = LLMBundle(tenant_id, LLMType.EMBEDDING.value, embd_id) v, c = embd_mdl.encode([doc.name, req["content_with_weight"] if not d["question_kwd"] else "\n".join(d["question_kwd"])]) v = 0.1 * v[0] + 0.9 * v[1] d["q_%d_vec" % len(v)] = v.tolist() settings.docStoreConn.insert([d], search.index_name(tenant_id), doc.kb_id) DocumentService.increment_chunk_num( doc.id, doc.kb_id, c, 1, 0) return get_json_result(data={"chunk_id": chunck_id}) return await asyncio.to_thread(_create_sync) except Exception as e: return server_error_response(e) @manager.route('/retrieval_test', methods=['POST']) # noqa: F821 @login_required @validate_request("kb_id", "question") async def retrieval_test(): req = await get_request_json() page = int(req.get("page", 1)) size = int(req.get("size", 30)) question = req["question"] kb_ids = req["kb_id"] if isinstance(kb_ids, str): kb_ids = [kb_ids] if not kb_ids: return get_json_result(data=False, message='Please specify dataset firstly.', code=RetCode.DATA_ERROR) doc_ids = req.get("doc_ids", []) use_kg = req.get("use_kg", False) top = int(req.get("top_k", 1024)) langs = req.get("cross_languages", []) user_id = current_user.id async def _retrieval(): local_doc_ids = list(doc_ids) if doc_ids else [] tenant_ids = [] meta_data_filter = {} chat_mdl = None if req.get("search_id", ""): search_config = SearchService.get_detail(req.get("search_id", "")).get("search_config", {}) meta_data_filter = search_config.get("meta_data_filter", {}) if meta_data_filter.get("method") in ["auto", "semi_auto"]: chat_mdl = LLMBundle(user_id, LLMType.CHAT, llm_name=search_config.get("chat_id", "")) else: meta_data_filter = req.get("meta_data_filter") or {} if meta_data_filter.get("method") in ["auto", "semi_auto"]: chat_mdl = LLMBundle(user_id, LLMType.CHAT) if meta_data_filter: metas = DocumentService.get_meta_by_kbs(kb_ids) local_doc_ids = await apply_meta_data_filter(meta_data_filter, metas, question, chat_mdl, local_doc_ids) tenants = UserTenantService.query(user_id=user_id) for kb_id in kb_ids: for tenant in tenants: if KnowledgebaseService.query( tenant_id=tenant.tenant_id, id=kb_id): tenant_ids.append(tenant.tenant_id) break else: return get_json_result( data=False, message='Only owner of dataset authorized for this operation.', code=RetCode.OPERATING_ERROR) e, kb = KnowledgebaseService.get_by_id(kb_ids[0]) if not e: return get_data_error_result(message="Knowledgebase not found!") _question = question if langs: _question = await cross_languages(kb.tenant_id, None, _question, langs) embd_mdl = LLMBundle(kb.tenant_id, LLMType.EMBEDDING.value, llm_name=kb.embd_id) rerank_mdl = None if req.get("rerank_id"): rerank_mdl = LLMBundle(kb.tenant_id, LLMType.RERANK.value, llm_name=req["rerank_id"]) if req.get("keyword", False): chat_mdl = LLMBundle(kb.tenant_id, LLMType.CHAT) _question += await keyword_extraction(chat_mdl, _question) labels = label_question(_question, [kb]) ranks = settings.retriever.retrieval(_question, embd_mdl, tenant_ids, kb_ids, page, size, float(req.get("similarity_threshold", 0.0)), float(req.get("vector_similarity_weight", 0.3)), top, local_doc_ids, rerank_mdl=rerank_mdl, highlight=req.get("highlight", False), rank_feature=labels ) if use_kg: ck = await settings.kg_retriever.retrieval(_question, tenant_ids, kb_ids, embd_mdl, LLMBundle(kb.tenant_id, LLMType.CHAT)) if ck["content_with_weight"]: ranks["chunks"].insert(0, ck) ranks["chunks"] = settings.retriever.retrieval_by_children(ranks["chunks"], tenant_ids) for c in ranks["chunks"]: c.pop("vector", None) ranks["labels"] = labels return get_json_result(data=ranks) try: return await _retrieval() except Exception as e: if str(e).find("not_found") > 0: return get_json_result(data=False, message='No chunk found! Check the chunk status please!', code=RetCode.DATA_ERROR) return server_error_response(e) @manager.route('/knowledge_graph', methods=['GET']) # noqa: F821 @login_required def knowledge_graph(): doc_id = request.args["doc_id"] tenant_id = DocumentService.get_tenant_id(doc_id) kb_ids = KnowledgebaseService.get_kb_ids(tenant_id) req = { "doc_ids": [doc_id], "knowledge_graph_kwd": ["graph", "mind_map"] } sres = settings.retriever.search(req, search.index_name(tenant_id), kb_ids) obj = {"graph": {}, "mind_map": {}} for id in sres.ids[:2]: ty = sres.field[id]["knowledge_graph_kwd"] try: content_json = json.loads(sres.field[id]["content_with_weight"]) except Exception: continue if ty == 'mind_map': node_dict = {} def repeat_deal(content_json, node_dict): if 'id' in content_json: if content_json['id'] in node_dict: node_name = content_json['id'] content_json['id'] += f"({node_dict[content_json['id']]})" node_dict[node_name] += 1 else: node_dict[content_json['id']] = 1 if 'children' in content_json and content_json['children']: for item in content_json['children']: repeat_deal(item, node_dict) repeat_deal(content_json, node_dict) obj[ty] = content_json return get_json_result(data=obj)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/llm_app.py
api/apps/llm_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import logging import json import os from quart import request from api.apps import login_required, current_user from api.db.services.tenant_llm_service import LLMFactoriesService, TenantLLMService from api.db.services.llm_service import LLMService from api.utils.api_utils import get_allowed_llm_factories, get_data_error_result, get_json_result, get_request_json, server_error_response, validate_request from common.constants import StatusEnum, LLMType from api.db.db_models import TenantLLM from rag.utils.base64_image import test_image from rag.llm import EmbeddingModel, ChatModel, RerankModel, CvModel, TTSModel, OcrModel, Seq2txtModel @manager.route("/factories", methods=["GET"]) # noqa: F821 @login_required def factories(): try: fac = get_allowed_llm_factories() fac = [f.to_dict() for f in fac if f.name not in ["Youdao", "FastEmbed", "BAAI", "Builtin"]] llms = LLMService.get_all() mdl_types = {} for m in llms: if m.status != StatusEnum.VALID.value: continue if m.fid not in mdl_types: mdl_types[m.fid] = set([]) mdl_types[m.fid].add(m.model_type) for f in fac: f["model_types"] = list( mdl_types.get( f["name"], [LLMType.CHAT, LLMType.EMBEDDING, LLMType.RERANK, LLMType.IMAGE2TEXT, LLMType.SPEECH2TEXT, LLMType.TTS, LLMType.OCR], ) ) return get_json_result(data=fac) except Exception as e: return server_error_response(e) @manager.route("/set_api_key", methods=["POST"]) # noqa: F821 @login_required @validate_request("llm_factory", "api_key") async def set_api_key(): req = await get_request_json() # test if api key works chat_passed, embd_passed, rerank_passed = False, False, False factory = req["llm_factory"] extra = {"provider": factory} msg = "" for llm in LLMService.query(fid=factory): if not embd_passed and llm.model_type == LLMType.EMBEDDING.value: assert factory in EmbeddingModel, f"Embedding model from {factory} is not supported yet." mdl = EmbeddingModel[factory](req["api_key"], llm.llm_name, base_url=req.get("base_url")) try: arr, tc = mdl.encode(["Test if the api key is available"]) if len(arr[0]) == 0: raise Exception("Fail") embd_passed = True except Exception as e: msg += f"\nFail to access embedding model({llm.llm_name}) using this api key." + str(e) elif not chat_passed and llm.model_type == LLMType.CHAT.value: assert factory in ChatModel, f"Chat model from {factory} is not supported yet." mdl = ChatModel[factory](req["api_key"], llm.llm_name, base_url=req.get("base_url"), **extra) try: m, tc = await mdl.async_chat(None, [{"role": "user", "content": "Hello! How are you doing!"}], {"temperature": 0.9, "max_tokens": 50}) if m.find("**ERROR**") >= 0: raise Exception(m) chat_passed = True except Exception as e: msg += f"\nFail to access model({llm.fid}/{llm.llm_name}) using this api key." + str(e) elif not rerank_passed and llm.model_type == LLMType.RERANK: assert factory in RerankModel, f"Re-rank model from {factory} is not supported yet." mdl = RerankModel[factory](req["api_key"], llm.llm_name, base_url=req.get("base_url")) try: arr, tc = mdl.similarity("What's the weather?", ["Is it sunny today?"]) if len(arr) == 0 or tc == 0: raise Exception("Fail") rerank_passed = True logging.debug(f"passed model rerank {llm.llm_name}") except Exception as e: msg += f"\nFail to access model({llm.fid}/{llm.llm_name}) using this api key." + str(e) if any([embd_passed, chat_passed, rerank_passed]): msg = "" break if msg: return get_data_error_result(message=msg) llm_config = {"api_key": req["api_key"], "api_base": req.get("base_url", "")} for n in ["model_type", "llm_name"]: if n in req: llm_config[n] = req[n] for llm in LLMService.query(fid=factory): llm_config["max_tokens"] = llm.max_tokens if not TenantLLMService.filter_update([TenantLLM.tenant_id == current_user.id, TenantLLM.llm_factory == factory, TenantLLM.llm_name == llm.llm_name], llm_config): TenantLLMService.save( tenant_id=current_user.id, llm_factory=factory, llm_name=llm.llm_name, model_type=llm.model_type, api_key=llm_config["api_key"], api_base=llm_config["api_base"], max_tokens=llm_config["max_tokens"], ) return get_json_result(data=True) @manager.route("/add_llm", methods=["POST"]) # noqa: F821 @login_required @validate_request("llm_factory") async def add_llm(): req = await get_request_json() factory = req["llm_factory"] api_key = req.get("api_key", "x") llm_name = req.get("llm_name") if factory not in [f.name for f in get_allowed_llm_factories()]: return get_data_error_result(message=f"LLM factory {factory} is not allowed") def apikey_json(keys): nonlocal req return json.dumps({k: req.get(k, "") for k in keys}) if factory == "VolcEngine": # For VolcEngine, due to its special authentication method # Assemble ark_api_key endpoint_id into api_key api_key = apikey_json(["ark_api_key", "endpoint_id"]) elif factory == "Tencent Hunyuan": req["api_key"] = apikey_json(["hunyuan_sid", "hunyuan_sk"]) return await set_api_key() elif factory == "Tencent Cloud": req["api_key"] = apikey_json(["tencent_cloud_sid", "tencent_cloud_sk"]) return await set_api_key() elif factory == "Bedrock": # For Bedrock, due to its special authentication method # Assemble bedrock_ak, bedrock_sk, bedrock_region api_key = apikey_json(["auth_mode", "bedrock_ak", "bedrock_sk", "bedrock_region", "aws_role_arn"]) elif factory == "LocalAI": llm_name += "___LocalAI" elif factory == "HuggingFace": llm_name += "___HuggingFace" elif factory == "OpenAI-API-Compatible": llm_name += "___OpenAI-API" elif factory == "VLLM": llm_name += "___VLLM" elif factory == "XunFei Spark": if req["model_type"] == "chat": api_key = req.get("spark_api_password", "") elif req["model_type"] == "tts": api_key = apikey_json(["spark_app_id", "spark_api_secret", "spark_api_key"]) elif factory == "BaiduYiyan": api_key = apikey_json(["yiyan_ak", "yiyan_sk"]) elif factory == "Fish Audio": api_key = apikey_json(["fish_audio_ak", "fish_audio_refid"]) elif factory == "Google Cloud": api_key = apikey_json(["google_project_id", "google_region", "google_service_account_key"]) elif factory == "Azure-OpenAI": api_key = apikey_json(["api_key", "api_version"]) elif factory == "OpenRouter": api_key = apikey_json(["api_key", "provider_order"]) elif factory == "MinerU": api_key = apikey_json(["api_key", "provider_order"]) llm = { "tenant_id": current_user.id, "llm_factory": factory, "model_type": req["model_type"], "llm_name": llm_name, "api_base": req.get("api_base", ""), "api_key": api_key, "max_tokens": req.get("max_tokens"), } msg = "" mdl_nm = llm["llm_name"].split("___")[0] extra = {"provider": factory} model_type = llm["model_type"] model_api_key = llm["api_key"] model_base_url = llm.get("api_base", "") match model_type: case LLMType.EMBEDDING.value: assert factory in EmbeddingModel, f"Embedding model from {factory} is not supported yet." mdl = EmbeddingModel[factory](key=model_api_key, model_name=mdl_nm, base_url=model_base_url) try: arr, tc = mdl.encode(["Test if the api key is available"]) if len(arr[0]) == 0: raise Exception("Fail") except Exception as e: msg += f"\nFail to access embedding model({mdl_nm})." + str(e) case LLMType.CHAT.value: assert factory in ChatModel, f"Chat model from {factory} is not supported yet." mdl = ChatModel[factory]( key=model_api_key, model_name=mdl_nm, base_url=model_base_url, **extra, ) try: m, tc = await mdl.async_chat(None, [{"role": "user", "content": "Hello! How are you doing!"}], {"temperature": 0.9}) if not tc and m.find("**ERROR**:") >= 0: raise Exception(m) except Exception as e: msg += f"\nFail to access model({factory}/{mdl_nm})." + str(e) case LLMType.RERANK.value: assert factory in RerankModel, f"RE-rank model from {factory} is not supported yet." try: mdl = RerankModel[factory](key=model_api_key, model_name=mdl_nm, base_url=model_base_url) arr, tc = mdl.similarity("Hello~ RAGFlower!", ["Hi, there!", "Ohh, my friend!"]) if len(arr) == 0: raise Exception("Not known.") except KeyError: msg += f"{factory} dose not support this model({factory}/{mdl_nm})" except Exception as e: msg += f"\nFail to access model({factory}/{mdl_nm})." + str(e) case LLMType.IMAGE2TEXT.value: assert factory in CvModel, f"Image to text model from {factory} is not supported yet." mdl = CvModel[factory](key=model_api_key, model_name=mdl_nm, base_url=model_base_url) try: image_data = test_image m, tc = mdl.describe(image_data) if not tc and m.find("**ERROR**:") >= 0: raise Exception(m) except Exception as e: msg += f"\nFail to access model({factory}/{mdl_nm})." + str(e) case LLMType.TTS.value: assert factory in TTSModel, f"TTS model from {factory} is not supported yet." mdl = TTSModel[factory](key=model_api_key, model_name=mdl_nm, base_url=model_base_url) try: for resp in mdl.tts("Hello~ RAGFlower!"): pass except RuntimeError as e: msg += f"\nFail to access model({factory}/{mdl_nm})." + str(e) case LLMType.OCR.value: assert factory in OcrModel, f"OCR model from {factory} is not supported yet." try: mdl = OcrModel[factory](key=model_api_key, model_name=mdl_nm, base_url=model_base_url) ok, reason = mdl.check_available() if not ok: raise RuntimeError(reason or "Model not available") except Exception as e: msg += f"\nFail to access model({factory}/{mdl_nm})." + str(e) case LLMType.SPEECH2TEXT: assert factory in Seq2txtModel, f"Speech model from {factory} is not supported yet." try: mdl = Seq2txtModel[factory](key=model_api_key, model_name=mdl_nm, base_url=model_base_url) # TODO: check the availability except Exception as e: msg += f"\nFail to access model({factory}/{mdl_nm})." + str(e) case _: raise RuntimeError(f"Unknown model type: {model_type}") if msg: return get_data_error_result(message=msg) if not TenantLLMService.filter_update([TenantLLM.tenant_id == current_user.id, TenantLLM.llm_factory == factory, TenantLLM.llm_name == llm["llm_name"]], llm): TenantLLMService.save(**llm) return get_json_result(data=True) @manager.route("/delete_llm", methods=["POST"]) # noqa: F821 @login_required @validate_request("llm_factory", "llm_name") async def delete_llm(): req = await get_request_json() TenantLLMService.filter_delete([TenantLLM.tenant_id == current_user.id, TenantLLM.llm_factory == req["llm_factory"], TenantLLM.llm_name == req["llm_name"]]) return get_json_result(data=True) @manager.route("/enable_llm", methods=["POST"]) # noqa: F821 @login_required @validate_request("llm_factory", "llm_name") async def enable_llm(): req = await get_request_json() TenantLLMService.filter_update( [TenantLLM.tenant_id == current_user.id, TenantLLM.llm_factory == req["llm_factory"], TenantLLM.llm_name == req["llm_name"]], {"status": str(req.get("status", "1"))} ) return get_json_result(data=True) @manager.route("/delete_factory", methods=["POST"]) # noqa: F821 @login_required @validate_request("llm_factory") async def delete_factory(): req = await get_request_json() TenantLLMService.filter_delete([TenantLLM.tenant_id == current_user.id, TenantLLM.llm_factory == req["llm_factory"]]) return get_json_result(data=True) @manager.route("/my_llms", methods=["GET"]) # noqa: F821 @login_required def my_llms(): try: TenantLLMService.ensure_mineru_from_env(current_user.id) include_details = request.args.get("include_details", "false").lower() == "true" if include_details: res = {} objs = TenantLLMService.query(tenant_id=current_user.id) factories = LLMFactoriesService.query(status=StatusEnum.VALID.value) for o in objs: o_dict = o.to_dict() factory_tags = None for f in factories: if f.name == o_dict["llm_factory"]: factory_tags = f.tags break if o_dict["llm_factory"] not in res: res[o_dict["llm_factory"]] = {"tags": factory_tags, "llm": []} res[o_dict["llm_factory"]]["llm"].append( { "type": o_dict["model_type"], "name": o_dict["llm_name"], "used_token": o_dict["used_tokens"], "api_base": o_dict["api_base"] or "", "max_tokens": o_dict["max_tokens"] or 8192, "status": o_dict["status"] or "1", } ) else: res = {} for o in TenantLLMService.get_my_llms(current_user.id): if o["llm_factory"] not in res: res[o["llm_factory"]] = {"tags": o["tags"], "llm": []} res[o["llm_factory"]]["llm"].append({"type": o["model_type"], "name": o["llm_name"], "used_token": o["used_tokens"], "status": o["status"]}) return get_json_result(data=res) except Exception as e: return server_error_response(e) @manager.route("/list", methods=["GET"]) # noqa: F821 @login_required def list_app(): self_deployed = ["FastEmbed", "Ollama", "Xinference", "LocalAI", "LM-Studio", "GPUStack"] weighted = [] model_type = request.args.get("model_type") try: TenantLLMService.ensure_mineru_from_env(current_user.id) objs = TenantLLMService.query(tenant_id=current_user.id) facts = set([o.to_dict()["llm_factory"] for o in objs if o.api_key and o.status == StatusEnum.VALID.value]) status = {(o.llm_name + "@" + o.llm_factory) for o in objs if o.status == StatusEnum.VALID.value} llms = LLMService.get_all() llms = [m.to_dict() for m in llms if m.status == StatusEnum.VALID.value and m.fid not in weighted and (m.fid == 'Builtin' or (m.llm_name + "@" + m.fid) in status)] for m in llms: m["available"] = m["fid"] in facts or m["llm_name"].lower() == "flag-embedding" or m["fid"] in self_deployed if "tei-" in os.getenv("COMPOSE_PROFILES", "") and m["model_type"] == LLMType.EMBEDDING and m["fid"] == "Builtin" and m["llm_name"] == os.getenv("TEI_MODEL", ""): m["available"] = True llm_set = set([m["llm_name"] + "@" + m["fid"] for m in llms]) for o in objs: if o.llm_name + "@" + o.llm_factory in llm_set: continue llms.append({"llm_name": o.llm_name, "model_type": o.model_type, "fid": o.llm_factory, "available": True, "status": StatusEnum.VALID.value}) res = {} for m in llms: if model_type and m["model_type"].find(model_type) < 0: continue if m["fid"] not in res: res[m["fid"]] = [] res[m["fid"]].append(m) return get_json_result(data=res) except Exception as e: return server_error_response(e)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/connector_app.py
api/apps/connector_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import asyncio import json import logging import time import uuid from html import escape from typing import Any from quart import request, make_response from google_auth_oauthlib.flow import Flow from api.db import InputType from api.db.services.connector_service import ConnectorService, SyncLogsService from api.utils.api_utils import get_data_error_result, get_json_result, get_request_json, validate_request from common.constants import RetCode, TaskStatus from common.data_source.config import GOOGLE_DRIVE_WEB_OAUTH_REDIRECT_URI, GMAIL_WEB_OAUTH_REDIRECT_URI, BOX_WEB_OAUTH_REDIRECT_URI, DocumentSource from common.data_source.google_util.constant import WEB_OAUTH_POPUP_TEMPLATE, GOOGLE_SCOPES from common.misc_utils import get_uuid from rag.utils.redis_conn import REDIS_CONN from api.apps import login_required, current_user from box_sdk_gen import BoxOAuth, OAuthConfig, GetAuthorizeUrlOptions @manager.route("/set", methods=["POST"]) # noqa: F821 @login_required async def set_connector(): req = await get_request_json() if req.get("id"): conn = {fld: req[fld] for fld in ["prune_freq", "refresh_freq", "config", "timeout_secs"] if fld in req} ConnectorService.update_by_id(req["id"], conn) else: req["id"] = get_uuid() conn = { "id": req["id"], "tenant_id": current_user.id, "name": req["name"], "source": req["source"], "input_type": InputType.POLL, "config": req["config"], "refresh_freq": int(req.get("refresh_freq", 30)), "prune_freq": int(req.get("prune_freq", 720)), "timeout_secs": int(req.get("timeout_secs", 60 * 29)), "status": TaskStatus.SCHEDULE, } ConnectorService.save(**conn) await asyncio.sleep(1) e, conn = ConnectorService.get_by_id(req["id"]) return get_json_result(data=conn.to_dict()) @manager.route("/list", methods=["GET"]) # noqa: F821 @login_required def list_connector(): return get_json_result(data=ConnectorService.list(current_user.id)) @manager.route("/<connector_id>", methods=["GET"]) # noqa: F821 @login_required def get_connector(connector_id): e, conn = ConnectorService.get_by_id(connector_id) if not e: return get_data_error_result(message="Can't find this Connector!") return get_json_result(data=conn.to_dict()) @manager.route("/<connector_id>/logs", methods=["GET"]) # noqa: F821 @login_required def list_logs(connector_id): req = request.args.to_dict(flat=True) arr, total = SyncLogsService.list_sync_tasks(connector_id, int(req.get("page", 1)), int(req.get("page_size", 15))) return get_json_result(data={"total": total, "logs": arr}) @manager.route("/<connector_id>/resume", methods=["PUT"]) # noqa: F821 @login_required async def resume(connector_id): req = await get_request_json() if req.get("resume"): ConnectorService.resume(connector_id, TaskStatus.SCHEDULE) else: ConnectorService.resume(connector_id, TaskStatus.CANCEL) return get_json_result(data=True) @manager.route("/<connector_id>/rebuild", methods=["PUT"]) # noqa: F821 @login_required @validate_request("kb_id") async def rebuild(connector_id): req = await get_request_json() err = ConnectorService.rebuild(req["kb_id"], connector_id, current_user.id) if err: return get_json_result(data=False, message=err, code=RetCode.SERVER_ERROR) return get_json_result(data=True) @manager.route("/<connector_id>/rm", methods=["POST"]) # noqa: F821 @login_required def rm_connector(connector_id): ConnectorService.resume(connector_id, TaskStatus.CANCEL) ConnectorService.delete_by_id(connector_id) return get_json_result(data=True) WEB_FLOW_TTL_SECS = 15 * 60 def _web_state_cache_key(flow_id: str, source_type: str | None = None) -> str: """Return Redis key for web OAuth state. The default prefix keeps backward compatibility for Google Drive. When source_type == "gmail", a different prefix is used so that Drive/Gmail flows don't clash in Redis. """ prefix = f"{source_type}_web_flow_state" return f"{prefix}:{flow_id}" def _web_result_cache_key(flow_id: str, source_type: str | None = None) -> str: """Return Redis key for web OAuth result. Mirrors _web_state_cache_key logic for result storage. """ prefix = f"{source_type}_web_flow_result" return f"{prefix}:{flow_id}" def _load_credentials(payload: str | dict[str, Any]) -> dict[str, Any]: if isinstance(payload, dict): return payload try: return json.loads(payload) except json.JSONDecodeError as exc: # pragma: no cover - defensive raise ValueError("Invalid Google credentials JSON.") from exc def _get_web_client_config(credentials: dict[str, Any]) -> dict[str, Any]: web_section = credentials.get("web") if not isinstance(web_section, dict): raise ValueError("Google OAuth JSON must include a 'web' client configuration to use browser-based authorization.") return {"web": web_section} async def _render_web_oauth_popup(flow_id: str, success: bool, message: str, source="drive"): status = "success" if success else "error" auto_close = "window.close();" if success else "" escaped_message = escape(message) # Drive: ragflow-google-drive-oauth # Gmail: ragflow-gmail-oauth payload_type = f"ragflow-{source}-oauth" payload_json = json.dumps( { "type": payload_type, "status": status, "flowId": flow_id or "", "message": message, } ) # TODO(google-oauth): title/heading/message may need to reflect drive/gmail based on cached type html = WEB_OAUTH_POPUP_TEMPLATE.format( title=f"Google {source.capitalize()} Authorization", heading="Authorization complete" if success else "Authorization failed", message=escaped_message, payload_json=payload_json, auto_close=auto_close, ) response = await make_response(html, 200) response.headers["Content-Type"] = "text/html; charset=utf-8" return response @manager.route("/google/oauth/web/start", methods=["POST"]) # noqa: F821 @login_required @validate_request("credentials") async def start_google_web_oauth(): source = request.args.get("type", "google-drive") if source not in ("google-drive", "gmail"): return get_json_result(code=RetCode.ARGUMENT_ERROR, message="Invalid Google OAuth type.") if source == "gmail": redirect_uri = GMAIL_WEB_OAUTH_REDIRECT_URI scopes = GOOGLE_SCOPES[DocumentSource.GMAIL] else: redirect_uri = GOOGLE_DRIVE_WEB_OAUTH_REDIRECT_URI scopes = GOOGLE_SCOPES[DocumentSource.GOOGLE_DRIVE] if not redirect_uri: return get_json_result( code=RetCode.SERVER_ERROR, message="Google OAuth redirect URI is not configured on the server.", ) req = await get_request_json() raw_credentials = req.get("credentials", "") try: credentials = _load_credentials(raw_credentials) print(credentials) except ValueError as exc: return get_json_result(code=RetCode.ARGUMENT_ERROR, message=str(exc)) if credentials.get("refresh_token"): return get_json_result( code=RetCode.ARGUMENT_ERROR, message="Uploaded credentials already include a refresh token.", ) try: client_config = _get_web_client_config(credentials) except ValueError as exc: return get_json_result(code=RetCode.ARGUMENT_ERROR, message=str(exc)) flow_id = str(uuid.uuid4()) try: flow = Flow.from_client_config(client_config, scopes=scopes) flow.redirect_uri = redirect_uri authorization_url, _ = flow.authorization_url( access_type="offline", include_granted_scopes="true", prompt="consent", state=flow_id, ) except Exception as exc: # pragma: no cover - defensive logging.exception("Failed to create Google OAuth flow: %s", exc) return get_json_result( code=RetCode.SERVER_ERROR, message="Failed to initialize Google OAuth flow. Please verify the uploaded client configuration.", ) cache_payload = { "user_id": current_user.id, "client_config": client_config, "created_at": int(time.time()), } REDIS_CONN.set_obj(_web_state_cache_key(flow_id, source), cache_payload, WEB_FLOW_TTL_SECS) return get_json_result( data={ "flow_id": flow_id, "authorization_url": authorization_url, "expires_in": WEB_FLOW_TTL_SECS, } ) @manager.route("/gmail/oauth/web/callback", methods=["GET"]) # noqa: F821 async def google_gmail_web_oauth_callback(): state_id = request.args.get("state") error = request.args.get("error") source = "gmail" error_description = request.args.get("error_description") or error if not state_id: return await _render_web_oauth_popup("", False, "Missing OAuth state parameter.", source) state_cache = REDIS_CONN.get(_web_state_cache_key(state_id, source)) if not state_cache: return await _render_web_oauth_popup(state_id, False, "Authorization session expired. Please restart from the main window.", source) state_obj = json.loads(state_cache) client_config = state_obj.get("client_config") if not client_config: REDIS_CONN.delete(_web_state_cache_key(state_id, source)) return await _render_web_oauth_popup(state_id, False, "Authorization session was invalid. Please retry.", source) if error: REDIS_CONN.delete(_web_state_cache_key(state_id, source)) return await _render_web_oauth_popup(state_id, False, error_description or "Authorization was cancelled.", source) code = request.args.get("code") if not code: return await _render_web_oauth_popup(state_id, False, "Missing authorization code from Google.", source) try: # TODO(google-oauth): branch scopes/redirect_uri based on source_type (drive vs gmail) flow = Flow.from_client_config(client_config, scopes=GOOGLE_SCOPES[DocumentSource.GMAIL]) flow.redirect_uri = GMAIL_WEB_OAUTH_REDIRECT_URI flow.fetch_token(code=code) except Exception as exc: # pragma: no cover - defensive logging.exception("Failed to exchange Google OAuth code: %s", exc) REDIS_CONN.delete(_web_state_cache_key(state_id, source)) return await _render_web_oauth_popup(state_id, False, "Failed to exchange tokens with Google. Please retry.", source) creds_json = flow.credentials.to_json() result_payload = { "user_id": state_obj.get("user_id"), "credentials": creds_json, } REDIS_CONN.set_obj(_web_result_cache_key(state_id, source), result_payload, WEB_FLOW_TTL_SECS) REDIS_CONN.delete(_web_state_cache_key(state_id, source)) return await _render_web_oauth_popup(state_id, True, "Authorization completed successfully.", source) @manager.route("/google-drive/oauth/web/callback", methods=["GET"]) # noqa: F821 async def google_drive_web_oauth_callback(): state_id = request.args.get("state") error = request.args.get("error") source = "google-drive" error_description = request.args.get("error_description") or error if not state_id: return await _render_web_oauth_popup("", False, "Missing OAuth state parameter.", source) state_cache = REDIS_CONN.get(_web_state_cache_key(state_id, source)) if not state_cache: return await _render_web_oauth_popup(state_id, False, "Authorization session expired. Please restart from the main window.", source) state_obj = json.loads(state_cache) client_config = state_obj.get("client_config") if not client_config: REDIS_CONN.delete(_web_state_cache_key(state_id, source)) return await _render_web_oauth_popup(state_id, False, "Authorization session was invalid. Please retry.", source) if error: REDIS_CONN.delete(_web_state_cache_key(state_id, source)) return await _render_web_oauth_popup(state_id, False, error_description or "Authorization was cancelled.", source) code = request.args.get("code") if not code: return await _render_web_oauth_popup(state_id, False, "Missing authorization code from Google.", source) try: # TODO(google-oauth): branch scopes/redirect_uri based on source_type (drive vs gmail) flow = Flow.from_client_config(client_config, scopes=GOOGLE_SCOPES[DocumentSource.GOOGLE_DRIVE]) flow.redirect_uri = GOOGLE_DRIVE_WEB_OAUTH_REDIRECT_URI flow.fetch_token(code=code) except Exception as exc: # pragma: no cover - defensive logging.exception("Failed to exchange Google OAuth code: %s", exc) REDIS_CONN.delete(_web_state_cache_key(state_id, source)) return await _render_web_oauth_popup(state_id, False, "Failed to exchange tokens with Google. Please retry.", source) creds_json = flow.credentials.to_json() result_payload = { "user_id": state_obj.get("user_id"), "credentials": creds_json, } REDIS_CONN.set_obj(_web_result_cache_key(state_id, source), result_payload, WEB_FLOW_TTL_SECS) REDIS_CONN.delete(_web_state_cache_key(state_id, source)) return await _render_web_oauth_popup(state_id, True, "Authorization completed successfully.", source) @manager.route("/google/oauth/web/result", methods=["POST"]) # noqa: F821 @login_required @validate_request("flow_id") async def poll_google_web_result(): req = await request.json or {} source = request.args.get("type") if source not in ("google-drive", "gmail"): return get_json_result(code=RetCode.ARGUMENT_ERROR, message="Invalid Google OAuth type.") flow_id = req.get("flow_id") cache_raw = REDIS_CONN.get(_web_result_cache_key(flow_id, source)) if not cache_raw: return get_json_result(code=RetCode.RUNNING, message="Authorization is still pending.") result = json.loads(cache_raw) if result.get("user_id") != current_user.id: return get_json_result(code=RetCode.PERMISSION_ERROR, message="You are not allowed to access this authorization result.") REDIS_CONN.delete(_web_result_cache_key(flow_id, source)) return get_json_result(data={"credentials": result.get("credentials")}) @manager.route("/box/oauth/web/start", methods=["POST"]) # noqa: F821 @login_required async def start_box_web_oauth(): req = await get_request_json() client_id = req.get("client_id") client_secret = req.get("client_secret") redirect_uri = req.get("redirect_uri", BOX_WEB_OAUTH_REDIRECT_URI) if not client_id or not client_secret: return get_json_result(code=RetCode.ARGUMENT_ERROR, message="Box client_id and client_secret are required.") flow_id = str(uuid.uuid4()) box_auth = BoxOAuth( OAuthConfig( client_id=client_id, client_secret=client_secret, ) ) auth_url = box_auth.get_authorize_url( options=GetAuthorizeUrlOptions( redirect_uri=redirect_uri, state=flow_id, ) ) cache_payload = { "user_id": current_user.id, "auth_url": auth_url, "client_id": client_id, "client_secret": client_secret, "created_at": int(time.time()), } REDIS_CONN.set_obj(_web_state_cache_key(flow_id, "box"), cache_payload, WEB_FLOW_TTL_SECS) return get_json_result( data = { "flow_id": flow_id, "authorization_url": auth_url, "expires_in": WEB_FLOW_TTL_SECS,} ) @manager.route("/box/oauth/web/callback", methods=["GET"]) # noqa: F821 async def box_web_oauth_callback(): flow_id = request.args.get("state") if not flow_id: return await _render_web_oauth_popup("", False, "Missing OAuth parameters.", "box") code = request.args.get("code") if not code: return await _render_web_oauth_popup(flow_id, False, "Missing authorization code from Box.", "box") cache_payload = json.loads(REDIS_CONN.get(_web_state_cache_key(flow_id, "box"))) if not cache_payload: return get_json_result(code=RetCode.ARGUMENT_ERROR, message="Box OAuth session expired or invalid.") error = request.args.get("error") error_description = request.args.get("error_description") or error if error: REDIS_CONN.delete(_web_state_cache_key(flow_id, "box")) return await _render_web_oauth_popup(flow_id, False, error_description or "Authorization failed.", "box") auth = BoxOAuth( OAuthConfig( client_id=cache_payload.get("client_id"), client_secret=cache_payload.get("client_secret"), ) ) auth.get_tokens_authorization_code_grant(code) token = auth.retrieve_token() result_payload = { "user_id": cache_payload.get("user_id"), "client_id": cache_payload.get("client_id"), "client_secret": cache_payload.get("client_secret"), "access_token": token.access_token, "refresh_token": token.refresh_token, } REDIS_CONN.set_obj(_web_result_cache_key(flow_id, "box"), result_payload, WEB_FLOW_TTL_SECS) REDIS_CONN.delete(_web_state_cache_key(flow_id, "box")) return await _render_web_oauth_popup(flow_id, True, "Authorization completed successfully.", "box") @manager.route("/box/oauth/web/result", methods=["POST"]) # noqa: F821 @login_required @validate_request("flow_id") async def poll_box_web_result(): req = await get_request_json() flow_id = req.get("flow_id") cache_blob = REDIS_CONN.get(_web_result_cache_key(flow_id, "box")) if not cache_blob: return get_json_result(code=RetCode.RUNNING, message="Authorization is still pending.") cache_raw = json.loads(cache_blob) if cache_raw.get("user_id") != current_user.id: return get_json_result(code=RetCode.PERMISSION_ERROR, message="You are not allowed to access this authorization result.") REDIS_CONN.delete(_web_result_cache_key(flow_id, "box")) return get_json_result(data={"credentials": cache_raw})
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false
infiniflow/ragflow
https://github.com/infiniflow/ragflow/blob/5ebe334a2f452cb35d4247a8c688bd3d3c76be4c/api/apps/dialog_app.py
api/apps/dialog_app.py
# # Copyright 2024 The InfiniFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from quart import request from api.db.services import duplicate_name from api.db.services.dialog_service import DialogService from common.constants import StatusEnum from api.db.services.tenant_llm_service import TenantLLMService from api.db.services.knowledgebase_service import KnowledgebaseService from api.db.services.user_service import TenantService, UserTenantService from api.utils.api_utils import get_data_error_result, get_json_result, get_request_json, server_error_response, validate_request from common.misc_utils import get_uuid from common.constants import RetCode from api.apps import login_required, current_user @manager.route('/set', methods=['POST']) # noqa: F821 @validate_request("prompt_config") @login_required async def set_dialog(): req = await get_request_json() dialog_id = req.get("dialog_id", "") is_create = not dialog_id name = req.get("name", "New Dialog") if not isinstance(name, str): return get_data_error_result(message="Dialog name must be string.") if name.strip() == "": return get_data_error_result(message="Dialog name can't be empty.") if len(name.encode("utf-8")) > 255: return get_data_error_result(message=f"Dialog name length is {len(name)} which is larger than 255") if is_create and DialogService.query(tenant_id=current_user.id, name=name.strip()): name = name.strip() name = duplicate_name( DialogService.query, name=name, tenant_id=current_user.id, status=StatusEnum.VALID.value) description = req.get("description", "A helpful dialog") icon = req.get("icon", "") top_n = req.get("top_n", 6) top_k = req.get("top_k", 1024) rerank_id = req.get("rerank_id", "") if not rerank_id: req["rerank_id"] = "" similarity_threshold = req.get("similarity_threshold", 0.1) vector_similarity_weight = req.get("vector_similarity_weight", 0.3) llm_setting = req.get("llm_setting", {}) meta_data_filter = req.get("meta_data_filter", {}) prompt_config = req["prompt_config"] if not is_create: if not req.get("kb_ids", []) and not prompt_config.get("tavily_api_key") and "{knowledge}" in prompt_config['system']: return get_data_error_result(message="Please remove `{knowledge}` in system prompt since no dataset / Tavily used here.") for p in prompt_config["parameters"]: if p["optional"]: continue if prompt_config["system"].find("{%s}" % p["key"]) < 0: return get_data_error_result( message="Parameter '{}' is not used".format(p["key"])) try: e, tenant = TenantService.get_by_id(current_user.id) if not e: return get_data_error_result(message="Tenant not found!") kbs = KnowledgebaseService.get_by_ids(req.get("kb_ids", [])) embd_ids = [TenantLLMService.split_model_name_and_factory(kb.embd_id)[0] for kb in kbs] # remove vendor suffix for comparison embd_count = len(set(embd_ids)) if embd_count > 1: return get_data_error_result(message=f'Datasets use different embedding models: {[kb.embd_id for kb in kbs]}"') llm_id = req.get("llm_id", tenant.llm_id) if not dialog_id: dia = { "id": get_uuid(), "tenant_id": current_user.id, "name": name, "kb_ids": req.get("kb_ids", []), "description": description, "llm_id": llm_id, "llm_setting": llm_setting, "prompt_config": prompt_config, "meta_data_filter": meta_data_filter, "top_n": top_n, "top_k": top_k, "rerank_id": rerank_id, "similarity_threshold": similarity_threshold, "vector_similarity_weight": vector_similarity_weight, "icon": icon } if not DialogService.save(**dia): return get_data_error_result(message="Fail to new a dialog!") return get_json_result(data=dia) else: del req["dialog_id"] if "kb_names" in req: del req["kb_names"] if not DialogService.update_by_id(dialog_id, req): return get_data_error_result(message="Dialog not found!") e, dia = DialogService.get_by_id(dialog_id) if not e: return get_data_error_result(message="Fail to update a dialog!") dia = dia.to_dict() dia.update(req) dia["kb_ids"], dia["kb_names"] = get_kb_names(dia["kb_ids"]) return get_json_result(data=dia) except Exception as e: return server_error_response(e) @manager.route('/get', methods=['GET']) # noqa: F821 @login_required def get(): dialog_id = request.args["dialog_id"] try: e, dia = DialogService.get_by_id(dialog_id) if not e: return get_data_error_result(message="Dialog not found!") dia = dia.to_dict() dia["kb_ids"], dia["kb_names"] = get_kb_names(dia["kb_ids"]) return get_json_result(data=dia) except Exception as e: return server_error_response(e) def get_kb_names(kb_ids): ids, nms = [], [] for kid in kb_ids: e, kb = KnowledgebaseService.get_by_id(kid) if not e or kb.status != StatusEnum.VALID.value: continue ids.append(kid) nms.append(kb.name) return ids, nms @manager.route('/list', methods=['GET']) # noqa: F821 @login_required def list_dialogs(): try: conversations = DialogService.query( tenant_id=current_user.id, status=StatusEnum.VALID.value, reverse=True, order_by=DialogService.model.create_time) conversations = [d.to_dict() for d in conversations] for conversation in conversations: conversation["kb_ids"], conversation["kb_names"] = get_kb_names(conversation["kb_ids"]) return get_json_result(data=conversations) except Exception as e: return server_error_response(e) @manager.route('/next', methods=['POST']) # noqa: F821 @login_required async def list_dialogs_next(): args = request.args keywords = args.get("keywords", "") page_number = int(args.get("page", 0)) items_per_page = int(args.get("page_size", 0)) parser_id = args.get("parser_id") orderby = args.get("orderby", "create_time") if args.get("desc", "true").lower() == "false": desc = False else: desc = True req = await get_request_json() owner_ids = req.get("owner_ids", []) try: if not owner_ids: # tenants = TenantService.get_joined_tenants_by_user_id(current_user.id) # tenants = [tenant["tenant_id"] for tenant in tenants] tenants = [] # keep it here dialogs, total = DialogService.get_by_tenant_ids( tenants, current_user.id, page_number, items_per_page, orderby, desc, keywords, parser_id) else: tenants = owner_ids dialogs, total = DialogService.get_by_tenant_ids( tenants, current_user.id, 0, 0, orderby, desc, keywords, parser_id) dialogs = [dialog for dialog in dialogs if dialog["tenant_id"] in tenants] total = len(dialogs) if page_number and items_per_page: dialogs = dialogs[(page_number-1)*items_per_page:page_number*items_per_page] return get_json_result(data={"dialogs": dialogs, "total": total}) except Exception as e: return server_error_response(e) @manager.route('/rm', methods=['POST']) # noqa: F821 @login_required @validate_request("dialog_ids") async def rm(): req = await get_request_json() dialog_list=[] tenants = UserTenantService.query(user_id=current_user.id) try: for id in req["dialog_ids"]: for tenant in tenants: if DialogService.query(tenant_id=tenant.tenant_id, id=id): break else: return get_json_result( data=False, message='Only owner of dialog authorized for this operation.', code=RetCode.OPERATING_ERROR) dialog_list.append({"id": id,"status":StatusEnum.INVALID.value}) DialogService.update_many_by_id(dialog_list) return get_json_result(data=True) except Exception as e: return server_error_response(e)
python
Apache-2.0
5ebe334a2f452cb35d4247a8c688bd3d3c76be4c
2026-01-04T14:38:19.006015Z
false