mirror of
https://git.mirrors.martin98.com/https://github.com/infiniflow/ragflow.git
synced 2025-04-19 12:39:59 +08:00
Test: Added test cases for Delete Documents HTTP API (#6175)
### What problem does this PR solve? cover [delete documents](https://ragflow.io/docs/dev/http_api_reference#delete-documents) endpoints ### Type of change - [x] add test cases
This commit is contained in:
parent
79482ff672
commit
503e5829bb
@ -150,3 +150,9 @@ def update_documnet(auth, dataset_id, document_id, payload):
|
||||
url = f"{HOST_ADDRESS}{FILE_API_URL}/{document_id}".format(dataset_id=dataset_id)
|
||||
res = requests.put(url=url, headers=HEADERS, auth=auth, json=payload)
|
||||
return res.json()
|
||||
|
||||
|
||||
def delete_documnet(auth, dataset_id, payload=None):
|
||||
url = f"{HOST_ADDRESS}{FILE_API_URL}".format(dataset_id=dataset_id)
|
||||
res = requests.delete(url=url, headers=HEADERS, auth=auth, json=payload)
|
||||
return res.json()
|
||||
|
@ -0,0 +1,183 @@
|
||||
#
|
||||
# Copyright 2025 The InfiniFlow Authors. All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
#
|
||||
from concurrent.futures import ThreadPoolExecutor
|
||||
|
||||
import pytest
|
||||
from common import (
|
||||
INVALID_API_TOKEN,
|
||||
batch_upload_documents,
|
||||
create_datasets,
|
||||
delete_documnet,
|
||||
list_documnet,
|
||||
)
|
||||
from libs.auth import RAGFlowHttpApiAuth
|
||||
|
||||
|
||||
class TestAuthorization:
|
||||
@pytest.mark.parametrize(
|
||||
"auth, expected_code, expected_message",
|
||||
[
|
||||
(None, 0, "`Authorization` can't be empty"),
|
||||
(
|
||||
RAGFlowHttpApiAuth(INVALID_API_TOKEN),
|
||||
109,
|
||||
"Authentication error: API key is invalid!",
|
||||
),
|
||||
],
|
||||
)
|
||||
def test_invalid_auth(
|
||||
self, get_http_api_auth, tmp_path, auth, expected_code, expected_message
|
||||
):
|
||||
ids = create_datasets(get_http_api_auth, 1)
|
||||
document_ids = batch_upload_documents(get_http_api_auth, ids[0], 1, tmp_path)
|
||||
res = delete_documnet(auth, ids[0], {"ids": document_ids[0]})
|
||||
assert res["code"] == expected_code
|
||||
assert res["message"] == expected_message
|
||||
|
||||
|
||||
class TestDocumentDeletion:
|
||||
@pytest.mark.parametrize(
|
||||
"payload, expected_code, expected_message, remaining",
|
||||
[
|
||||
(None, 0, "", 0),
|
||||
({"ids": []}, 0, "", 0),
|
||||
({"ids": ["invalid_id"]}, 102, "Document not found!", 3),
|
||||
(
|
||||
{"ids": ["\n!?。;!?\"'"]},
|
||||
102,
|
||||
"Document not found!",
|
||||
3,
|
||||
),
|
||||
(
|
||||
"not json",
|
||||
100,
|
||||
"AttributeError(\"'str' object has no attribute 'get'\")",
|
||||
3,
|
||||
),
|
||||
(lambda r: {"ids": r[:1]}, 0, "", 2),
|
||||
(lambda r: {"ids": r}, 0, "", 0),
|
||||
],
|
||||
)
|
||||
def test_basic_scenarios(
|
||||
self,
|
||||
get_http_api_auth,
|
||||
tmp_path,
|
||||
payload,
|
||||
expected_code,
|
||||
expected_message,
|
||||
remaining,
|
||||
):
|
||||
ids = create_datasets(get_http_api_auth, 1)
|
||||
document_ids = batch_upload_documents(get_http_api_auth, ids[0], 3, tmp_path)
|
||||
if callable(payload):
|
||||
payload = payload(document_ids)
|
||||
res = delete_documnet(get_http_api_auth, ids[0], payload)
|
||||
assert res["code"] == expected_code
|
||||
if res["code"] != 0:
|
||||
assert res["message"] == expected_message
|
||||
|
||||
res = list_documnet(get_http_api_auth, ids[0])
|
||||
assert len(res["data"]["docs"]) == remaining
|
||||
assert res["data"]["total"] == remaining
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
"dataset_id, expected_code, expected_message",
|
||||
[
|
||||
("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"),
|
||||
(
|
||||
"invalid_dataset_id",
|
||||
102,
|
||||
"You don't own the dataset invalid_dataset_id. ",
|
||||
),
|
||||
],
|
||||
)
|
||||
def test_invalid_dataset_id(
|
||||
self, get_http_api_auth, tmp_path, dataset_id, expected_code, expected_message
|
||||
):
|
||||
ids = create_datasets(get_http_api_auth, 1)
|
||||
document_ids = batch_upload_documents(get_http_api_auth, ids[0], 3, tmp_path)
|
||||
res = delete_documnet(get_http_api_auth, dataset_id, {"ids": document_ids[:1]})
|
||||
assert res["code"] == expected_code
|
||||
assert res["message"] == expected_message
|
||||
|
||||
@pytest.mark.xfail(reason="issues/6174")
|
||||
@pytest.mark.parametrize(
|
||||
"payload",
|
||||
[
|
||||
lambda r: {"ids": ["invalid_id"] + r},
|
||||
lambda r: {"ids": r[:1] + ["invalid_id"] + r[1:3]},
|
||||
lambda r: {"ids": r + ["invalid_id"]},
|
||||
],
|
||||
)
|
||||
def test_delete_partial_invalid_id(self, get_http_api_auth, tmp_path, payload):
|
||||
ids = create_datasets(get_http_api_auth, 1)
|
||||
document_ids = batch_upload_documents(get_http_api_auth, ids[0], 3, tmp_path)
|
||||
if callable(payload):
|
||||
payload = payload(document_ids)
|
||||
res = delete_documnet(get_http_api_auth, ids[0], payload)
|
||||
assert res["code"] == 0
|
||||
assert res["data"]["errors"][0] == "You don't own the dataset invalid_id"
|
||||
assert res["data"]["success_count"] == 3
|
||||
|
||||
res = list_documnet(get_http_api_auth, ids[0])
|
||||
assert len(res["data"]["docs"]) == 0
|
||||
assert res["data"]["total"] == 0
|
||||
|
||||
def test_repeated_deletion(self, get_http_api_auth, tmp_path):
|
||||
ids = create_datasets(get_http_api_auth, 1)
|
||||
document_ids = batch_upload_documents(get_http_api_auth, ids[0], 1, tmp_path)
|
||||
res = delete_documnet(get_http_api_auth, ids[0], {"ids": document_ids})
|
||||
assert res["code"] == 0
|
||||
|
||||
res = delete_documnet(get_http_api_auth, ids[0], {"ids": document_ids})
|
||||
assert res["code"] == 102
|
||||
assert res["message"] == "Document not found!"
|
||||
|
||||
def test_concurrent_deletion(self, get_http_api_auth, tmp_path):
|
||||
documnets_num = 100
|
||||
ids = create_datasets(get_http_api_auth, 1)
|
||||
document_ids = batch_upload_documents(
|
||||
get_http_api_auth, ids[0], documnets_num, tmp_path
|
||||
)
|
||||
|
||||
with ThreadPoolExecutor(max_workers=5) as executor:
|
||||
futures = [
|
||||
executor.submit(
|
||||
delete_documnet,
|
||||
get_http_api_auth,
|
||||
ids[0],
|
||||
{"ids": document_ids[i : i + 1]},
|
||||
)
|
||||
for i in range(documnets_num)
|
||||
]
|
||||
responses = [f.result() for f in futures]
|
||||
assert all(r["code"] == 0 for r in responses)
|
||||
|
||||
@pytest.mark.slow
|
||||
def test_delete_1k(self, get_http_api_auth, tmp_path):
|
||||
documnets_num = 1_000
|
||||
ids = create_datasets(get_http_api_auth, 1)
|
||||
document_ids = batch_upload_documents(
|
||||
get_http_api_auth, ids[0], documnets_num, tmp_path
|
||||
)
|
||||
res = list_documnet(get_http_api_auth, ids[0])
|
||||
assert res["data"]["total"] == documnets_num
|
||||
|
||||
res = delete_documnet(get_http_api_auth, ids[0], {"ids": document_ids})
|
||||
assert res["code"] == 0
|
||||
|
||||
res = list_documnet(get_http_api_auth, ids[0])
|
||||
assert res["data"]["total"] == 0
|
Loading…
x
Reference in New Issue
Block a user