Test: Added test cases for Stop Parse Documents HTTP API (#6285)

### What problem does this PR solve?

cover [stop parse
documents](https://ragflow.io/docs/dev/http_api_reference#stop-parsing-documents)
endpoints

### Type of change

- [x] Add test cases
This commit is contained in:
liu an 2025-03-20 09:42:50 +08:00 committed by GitHub
parent 1d6760dd84
commit dbf2ee56c6
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
3 changed files with 236 additions and 1 deletions

View File

@ -163,3 +163,9 @@ def parse_documnet(auth, dataset_id, payload=None):
url = f"{HOST_ADDRESS}{FILE_CHUNK_API_URL}".format(dataset_id=dataset_id)
res = requests.post(url=url, headers=HEADERS, auth=auth, json=payload)
return res.json()
def stop_parse_documnet(auth, dataset_id, payload=None):
url = f"{HOST_ADDRESS}{FILE_CHUNK_API_URL}".format(dataset_id=dataset_id)
res = requests.delete(url=url, headers=HEADERS, auth=auth, json=payload)
return res.json()

View File

@ -60,7 +60,7 @@ class TestAuthorization:
assert res["message"] == expected_message
class TestDatasetParse:
class TestDocumentsParse:
@pytest.mark.parametrize(
"payload, expected_code, expected_message",
[

View File

@ -0,0 +1,229 @@
#
# Copyright 2025 The InfiniFlow Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
from concurrent.futures import ThreadPoolExecutor
import pytest
from common import (
INVALID_API_TOKEN,
batch_upload_documents,
create_datasets,
list_documnet,
parse_documnet,
stop_parse_documnet,
)
from libs.auth import RAGFlowHttpApiAuth
from libs.utils import wait_for
def validate_document_parse_done(auth, dataset_id, document_ids):
for document_id in document_ids:
res = list_documnet(auth, dataset_id, params={"id": document_id})
doc = res["data"]["docs"][0]
assert doc["run"] == "DONE"
assert len(doc["process_begin_at"]) > 0
assert doc["process_duation"] > 0
assert doc["progress"] > 0
assert "Task done" in doc["progress_msg"]
def validate_document_parse_cancel(auth, dataset_id, document_ids):
for document_id in document_ids:
res = list_documnet(auth, dataset_id, params={"id": document_id})
doc = res["data"]["docs"][0]
assert doc["run"] == "CANCEL"
assert len(doc["process_begin_at"]) > 0
assert doc["progress"] == 0.0
class TestAuthorization:
@pytest.mark.parametrize(
"auth, expected_code, expected_message",
[
(None, 0, "`Authorization` can't be empty"),
(
RAGFlowHttpApiAuth(INVALID_API_TOKEN),
109,
"Authentication error: API key is invalid!",
),
],
)
def test_invalid_auth(self, get_http_api_auth, auth, expected_code, expected_message):
ids = create_datasets(get_http_api_auth, 1)
res = stop_parse_documnet(auth, ids[0])
assert res["code"] == expected_code
assert res["message"] == expected_message
class TestDocumentsParseStop:
@pytest.mark.parametrize(
"payload, expected_code, expected_message",
[
pytest.param(
None,
102,
"""AttributeError("\'NoneType\' object has no attribute \'get\'")""",
marks=pytest.mark.xfail,
),
({"document_ids": []}, 102, "`document_ids` is required"),
(
{"document_ids": ["invalid_id"]},
102,
"You don't own the document invalid_id.",
),
(
{"document_ids": ["\n!?。;!?\"'"]},
102,
"""You don\'t own the document \n!?。;!?"\'.""",
),
pytest.param(
"not json",
102,
"AttributeError(\"'str' object has no attribute 'get'\")",
marks=pytest.mark.xfail,
),
(lambda r: {"document_ids": r[:1]}, 0, ""),
(lambda r: {"document_ids": r}, 0, ""),
],
)
def test_basic_scenarios(self, get_http_api_auth, tmp_path, payload, expected_code, expected_message):
@wait_for(10, 1, "Document parsing timeout")
def condition(_auth, _dataset_id, _document_ids):
for _document_id in _document_ids:
res = list_documnet(_auth, _dataset_id, {"id": _document_id})
if res["data"]["docs"][0]["run"] != "DONE":
return False
return True
ids = create_datasets(get_http_api_auth, 1)
dataset_id = ids[0]
document_ids = batch_upload_documents(get_http_api_auth, dataset_id, 3, tmp_path)
parse_documnet(get_http_api_auth, dataset_id, {"document_ids": document_ids})
if callable(payload):
payload = payload(document_ids)
res = stop_parse_documnet(get_http_api_auth, dataset_id, payload)
assert res["code"] == expected_code
if expected_code != 0:
assert res["message"] == expected_message
else:
completed_document_ids = list(set(document_ids) - set(payload["document_ids"]))
condition(get_http_api_auth, dataset_id, completed_document_ids)
validate_document_parse_cancel(get_http_api_auth, dataset_id, payload["document_ids"])
validate_document_parse_done(get_http_api_auth, dataset_id, completed_document_ids)
@pytest.mark.parametrize(
"dataset_id, expected_code, expected_message",
[
("", 100, "<MethodNotAllowed '405: Method Not Allowed'>"),
(
"invalid_dataset_id",
102,
"You don't own the dataset invalid_dataset_id.",
),
],
)
def test_invalid_dataset_id(
self,
get_http_api_auth,
tmp_path,
dataset_id,
expected_code,
expected_message,
):
ids = create_datasets(get_http_api_auth, 1)
document_ids = batch_upload_documents(get_http_api_auth, ids[0], 1, tmp_path)
res = stop_parse_documnet(get_http_api_auth, dataset_id, {"document_ids": document_ids})
assert res["code"] == expected_code
assert res["message"] == expected_message
@pytest.mark.xfail
@pytest.mark.parametrize(
"payload",
[
lambda r: {"document_ids": ["invalid_id"] + r},
lambda r: {"document_ids": r[:1] + ["invalid_id"] + r[1:3]},
lambda r: {"document_ids": r + ["invalid_id"]},
],
)
def test_stop_parse_partial_invalid_document_id(self, get_http_api_auth, tmp_path, payload):
ids = create_datasets(get_http_api_auth, 1)
dataset_id = ids[0]
document_ids = batch_upload_documents(get_http_api_auth, dataset_id, 3, tmp_path)
parse_documnet(get_http_api_auth, dataset_id, {"document_ids": document_ids})
if callable(payload):
payload = payload(document_ids)
res = stop_parse_documnet(get_http_api_auth, dataset_id, payload)
assert res["code"] == 102
validate_document_parse_cancel(get_http_api_auth, dataset_id, document_ids)
def test_repeated_stop_parse(self, get_http_api_auth, tmp_path):
ids = create_datasets(get_http_api_auth, 1)
dataset_id = ids[0]
document_ids = batch_upload_documents(get_http_api_auth, dataset_id, 1, tmp_path)
parse_documnet(get_http_api_auth, dataset_id, {"document_ids": document_ids})
res = stop_parse_documnet(get_http_api_auth, dataset_id, {"document_ids": document_ids})
assert res["code"] == 0
res = stop_parse_documnet(get_http_api_auth, dataset_id, {"document_ids": document_ids})
assert res["code"] == 102
assert res["message"] == "Can't stop parsing document with progress at 0 or 1"
@pytest.mark.xfail
def test_duplicate_stop_parse(self, get_http_api_auth, tmp_path):
ids = create_datasets(get_http_api_auth, 1)
dataset_id = ids[0]
document_ids = batch_upload_documents(get_http_api_auth, dataset_id, 1, tmp_path)
parse_documnet(get_http_api_auth, dataset_id, {"document_ids": document_ids})
res = stop_parse_documnet(get_http_api_auth, dataset_id, {"document_ids": document_ids + document_ids})
assert res["code"] == 0
assert res["success_count"] == 1
assert f"Duplicate document ids: {document_ids[0]}" in res["data"]["errors"]
@pytest.mark.slow
def test_stop_parse_100_files(self, get_http_api_auth, tmp_path):
document_num = 100
ids = create_datasets(get_http_api_auth, 1)
dataset_id = ids[0]
document_ids = batch_upload_documents(get_http_api_auth, dataset_id, document_num, tmp_path)
parse_documnet(get_http_api_auth, dataset_id, {"document_ids": document_ids})
res = stop_parse_documnet(get_http_api_auth, dataset_id, {"document_ids": document_ids})
assert res["code"] == 0
validate_document_parse_cancel(get_http_api_auth, dataset_id, document_ids)
@pytest.mark.slow
def test_concurrent_parse(self, get_http_api_auth, tmp_path):
document_num = 50
ids = create_datasets(get_http_api_auth, 1)
dataset_id = ids[0]
document_ids = batch_upload_documents(get_http_api_auth, dataset_id, document_num, tmp_path)
parse_documnet(get_http_api_auth, dataset_id, {"document_ids": document_ids})
with ThreadPoolExecutor(max_workers=5) as executor:
futures = [
executor.submit(
stop_parse_documnet,
get_http_api_auth,
dataset_id,
{"document_ids": document_ids[i : i + 1]},
)
for i in range(document_num)
]
responses = [f.result() for f in futures]
assert all(r["code"] == 0 for r in responses)
validate_document_parse_cancel(get_http_api_auth, dataset_id, document_ids)