| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542 | 
							- from typing import Literal
 - 
 - from flask import request
 - from flask_restful import marshal, marshal_with, reqparse
 - from werkzeug.exceptions import Forbidden, NotFound
 - 
 - import services.dataset_service
 - from controllers.service_api import api
 - from controllers.service_api.dataset.error import DatasetInUseError, DatasetNameDuplicateError, InvalidActionError
 - from controllers.service_api.wraps import (
 -     DatasetApiResource,
 -     cloud_edition_billing_rate_limit_check,
 -     validate_dataset_token,
 - )
 - from core.model_runtime.entities.model_entities import ModelType
 - from core.plugin.entities.plugin import ModelProviderID
 - from core.provider_manager import ProviderManager
 - from fields.dataset_fields import dataset_detail_fields
 - from fields.tag_fields import tag_fields
 - from libs.login import current_user
 - from models.dataset import Dataset, DatasetPermissionEnum
 - from services.dataset_service import DatasetPermissionService, DatasetService, DocumentService
 - from services.entities.knowledge_entities.knowledge_entities import RetrievalModel
 - from services.tag_service import TagService
 - 
 - 
 - def _validate_name(name):
 -     if not name or len(name) < 1 or len(name) > 40:
 -         raise ValueError("Name must be between 1 to 40 characters.")
 -     return name
 - 
 - 
 - def _validate_description_length(description):
 -     if description and len(description) > 400:
 -         raise ValueError("Description cannot exceed 400 characters.")
 -     return description
 - 
 - 
 - class DatasetListApi(DatasetApiResource):
 -     """Resource for datasets."""
 - 
 -     def get(self, tenant_id):
 -         """Resource for getting datasets."""
 - 
 -         page = request.args.get("page", default=1, type=int)
 -         limit = request.args.get("limit", default=20, type=int)
 -         # provider = request.args.get("provider", default="vendor")
 -         search = request.args.get("keyword", default=None, type=str)
 -         tag_ids = request.args.getlist("tag_ids")
 -         include_all = request.args.get("include_all", default="false").lower() == "true"
 - 
 -         datasets, total = DatasetService.get_datasets(
 -             page, limit, tenant_id, current_user, search, tag_ids, include_all
 -         )
 -         # check embedding setting
 -         provider_manager = ProviderManager()
 -         configurations = provider_manager.get_configurations(tenant_id=current_user.current_tenant_id)
 - 
 -         embedding_models = configurations.get_models(model_type=ModelType.TEXT_EMBEDDING, only_active=True)
 - 
 -         model_names = []
 -         for embedding_model in embedding_models:
 -             model_names.append(f"{embedding_model.model}:{embedding_model.provider.provider}")
 - 
 -         data = marshal(datasets, dataset_detail_fields)
 -         for item in data:
 -             if item["indexing_technique"] == "high_quality" and item["embedding_model_provider"]:
 -                 item["embedding_model_provider"] = str(ModelProviderID(item["embedding_model_provider"]))
 -                 item_model = f"{item['embedding_model']}:{item['embedding_model_provider']}"
 -                 if item_model in model_names:
 -                     item["embedding_available"] = True
 -                 else:
 -                     item["embedding_available"] = False
 -             else:
 -                 item["embedding_available"] = True
 -         response = {"data": data, "has_more": len(datasets) == limit, "limit": limit, "total": total, "page": page}
 -         return response, 200
 - 
 -     @cloud_edition_billing_rate_limit_check("knowledge", "dataset")
 -     def post(self, tenant_id):
 -         """Resource for creating datasets."""
 -         parser = reqparse.RequestParser()
 -         parser.add_argument(
 -             "name",
 -             nullable=False,
 -             required=True,
 -             help="type is required. Name must be between 1 to 40 characters.",
 -             type=_validate_name,
 -         )
 -         parser.add_argument(
 -             "description",
 -             type=_validate_description_length,
 -             nullable=True,
 -             required=False,
 -             default="",
 -         )
 -         parser.add_argument(
 -             "indexing_technique",
 -             type=str,
 -             location="json",
 -             choices=Dataset.INDEXING_TECHNIQUE_LIST,
 -             help="Invalid indexing technique.",
 -         )
 -         parser.add_argument(
 -             "permission",
 -             type=str,
 -             location="json",
 -             choices=(DatasetPermissionEnum.ONLY_ME, DatasetPermissionEnum.ALL_TEAM, DatasetPermissionEnum.PARTIAL_TEAM),
 -             help="Invalid permission.",
 -             required=False,
 -             nullable=False,
 -         )
 -         parser.add_argument(
 -             "external_knowledge_api_id",
 -             type=str,
 -             nullable=True,
 -             required=False,
 -             default="_validate_name",
 -         )
 -         parser.add_argument(
 -             "provider",
 -             type=str,
 -             nullable=True,
 -             required=False,
 -             default="vendor",
 -         )
 -         parser.add_argument(
 -             "external_knowledge_id",
 -             type=str,
 -             nullable=True,
 -             required=False,
 -         )
 -         parser.add_argument("retrieval_model", type=dict, required=False, nullable=True, location="json")
 -         parser.add_argument("embedding_model", type=str, required=False, nullable=True, location="json")
 -         parser.add_argument("embedding_model_provider", type=str, required=False, nullable=True, location="json")
 - 
 -         args = parser.parse_args()
 - 
 -         if args.get("embedding_model_provider"):
 -             DatasetService.check_embedding_model_setting(
 -                 tenant_id, args.get("embedding_model_provider"), args.get("embedding_model")
 -             )
 -         if (
 -             args.get("retrieval_model")
 -             and args.get("retrieval_model").get("reranking_model")
 -             and args.get("retrieval_model").get("reranking_model").get("reranking_provider_name")
 -         ):
 -             DatasetService.check_reranking_model_setting(
 -                 tenant_id,
 -                 args.get("retrieval_model").get("reranking_model").get("reranking_provider_name"),
 -                 args.get("retrieval_model").get("reranking_model").get("reranking_model_name"),
 -             )
 - 
 -         try:
 -             dataset = DatasetService.create_empty_dataset(
 -                 tenant_id=tenant_id,
 -                 name=args["name"],
 -                 description=args["description"],
 -                 indexing_technique=args["indexing_technique"],
 -                 account=current_user,
 -                 permission=args["permission"],
 -                 provider=args["provider"],
 -                 external_knowledge_api_id=args["external_knowledge_api_id"],
 -                 external_knowledge_id=args["external_knowledge_id"],
 -                 embedding_model_provider=args["embedding_model_provider"],
 -                 embedding_model_name=args["embedding_model"],
 -                 retrieval_model=RetrievalModel(**args["retrieval_model"])
 -                 if args["retrieval_model"] is not None
 -                 else None,
 -             )
 -         except services.errors.dataset.DatasetNameDuplicateError:
 -             raise DatasetNameDuplicateError()
 - 
 -         return marshal(dataset, dataset_detail_fields), 200
 - 
 - 
 - class DatasetApi(DatasetApiResource):
 -     """Resource for dataset."""
 - 
 -     def get(self, _, dataset_id):
 -         dataset_id_str = str(dataset_id)
 -         dataset = DatasetService.get_dataset(dataset_id_str)
 -         if dataset is None:
 -             raise NotFound("Dataset not found.")
 -         try:
 -             DatasetService.check_dataset_permission(dataset, current_user)
 -         except services.errors.account.NoPermissionError as e:
 -             raise Forbidden(str(e))
 -         data = marshal(dataset, dataset_detail_fields)
 -         if data.get("permission") == "partial_members":
 -             part_users_list = DatasetPermissionService.get_dataset_partial_member_list(dataset_id_str)
 -             data.update({"partial_member_list": part_users_list})
 - 
 -         # check embedding setting
 -         provider_manager = ProviderManager()
 -         configurations = provider_manager.get_configurations(tenant_id=current_user.current_tenant_id)
 - 
 -         embedding_models = configurations.get_models(model_type=ModelType.TEXT_EMBEDDING, only_active=True)
 - 
 -         model_names = []
 -         for embedding_model in embedding_models:
 -             model_names.append(f"{embedding_model.model}:{embedding_model.provider.provider}")
 - 
 -         if data["indexing_technique"] == "high_quality":
 -             item_model = f"{data['embedding_model']}:{data['embedding_model_provider']}"
 -             if item_model in model_names:
 -                 data["embedding_available"] = True
 -             else:
 -                 data["embedding_available"] = False
 -         else:
 -             data["embedding_available"] = True
 - 
 -         if data.get("permission") == "partial_members":
 -             part_users_list = DatasetPermissionService.get_dataset_partial_member_list(dataset_id_str)
 -             data.update({"partial_member_list": part_users_list})
 - 
 -         return data, 200
 - 
 -     @cloud_edition_billing_rate_limit_check("knowledge", "dataset")
 -     def patch(self, _, dataset_id):
 -         dataset_id_str = str(dataset_id)
 -         dataset = DatasetService.get_dataset(dataset_id_str)
 -         if dataset is None:
 -             raise NotFound("Dataset not found.")
 - 
 -         parser = reqparse.RequestParser()
 -         parser.add_argument(
 -             "name",
 -             nullable=False,
 -             help="type is required. Name must be between 1 to 40 characters.",
 -             type=_validate_name,
 -         )
 -         parser.add_argument("description", location="json", store_missing=False, type=_validate_description_length)
 -         parser.add_argument(
 -             "indexing_technique",
 -             type=str,
 -             location="json",
 -             choices=Dataset.INDEXING_TECHNIQUE_LIST,
 -             nullable=True,
 -             help="Invalid indexing technique.",
 -         )
 -         parser.add_argument(
 -             "permission",
 -             type=str,
 -             location="json",
 -             choices=(DatasetPermissionEnum.ONLY_ME, DatasetPermissionEnum.ALL_TEAM, DatasetPermissionEnum.PARTIAL_TEAM),
 -             help="Invalid permission.",
 -         )
 -         parser.add_argument("embedding_model", type=str, location="json", help="Invalid embedding model.")
 -         parser.add_argument(
 -             "embedding_model_provider", type=str, location="json", help="Invalid embedding model provider."
 -         )
 -         parser.add_argument("retrieval_model", type=dict, location="json", help="Invalid retrieval model.")
 -         parser.add_argument("partial_member_list", type=list, location="json", help="Invalid parent user list.")
 - 
 -         parser.add_argument(
 -             "external_retrieval_model",
 -             type=dict,
 -             required=False,
 -             nullable=True,
 -             location="json",
 -             help="Invalid external retrieval model.",
 -         )
 - 
 -         parser.add_argument(
 -             "external_knowledge_id",
 -             type=str,
 -             required=False,
 -             nullable=True,
 -             location="json",
 -             help="Invalid external knowledge id.",
 -         )
 - 
 -         parser.add_argument(
 -             "external_knowledge_api_id",
 -             type=str,
 -             required=False,
 -             nullable=True,
 -             location="json",
 -             help="Invalid external knowledge api id.",
 -         )
 -         args = parser.parse_args()
 -         data = request.get_json()
 - 
 -         # check embedding model setting
 -         if data.get("indexing_technique") == "high_quality" or data.get("embedding_model_provider"):
 -             DatasetService.check_embedding_model_setting(
 -                 dataset.tenant_id, data.get("embedding_model_provider"), data.get("embedding_model")
 -             )
 -         if (
 -             data.get("retrieval_model")
 -             and data.get("retrieval_model").get("reranking_model")
 -             and data.get("retrieval_model").get("reranking_model").get("reranking_provider_name")
 -         ):
 -             DatasetService.check_reranking_model_setting(
 -                 dataset.tenant_id,
 -                 data.get("retrieval_model").get("reranking_model").get("reranking_provider_name"),
 -                 data.get("retrieval_model").get("reranking_model").get("reranking_model_name"),
 -             )
 - 
 -         # The role of the current user in the ta table must be admin, owner, editor, or dataset_operator
 -         DatasetPermissionService.check_permission(
 -             current_user, dataset, data.get("permission"), data.get("partial_member_list")
 -         )
 - 
 -         dataset = DatasetService.update_dataset(dataset_id_str, args, current_user)
 - 
 -         if dataset is None:
 -             raise NotFound("Dataset not found.")
 - 
 -         result_data = marshal(dataset, dataset_detail_fields)
 -         tenant_id = current_user.current_tenant_id
 - 
 -         if data.get("partial_member_list") and data.get("permission") == "partial_members":
 -             DatasetPermissionService.update_partial_member_list(
 -                 tenant_id, dataset_id_str, data.get("partial_member_list")
 -             )
 -         # clear partial member list when permission is only_me or all_team_members
 -         elif (
 -             data.get("permission") == DatasetPermissionEnum.ONLY_ME
 -             or data.get("permission") == DatasetPermissionEnum.ALL_TEAM
 -         ):
 -             DatasetPermissionService.clear_partial_member_list(dataset_id_str)
 - 
 -         partial_member_list = DatasetPermissionService.get_dataset_partial_member_list(dataset_id_str)
 -         result_data.update({"partial_member_list": partial_member_list})
 - 
 -         return result_data, 200
 - 
 -     @cloud_edition_billing_rate_limit_check("knowledge", "dataset")
 -     def delete(self, _, dataset_id):
 -         """
 -         Deletes a dataset given its ID.
 - 
 -         Args:
 -             _: ignore
 -             dataset_id (UUID): The ID of the dataset to be deleted.
 - 
 -         Returns:
 -             dict: A dictionary with a key 'result' and a value 'success'
 -                   if the dataset was successfully deleted. Omitted in HTTP response.
 -             int: HTTP status code 204 indicating that the operation was successful.
 - 
 -         Raises:
 -             NotFound: If the dataset with the given ID does not exist.
 -         """
 - 
 -         dataset_id_str = str(dataset_id)
 - 
 -         try:
 -             if DatasetService.delete_dataset(dataset_id_str, current_user):
 -                 DatasetPermissionService.clear_partial_member_list(dataset_id_str)
 -                 return 204
 -             else:
 -                 raise NotFound("Dataset not found.")
 -         except services.errors.dataset.DatasetInUseError:
 -             raise DatasetInUseError()
 - 
 - 
 - class DocumentStatusApi(DatasetApiResource):
 -     """Resource for batch document status operations."""
 - 
 -     def patch(self, tenant_id, dataset_id, action: Literal["enable", "disable", "archive", "un_archive"]):
 -         """
 -         Batch update document status.
 - 
 -         Args:
 -             tenant_id: tenant id
 -             dataset_id: dataset id
 -             action: action to perform (Literal["enable", "disable", "archive", "un_archive"])
 - 
 -         Returns:
 -             dict: A dictionary with a key 'result' and a value 'success'
 -             int: HTTP status code 200 indicating that the operation was successful.
 - 
 -         Raises:
 -             NotFound: If the dataset with the given ID does not exist.
 -             Forbidden: If the user does not have permission.
 -             InvalidActionError: If the action is invalid or cannot be performed.
 -         """
 -         dataset_id_str = str(dataset_id)
 -         dataset = DatasetService.get_dataset(dataset_id_str)
 - 
 -         if dataset is None:
 -             raise NotFound("Dataset not found.")
 - 
 -         # Check user's permission
 -         try:
 -             DatasetService.check_dataset_permission(dataset, current_user)
 -         except services.errors.account.NoPermissionError as e:
 -             raise Forbidden(str(e))
 - 
 -         # Check dataset model setting
 -         DatasetService.check_dataset_model_setting(dataset)
 - 
 -         # Get document IDs from request body
 -         data = request.get_json()
 -         document_ids = data.get("document_ids", [])
 - 
 -         try:
 -             DocumentService.batch_update_document_status(dataset, document_ids, action, current_user)
 -         except services.errors.document.DocumentIndexingError as e:
 -             raise InvalidActionError(str(e))
 -         except ValueError as e:
 -             raise InvalidActionError(str(e))
 - 
 -         return {"result": "success"}, 200
 - 
 - 
 - class DatasetTagsApi(DatasetApiResource):
 -     @validate_dataset_token
 -     @marshal_with(tag_fields)
 -     def get(self, _, dataset_id):
 -         """Get all knowledge type tags."""
 -         tags = TagService.get_tags("knowledge", current_user.current_tenant_id)
 - 
 -         return tags, 200
 - 
 -     @validate_dataset_token
 -     def post(self, _, dataset_id):
 -         """Add a knowledge type tag."""
 -         if not (current_user.is_editor or current_user.is_dataset_editor):
 -             raise Forbidden()
 - 
 -         parser = reqparse.RequestParser()
 -         parser.add_argument(
 -             "name",
 -             nullable=False,
 -             required=True,
 -             help="Name must be between 1 to 50 characters.",
 -             type=DatasetTagsApi._validate_tag_name,
 -         )
 - 
 -         args = parser.parse_args()
 -         args["type"] = "knowledge"
 -         tag = TagService.save_tags(args)
 - 
 -         response = {"id": tag.id, "name": tag.name, "type": tag.type, "binding_count": 0}
 - 
 -         return response, 200
 - 
 -     @validate_dataset_token
 -     def patch(self, _, dataset_id):
 -         if not (current_user.is_editor or current_user.is_dataset_editor):
 -             raise Forbidden()
 - 
 -         parser = reqparse.RequestParser()
 -         parser.add_argument(
 -             "name",
 -             nullable=False,
 -             required=True,
 -             help="Name must be between 1 to 50 characters.",
 -             type=DatasetTagsApi._validate_tag_name,
 -         )
 -         parser.add_argument("tag_id", nullable=False, required=True, help="Id of a tag.", type=str)
 -         args = parser.parse_args()
 -         args["type"] = "knowledge"
 -         tag = TagService.update_tags(args, args.get("tag_id"))
 - 
 -         binding_count = TagService.get_tag_binding_count(args.get("tag_id"))
 - 
 -         response = {"id": tag.id, "name": tag.name, "type": tag.type, "binding_count": binding_count}
 - 
 -         return response, 200
 - 
 -     @validate_dataset_token
 -     def delete(self, _, dataset_id):
 -         """Delete a knowledge type tag."""
 -         if not current_user.is_editor:
 -             raise Forbidden()
 -         parser = reqparse.RequestParser()
 -         parser.add_argument("tag_id", nullable=False, required=True, help="Id of a tag.", type=str)
 -         args = parser.parse_args()
 -         TagService.delete_tag(args.get("tag_id"))
 - 
 -         return 204
 - 
 -     @staticmethod
 -     def _validate_tag_name(name):
 -         if not name or len(name) < 1 or len(name) > 50:
 -             raise ValueError("Name must be between 1 to 50 characters.")
 -         return name
 - 
 - 
 - class DatasetTagBindingApi(DatasetApiResource):
 -     @validate_dataset_token
 -     def post(self, _, dataset_id):
 -         # The role of the current user in the ta table must be admin, owner, editor, or dataset_operator
 -         if not (current_user.is_editor or current_user.is_dataset_editor):
 -             raise Forbidden()
 - 
 -         parser = reqparse.RequestParser()
 -         parser.add_argument(
 -             "tag_ids", type=list, nullable=False, required=True, location="json", help="Tag IDs is required."
 -         )
 -         parser.add_argument(
 -             "target_id", type=str, nullable=False, required=True, location="json", help="Target Dataset ID is required."
 -         )
 - 
 -         args = parser.parse_args()
 -         args["type"] = "knowledge"
 -         TagService.save_tag_binding(args)
 - 
 -         return 204
 - 
 - 
 - class DatasetTagUnbindingApi(DatasetApiResource):
 -     @validate_dataset_token
 -     def post(self, _, dataset_id):
 -         # The role of the current user in the ta table must be admin, owner, editor, or dataset_operator
 -         if not (current_user.is_editor or current_user.is_dataset_editor):
 -             raise Forbidden()
 - 
 -         parser = reqparse.RequestParser()
 -         parser.add_argument("tag_id", type=str, nullable=False, required=True, help="Tag ID is required.")
 -         parser.add_argument("target_id", type=str, nullable=False, required=True, help="Target ID is required.")
 - 
 -         args = parser.parse_args()
 -         args["type"] = "knowledge"
 -         TagService.delete_tag_binding(args)
 - 
 -         return 204
 - 
 - 
 - class DatasetTagsBindingStatusApi(DatasetApiResource):
 -     @validate_dataset_token
 -     def get(self, _, *args, **kwargs):
 -         """Get all knowledge type tags."""
 -         dataset_id = kwargs.get("dataset_id")
 -         tags = TagService.get_tags_by_target_id("knowledge", current_user.current_tenant_id, str(dataset_id))
 -         tags_list = [{"id": tag.id, "name": tag.name} for tag in tags]
 -         response = {"data": tags_list, "total": len(tags)}
 -         return response, 200
 - 
 - 
 - api.add_resource(DatasetListApi, "/datasets")
 - api.add_resource(DatasetApi, "/datasets/<uuid:dataset_id>")
 - api.add_resource(DocumentStatusApi, "/datasets/<uuid:dataset_id>/documents/status/<string:action>")
 - api.add_resource(DatasetTagsApi, "/datasets/tags")
 - api.add_resource(DatasetTagBindingApi, "/datasets/tags/binding")
 - api.add_resource(DatasetTagUnbindingApi, "/datasets/tags/unbinding")
 - api.add_resource(DatasetTagsBindingStatusApi, "/datasets/<uuid:dataset_id>/tags")
 
 
  |