Du kannst nicht mehr als 25 Themen auswählen Themen müssen mit entweder einem Buchstaben oder einer Ziffer beginnen. Sie können Bindestriche („-“) enthalten und bis zu 35 Zeichen lang sein.

dataset.py 18KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535
  1. #
  2. # Copyright 2024 The InfiniFlow Authors. All Rights Reserved.
  3. #
  4. # Licensed under the Apache License, Version 2.0 (the "License");
  5. # you may not use this file except in compliance with the License.
  6. # You may obtain a copy of the License at
  7. #
  8. # http://www.apache.org/licenses/LICENSE-2.0
  9. #
  10. # Unless required by applicable law or agreed to in writing, software
  11. # distributed under the License is distributed on an "AS IS" BASIS,
  12. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. # See the License for the specific language governing permissions and
  14. # limitations under the License.
  15. #
  16. from flask import request
  17. from api.db import StatusEnum, FileSource
  18. from api.db.db_models import File
  19. from api.db.services.document_service import DocumentService
  20. from api.db.services.file2document_service import File2DocumentService
  21. from api.db.services.file_service import FileService
  22. from api.db.services.knowledgebase_service import KnowledgebaseService
  23. from api.db.services.llm_service import TenantLLMService, LLMService
  24. from api.db.services.user_service import TenantService
  25. from api import settings
  26. from api.utils import get_uuid
  27. from api.utils.api_utils import (
  28. get_result,
  29. token_required,
  30. get_error_data_result,
  31. valid,
  32. get_parser_config,
  33. )
  34. @manager.route("/datasets", methods=["POST"]) # noqa: F821
  35. @token_required
  36. def create(tenant_id):
  37. """
  38. Create a new dataset.
  39. ---
  40. tags:
  41. - Datasets
  42. security:
  43. - ApiKeyAuth: []
  44. parameters:
  45. - in: header
  46. name: Authorization
  47. type: string
  48. required: true
  49. description: Bearer token for authentication.
  50. - in: body
  51. name: body
  52. description: Dataset creation parameters.
  53. required: true
  54. schema:
  55. type: object
  56. required:
  57. - name
  58. properties:
  59. name:
  60. type: string
  61. description: Name of the dataset.
  62. permission:
  63. type: string
  64. enum: ['me', 'team']
  65. description: Dataset permission.
  66. language:
  67. type: string
  68. enum: ['Chinese', 'English']
  69. description: Language of the dataset.
  70. chunk_method:
  71. type: string
  72. enum: ["naive", "manual", "qa", "table", "paper", "book", "laws",
  73. "presentation", "picture", "one", "knowledge_graph", "email", "tag"
  74. ]
  75. description: Chunking method.
  76. parser_config:
  77. type: object
  78. description: Parser configuration.
  79. responses:
  80. 200:
  81. description: Successful operation.
  82. schema:
  83. type: object
  84. properties:
  85. data:
  86. type: object
  87. """
  88. req = request.json
  89. e, t = TenantService.get_by_id(tenant_id)
  90. permission = req.get("permission")
  91. language = req.get("language")
  92. chunk_method = req.get("chunk_method")
  93. parser_config = req.get("parser_config")
  94. valid_permission = ["me", "team"]
  95. valid_language = ["Chinese", "English"]
  96. valid_chunk_method = [
  97. "naive",
  98. "manual",
  99. "qa",
  100. "table",
  101. "paper",
  102. "book",
  103. "laws",
  104. "presentation",
  105. "picture",
  106. "one",
  107. "knowledge_graph",
  108. "email",
  109. "tag"
  110. ]
  111. check_validation = valid(
  112. permission,
  113. valid_permission,
  114. language,
  115. valid_language,
  116. chunk_method,
  117. valid_chunk_method,
  118. )
  119. if check_validation:
  120. return check_validation
  121. req["parser_config"] = get_parser_config(chunk_method, parser_config)
  122. if "tenant_id" in req:
  123. return get_error_data_result(message="`tenant_id` must not be provided")
  124. if "chunk_count" in req or "document_count" in req:
  125. return get_error_data_result(
  126. message="`chunk_count` or `document_count` must not be provided"
  127. )
  128. if "name" not in req:
  129. return get_error_data_result(message="`name` is not empty!")
  130. req["id"] = get_uuid()
  131. req["name"] = req["name"].strip()
  132. if req["name"] == "":
  133. return get_error_data_result(message="`name` is not empty string!")
  134. if KnowledgebaseService.query(
  135. name=req["name"], tenant_id=tenant_id, status=StatusEnum.VALID.value
  136. ):
  137. return get_error_data_result(
  138. message="Duplicated dataset name in creating dataset."
  139. )
  140. req["tenant_id"] = req["created_by"] = tenant_id
  141. if not req.get("embedding_model"):
  142. req["embedding_model"] = t.embd_id
  143. else:
  144. valid_embedding_models = [
  145. "BAAI/bge-large-zh-v1.5",
  146. "BAAI/bge-base-en-v1.5",
  147. "BAAI/bge-large-en-v1.5",
  148. "BAAI/bge-small-en-v1.5",
  149. "BAAI/bge-small-zh-v1.5",
  150. "jinaai/jina-embeddings-v2-base-en",
  151. "jinaai/jina-embeddings-v2-small-en",
  152. "nomic-ai/nomic-embed-text-v1.5",
  153. "sentence-transformers/all-MiniLM-L6-v2",
  154. "text-embedding-v2",
  155. "text-embedding-v3",
  156. "maidalun1020/bce-embedding-base_v1",
  157. ]
  158. embd_model = LLMService.query(
  159. llm_name=req["embedding_model"], model_type="embedding"
  160. )
  161. if embd_model:
  162. if req["embedding_model"] not in valid_embedding_models and not TenantLLMService.query(tenant_id=tenant_id,model_type="embedding",llm_name=req.get("embedding_model"),):
  163. return get_error_data_result(f"`embedding_model` {req.get('embedding_model')} doesn't exist")
  164. if not embd_model:
  165. embd_model=TenantLLMService.query(tenant_id=tenant_id,model_type="embedding", llm_name=req.get("embedding_model"))
  166. if not embd_model:
  167. return get_error_data_result(
  168. f"`embedding_model` {req.get('embedding_model')} doesn't exist"
  169. )
  170. key_mapping = {
  171. "chunk_num": "chunk_count",
  172. "doc_num": "document_count",
  173. "parser_id": "chunk_method",
  174. "embd_id": "embedding_model",
  175. }
  176. mapped_keys = {
  177. new_key: req[old_key]
  178. for new_key, old_key in key_mapping.items()
  179. if old_key in req
  180. }
  181. req.update(mapped_keys)
  182. if not KnowledgebaseService.save(**req):
  183. return get_error_data_result(message="Create dataset error.(Database error)")
  184. renamed_data = {}
  185. e, k = KnowledgebaseService.get_by_id(req["id"])
  186. for key, value in k.to_dict().items():
  187. new_key = key_mapping.get(key, key)
  188. renamed_data[new_key] = value
  189. return get_result(data=renamed_data)
  190. @manager.route("/datasets", methods=["DELETE"]) # noqa: F821
  191. @token_required
  192. def delete(tenant_id):
  193. """
  194. Delete datasets.
  195. ---
  196. tags:
  197. - Datasets
  198. security:
  199. - ApiKeyAuth: []
  200. parameters:
  201. - in: header
  202. name: Authorization
  203. type: string
  204. required: true
  205. description: Bearer token for authentication.
  206. - in: body
  207. name: body
  208. description: Dataset deletion parameters.
  209. required: true
  210. schema:
  211. type: object
  212. properties:
  213. ids:
  214. type: array
  215. items:
  216. type: string
  217. description: List of dataset IDs to delete.
  218. responses:
  219. 200:
  220. description: Successful operation.
  221. schema:
  222. type: object
  223. """
  224. req = request.json
  225. if not req:
  226. ids = None
  227. else:
  228. ids = req.get("ids")
  229. if not ids:
  230. id_list = []
  231. kbs = KnowledgebaseService.query(tenant_id=tenant_id)
  232. for kb in kbs:
  233. id_list.append(kb.id)
  234. else:
  235. id_list = ids
  236. for id in id_list:
  237. kbs = KnowledgebaseService.query(id=id, tenant_id=tenant_id)
  238. if not kbs:
  239. return get_error_data_result(message=f"You don't own the dataset {id}")
  240. for doc in DocumentService.query(kb_id=id):
  241. if not DocumentService.remove_document(doc, tenant_id):
  242. return get_error_data_result(
  243. message="Remove document error.(Database error)"
  244. )
  245. f2d = File2DocumentService.get_by_document_id(doc.id)
  246. FileService.filter_delete(
  247. [
  248. File.source_type == FileSource.KNOWLEDGEBASE,
  249. File.id == f2d[0].file_id,
  250. ]
  251. )
  252. File2DocumentService.delete_by_document_id(doc.id)
  253. FileService.filter_delete(
  254. [File.source_type == FileSource.KNOWLEDGEBASE, File.type == "folder", File.name == kbs[0].name])
  255. if not KnowledgebaseService.delete_by_id(id):
  256. return get_error_data_result(message="Delete dataset error.(Database error)")
  257. return get_result(code=settings.RetCode.SUCCESS)
  258. @manager.route("/datasets/<dataset_id>", methods=["PUT"]) # noqa: F821
  259. @token_required
  260. def update(tenant_id, dataset_id):
  261. """
  262. Update a dataset.
  263. ---
  264. tags:
  265. - Datasets
  266. security:
  267. - ApiKeyAuth: []
  268. parameters:
  269. - in: path
  270. name: dataset_id
  271. type: string
  272. required: true
  273. description: ID of the dataset to update.
  274. - in: header
  275. name: Authorization
  276. type: string
  277. required: true
  278. description: Bearer token for authentication.
  279. - in: body
  280. name: body
  281. description: Dataset update parameters.
  282. required: true
  283. schema:
  284. type: object
  285. properties:
  286. name:
  287. type: string
  288. description: New name of the dataset.
  289. permission:
  290. type: string
  291. enum: ['me', 'team']
  292. description: Updated permission.
  293. language:
  294. type: string
  295. enum: ['Chinese', 'English']
  296. description: Updated language.
  297. chunk_method:
  298. type: string
  299. enum: ["naive", "manual", "qa", "table", "paper", "book", "laws",
  300. "presentation", "picture", "one", "knowledge_graph", "email", "tag"
  301. ]
  302. description: Updated chunking method.
  303. parser_config:
  304. type: object
  305. description: Updated parser configuration.
  306. responses:
  307. 200:
  308. description: Successful operation.
  309. schema:
  310. type: object
  311. """
  312. if not KnowledgebaseService.query(id=dataset_id, tenant_id=tenant_id):
  313. return get_error_data_result(message="You don't own the dataset")
  314. req = request.json
  315. e, t = TenantService.get_by_id(tenant_id)
  316. invalid_keys = {"id", "embd_id", "chunk_num", "doc_num", "parser_id"}
  317. if any(key in req for key in invalid_keys):
  318. return get_error_data_result(message="The input parameters are invalid.")
  319. permission = req.get("permission")
  320. language = req.get("language")
  321. chunk_method = req.get("chunk_method")
  322. parser_config = req.get("parser_config")
  323. valid_permission = ["me", "team"]
  324. valid_language = ["Chinese", "English"]
  325. valid_chunk_method = [
  326. "naive",
  327. "manual",
  328. "qa",
  329. "table",
  330. "paper",
  331. "book",
  332. "laws",
  333. "presentation",
  334. "picture",
  335. "one",
  336. "knowledge_graph",
  337. "email",
  338. "tag"
  339. ]
  340. check_validation = valid(
  341. permission,
  342. valid_permission,
  343. language,
  344. valid_language,
  345. chunk_method,
  346. valid_chunk_method,
  347. )
  348. if check_validation:
  349. return check_validation
  350. if "tenant_id" in req:
  351. if req["tenant_id"] != tenant_id:
  352. return get_error_data_result(message="Can't change `tenant_id`.")
  353. e, kb = KnowledgebaseService.get_by_id(dataset_id)
  354. if "parser_config" in req:
  355. temp_dict = kb.parser_config
  356. temp_dict.update(req["parser_config"])
  357. req["parser_config"] = temp_dict
  358. if "chunk_count" in req:
  359. if req["chunk_count"] != kb.chunk_num:
  360. return get_error_data_result(message="Can't change `chunk_count`.")
  361. req.pop("chunk_count")
  362. if "document_count" in req:
  363. if req["document_count"] != kb.doc_num:
  364. return get_error_data_result(message="Can't change `document_count`.")
  365. req.pop("document_count")
  366. if "chunk_method" in req:
  367. if kb.chunk_num != 0 and req["chunk_method"] != kb.parser_id:
  368. return get_error_data_result(
  369. message="If `chunk_count` is not 0, `chunk_method` is not changeable."
  370. )
  371. req["parser_id"] = req.pop("chunk_method")
  372. if req["parser_id"] != kb.parser_id:
  373. if not req.get("parser_config"):
  374. req["parser_config"] = get_parser_config(chunk_method, parser_config)
  375. if "embedding_model" in req:
  376. if kb.chunk_num != 0 and req["embedding_model"] != kb.embd_id:
  377. return get_error_data_result(
  378. message="If `chunk_count` is not 0, `embedding_model` is not changeable."
  379. )
  380. if not req.get("embedding_model"):
  381. return get_error_data_result("`embedding_model` can't be empty")
  382. valid_embedding_models = [
  383. "BAAI/bge-large-zh-v1.5",
  384. "BAAI/bge-base-en-v1.5",
  385. "BAAI/bge-large-en-v1.5",
  386. "BAAI/bge-small-en-v1.5",
  387. "BAAI/bge-small-zh-v1.5",
  388. "jinaai/jina-embeddings-v2-base-en",
  389. "jinaai/jina-embeddings-v2-small-en",
  390. "nomic-ai/nomic-embed-text-v1.5",
  391. "sentence-transformers/all-MiniLM-L6-v2",
  392. "text-embedding-v2",
  393. "text-embedding-v3",
  394. "maidalun1020/bce-embedding-base_v1",
  395. ]
  396. embd_model = LLMService.query(
  397. llm_name=req["embedding_model"], model_type="embedding"
  398. )
  399. if embd_model:
  400. if req["embedding_model"] not in valid_embedding_models and not TenantLLMService.query(tenant_id=tenant_id,model_type="embedding",llm_name=req.get("embedding_model"),):
  401. return get_error_data_result(f"`embedding_model` {req.get('embedding_model')} doesn't exist")
  402. if not embd_model:
  403. embd_model=TenantLLMService.query(tenant_id=tenant_id,model_type="embedding", llm_name=req.get("embedding_model"))
  404. if not embd_model:
  405. return get_error_data_result(
  406. f"`embedding_model` {req.get('embedding_model')} doesn't exist"
  407. )
  408. req["embd_id"] = req.pop("embedding_model")
  409. if "name" in req:
  410. req["name"] = req["name"].strip()
  411. if (
  412. req["name"].lower() != kb.name.lower()
  413. and len(
  414. KnowledgebaseService.query(
  415. name=req["name"], tenant_id=tenant_id, status=StatusEnum.VALID.value
  416. )
  417. )
  418. > 0
  419. ):
  420. return get_error_data_result(
  421. message="Duplicated dataset name in updating dataset."
  422. )
  423. if not KnowledgebaseService.update_by_id(kb.id, req):
  424. return get_error_data_result(message="Update dataset error.(Database error)")
  425. return get_result(code=settings.RetCode.SUCCESS)
  426. @manager.route("/datasets", methods=["GET"]) # noqa: F821
  427. @token_required
  428. def list(tenant_id):
  429. """
  430. List datasets.
  431. ---
  432. tags:
  433. - Datasets
  434. security:
  435. - ApiKeyAuth: []
  436. parameters:
  437. - in: query
  438. name: id
  439. type: string
  440. required: false
  441. description: Dataset ID to filter.
  442. - in: query
  443. name: name
  444. type: string
  445. required: false
  446. description: Dataset name to filter.
  447. - in: query
  448. name: page
  449. type: integer
  450. required: false
  451. default: 1
  452. description: Page number.
  453. - in: query
  454. name: page_size
  455. type: integer
  456. required: false
  457. default: 1024
  458. description: Number of items per page.
  459. - in: query
  460. name: orderby
  461. type: string
  462. required: false
  463. default: "create_time"
  464. description: Field to order by.
  465. - in: query
  466. name: desc
  467. type: boolean
  468. required: false
  469. default: true
  470. description: Order in descending.
  471. - in: header
  472. name: Authorization
  473. type: string
  474. required: true
  475. description: Bearer token for authentication.
  476. responses:
  477. 200:
  478. description: Successful operation.
  479. schema:
  480. type: array
  481. items:
  482. type: object
  483. """
  484. id = request.args.get("id")
  485. name = request.args.get("name")
  486. if id:
  487. kbs = KnowledgebaseService.get_kb_by_id(id,tenant_id)
  488. if not kbs:
  489. return get_error_data_result(f"You don't own the dataset {id}")
  490. if name:
  491. kbs = KnowledgebaseService.get_kb_by_name(name,tenant_id)
  492. if not kbs:
  493. return get_error_data_result(f"You don't own the dataset {name}")
  494. page_number = int(request.args.get("page", 1))
  495. items_per_page = int(request.args.get("page_size", 30))
  496. orderby = request.args.get("orderby", "create_time")
  497. if request.args.get("desc") == "False" or request.args.get("desc") == "false":
  498. desc = False
  499. else:
  500. desc = True
  501. tenants = TenantService.get_joined_tenants_by_user_id(tenant_id)
  502. kbs = KnowledgebaseService.get_list(
  503. [m["tenant_id"] for m in tenants],
  504. tenant_id,
  505. page_number,
  506. items_per_page,
  507. orderby,
  508. desc,
  509. id,
  510. name,
  511. )
  512. renamed_list = []
  513. for kb in kbs:
  514. key_mapping = {
  515. "chunk_num": "chunk_count",
  516. "doc_num": "document_count",
  517. "parser_id": "chunk_method",
  518. "embd_id": "embedding_model",
  519. }
  520. renamed_data = {}
  521. for key, value in kb.items():
  522. new_key = key_mapping.get(key, key)
  523. renamed_data[new_key] = value
  524. renamed_list.append(renamed_data)
  525. return get_result(data=renamed_list)