You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

settings.py 2.8KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485
  1. #
  2. # Copyright 2024 The InfiniFlow Authors. All Rights Reserved.
  3. #
  4. # Licensed under the Apache License, Version 2.0 (the "License");
  5. # you may not use this file except in compliance with the License.
  6. # You may obtain a copy of the License at
  7. #
  8. # http://www.apache.org/licenses/LICENSE-2.0
  9. #
  10. # Unless required by applicable law or agreed to in writing, software
  11. # distributed under the License is distributed on an "AS IS" BASIS,
  12. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. # See the License for the specific language governing permissions and
  14. # limitations under the License.
  15. #
  16. import os
  17. import logging
  18. from api.utils import get_base_config, decrypt_database_config
  19. from api.utils.file_utils import get_project_base_directory
  20. # Server
  21. RAG_CONF_PATH = os.path.join(get_project_base_directory(), "conf")
  22. # Get storage type and document engine from system environment variables
  23. STORAGE_IMPL_TYPE = os.getenv('STORAGE_IMPL', 'MINIO')
  24. DOC_ENGINE = os.getenv('DOC_ENGINE', 'elasticsearch')
  25. ES = {}
  26. INFINITY = {}
  27. AZURE = {}
  28. S3 = {}
  29. MINIO = {}
  30. OSS = {}
  31. OS = {}
  32. # Initialize the selected configuration data based on environment variables to solve the problem of initialization errors due to lack of configuration
  33. if DOC_ENGINE == 'elasticsearch':
  34. ES = get_base_config("es", {})
  35. elif DOC_ENGINE == 'opensearch':
  36. OS = get_base_config("os", {})
  37. elif DOC_ENGINE == 'infinity':
  38. INFINITY = get_base_config("infinity", {"uri": "infinity:23817"})
  39. if STORAGE_IMPL_TYPE in ['AZURE_SPN', 'AZURE_SAS']:
  40. AZURE = get_base_config("azure", {})
  41. elif STORAGE_IMPL_TYPE == 'AWS_S3':
  42. S3 = get_base_config("s3", {})
  43. elif STORAGE_IMPL_TYPE == 'MINIO':
  44. MINIO = decrypt_database_config(name="minio")
  45. elif STORAGE_IMPL_TYPE == 'OSS':
  46. OSS = get_base_config("oss", {})
  47. try:
  48. REDIS = decrypt_database_config(name="redis")
  49. except Exception:
  50. REDIS = {}
  51. pass
  52. DOC_MAXIMUM_SIZE = int(os.environ.get("MAX_CONTENT_LENGTH", 128 * 1024 * 1024))
  53. DOC_BULK_SIZE = int(os.environ.get("DOC_BULK_SIZE", 4))
  54. EMBEDDING_BATCH_SIZE = int(os.environ.get("EMBEDDING_BATCH_SIZE", 16))
  55. SVR_QUEUE_NAME = "rag_flow_svr_queue"
  56. SVR_CONSUMER_GROUP_NAME = "rag_flow_svr_task_broker"
  57. PAGERANK_FLD = "pagerank_fea"
  58. TAG_FLD = "tag_feas"
  59. PARALLEL_DEVICES = 0
  60. try:
  61. import torch.cuda
  62. PARALLEL_DEVICES = torch.cuda.device_count()
  63. logging.info(f"found {PARALLEL_DEVICES} gpus")
  64. except Exception:
  65. logging.info("can't import package 'torch'")
  66. def print_rag_settings():
  67. logging.info(f"MAX_CONTENT_LENGTH: {DOC_MAXIMUM_SIZE}")
  68. logging.info(f"MAX_FILE_COUNT_PER_USER: {int(os.environ.get('MAX_FILE_NUM_PER_USER', 0))}")
  69. def get_svr_queue_name(priority: int) -> str:
  70. if priority == 0:
  71. return SVR_QUEUE_NAME
  72. return f"{SVR_QUEUE_NAME}_{priority}"
  73. def get_svr_queue_names():
  74. return [get_svr_queue_name(priority) for priority in [1, 0]]