You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

azure_blob_storage.py 3.4KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485
  1. from collections.abc import Generator
  2. from datetime import timedelta
  3. from typing import Optional
  4. from azure.identity import ChainedTokenCredential, DefaultAzureCredential
  5. from azure.storage.blob import AccountSasPermissions, BlobServiceClient, ResourceTypes, generate_account_sas
  6. from configs import dify_config
  7. from extensions.ext_redis import redis_client
  8. from extensions.storage.base_storage import BaseStorage
  9. from libs.datetime_utils import naive_utc_now
  10. class AzureBlobStorage(BaseStorage):
  11. """Implementation for Azure Blob storage."""
  12. def __init__(self):
  13. super().__init__()
  14. self.bucket_name = dify_config.AZURE_BLOB_CONTAINER_NAME
  15. self.account_url = dify_config.AZURE_BLOB_ACCOUNT_URL
  16. self.account_name = dify_config.AZURE_BLOB_ACCOUNT_NAME
  17. self.account_key = dify_config.AZURE_BLOB_ACCOUNT_KEY
  18. self.credential: Optional[ChainedTokenCredential] = None
  19. if self.account_key == "managedidentity":
  20. self.credential = DefaultAzureCredential()
  21. else:
  22. self.credential = None
  23. def save(self, filename, data):
  24. client = self._sync_client()
  25. blob_container = client.get_container_client(container=self.bucket_name)
  26. blob_container.upload_blob(filename, data)
  27. def load_once(self, filename: str) -> bytes:
  28. client = self._sync_client()
  29. blob = client.get_container_client(container=self.bucket_name)
  30. blob = blob.get_blob_client(blob=filename)
  31. data: bytes = blob.download_blob().readall()
  32. return data
  33. def load_stream(self, filename: str) -> Generator:
  34. client = self._sync_client()
  35. blob = client.get_blob_client(container=self.bucket_name, blob=filename)
  36. blob_data = blob.download_blob()
  37. yield from blob_data.chunks()
  38. def download(self, filename, target_filepath):
  39. client = self._sync_client()
  40. blob = client.get_blob_client(container=self.bucket_name, blob=filename)
  41. with open(target_filepath, "wb") as my_blob:
  42. blob_data = blob.download_blob()
  43. blob_data.readinto(my_blob)
  44. def exists(self, filename):
  45. client = self._sync_client()
  46. blob = client.get_blob_client(container=self.bucket_name, blob=filename)
  47. return blob.exists()
  48. def delete(self, filename):
  49. client = self._sync_client()
  50. blob_container = client.get_container_client(container=self.bucket_name)
  51. blob_container.delete_blob(filename)
  52. def _sync_client(self):
  53. if self.account_key == "managedidentity":
  54. return BlobServiceClient(account_url=self.account_url, credential=self.credential) # type: ignore
  55. cache_key = "azure_blob_sas_token_{}_{}".format(self.account_name, self.account_key)
  56. cache_result = redis_client.get(cache_key)
  57. if cache_result is not None:
  58. sas_token = cache_result.decode("utf-8")
  59. else:
  60. sas_token = generate_account_sas(
  61. account_name=self.account_name or "",
  62. account_key=self.account_key or "",
  63. resource_types=ResourceTypes(service=True, container=True, object=True),
  64. permission=AccountSasPermissions(read=True, write=True, delete=True, list=True, add=True, create=True),
  65. expiry=naive_utc_now() + timedelta(hours=1),
  66. )
  67. redis_client.set(cache_key, sas_token, ex=3000)
  68. return BlobServiceClient(account_url=self.account_url or "", credential=sas_token)