mirror of
https://github.com/langgenius/dify.git
synced 2026-02-13 16:00:55 -05:00
Implement phase 1 of the file module migration by moving workflow-facing file primitives to core.workflow.file while keeping core.file as a temporary compatibility layer. What this commit changes - Add core.workflow.file package (constants/enums/models/helpers/file_manager/tool_file_parser). - Add protocol-based runtime binding in core.workflow.file.runtime and core.workflow.file.protocols. - Add application adapter core.app.workflow.file_runtime and bind runtime in extensions.ext_storage.init_app. - Bind runtime in tests via tests/conftest.py. - Migrate workflow-only imports from core.file.* to core.workflow.file.* across workflow runtime/nodes/entry/encoder and workflow node factory. - Update workflow unit tests to patch/import the new workflow file namespace. - Remove workflow-external-imports ignore_imports entries related to core.file from .importlinter. Compatibility guarantee for phase split - Keep core.file import path available in this phase by replacing core/file/*.py with forwarding bridge modules that re-export core.workflow.file symbols. - Preserve runtime behavior and isinstance(File) identity consistency while non-workflow modules are still on legacy import paths. Notes - This commit intentionally does not remove core.file. Full repository replacement and bridge removal are handled in phase 2.
130 lines
4.8 KiB
Python
130 lines
4.8 KiB
Python
import logging
|
|
from collections.abc import Callable, Generator
|
|
from typing import Literal, Union, overload
|
|
|
|
from flask import Flask
|
|
|
|
from configs import dify_config
|
|
from dify_app import DifyApp
|
|
from extensions.storage.base_storage import BaseStorage
|
|
from extensions.storage.storage_type import StorageType
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
class Storage:
|
|
def init_app(self, app: Flask):
|
|
storage_factory = self.get_storage_factory(dify_config.STORAGE_TYPE)
|
|
with app.app_context():
|
|
self.storage_runner = storage_factory()
|
|
|
|
@staticmethod
|
|
def get_storage_factory(storage_type: str) -> Callable[[], BaseStorage]:
|
|
match storage_type:
|
|
case StorageType.S3:
|
|
from extensions.storage.aws_s3_storage import AwsS3Storage
|
|
|
|
return AwsS3Storage
|
|
case StorageType.OPENDAL:
|
|
from extensions.storage.opendal_storage import OpenDALStorage
|
|
|
|
return lambda: OpenDALStorage(dify_config.OPENDAL_SCHEME)
|
|
case StorageType.LOCAL:
|
|
from extensions.storage.opendal_storage import OpenDALStorage
|
|
|
|
return lambda: OpenDALStorage(scheme="fs", root=dify_config.STORAGE_LOCAL_PATH)
|
|
case StorageType.AZURE_BLOB:
|
|
from extensions.storage.azure_blob_storage import AzureBlobStorage
|
|
|
|
return AzureBlobStorage
|
|
case StorageType.ALIYUN_OSS:
|
|
from extensions.storage.aliyun_oss_storage import AliyunOssStorage
|
|
|
|
return AliyunOssStorage
|
|
case StorageType.GOOGLE_STORAGE:
|
|
from extensions.storage.google_cloud_storage import GoogleCloudStorage
|
|
|
|
return GoogleCloudStorage
|
|
case StorageType.TENCENT_COS:
|
|
from extensions.storage.tencent_cos_storage import TencentCosStorage
|
|
|
|
return TencentCosStorage
|
|
case StorageType.OCI_STORAGE:
|
|
from extensions.storage.oracle_oci_storage import OracleOCIStorage
|
|
|
|
return OracleOCIStorage
|
|
case StorageType.HUAWEI_OBS:
|
|
from extensions.storage.huawei_obs_storage import HuaweiObsStorage
|
|
|
|
return HuaweiObsStorage
|
|
case StorageType.BAIDU_OBS:
|
|
from extensions.storage.baidu_obs_storage import BaiduObsStorage
|
|
|
|
return BaiduObsStorage
|
|
case StorageType.VOLCENGINE_TOS:
|
|
from extensions.storage.volcengine_tos_storage import VolcengineTosStorage
|
|
|
|
return VolcengineTosStorage
|
|
case StorageType.SUPABASE:
|
|
from extensions.storage.supabase_storage import SupabaseStorage
|
|
|
|
return SupabaseStorage
|
|
case StorageType.CLICKZETTA_VOLUME:
|
|
from extensions.storage.clickzetta_volume.clickzetta_volume_storage import (
|
|
ClickZettaVolumeConfig,
|
|
ClickZettaVolumeStorage,
|
|
)
|
|
|
|
def create_clickzetta_volume_storage():
|
|
# ClickZettaVolumeConfig will automatically read from environment variables
|
|
# and fallback to CLICKZETTA_* config if CLICKZETTA_VOLUME_* is not set
|
|
volume_config = ClickZettaVolumeConfig()
|
|
return ClickZettaVolumeStorage(volume_config)
|
|
|
|
return create_clickzetta_volume_storage
|
|
case _:
|
|
raise ValueError(f"unsupported storage type {storage_type}")
|
|
|
|
def save(self, filename: str, data: bytes):
|
|
self.storage_runner.save(filename, data)
|
|
|
|
@overload
|
|
def load(self, filename: str, /, *, stream: Literal[False] = False) -> bytes: ...
|
|
|
|
@overload
|
|
def load(self, filename: str, /, *, stream: Literal[True]) -> Generator: ...
|
|
|
|
def load(self, filename: str, /, *, stream: bool = False) -> Union[bytes, Generator]:
|
|
if stream:
|
|
return self.load_stream(filename)
|
|
else:
|
|
return self.load_once(filename)
|
|
|
|
def load_once(self, filename: str) -> bytes:
|
|
return self.storage_runner.load_once(filename)
|
|
|
|
def load_stream(self, filename: str) -> Generator:
|
|
return self.storage_runner.load_stream(filename)
|
|
|
|
def download(self, filename, target_filepath):
|
|
self.storage_runner.download(filename, target_filepath)
|
|
|
|
def exists(self, filename):
|
|
return self.storage_runner.exists(filename)
|
|
|
|
def delete(self, filename: str):
|
|
return self.storage_runner.delete(filename)
|
|
|
|
def scan(self, path: str, files: bool = True, directories: bool = False) -> list[str]:
|
|
return self.storage_runner.scan(path, files=files, directories=directories)
|
|
|
|
|
|
storage = Storage()
|
|
|
|
|
|
def init_app(app: DifyApp):
|
|
storage.init_app(app)
|
|
from core.app.workflow.file_runtime import bind_dify_workflow_file_runtime
|
|
|
|
bind_dify_workflow_file_runtime()
|