| import base64 |
| import json |
| import logging |
| import secrets |
| from typing import Optional |
|
|
| import click |
| from flask import current_app |
| from werkzeug.exceptions import NotFound |
|
|
| from configs import dify_config |
| from constants.languages import languages |
| from core.rag.datasource.vdb.vector_factory import Vector |
| from core.rag.datasource.vdb.vector_type import VectorType |
| from core.rag.models.document import Document |
| from events.app_event import app_was_created |
| from extensions.ext_database import db |
| from extensions.ext_redis import redis_client |
| from libs.helper import email as email_validate |
| from libs.password import hash_password, password_pattern, valid_password |
| from libs.rsa import generate_key_pair |
| from models import Tenant |
| from models.dataset import Dataset, DatasetCollectionBinding, DocumentSegment |
| from models.dataset import Document as DatasetDocument |
| from models.model import Account, App, AppAnnotationSetting, AppMode, Conversation, MessageAnnotation |
| from models.provider import Provider, ProviderModel |
| from services.account_service import RegisterService, TenantService |
|
|
|
|
| @click.command("reset-password", help="Reset the account password.") |
| @click.option("--email", prompt=True, help="Account email to reset password for") |
| @click.option("--new-password", prompt=True, help="New password") |
| @click.option("--password-confirm", prompt=True, help="Confirm new password") |
| def reset_password(email, new_password, password_confirm): |
| """ |
| Reset password of owner account |
| Only available in SELF_HOSTED mode |
| """ |
| if str(new_password).strip() != str(password_confirm).strip(): |
| click.echo(click.style("Passwords do not match.", fg="red")) |
| return |
|
|
| account = db.session.query(Account).filter(Account.email == email).one_or_none() |
|
|
| if not account: |
| click.echo(click.style("Account not found for email: {}".format(email), fg="red")) |
| return |
|
|
| try: |
| valid_password(new_password) |
| except: |
| click.echo(click.style("Invalid password. Must match {}".format(password_pattern), fg="red")) |
| return |
|
|
| |
| salt = secrets.token_bytes(16) |
| base64_salt = base64.b64encode(salt).decode() |
|
|
| |
| password_hashed = hash_password(new_password, salt) |
| base64_password_hashed = base64.b64encode(password_hashed).decode() |
| account.password = base64_password_hashed |
| account.password_salt = base64_salt |
| db.session.commit() |
| click.echo(click.style("Password reset successfully.", fg="green")) |
|
|
|
|
| @click.command("reset-email", help="Reset the account email.") |
| @click.option("--email", prompt=True, help="Current account email") |
| @click.option("--new-email", prompt=True, help="New email") |
| @click.option("--email-confirm", prompt=True, help="Confirm new email") |
| def reset_email(email, new_email, email_confirm): |
| """ |
| Replace account email |
| :return: |
| """ |
| if str(new_email).strip() != str(email_confirm).strip(): |
| click.echo(click.style("New emails do not match.", fg="red")) |
| return |
|
|
| account = db.session.query(Account).filter(Account.email == email).one_or_none() |
|
|
| if not account: |
| click.echo(click.style("Account not found for email: {}".format(email), fg="red")) |
| return |
|
|
| try: |
| email_validate(new_email) |
| except: |
| click.echo(click.style("Invalid email: {}".format(new_email), fg="red")) |
| return |
|
|
| account.email = new_email |
| db.session.commit() |
| click.echo(click.style("Email updated successfully.", fg="green")) |
|
|
|
|
| @click.command( |
| "reset-encrypt-key-pair", |
| help="Reset the asymmetric key pair of workspace for encrypt LLM credentials. " |
| "After the reset, all LLM credentials will become invalid, " |
| "requiring re-entry." |
| "Only support SELF_HOSTED mode.", |
| ) |
| @click.confirmation_option( |
| prompt=click.style( |
| "Are you sure you want to reset encrypt key pair? This operation cannot be rolled back!", fg="red" |
| ) |
| ) |
| def reset_encrypt_key_pair(): |
| """ |
| Reset the encrypted key pair of workspace for encrypt LLM credentials. |
| After the reset, all LLM credentials will become invalid, requiring re-entry. |
| Only support SELF_HOSTED mode. |
| """ |
| if dify_config.EDITION != "SELF_HOSTED": |
| click.echo(click.style("This command is only for SELF_HOSTED installations.", fg="red")) |
| return |
|
|
| tenants = db.session.query(Tenant).all() |
| for tenant in tenants: |
| if not tenant: |
| click.echo(click.style("No workspaces found. Run /install first.", fg="red")) |
| return |
|
|
| tenant.encrypt_public_key = generate_key_pair(tenant.id) |
|
|
| db.session.query(Provider).filter(Provider.provider_type == "custom", Provider.tenant_id == tenant.id).delete() |
| db.session.query(ProviderModel).filter(ProviderModel.tenant_id == tenant.id).delete() |
| db.session.commit() |
|
|
| click.echo( |
| click.style( |
| "Congratulations! The asymmetric key pair of workspace {} has been reset.".format(tenant.id), |
| fg="green", |
| ) |
| ) |
|
|
|
|
| @click.command("vdb-migrate", help="Migrate vector db.") |
| @click.option("--scope", default="all", prompt=False, help="The scope of vector database to migrate, Default is All.") |
| def vdb_migrate(scope: str): |
| if scope in {"knowledge", "all"}: |
| migrate_knowledge_vector_database() |
| if scope in {"annotation", "all"}: |
| migrate_annotation_vector_database() |
|
|
|
|
| def migrate_annotation_vector_database(): |
| """ |
| Migrate annotation datas to target vector database . |
| """ |
| click.echo(click.style("Starting annotation data migration.", fg="green")) |
| create_count = 0 |
| skipped_count = 0 |
| total_count = 0 |
| page = 1 |
| while True: |
| try: |
| |
| apps = ( |
| db.session.query(App) |
| .filter(App.status == "normal") |
| .order_by(App.created_at.desc()) |
| .paginate(page=page, per_page=50) |
| ) |
| except NotFound: |
| break |
|
|
| page += 1 |
| for app in apps: |
| total_count = total_count + 1 |
| click.echo( |
| f"Processing the {total_count} app {app.id}. " + f"{create_count} created, {skipped_count} skipped." |
| ) |
| try: |
| click.echo("Creating app annotation index: {}".format(app.id)) |
| app_annotation_setting = ( |
| db.session.query(AppAnnotationSetting).filter(AppAnnotationSetting.app_id == app.id).first() |
| ) |
|
|
| if not app_annotation_setting: |
| skipped_count = skipped_count + 1 |
| click.echo("App annotation setting disabled: {}".format(app.id)) |
| continue |
| |
| dataset_collection_binding = ( |
| db.session.query(DatasetCollectionBinding) |
| .filter(DatasetCollectionBinding.id == app_annotation_setting.collection_binding_id) |
| .first() |
| ) |
| if not dataset_collection_binding: |
| click.echo("App annotation collection binding not found: {}".format(app.id)) |
| continue |
| annotations = db.session.query(MessageAnnotation).filter(MessageAnnotation.app_id == app.id).all() |
| dataset = Dataset( |
| id=app.id, |
| tenant_id=app.tenant_id, |
| indexing_technique="high_quality", |
| embedding_model_provider=dataset_collection_binding.provider_name, |
| embedding_model=dataset_collection_binding.model_name, |
| collection_binding_id=dataset_collection_binding.id, |
| ) |
| documents = [] |
| if annotations: |
| for annotation in annotations: |
| document = Document( |
| page_content=annotation.question, |
| metadata={"annotation_id": annotation.id, "app_id": app.id, "doc_id": annotation.id}, |
| ) |
| documents.append(document) |
|
|
| vector = Vector(dataset, attributes=["doc_id", "annotation_id", "app_id"]) |
| click.echo(f"Migrating annotations for app: {app.id}.") |
|
|
| try: |
| vector.delete() |
| click.echo(click.style(f"Deleted vector index for app {app.id}.", fg="green")) |
| except Exception as e: |
| click.echo(click.style(f"Failed to delete vector index for app {app.id}.", fg="red")) |
| raise e |
| if documents: |
| try: |
| click.echo( |
| click.style( |
| f"Creating vector index with {len(documents)} annotations for app {app.id}.", |
| fg="green", |
| ) |
| ) |
| vector.create(documents) |
| click.echo(click.style(f"Created vector index for app {app.id}.", fg="green")) |
| except Exception as e: |
| click.echo(click.style(f"Failed to created vector index for app {app.id}.", fg="red")) |
| raise e |
| click.echo(f"Successfully migrated app annotation {app.id}.") |
| create_count += 1 |
| except Exception as e: |
| click.echo( |
| click.style( |
| "Error creating app annotation index: {} {}".format(e.__class__.__name__, str(e)), fg="red" |
| ) |
| ) |
| continue |
|
|
| click.echo( |
| click.style( |
| f"Migration complete. Created {create_count} app annotation indexes. Skipped {skipped_count} apps.", |
| fg="green", |
| ) |
| ) |
|
|
|
|
| def migrate_knowledge_vector_database(): |
| """ |
| Migrate vector database datas to target vector database . |
| """ |
| click.echo(click.style("Starting vector database migration.", fg="green")) |
| create_count = 0 |
| skipped_count = 0 |
| total_count = 0 |
| vector_type = dify_config.VECTOR_STORE |
| upper_colletion_vector_types = { |
| VectorType.MILVUS, |
| VectorType.PGVECTOR, |
| VectorType.RELYT, |
| VectorType.WEAVIATE, |
| VectorType.ORACLE, |
| VectorType.ELASTICSEARCH, |
| } |
| lower_colletion_vector_types = { |
| VectorType.ANALYTICDB, |
| VectorType.CHROMA, |
| VectorType.MYSCALE, |
| VectorType.PGVECTO_RS, |
| VectorType.TIDB_VECTOR, |
| VectorType.OPENSEARCH, |
| VectorType.TENCENT, |
| VectorType.BAIDU, |
| VectorType.VIKINGDB, |
| VectorType.UPSTASH, |
| VectorType.COUCHBASE, |
| VectorType.OCEANBASE, |
| } |
| page = 1 |
| while True: |
| try: |
| datasets = ( |
| db.session.query(Dataset) |
| .filter(Dataset.indexing_technique == "high_quality") |
| .order_by(Dataset.created_at.desc()) |
| .paginate(page=page, per_page=50) |
| ) |
| except NotFound: |
| break |
|
|
| page += 1 |
| for dataset in datasets: |
| total_count = total_count + 1 |
| click.echo( |
| f"Processing the {total_count} dataset {dataset.id}. {create_count} created, {skipped_count} skipped." |
| ) |
| try: |
| click.echo("Creating dataset vector database index: {}".format(dataset.id)) |
| if dataset.index_struct_dict: |
| if dataset.index_struct_dict["type"] == vector_type: |
| skipped_count = skipped_count + 1 |
| continue |
| collection_name = "" |
| dataset_id = dataset.id |
| if vector_type in upper_colletion_vector_types: |
| collection_name = Dataset.gen_collection_name_by_id(dataset_id) |
| elif vector_type == VectorType.QDRANT: |
| if dataset.collection_binding_id: |
| dataset_collection_binding = ( |
| db.session.query(DatasetCollectionBinding) |
| .filter(DatasetCollectionBinding.id == dataset.collection_binding_id) |
| .one_or_none() |
| ) |
| if dataset_collection_binding: |
| collection_name = dataset_collection_binding.collection_name |
| else: |
| raise ValueError("Dataset Collection Binding not found") |
| else: |
| collection_name = Dataset.gen_collection_name_by_id(dataset_id) |
|
|
| elif vector_type in lower_colletion_vector_types: |
| collection_name = Dataset.gen_collection_name_by_id(dataset_id).lower() |
| else: |
| raise ValueError(f"Vector store {vector_type} is not supported.") |
|
|
| index_struct_dict = {"type": vector_type, "vector_store": {"class_prefix": collection_name}} |
| dataset.index_struct = json.dumps(index_struct_dict) |
| vector = Vector(dataset) |
| click.echo(f"Migrating dataset {dataset.id}.") |
|
|
| try: |
| vector.delete() |
| click.echo( |
| click.style(f"Deleted vector index {collection_name} for dataset {dataset.id}.", fg="green") |
| ) |
| except Exception as e: |
| click.echo( |
| click.style( |
| f"Failed to delete vector index {collection_name} for dataset {dataset.id}.", fg="red" |
| ) |
| ) |
| raise e |
|
|
| dataset_documents = ( |
| db.session.query(DatasetDocument) |
| .filter( |
| DatasetDocument.dataset_id == dataset.id, |
| DatasetDocument.indexing_status == "completed", |
| DatasetDocument.enabled == True, |
| DatasetDocument.archived == False, |
| ) |
| .all() |
| ) |
|
|
| documents = [] |
| segments_count = 0 |
| for dataset_document in dataset_documents: |
| segments = ( |
| db.session.query(DocumentSegment) |
| .filter( |
| DocumentSegment.document_id == dataset_document.id, |
| DocumentSegment.status == "completed", |
| DocumentSegment.enabled == True, |
| ) |
| .all() |
| ) |
|
|
| for segment in segments: |
| document = Document( |
| page_content=segment.content, |
| metadata={ |
| "doc_id": segment.index_node_id, |
| "doc_hash": segment.index_node_hash, |
| "document_id": segment.document_id, |
| "dataset_id": segment.dataset_id, |
| }, |
| ) |
|
|
| documents.append(document) |
| segments_count = segments_count + 1 |
|
|
| if documents: |
| try: |
| click.echo( |
| click.style( |
| f"Creating vector index with {len(documents)} documents of {segments_count}" |
| f" segments for dataset {dataset.id}.", |
| fg="green", |
| ) |
| ) |
| vector.create(documents) |
| click.echo(click.style(f"Created vector index for dataset {dataset.id}.", fg="green")) |
| except Exception as e: |
| click.echo(click.style(f"Failed to created vector index for dataset {dataset.id}.", fg="red")) |
| raise e |
| db.session.add(dataset) |
| db.session.commit() |
| click.echo(f"Successfully migrated dataset {dataset.id}.") |
| create_count += 1 |
| except Exception as e: |
| db.session.rollback() |
| click.echo( |
| click.style("Error creating dataset index: {} {}".format(e.__class__.__name__, str(e)), fg="red") |
| ) |
| continue |
|
|
| click.echo( |
| click.style( |
| f"Migration complete. Created {create_count} dataset indexes. Skipped {skipped_count} datasets.", fg="green" |
| ) |
| ) |
|
|
|
|
| @click.command("convert-to-agent-apps", help="Convert Agent Assistant to Agent App.") |
| def convert_to_agent_apps(): |
| """ |
| Convert Agent Assistant to Agent App. |
| """ |
| click.echo(click.style("Starting convert to agent apps.", fg="green")) |
|
|
| proceeded_app_ids = [] |
|
|
| while True: |
| |
| sql_query = """SELECT a.id AS id FROM apps a |
| INNER JOIN app_model_configs am ON a.app_model_config_id=am.id |
| WHERE a.mode = 'chat' |
| AND am.agent_mode is not null |
| AND ( |
| am.agent_mode like '%"strategy": "function_call"%' |
| OR am.agent_mode like '%"strategy": "react"%' |
| ) |
| AND ( |
| am.agent_mode like '{"enabled": true%' |
| OR am.agent_mode like '{"max_iteration": %' |
| ) ORDER BY a.created_at DESC LIMIT 1000 |
| """ |
|
|
| with db.engine.begin() as conn: |
| rs = conn.execute(db.text(sql_query)) |
|
|
| apps = [] |
| for i in rs: |
| app_id = str(i.id) |
| if app_id not in proceeded_app_ids: |
| proceeded_app_ids.append(app_id) |
| app = db.session.query(App).filter(App.id == app_id).first() |
| apps.append(app) |
|
|
| if len(apps) == 0: |
| break |
|
|
| for app in apps: |
| click.echo("Converting app: {}".format(app.id)) |
|
|
| try: |
| app.mode = AppMode.AGENT_CHAT.value |
| db.session.commit() |
|
|
| |
| db.session.query(Conversation).filter(Conversation.app_id == app.id).update( |
| {Conversation.mode: AppMode.AGENT_CHAT.value} |
| ) |
|
|
| db.session.commit() |
| click.echo(click.style("Converted app: {}".format(app.id), fg="green")) |
| except Exception as e: |
| click.echo(click.style("Convert app error: {} {}".format(e.__class__.__name__, str(e)), fg="red")) |
|
|
| click.echo(click.style("Conversion complete. Converted {} agent apps.".format(len(proceeded_app_ids)), fg="green")) |
|
|
|
|
| @click.command("add-qdrant-doc-id-index", help="Add Qdrant doc_id index.") |
| @click.option("--field", default="metadata.doc_id", prompt=False, help="Index field , default is metadata.doc_id.") |
| def add_qdrant_doc_id_index(field: str): |
| click.echo(click.style("Starting Qdrant doc_id index creation.", fg="green")) |
| vector_type = dify_config.VECTOR_STORE |
| if vector_type != "qdrant": |
| click.echo(click.style("This command only supports Qdrant vector store.", fg="red")) |
| return |
| create_count = 0 |
|
|
| try: |
| bindings = db.session.query(DatasetCollectionBinding).all() |
| if not bindings: |
| click.echo(click.style("No dataset collection bindings found.", fg="red")) |
| return |
| import qdrant_client |
| from qdrant_client.http.exceptions import UnexpectedResponse |
| from qdrant_client.http.models import PayloadSchemaType |
|
|
| from core.rag.datasource.vdb.qdrant.qdrant_vector import QdrantConfig |
|
|
| for binding in bindings: |
| if dify_config.QDRANT_URL is None: |
| raise ValueError("Qdrant URL is required.") |
| qdrant_config = QdrantConfig( |
| endpoint=dify_config.QDRANT_URL, |
| api_key=dify_config.QDRANT_API_KEY, |
| root_path=current_app.root_path, |
| timeout=dify_config.QDRANT_CLIENT_TIMEOUT, |
| grpc_port=dify_config.QDRANT_GRPC_PORT, |
| prefer_grpc=dify_config.QDRANT_GRPC_ENABLED, |
| ) |
| try: |
| client = qdrant_client.QdrantClient(**qdrant_config.to_qdrant_params()) |
| |
| client.create_payload_index(binding.collection_name, field, field_schema=PayloadSchemaType.KEYWORD) |
| create_count += 1 |
| except UnexpectedResponse as e: |
| |
| if e.status_code == 404: |
| click.echo(click.style(f"Collection not found: {binding.collection_name}.", fg="red")) |
| continue |
| |
| else: |
| click.echo( |
| click.style( |
| f"Failed to create Qdrant index for collection: {binding.collection_name}.", fg="red" |
| ) |
| ) |
|
|
| except Exception as e: |
| click.echo(click.style("Failed to create Qdrant client.", fg="red")) |
|
|
| click.echo(click.style(f"Index creation complete. Created {create_count} collection indexes.", fg="green")) |
|
|
|
|
| @click.command("create-tenant", help="Create account and tenant.") |
| @click.option("--email", prompt=True, help="Tenant account email.") |
| @click.option("--name", prompt=True, help="Workspace name.") |
| @click.option("--language", prompt=True, help="Account language, default: en-US.") |
| def create_tenant(email: str, language: Optional[str] = None, name: Optional[str] = None): |
| """ |
| Create tenant account |
| """ |
| if not email: |
| click.echo(click.style("Email is required.", fg="red")) |
| return |
|
|
| |
| email = email.strip() |
|
|
| if "@" not in email: |
| click.echo(click.style("Invalid email address.", fg="red")) |
| return |
|
|
| account_name = email.split("@")[0] |
|
|
| if language not in languages: |
| language = "en-US" |
|
|
| name = name.strip() |
|
|
| |
| new_password = secrets.token_urlsafe(16) |
|
|
| |
| account = RegisterService.register(email=email, name=account_name, password=new_password, language=language) |
|
|
| TenantService.create_owner_tenant_if_not_exist(account, name) |
|
|
| click.echo( |
| click.style( |
| "Account and tenant created.\nAccount: {}\nPassword: {}".format(email, new_password), |
| fg="green", |
| ) |
| ) |
|
|
|
|
| @click.command("upgrade-db", help="Upgrade the database") |
| def upgrade_db(): |
| click.echo("Preparing database migration...") |
| lock = redis_client.lock(name="db_upgrade_lock", timeout=60) |
| if lock.acquire(blocking=False): |
| try: |
| click.echo(click.style("Starting database migration.", fg="green")) |
|
|
| |
| import flask_migrate |
|
|
| flask_migrate.upgrade() |
|
|
| click.echo(click.style("Database migration successful!", fg="green")) |
|
|
| except Exception as e: |
| logging.exception(f"Database migration failed: {e}") |
| finally: |
| lock.release() |
| else: |
| click.echo("Database migration skipped") |
|
|
|
|
| @click.command("fix-app-site-missing", help="Fix app related site missing issue.") |
| def fix_app_site_missing(): |
| """ |
| Fix app related site missing issue. |
| """ |
| click.echo(click.style("Starting fix for missing app-related sites.", fg="green")) |
|
|
| failed_app_ids = [] |
| while True: |
| sql = """select apps.id as id from apps left join sites on sites.app_id=apps.id |
| where sites.id is null limit 1000""" |
| with db.engine.begin() as conn: |
| rs = conn.execute(db.text(sql)) |
|
|
| processed_count = 0 |
| for i in rs: |
| processed_count += 1 |
| app_id = str(i.id) |
|
|
| if app_id in failed_app_ids: |
| continue |
|
|
| try: |
| app = db.session.query(App).filter(App.id == app_id).first() |
| tenant = app.tenant |
| if tenant: |
| accounts = tenant.get_accounts() |
| if not accounts: |
| print("Fix failed for app {}".format(app.id)) |
| continue |
|
|
| account = accounts[0] |
| print("Fixing missing site for app {}".format(app.id)) |
| app_was_created.send(app, account=account) |
| except Exception as e: |
| failed_app_ids.append(app_id) |
| click.echo(click.style("Failed to fix missing site for app {}".format(app_id), fg="red")) |
| logging.exception(f"Fix app related site missing issue failed, error: {e}") |
| continue |
|
|
| if not processed_count: |
| break |
|
|
| click.echo(click.style("Fix for missing app-related sites completed successfully!", fg="green")) |
|
|
|
|
| def register_commands(app): |
| app.cli.add_command(reset_password) |
| app.cli.add_command(reset_email) |
| app.cli.add_command(reset_encrypt_key_pair) |
| app.cli.add_command(vdb_migrate) |
| app.cli.add_command(convert_to_agent_apps) |
| app.cli.add_command(add_qdrant_doc_id_index) |
| app.cli.add_command(create_tenant) |
| app.cli.add_command(upgrade_db) |
| app.cli.add_command(fix_app_site_missing) |
|
|