|
import base64 |
|
import json |
|
import logging |
|
import secrets |
|
from typing import Optional |
|
|
|
import click |
|
from flask import current_app |
|
from werkzeug.exceptions import NotFound |
|
|
|
from configs import dify_config |
|
from constants.languages import languages |
|
from core.rag.datasource.vdb.vector_factory import Vector |
|
from core.rag.datasource.vdb.vector_type import VectorType |
|
from core.rag.models.document import Document |
|
from events.app_event import app_was_created |
|
from extensions.ext_database import db |
|
from extensions.ext_redis import redis_client |
|
from libs.helper import email as email_validate |
|
from libs.password import hash_password, password_pattern, valid_password |
|
from libs.rsa import generate_key_pair |
|
from models import Tenant |
|
from models.dataset import Dataset, DatasetCollectionBinding, DocumentSegment |
|
from models.dataset import Document as DatasetDocument |
|
from models.model import Account, App, AppAnnotationSetting, AppMode, Conversation, MessageAnnotation |
|
from models.provider import Provider, ProviderModel |
|
from services.account_service import RegisterService, TenantService |
|
|
|
|
|
@click.command("reset-password", help="Reset the account password.") |
|
@click.option("--email", prompt=True, help="Account email to reset password for") |
|
@click.option("--new-password", prompt=True, help="New password") |
|
@click.option("--password-confirm", prompt=True, help="Confirm new password") |
|
def reset_password(email, new_password, password_confirm): |
|
""" |
|
Reset password of owner account |
|
Only available in SELF_HOSTED mode |
|
""" |
|
if str(new_password).strip() != str(password_confirm).strip(): |
|
click.echo(click.style("Passwords do not match.", fg="red")) |
|
return |
|
|
|
account = db.session.query(Account).filter(Account.email == email).one_or_none() |
|
|
|
if not account: |
|
click.echo(click.style("Account not found for email: {}".format(email), fg="red")) |
|
return |
|
|
|
try: |
|
valid_password(new_password) |
|
except: |
|
click.echo(click.style("Invalid password. Must match {}".format(password_pattern), fg="red")) |
|
return |
|
|
|
|
|
salt = secrets.token_bytes(16) |
|
base64_salt = base64.b64encode(salt).decode() |
|
|
|
|
|
password_hashed = hash_password(new_password, salt) |
|
base64_password_hashed = base64.b64encode(password_hashed).decode() |
|
account.password = base64_password_hashed |
|
account.password_salt = base64_salt |
|
db.session.commit() |
|
click.echo(click.style("Password reset successfully.", fg="green")) |
|
|
|
|
|
@click.command("reset-email", help="Reset the account email.") |
|
@click.option("--email", prompt=True, help="Current account email") |
|
@click.option("--new-email", prompt=True, help="New email") |
|
@click.option("--email-confirm", prompt=True, help="Confirm new email") |
|
def reset_email(email, new_email, email_confirm): |
|
""" |
|
Replace account email |
|
:return: |
|
""" |
|
if str(new_email).strip() != str(email_confirm).strip(): |
|
click.echo(click.style("New emails do not match.", fg="red")) |
|
return |
|
|
|
account = db.session.query(Account).filter(Account.email == email).one_or_none() |
|
|
|
if not account: |
|
click.echo(click.style("Account not found for email: {}".format(email), fg="red")) |
|
return |
|
|
|
try: |
|
email_validate(new_email) |
|
except: |
|
click.echo(click.style("Invalid email: {}".format(new_email), fg="red")) |
|
return |
|
|
|
account.email = new_email |
|
db.session.commit() |
|
click.echo(click.style("Email updated successfully.", fg="green")) |
|
|
|
|
|
@click.command( |
|
"reset-encrypt-key-pair", |
|
help="Reset the asymmetric key pair of workspace for encrypt LLM credentials. " |
|
"After the reset, all LLM credentials will become invalid, " |
|
"requiring re-entry." |
|
"Only support SELF_HOSTED mode.", |
|
) |
|
@click.confirmation_option( |
|
prompt=click.style( |
|
"Are you sure you want to reset encrypt key pair? This operation cannot be rolled back!", fg="red" |
|
) |
|
) |
|
def reset_encrypt_key_pair(): |
|
""" |
|
Reset the encrypted key pair of workspace for encrypt LLM credentials. |
|
After the reset, all LLM credentials will become invalid, requiring re-entry. |
|
Only support SELF_HOSTED mode. |
|
""" |
|
if dify_config.EDITION != "SELF_HOSTED": |
|
click.echo(click.style("This command is only for SELF_HOSTED installations.", fg="red")) |
|
return |
|
|
|
tenants = db.session.query(Tenant).all() |
|
for tenant in tenants: |
|
if not tenant: |
|
click.echo(click.style("No workspaces found. Run /install first.", fg="red")) |
|
return |
|
|
|
tenant.encrypt_public_key = generate_key_pair(tenant.id) |
|
|
|
db.session.query(Provider).filter(Provider.provider_type == "custom", Provider.tenant_id == tenant.id).delete() |
|
db.session.query(ProviderModel).filter(ProviderModel.tenant_id == tenant.id).delete() |
|
db.session.commit() |
|
|
|
click.echo( |
|
click.style( |
|
"Congratulations! The asymmetric key pair of workspace {} has been reset.".format(tenant.id), |
|
fg="green", |
|
) |
|
) |
|
|
|
|
|
@click.command("vdb-migrate", help="Migrate vector db.") |
|
@click.option("--scope", default="all", prompt=False, help="The scope of vector database to migrate, Default is All.") |
|
def vdb_migrate(scope: str): |
|
if scope in {"knowledge", "all"}: |
|
migrate_knowledge_vector_database() |
|
if scope in {"annotation", "all"}: |
|
migrate_annotation_vector_database() |
|
|
|
|
|
def migrate_annotation_vector_database(): |
|
""" |
|
Migrate annotation datas to target vector database . |
|
""" |
|
click.echo(click.style("Starting annotation data migration.", fg="green")) |
|
create_count = 0 |
|
skipped_count = 0 |
|
total_count = 0 |
|
page = 1 |
|
while True: |
|
try: |
|
|
|
apps = ( |
|
db.session.query(App) |
|
.filter(App.status == "normal") |
|
.order_by(App.created_at.desc()) |
|
.paginate(page=page, per_page=50) |
|
) |
|
except NotFound: |
|
break |
|
|
|
page += 1 |
|
for app in apps: |
|
total_count = total_count + 1 |
|
click.echo( |
|
f"Processing the {total_count} app {app.id}. " + f"{create_count} created, {skipped_count} skipped." |
|
) |
|
try: |
|
click.echo("Creating app annotation index: {}".format(app.id)) |
|
app_annotation_setting = ( |
|
db.session.query(AppAnnotationSetting).filter(AppAnnotationSetting.app_id == app.id).first() |
|
) |
|
|
|
if not app_annotation_setting: |
|
skipped_count = skipped_count + 1 |
|
click.echo("App annotation setting disabled: {}".format(app.id)) |
|
continue |
|
|
|
dataset_collection_binding = ( |
|
db.session.query(DatasetCollectionBinding) |
|
.filter(DatasetCollectionBinding.id == app_annotation_setting.collection_binding_id) |
|
.first() |
|
) |
|
if not dataset_collection_binding: |
|
click.echo("App annotation collection binding not found: {}".format(app.id)) |
|
continue |
|
annotations = db.session.query(MessageAnnotation).filter(MessageAnnotation.app_id == app.id).all() |
|
dataset = Dataset( |
|
id=app.id, |
|
tenant_id=app.tenant_id, |
|
indexing_technique="high_quality", |
|
embedding_model_provider=dataset_collection_binding.provider_name, |
|
embedding_model=dataset_collection_binding.model_name, |
|
collection_binding_id=dataset_collection_binding.id, |
|
) |
|
documents = [] |
|
if annotations: |
|
for annotation in annotations: |
|
document = Document( |
|
page_content=annotation.question, |
|
metadata={"annotation_id": annotation.id, "app_id": app.id, "doc_id": annotation.id}, |
|
) |
|
documents.append(document) |
|
|
|
vector = Vector(dataset, attributes=["doc_id", "annotation_id", "app_id"]) |
|
click.echo(f"Migrating annotations for app: {app.id}.") |
|
|
|
try: |
|
vector.delete() |
|
click.echo(click.style(f"Deleted vector index for app {app.id}.", fg="green")) |
|
except Exception as e: |
|
click.echo(click.style(f"Failed to delete vector index for app {app.id}.", fg="red")) |
|
raise e |
|
if documents: |
|
try: |
|
click.echo( |
|
click.style( |
|
f"Creating vector index with {len(documents)} annotations for app {app.id}.", |
|
fg="green", |
|
) |
|
) |
|
vector.create(documents) |
|
click.echo(click.style(f"Created vector index for app {app.id}.", fg="green")) |
|
except Exception as e: |
|
click.echo(click.style(f"Failed to created vector index for app {app.id}.", fg="red")) |
|
raise e |
|
click.echo(f"Successfully migrated app annotation {app.id}.") |
|
create_count += 1 |
|
except Exception as e: |
|
click.echo( |
|
click.style( |
|
"Error creating app annotation index: {} {}".format(e.__class__.__name__, str(e)), fg="red" |
|
) |
|
) |
|
continue |
|
|
|
click.echo( |
|
click.style( |
|
f"Migration complete. Created {create_count} app annotation indexes. Skipped {skipped_count} apps.", |
|
fg="green", |
|
) |
|
) |
|
|
|
|
|
def migrate_knowledge_vector_database(): |
|
""" |
|
Migrate vector database datas to target vector database . |
|
""" |
|
click.echo(click.style("Starting vector database migration.", fg="green")) |
|
create_count = 0 |
|
skipped_count = 0 |
|
total_count = 0 |
|
vector_type = dify_config.VECTOR_STORE |
|
upper_colletion_vector_types = { |
|
VectorType.MILVUS, |
|
VectorType.PGVECTOR, |
|
VectorType.RELYT, |
|
VectorType.WEAVIATE, |
|
VectorType.ORACLE, |
|
VectorType.ELASTICSEARCH, |
|
} |
|
lower_colletion_vector_types = { |
|
VectorType.ANALYTICDB, |
|
VectorType.CHROMA, |
|
VectorType.MYSCALE, |
|
VectorType.PGVECTO_RS, |
|
VectorType.TIDB_VECTOR, |
|
VectorType.OPENSEARCH, |
|
VectorType.TENCENT, |
|
VectorType.BAIDU, |
|
VectorType.VIKINGDB, |
|
VectorType.UPSTASH, |
|
VectorType.COUCHBASE, |
|
VectorType.OCEANBASE, |
|
} |
|
page = 1 |
|
while True: |
|
try: |
|
datasets = ( |
|
db.session.query(Dataset) |
|
.filter(Dataset.indexing_technique == "high_quality") |
|
.order_by(Dataset.created_at.desc()) |
|
.paginate(page=page, per_page=50) |
|
) |
|
except NotFound: |
|
break |
|
|
|
page += 1 |
|
for dataset in datasets: |
|
total_count = total_count + 1 |
|
click.echo( |
|
f"Processing the {total_count} dataset {dataset.id}. {create_count} created, {skipped_count} skipped." |
|
) |
|
try: |
|
click.echo("Creating dataset vector database index: {}".format(dataset.id)) |
|
if dataset.index_struct_dict: |
|
if dataset.index_struct_dict["type"] == vector_type: |
|
skipped_count = skipped_count + 1 |
|
continue |
|
collection_name = "" |
|
dataset_id = dataset.id |
|
if vector_type in upper_colletion_vector_types: |
|
collection_name = Dataset.gen_collection_name_by_id(dataset_id) |
|
elif vector_type == VectorType.QDRANT: |
|
if dataset.collection_binding_id: |
|
dataset_collection_binding = ( |
|
db.session.query(DatasetCollectionBinding) |
|
.filter(DatasetCollectionBinding.id == dataset.collection_binding_id) |
|
.one_or_none() |
|
) |
|
if dataset_collection_binding: |
|
collection_name = dataset_collection_binding.collection_name |
|
else: |
|
raise ValueError("Dataset Collection Binding not found") |
|
else: |
|
collection_name = Dataset.gen_collection_name_by_id(dataset_id) |
|
|
|
elif vector_type in lower_colletion_vector_types: |
|
collection_name = Dataset.gen_collection_name_by_id(dataset_id).lower() |
|
else: |
|
raise ValueError(f"Vector store {vector_type} is not supported.") |
|
|
|
index_struct_dict = {"type": vector_type, "vector_store": {"class_prefix": collection_name}} |
|
dataset.index_struct = json.dumps(index_struct_dict) |
|
vector = Vector(dataset) |
|
click.echo(f"Migrating dataset {dataset.id}.") |
|
|
|
try: |
|
vector.delete() |
|
click.echo( |
|
click.style(f"Deleted vector index {collection_name} for dataset {dataset.id}.", fg="green") |
|
) |
|
except Exception as e: |
|
click.echo( |
|
click.style( |
|
f"Failed to delete vector index {collection_name} for dataset {dataset.id}.", fg="red" |
|
) |
|
) |
|
raise e |
|
|
|
dataset_documents = ( |
|
db.session.query(DatasetDocument) |
|
.filter( |
|
DatasetDocument.dataset_id == dataset.id, |
|
DatasetDocument.indexing_status == "completed", |
|
DatasetDocument.enabled == True, |
|
DatasetDocument.archived == False, |
|
) |
|
.all() |
|
) |
|
|
|
documents = [] |
|
segments_count = 0 |
|
for dataset_document in dataset_documents: |
|
segments = ( |
|
db.session.query(DocumentSegment) |
|
.filter( |
|
DocumentSegment.document_id == dataset_document.id, |
|
DocumentSegment.status == "completed", |
|
DocumentSegment.enabled == True, |
|
) |
|
.all() |
|
) |
|
|
|
for segment in segments: |
|
document = Document( |
|
page_content=segment.content, |
|
metadata={ |
|
"doc_id": segment.index_node_id, |
|
"doc_hash": segment.index_node_hash, |
|
"document_id": segment.document_id, |
|
"dataset_id": segment.dataset_id, |
|
}, |
|
) |
|
|
|
documents.append(document) |
|
segments_count = segments_count + 1 |
|
|
|
if documents: |
|
try: |
|
click.echo( |
|
click.style( |
|
f"Creating vector index with {len(documents)} documents of {segments_count}" |
|
f" segments for dataset {dataset.id}.", |
|
fg="green", |
|
) |
|
) |
|
vector.create(documents) |
|
click.echo(click.style(f"Created vector index for dataset {dataset.id}.", fg="green")) |
|
except Exception as e: |
|
click.echo(click.style(f"Failed to created vector index for dataset {dataset.id}.", fg="red")) |
|
raise e |
|
db.session.add(dataset) |
|
db.session.commit() |
|
click.echo(f"Successfully migrated dataset {dataset.id}.") |
|
create_count += 1 |
|
except Exception as e: |
|
db.session.rollback() |
|
click.echo( |
|
click.style("Error creating dataset index: {} {}".format(e.__class__.__name__, str(e)), fg="red") |
|
) |
|
continue |
|
|
|
click.echo( |
|
click.style( |
|
f"Migration complete. Created {create_count} dataset indexes. Skipped {skipped_count} datasets.", fg="green" |
|
) |
|
) |
|
|
|
|
|
@click.command("convert-to-agent-apps", help="Convert Agent Assistant to Agent App.") |
|
def convert_to_agent_apps(): |
|
""" |
|
Convert Agent Assistant to Agent App. |
|
""" |
|
click.echo(click.style("Starting convert to agent apps.", fg="green")) |
|
|
|
proceeded_app_ids = [] |
|
|
|
while True: |
|
|
|
sql_query = """SELECT a.id AS id FROM apps a |
|
INNER JOIN app_model_configs am ON a.app_model_config_id=am.id |
|
WHERE a.mode = 'chat' |
|
AND am.agent_mode is not null |
|
AND ( |
|
am.agent_mode like '%"strategy": "function_call"%' |
|
OR am.agent_mode like '%"strategy": "react"%' |
|
) |
|
AND ( |
|
am.agent_mode like '{"enabled": true%' |
|
OR am.agent_mode like '{"max_iteration": %' |
|
) ORDER BY a.created_at DESC LIMIT 1000 |
|
""" |
|
|
|
with db.engine.begin() as conn: |
|
rs = conn.execute(db.text(sql_query)) |
|
|
|
apps = [] |
|
for i in rs: |
|
app_id = str(i.id) |
|
if app_id not in proceeded_app_ids: |
|
proceeded_app_ids.append(app_id) |
|
app = db.session.query(App).filter(App.id == app_id).first() |
|
apps.append(app) |
|
|
|
if len(apps) == 0: |
|
break |
|
|
|
for app in apps: |
|
click.echo("Converting app: {}".format(app.id)) |
|
|
|
try: |
|
app.mode = AppMode.AGENT_CHAT.value |
|
db.session.commit() |
|
|
|
|
|
db.session.query(Conversation).filter(Conversation.app_id == app.id).update( |
|
{Conversation.mode: AppMode.AGENT_CHAT.value} |
|
) |
|
|
|
db.session.commit() |
|
click.echo(click.style("Converted app: {}".format(app.id), fg="green")) |
|
except Exception as e: |
|
click.echo(click.style("Convert app error: {} {}".format(e.__class__.__name__, str(e)), fg="red")) |
|
|
|
click.echo(click.style("Conversion complete. Converted {} agent apps.".format(len(proceeded_app_ids)), fg="green")) |
|
|
|
|
|
@click.command("add-qdrant-doc-id-index", help="Add Qdrant doc_id index.") |
|
@click.option("--field", default="metadata.doc_id", prompt=False, help="Index field , default is metadata.doc_id.") |
|
def add_qdrant_doc_id_index(field: str): |
|
click.echo(click.style("Starting Qdrant doc_id index creation.", fg="green")) |
|
vector_type = dify_config.VECTOR_STORE |
|
if vector_type != "qdrant": |
|
click.echo(click.style("This command only supports Qdrant vector store.", fg="red")) |
|
return |
|
create_count = 0 |
|
|
|
try: |
|
bindings = db.session.query(DatasetCollectionBinding).all() |
|
if not bindings: |
|
click.echo(click.style("No dataset collection bindings found.", fg="red")) |
|
return |
|
import qdrant_client |
|
from qdrant_client.http.exceptions import UnexpectedResponse |
|
from qdrant_client.http.models import PayloadSchemaType |
|
|
|
from core.rag.datasource.vdb.qdrant.qdrant_vector import QdrantConfig |
|
|
|
for binding in bindings: |
|
if dify_config.QDRANT_URL is None: |
|
raise ValueError("Qdrant URL is required.") |
|
qdrant_config = QdrantConfig( |
|
endpoint=dify_config.QDRANT_URL, |
|
api_key=dify_config.QDRANT_API_KEY, |
|
root_path=current_app.root_path, |
|
timeout=dify_config.QDRANT_CLIENT_TIMEOUT, |
|
grpc_port=dify_config.QDRANT_GRPC_PORT, |
|
prefer_grpc=dify_config.QDRANT_GRPC_ENABLED, |
|
) |
|
try: |
|
client = qdrant_client.QdrantClient(**qdrant_config.to_qdrant_params()) |
|
|
|
client.create_payload_index(binding.collection_name, field, field_schema=PayloadSchemaType.KEYWORD) |
|
create_count += 1 |
|
except UnexpectedResponse as e: |
|
|
|
if e.status_code == 404: |
|
click.echo(click.style(f"Collection not found: {binding.collection_name}.", fg="red")) |
|
continue |
|
|
|
else: |
|
click.echo( |
|
click.style( |
|
f"Failed to create Qdrant index for collection: {binding.collection_name}.", fg="red" |
|
) |
|
) |
|
|
|
except Exception as e: |
|
click.echo(click.style("Failed to create Qdrant client.", fg="red")) |
|
|
|
click.echo(click.style(f"Index creation complete. Created {create_count} collection indexes.", fg="green")) |
|
|
|
|
|
@click.command("create-tenant", help="Create account and tenant.") |
|
@click.option("--email", prompt=True, help="Tenant account email.") |
|
@click.option("--name", prompt=True, help="Workspace name.") |
|
@click.option("--language", prompt=True, help="Account language, default: en-US.") |
|
def create_tenant(email: str, language: Optional[str] = None, name: Optional[str] = None): |
|
""" |
|
Create tenant account |
|
""" |
|
if not email: |
|
click.echo(click.style("Email is required.", fg="red")) |
|
return |
|
|
|
|
|
email = email.strip() |
|
|
|
if "@" not in email: |
|
click.echo(click.style("Invalid email address.", fg="red")) |
|
return |
|
|
|
account_name = email.split("@")[0] |
|
|
|
if language not in languages: |
|
language = "en-US" |
|
|
|
name = name.strip() |
|
|
|
|
|
new_password = secrets.token_urlsafe(16) |
|
|
|
|
|
account = RegisterService.register(email=email, name=account_name, password=new_password, language=language) |
|
|
|
TenantService.create_owner_tenant_if_not_exist(account, name) |
|
|
|
click.echo( |
|
click.style( |
|
"Account and tenant created.\nAccount: {}\nPassword: {}".format(email, new_password), |
|
fg="green", |
|
) |
|
) |
|
|
|
|
|
@click.command("upgrade-db", help="Upgrade the database") |
|
def upgrade_db(): |
|
click.echo("Preparing database migration...") |
|
lock = redis_client.lock(name="db_upgrade_lock", timeout=60) |
|
if lock.acquire(blocking=False): |
|
try: |
|
click.echo(click.style("Starting database migration.", fg="green")) |
|
|
|
|
|
import flask_migrate |
|
|
|
flask_migrate.upgrade() |
|
|
|
click.echo(click.style("Database migration successful!", fg="green")) |
|
|
|
except Exception as e: |
|
logging.exception(f"Database migration failed: {e}") |
|
finally: |
|
lock.release() |
|
else: |
|
click.echo("Database migration skipped") |
|
|
|
|
|
@click.command("fix-app-site-missing", help="Fix app related site missing issue.") |
|
def fix_app_site_missing(): |
|
""" |
|
Fix app related site missing issue. |
|
""" |
|
click.echo(click.style("Starting fix for missing app-related sites.", fg="green")) |
|
|
|
failed_app_ids = [] |
|
while True: |
|
sql = """select apps.id as id from apps left join sites on sites.app_id=apps.id |
|
where sites.id is null limit 1000""" |
|
with db.engine.begin() as conn: |
|
rs = conn.execute(db.text(sql)) |
|
|
|
processed_count = 0 |
|
for i in rs: |
|
processed_count += 1 |
|
app_id = str(i.id) |
|
|
|
if app_id in failed_app_ids: |
|
continue |
|
|
|
try: |
|
app = db.session.query(App).filter(App.id == app_id).first() |
|
tenant = app.tenant |
|
if tenant: |
|
accounts = tenant.get_accounts() |
|
if not accounts: |
|
print("Fix failed for app {}".format(app.id)) |
|
continue |
|
|
|
account = accounts[0] |
|
print("Fixing missing site for app {}".format(app.id)) |
|
app_was_created.send(app, account=account) |
|
except Exception as e: |
|
failed_app_ids.append(app_id) |
|
click.echo(click.style("Failed to fix missing site for app {}".format(app_id), fg="red")) |
|
logging.exception(f"Fix app related site missing issue failed, error: {e}") |
|
continue |
|
|
|
if not processed_count: |
|
break |
|
|
|
click.echo(click.style("Fix for missing app-related sites completed successfully!", fg="green")) |
|
|
|
|
|
def register_commands(app): |
|
app.cli.add_command(reset_password) |
|
app.cli.add_command(reset_email) |
|
app.cli.add_command(reset_encrypt_key_pair) |
|
app.cli.add_command(vdb_migrate) |
|
app.cli.add_command(convert_to_agent_apps) |
|
app.cli.add_command(add_qdrant_doc_id_index) |
|
app.cli.add_command(create_tenant) |
|
app.cli.add_command(upgrade_db) |
|
app.cli.add_command(fix_app_site_missing) |
|
|