id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
16,901
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('tool_api_providers', sa.Column('id', postgresql.UUID(), server_default=sa.text('uuid_generate_v4()'), nullable=False), sa.Column('name', sa.String(length=40), nullable=False), sa.Column('schema', sa.Text(), nullable=False), sa.Column('schema_type_str', sa.String(length=40), nullable=False), sa.Column('user_id', postgresql.UUID(), nullable=False), sa.Column('tenant_id', postgresql.UUID(), nullable=False), sa.Column('description_str', sa.Text(), nullable=False), sa.Column('tools_str', sa.Text(), nullable=False), sa.PrimaryKeyConstraint('id', name='tool_api_provider_pkey') ) op.create_table('tool_builtin_providers', sa.Column('id', postgresql.UUID(), server_default=sa.text('uuid_generate_v4()'), nullable=False), sa.Column('tenant_id', postgresql.UUID(), nullable=True), sa.Column('user_id', postgresql.UUID(), nullable=False), sa.Column('provider', sa.String(length=40), nullable=False), sa.Column('encrypted_credentials', sa.Text(), nullable=True), sa.Column('created_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.Column('updated_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.PrimaryKeyConstraint('id', name='tool_builtin_provider_pkey'), sa.UniqueConstraint('tenant_id', 'provider', name='unique_builtin_tool_provider') ) op.create_table('tool_published_apps', sa.Column('id', postgresql.UUID(), server_default=sa.text('uuid_generate_v4()'), nullable=False), sa.Column('app_id', postgresql.UUID(), nullable=False), sa.Column('user_id', postgresql.UUID(), nullable=False), sa.Column('description', sa.Text(), nullable=False), sa.Column('llm_description', sa.Text(), nullable=False), sa.Column('query_description', sa.Text(), nullable=False), sa.Column('query_name', sa.String(length=40), nullable=False), sa.Column('tool_name', sa.String(length=40), nullable=False), sa.Column('author', sa.String(length=40), nullable=False), sa.Column('created_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.Column('updated_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.ForeignKeyConstraint(['app_id'], ['apps.id'], ), sa.PrimaryKeyConstraint('id', name='published_app_tool_pkey'), sa.UniqueConstraint('app_id', 'user_id', name='unique_published_app_tool') ) # ### end Alembic commands ###
null
16,902
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_table('tool_published_apps') op.drop_table('tool_builtin_providers') op.drop_table('tool_api_providers') # ### end Alembic commands ###
null
16,903
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('document_segments', schema=None) as batch_op: batch_op.add_column(sa.Column('answer', sa.Text(), nullable=True)) batch_op.add_column(sa.Column('updated_by', postgresql.UUID(), nullable=True)) batch_op.add_column(sa.Column('updated_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False)) with op.batch_alter_table('documents', schema=None) as batch_op: batch_op.add_column(sa.Column('doc_form', sa.String(length=255), server_default=sa.text("'text_model'::character varying"), nullable=False)) # ### end Alembic commands ###
null
16,904
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('documents', schema=None) as batch_op: batch_op.drop_column('doc_form') with op.batch_alter_table('document_segments', schema=None) as batch_op: batch_op.drop_column('updated_at') batch_op.drop_column('updated_by') batch_op.drop_column('answer') # ### end Alembic commands ###
null
16,905
from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('message_agent_thoughts', schema=None) as batch_op: batch_op.alter_column('message_chain_id', existing_type=postgresql.UUID(), nullable=True) # ### end Alembic commands ###
null
16,906
from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('message_agent_thoughts', schema=None) as batch_op: batch_op.alter_column('message_chain_id', existing_type=postgresql.UUID(), nullable=False) # ### end Alembic commands ###
null
16,907
from json import dumps, loads import sqlalchemy as sa from alembic import op The provided code snippet includes necessary dependencies for implementing the `upgrade` function. Write a Python function `def upgrade()` to solve the following problem: 1. select all tool_providers 2. insert api_key to tool_provider_configs tool_providers - id - tenant_id - tool_name - encrypted_credentials {"api_key": "$KEY"} - created_at - updated_at tool_builtin_providers - id <- tool_providers.id - tenant_id <- tool_providers.tenant_id - user_id <- tenant_account_joins.account_id (tenant_account_joins.tenant_id = tool_providers.tenant_id and tenant_account_joins.role = 'owner') - encrypted_credentials <- tool_providers.encrypted_credentials {"serpapi_api_key": "$KEY"} - created_at <- tool_providers.created_at - updated_at <- tool_providers.updated_at Here is the function: def upgrade(): # ### commands auto generated by Alembic - please adjust! ### """ 1. select all tool_providers 2. insert api_key to tool_provider_configs tool_providers - id - tenant_id - tool_name - encrypted_credentials {"api_key": "$KEY"} - created_at - updated_at tool_builtin_providers - id <- tool_providers.id - tenant_id <- tool_providers.tenant_id - user_id <- tenant_account_joins.account_id (tenant_account_joins.tenant_id = tool_providers.tenant_id and tenant_account_joins.role = 'owner') - encrypted_credentials <- tool_providers.encrypted_credentials {"serpapi_api_key": "$KEY"} - created_at <- tool_providers.created_at - updated_at <- tool_providers.updated_at """ # select all tool_providers tool_providers = op.get_bind().execute( sa.text( "SELECT * FROM tool_providers WHERE tool_name = 'serpapi'" ) ).fetchall() # insert api_key to tool_provider_configs for tool_provider in tool_providers: id = tool_provider['id'] tenant_id = tool_provider['tenant_id'] encrypted_credentials = tool_provider['encrypted_credentials'] try: credentials = loads(encrypted_credentials) api_key = credentials['api_key'] credentials['serpapi_api_key'] = api_key credentials.pop('api_key') encrypted_credentials = dumps(credentials) except Exception as e: print(e) continue # get user_id user_id = op.get_bind().execute( sa.text( "SELECT account_id FROM tenant_account_joins WHERE tenant_id = :tenant_id AND role = 'owner'" ), tenant_id=tenant_id ).fetchone()['account_id'] created_at = tool_provider['created_at'] updated_at = tool_provider['updated_at'] # insert to tool_builtin_providers # check if exists exists = op.get_bind().execute( sa.text( "SELECT * FROM tool_builtin_providers WHERE tenant_id = :tenant_id AND provider = 'google'" ), tenant_id=tenant_id ).fetchone() if exists: continue op.get_bind().execute( sa.text( "INSERT INTO tool_builtin_providers (id, tenant_id, user_id, provider, encrypted_credentials, created_at, updated_at) VALUES (:id, :tenant_id, :user_id, :provider, :encrypted_credentials, :created_at, :updated_at)" ), id=id, tenant_id=tenant_id, user_id=user_id, provider='google', encrypted_credentials=encrypted_credentials, created_at=created_at, updated_at=updated_at ) # ### end Alembic commands ###
1. select all tool_providers 2. insert api_key to tool_provider_configs tool_providers - id - tenant_id - tool_name - encrypted_credentials {"api_key": "$KEY"} - created_at - updated_at tool_builtin_providers - id <- tool_providers.id - tenant_id <- tool_providers.tenant_id - user_id <- tenant_account_joins.account_id (tenant_account_joins.tenant_id = tool_providers.tenant_id and tenant_account_joins.role = 'owner') - encrypted_credentials <- tool_providers.encrypted_credentials {"serpapi_api_key": "$KEY"} - created_at <- tool_providers.created_at - updated_at <- tool_providers.updated_at
16,908
from json import dumps, loads import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### pass # ### end Alembic commands ###
null
16,909
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('dataset_retriever_resources', sa.Column('id', postgresql.UUID(), server_default=sa.text('uuid_generate_v4()'), nullable=False), sa.Column('message_id', postgresql.UUID(), nullable=False), sa.Column('position', sa.Integer(), nullable=False), sa.Column('dataset_id', postgresql.UUID(), nullable=False), sa.Column('dataset_name', sa.Text(), nullable=False), sa.Column('document_id', postgresql.UUID(), nullable=False), sa.Column('document_name', sa.Text(), nullable=False), sa.Column('data_source_type', sa.Text(), nullable=False), sa.Column('segment_id', postgresql.UUID(), nullable=False), sa.Column('score', sa.Float(), nullable=True), sa.Column('content', sa.Text(), nullable=False), sa.Column('hit_count', sa.Integer(), nullable=True), sa.Column('word_count', sa.Integer(), nullable=True), sa.Column('segment_position', sa.Integer(), nullable=True), sa.Column('index_node_hash', sa.Text(), nullable=True), sa.Column('retriever_from', sa.Text(), nullable=False), sa.Column('created_by', postgresql.UUID(), nullable=False), sa.Column('created_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP'), nullable=False), sa.PrimaryKeyConstraint('id', name='dataset_retriever_resource_pkey') ) with op.batch_alter_table('dataset_retriever_resources', schema=None) as batch_op: batch_op.create_index('dataset_retriever_resource_message_id_idx', ['message_id'], unique=False) # ### end Alembic commands ###
null
16,910
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('dataset_retriever_resources', schema=None) as batch_op: batch_op.drop_index('dataset_retriever_resource_message_id_idx') op.drop_table('dataset_retriever_resources') # ### end Alembic commands ###
null
16,911
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_annotation_hit_histories', schema=None) as batch_op: batch_op.add_column(sa.Column('annotation_question', sa.Text(), nullable=False)) batch_op.add_column(sa.Column('annotation_content', sa.Text(), nullable=False)) # ### end Alembic commands ###
null
16,912
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_annotation_hit_histories', schema=None) as batch_op: batch_op.drop_column('annotation_content') batch_op.drop_column('annotation_question') # ### end Alembic commands ###
null
16,913
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('apps', schema=None) as batch_op: batch_op.add_column(sa.Column('is_universal', sa.Boolean(), server_default=sa.text('false'), nullable=False)) # ### end Alembic commands ###
null
16,914
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('apps', schema=None) as batch_op: batch_op.drop_column('is_universal') # ### end Alembic commands ###
null
16,915
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('api_based_extensions', sa.Column('id', postgresql.UUID(), server_default=sa.text('uuid_generate_v4()'), nullable=False), sa.Column('tenant_id', postgresql.UUID(), nullable=False), sa.Column('name', sa.String(length=255), nullable=False), sa.Column('api_endpoint', sa.String(length=255), nullable=False), sa.Column('api_key', sa.Text(), nullable=False), sa.Column('created_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.PrimaryKeyConstraint('id', name='api_based_extension_pkey') ) with op.batch_alter_table('api_based_extensions', schema=None) as batch_op: batch_op.create_index('api_based_extension_tenant_idx', ['tenant_id'], unique=False) # ### end Alembic commands ###
null
16,916
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('api_based_extensions', schema=None) as batch_op: batch_op.drop_index('api_based_extension_tenant_idx') op.drop_table('api_based_extensions') # ### end Alembic commands ###
null
16,917
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.add_column(sa.Column('dataset_query_variable', sa.String(length=255), nullable=True)) # ### end Alembic commands ###
null
16,918
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.drop_column('dataset_query_variable') # ### end Alembic commands ###
null
16,919
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('api_tokens', schema=None) as batch_op: batch_op.add_column(sa.Column('tenant_id', postgresql.UUID(), nullable=True)) batch_op.create_index('api_token_tenant_idx', ['tenant_id', 'type'], unique=False) batch_op.drop_column('dataset_id') # ### end Alembic commands ###
null
16,920
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('api_tokens', schema=None) as batch_op: batch_op.add_column(sa.Column('dataset_id', postgresql.UUID(), autoincrement=False, nullable=True)) batch_op.drop_index('api_token_tenant_idx') batch_op.drop_column('tenant_id') # ### end Alembic commands ###
null
16,921
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('recommended_apps', schema=None) as batch_op: batch_op.add_column(sa.Column('language', sa.String(length=255), server_default=sa.text("'en-US'::character varying"), nullable=False)) batch_op.drop_index('recommended_app_is_listed_idx') batch_op.create_index('recommended_app_is_listed_idx', ['is_listed', 'language'], unique=False) # ### end Alembic commands ###
null
16,922
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('recommended_apps', schema=None) as batch_op: batch_op.drop_index('recommended_app_is_listed_idx') batch_op.create_index('recommended_app_is_listed_idx', ['is_listed'], unique=False) batch_op.drop_column('language') # ### end Alembic commands ###
null
16,923
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('message_files', schema=None) as batch_op: batch_op.add_column(sa.Column('belongs_to', sa.String(length=255), nullable=True)) # ### end Alembic commands ###
null
16,924
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('message_files', schema=None) as batch_op: batch_op.drop_column('belongs_to') # ### end Alembic commands ###
null
16,925
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tool_api_providers', schema=None) as batch_op: batch_op.add_column(sa.Column('credentials_str', sa.Text(), nullable=False)) # ### end Alembic commands ###
null
16,926
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tool_api_providers', schema=None) as batch_op: batch_op.drop_column('credentials_str') # ### end Alembic commands ###
null
16,927
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tool_api_providers', schema=None) as batch_op: batch_op.add_column(sa.Column('description', sa.Text(), nullable=False)) batch_op.drop_column('description_str') # ### end Alembic commands ###
null
16,928
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tool_api_providers', schema=None) as batch_op: batch_op.add_column(sa.Column('description_str', sa.TEXT(), autoincrement=False, nullable=False)) batch_op.drop_column('description') # ### end Alembic commands ###
null
16,929
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('provider_orders', sa.Column('id', postgresql.UUID(), server_default=sa.text('uuid_generate_v4()'), nullable=False), sa.Column('tenant_id', postgresql.UUID(), nullable=False), sa.Column('provider_name', sa.String(length=40), nullable=False), sa.Column('account_id', postgresql.UUID(), nullable=False), sa.Column('payment_product_id', sa.String(length=191), nullable=False), sa.Column('payment_id', sa.String(length=191), nullable=True), sa.Column('transaction_id', sa.String(length=191), nullable=True), sa.Column('quantity', sa.Integer(), server_default=sa.text('1'), nullable=False), sa.Column('currency', sa.String(length=40), nullable=True), sa.Column('total_amount', sa.Integer(), nullable=True), sa.Column('payment_status', sa.String(length=40), server_default=sa.text("'wait_pay'::character varying"), nullable=False), sa.Column('paid_at', sa.DateTime(), nullable=True), sa.Column('pay_failed_at', sa.DateTime(), nullable=True), sa.Column('refunded_at', sa.DateTime(), nullable=True), sa.Column('created_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.Column('updated_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.PrimaryKeyConstraint('id', name='provider_order_pkey') ) with op.batch_alter_table('provider_orders', schema=None) as batch_op: batch_op.create_index('provider_order_tenant_provider_idx', ['tenant_id', 'provider_name'], unique=False) # ### end Alembic commands ###
null
16,930
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('provider_orders', schema=None) as batch_op: batch_op.drop_index('provider_order_tenant_provider_idx') op.drop_table('provider_orders') # ### end Alembic commands ###
null
16,931
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('message_agent_thoughts', schema=None) as batch_op: batch_op.add_column(sa.Column('tool_labels_str', sa.Text(), server_default=sa.text("'{}'::text"), nullable=False)) # ### end Alembic commands ###
null
16,932
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('message_agent_thoughts', schema=None) as batch_op: batch_op.drop_column('tool_labels_str') # ### end Alembic commands ###
null
16,933
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_annotation_hit_histories', schema=None) as batch_op: batch_op.add_column(sa.Column('score', sa.Float(), server_default=sa.text('0'), nullable=False)) # ### end Alembic commands ###
null
16,934
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_annotation_hit_histories', schema=None) as batch_op: batch_op.drop_column('score') # ### end Alembic commands ###
null
16,935
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('message_agent_thoughts', schema=None) as batch_op: batch_op.add_column(sa.Column('message_price_unit', sa.Numeric(precision=10, scale=7), server_default=sa.text('0.001'), nullable=False)) batch_op.add_column(sa.Column('answer_price_unit', sa.Numeric(precision=10, scale=7), server_default=sa.text('0.001'), nullable=False)) with op.batch_alter_table('messages', schema=None) as batch_op: batch_op.add_column(sa.Column('message_price_unit', sa.Numeric(precision=10, scale=7), server_default=sa.text('0.001'), nullable=False)) batch_op.add_column(sa.Column('answer_price_unit', sa.Numeric(precision=10, scale=7), server_default=sa.text('0.001'), nullable=False)) # ### end Alembic commands ###
null
16,936
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('messages', schema=None) as batch_op: batch_op.drop_column('answer_price_unit') batch_op.drop_column('message_price_unit') with op.batch_alter_table('message_agent_thoughts', schema=None) as batch_op: batch_op.drop_column('answer_price_unit') batch_op.drop_column('message_price_unit') # ### end Alembic commands ###
null
16,937
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('documents', schema=None) as batch_op: batch_op.add_column(sa.Column('doc_language', sa.String(length=255), nullable=True)) # ### end Alembic commands ###
null
16,938
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('documents', schema=None) as batch_op: batch_op.drop_column('doc_language') # ### end Alembic commands ###
null
16,939
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.add_column(sa.Column('text_to_speech', sa.Text(), nullable=True)) # ### end Alembic commands ###
null
16,940
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.drop_column('text_to_speech') # ### end Alembic commands ###
null
16,941
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('tool_model_invokes', sa.Column('id', postgresql.UUID(), server_default=sa.text('uuid_generate_v4()'), nullable=False), sa.Column('user_id', postgresql.UUID(), nullable=False), sa.Column('tenant_id', postgresql.UUID(), nullable=False), sa.Column('provider', sa.String(length=40), nullable=False), sa.Column('tool_type', sa.String(length=40), nullable=False), sa.Column('tool_name', sa.String(length=40), nullable=False), sa.Column('tool_id', postgresql.UUID(), nullable=False), sa.Column('model_parameters', sa.Text(), nullable=False), sa.Column('prompt_messages', sa.Text(), nullable=False), sa.Column('model_response', sa.Text(), nullable=False), sa.Column('prompt_tokens', sa.Integer(), server_default=sa.text('0'), nullable=False), sa.Column('answer_tokens', sa.Integer(), server_default=sa.text('0'), nullable=False), sa.Column('answer_unit_price', sa.Numeric(precision=10, scale=4), nullable=False), sa.Column('answer_price_unit', sa.Numeric(precision=10, scale=7), server_default=sa.text('0.001'), nullable=False), sa.Column('provider_response_latency', sa.Float(), server_default=sa.text('0'), nullable=False), sa.Column('total_price', sa.Numeric(precision=10, scale=7), nullable=True), sa.Column('currency', sa.String(length=255), nullable=False), sa.Column('created_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.Column('updated_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.PrimaryKeyConstraint('id', name='tool_model_invoke_pkey') ) # ### end Alembic commands ###
null
16,942
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_table('tool_model_invokes') # ### end Alembic commands ###
null
16,943
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_table('sessions') with op.batch_alter_table('datasets', schema=None) as batch_op: batch_op.add_column(sa.Column('retrieval_model', postgresql.JSONB(astext_type=sa.Text()), nullable=True)) batch_op.create_index('retrieval_model_idx', ['retrieval_model'], unique=False, postgresql_using='gin') # ### end Alembic commands ###
null
16,944
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('datasets', schema=None) as batch_op: batch_op.drop_index('retrieval_model_idx', postgresql_using='gin') batch_op.drop_column('retrieval_model') op.create_table('sessions', sa.Column('id', sa.INTEGER(), autoincrement=True, nullable=False), sa.Column('session_id', sa.VARCHAR(length=255), autoincrement=False, nullable=True), sa.Column('data', postgresql.BYTEA(), autoincrement=False, nullable=True), sa.Column('expiry', postgresql.TIMESTAMP(), autoincrement=False, nullable=True), sa.PrimaryKeyConstraint('id', name='sessions_pkey'), sa.UniqueConstraint('session_id', name='sessions_session_id_key') ) # ### end Alembic commands ###
null
16,945
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tool_model_invokes', schema=None) as batch_op: batch_op.drop_column('tool_id') # ### end Alembic commands ###
null
16,946
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tool_model_invokes', schema=None) as batch_op: batch_op.add_column(sa.Column('tool_id', postgresql.UUID(), autoincrement=False, nullable=False)) # ### end Alembic commands ###
null
16,947
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('app_annotation_settings', sa.Column('id', postgresql.UUID(), server_default=sa.text('uuid_generate_v4()'), nullable=False), sa.Column('app_id', postgresql.UUID(), nullable=False), sa.Column('score_threshold', sa.Float(), server_default=sa.text('0'), nullable=False), sa.Column('collection_binding_id', postgresql.UUID(), nullable=False), sa.Column('created_user_id', postgresql.UUID(), nullable=False), sa.Column('created_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.Column('updated_user_id', postgresql.UUID(), nullable=False), sa.Column('updated_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.PrimaryKeyConstraint('id', name='app_annotation_settings_pkey') ) with op.batch_alter_table('app_annotation_settings', schema=None) as batch_op: batch_op.create_index('app_annotation_settings_app_idx', ['app_id'], unique=False) with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.drop_column('annotation_reply') # ### end Alembic commands ###
null
16,948
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.add_column(sa.Column('annotation_reply', sa.TEXT(), autoincrement=False, nullable=True)) with op.batch_alter_table('app_annotation_settings', schema=None) as batch_op: batch_op.drop_index('app_annotation_settings_app_idx') op.drop_table('app_annotation_settings') # ### end Alembic commands ###
null
16,949
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.add_column(sa.Column('speech_to_text', sa.Text(), nullable=True)) # ### end Alembic commands ###
null
16,950
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.drop_column('speech_to_text') # ### end Alembic commands ###
null
16,951
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('message_agent_thoughts', schema=None) as batch_op: batch_op.add_column(sa.Column('message_files', sa.Text(), nullable=True)) # ### end Alembic commands ###
null
16,952
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('message_agent_thoughts', schema=None) as batch_op: batch_op.drop_column('message_files') # ### end Alembic commands ###
null
16,953
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('app_annotation_hit_histories', sa.Column('id', postgresql.UUID(), server_default=sa.text('uuid_generate_v4()'), nullable=False), sa.Column('app_id', postgresql.UUID(), nullable=False), sa.Column('annotation_id', postgresql.UUID(), nullable=False), sa.Column('source', sa.Text(), nullable=False), sa.Column('question', sa.Text(), nullable=False), sa.Column('account_id', postgresql.UUID(), nullable=False), sa.Column('created_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.PrimaryKeyConstraint('id', name='app_annotation_hit_histories_pkey') ) with op.batch_alter_table('app_annotation_hit_histories', schema=None) as batch_op: batch_op.create_index('app_annotation_hit_histories_account_idx', ['account_id'], unique=False) batch_op.create_index('app_annotation_hit_histories_annotation_idx', ['annotation_id'], unique=False) batch_op.create_index('app_annotation_hit_histories_app_idx', ['app_id'], unique=False) with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.add_column(sa.Column('annotation_reply', sa.Text(), nullable=True)) with op.batch_alter_table('dataset_collection_bindings', schema=None) as batch_op: batch_op.add_column(sa.Column('type', sa.String(length=40), server_default=sa.text("'dataset'::character varying"), nullable=False)) with op.batch_alter_table('message_annotations', schema=None) as batch_op: batch_op.add_column(sa.Column('question', sa.Text(), nullable=True)) batch_op.add_column(sa.Column('hit_count', sa.Integer(), server_default=sa.text('0'), nullable=False)) batch_op.alter_column('conversation_id', existing_type=postgresql.UUID(), nullable=True) batch_op.alter_column('message_id', existing_type=postgresql.UUID(), nullable=True) # ### end Alembic commands ###
null
16,954
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('message_annotations', schema=None) as batch_op: batch_op.alter_column('message_id', existing_type=postgresql.UUID(), nullable=False) batch_op.alter_column('conversation_id', existing_type=postgresql.UUID(), nullable=False) batch_op.drop_column('hit_count') batch_op.drop_column('question') with op.batch_alter_table('dataset_collection_bindings', schema=None) as batch_op: batch_op.drop_column('type') with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.drop_column('annotation_reply') with op.batch_alter_table('app_annotation_hit_histories', schema=None) as batch_op: batch_op.drop_index('app_annotation_hit_histories_app_idx') batch_op.drop_index('app_annotation_hit_histories_annotation_idx') batch_op.drop_index('app_annotation_hit_histories_account_idx') op.drop_table('app_annotation_hit_histories') # ### end Alembic commands ###
null
16,955
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('conversations', schema=None) as batch_op: batch_op.add_column(sa.Column('is_deleted', sa.Boolean(), server_default=sa.text('false'), nullable=False)) # ### end Alembic commands ###
null
16,956
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('conversations', schema=None) as batch_op: batch_op.drop_column('is_deleted') # ### end Alembic commands ###
null
16,957
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('datasets', schema=None) as batch_op: batch_op.add_column(sa.Column('embedding_model', sa.String(length=255), server_default=sa.text("'text-embedding-ada-002'::character varying"), nullable=False)) batch_op.add_column(sa.Column('embedding_model_provider', sa.String(length=255), server_default=sa.text("'openai'::character varying"), nullable=False)) # ### end Alembic commands ###
null
16,958
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('datasets', schema=None) as batch_op: batch_op.drop_column('embedding_model_provider') batch_op.drop_column('embedding_model') # ### end Alembic commands ###
null
16,959
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('message_files', sa.Column('id', postgresql.UUID(), server_default=sa.text('uuid_generate_v4()'), nullable=False), sa.Column('message_id', postgresql.UUID(), nullable=False), sa.Column('type', sa.String(length=255), nullable=False), sa.Column('transfer_method', sa.String(length=255), nullable=False), sa.Column('url', sa.Text(), nullable=True), sa.Column('upload_file_id', postgresql.UUID(), nullable=True), sa.Column('created_by_role', sa.String(length=255), nullable=False), sa.Column('created_by', postgresql.UUID(), nullable=False), sa.Column('created_at', sa.DateTime(), server_default=sa.text('CURRENT_TIMESTAMP(0)'), nullable=False), sa.PrimaryKeyConstraint('id', name='message_file_pkey') ) with op.batch_alter_table('message_files', schema=None) as batch_op: batch_op.create_index('message_file_created_by_idx', ['created_by'], unique=False) batch_op.create_index('message_file_message_idx', ['message_id'], unique=False) with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.add_column(sa.Column('file_upload', sa.Text(), nullable=True)) with op.batch_alter_table('upload_files', schema=None) as batch_op: batch_op.add_column(sa.Column('created_by_role', sa.String(length=255), server_default=sa.text("'account'::character varying"), nullable=False)) # ### end Alembic commands ###
null
16,960
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('upload_files', schema=None) as batch_op: batch_op.drop_column('created_by_role') with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.drop_column('file_upload') with op.batch_alter_table('message_files', schema=None) as batch_op: batch_op.drop_index('message_file_message_idx') batch_op.drop_index('message_file_created_by_idx') op.drop_table('message_files') # ### end Alembic commands ###
null
16,961
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_annotation_hit_histories', schema=None) as batch_op: batch_op.add_column(sa.Column('message_id', postgresql.UUID(), nullable=False)) batch_op.create_index('app_annotation_hit_histories_message_idx', ['message_id'], unique=False) # ### end Alembic commands ###
null
16,962
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_annotation_hit_histories', schema=None) as batch_op: batch_op.drop_index('app_annotation_hit_histories_message_idx') batch_op.drop_column('message_id') # ### end Alembic commands ###
null
16,963
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tenant_account_joins', schema=None) as batch_op: batch_op.add_column(sa.Column('current', sa.Boolean(), server_default=sa.text('false'), nullable=False)) # ### end Alembic commands ###
null
16,964
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tenant_account_joins', schema=None) as batch_op: batch_op.drop_column('current') # ### end Alembic commands ###
null
16,965
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tool_api_providers', schema=None) as batch_op: batch_op.add_column(sa.Column('icon', sa.String(length=256), nullable=False)) # ### end Alembic commands ###
null
16,966
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tool_api_providers', schema=None) as batch_op: batch_op.drop_column('icon') # ### end Alembic commands ###
null
16,967
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.add_column(sa.Column('retriever_resource', sa.Text(), nullable=True)) # ### end Alembic commands ###
null
16,968
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('app_model_configs', schema=None) as batch_op: batch_op.drop_column('retriever_resource') # ### end Alembic commands ###
null
16,969
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tenants', schema=None) as batch_op: batch_op.add_column(sa.Column('custom_config', sa.Text(), nullable=True)) # ### end Alembic commands ###
null
16,970
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tenants', schema=None) as batch_op: batch_op.drop_column('custom_config') # ### end Alembic commands ###
null
16,971
from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tool_conversation_variables', schema=None) as batch_op: batch_op.create_index('conversation_id_idx', ['conversation_id'], unique=False) batch_op.create_index('user_id_idx', ['user_id'], unique=False) # ### end Alembic commands ###
null
16,972
from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('tool_conversation_variables', schema=None) as batch_op: batch_op.drop_index('user_id_idx') batch_op.drop_index('conversation_id_idx') # ### end Alembic commands ###
null
16,973
import sqlalchemy as sa from alembic import op def upgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('pinned_conversations', schema=None) as batch_op: batch_op.add_column(sa.Column('created_by_role', sa.String(length=255), server_default=sa.text("'end_user'::character varying"), nullable=False)) batch_op.drop_index('pinned_conversation_conversation_idx') batch_op.create_index('pinned_conversation_conversation_idx', ['app_id', 'conversation_id', 'created_by_role', 'created_by'], unique=False) with op.batch_alter_table('saved_messages', schema=None) as batch_op: batch_op.add_column(sa.Column('created_by_role', sa.String(length=255), server_default=sa.text("'end_user'::character varying"), nullable=False)) batch_op.drop_index('saved_message_message_idx') batch_op.create_index('saved_message_message_idx', ['app_id', 'message_id', 'created_by_role', 'created_by'], unique=False) # ### end Alembic commands ###
null
16,974
import sqlalchemy as sa from alembic import op def downgrade(): # ### commands auto generated by Alembic - please adjust! ### with op.batch_alter_table('saved_messages', schema=None) as batch_op: batch_op.drop_index('saved_message_message_idx') batch_op.create_index('saved_message_message_idx', ['app_id', 'message_id', 'created_by'], unique=False) batch_op.drop_column('created_by_role') with op.batch_alter_table('pinned_conversations', schema=None) as batch_op: batch_op.drop_index('pinned_conversation_conversation_idx') batch_op.create_index('pinned_conversation_conversation_idx', ['app_id', 'conversation_id', 'created_by'], unique=False) batch_op.drop_column('created_by_role') # ### end Alembic commands ###
null
16,975
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('tool_files', sa.Column('id', postgresql.UUID(), server_default=sa.text('uuid_generate_v4()'), nullable=False), sa.Column('user_id', postgresql.UUID(), nullable=False), sa.Column('tenant_id', postgresql.UUID(), nullable=False), sa.Column('conversation_id', postgresql.UUID(), nullable=False), sa.Column('file_key', sa.String(length=255), nullable=False), sa.Column('mimetype', sa.String(length=255), nullable=False), sa.Column('original_url', sa.String(length=255), nullable=True), sa.PrimaryKeyConstraint('id', name='tool_file_pkey') ) # ### end Alembic commands ###
null
16,976
import sqlalchemy as sa from alembic import op from sqlalchemy.dialects import postgresql def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_table('tool_files') # ### end Alembic commands ###
null
16,977
import logging from logging.config import fileConfig from alembic import context from flask import current_app def get_engine(): return current_app.extensions['migrate'].db.engine def get_engine_url(): try: return get_engine().url.render_as_string(hide_password=False).replace( '%', '%%') except AttributeError: return str(get_engine().url).replace('%', '%%')
null
16,978
import logging from logging.config import fileConfig from alembic import context from flask import current_app config = context.config config.set_main_option('sqlalchemy.url', get_engine_url()) def get_metadata(): if hasattr(target_db, 'metadatas'): return target_db.metadatas[None] return target_db.metadata if context.is_offline_mode(): run_migrations_offline() else: run_migrations_online() The provided code snippet includes necessary dependencies for implementing the `run_migrations_offline` function. Write a Python function `def run_migrations_offline()` to solve the following problem: Run migrations in 'offline' mode. This configures the context with just a URL and not an Engine, though an Engine is acceptable here as well. By skipping the Engine creation we don't even need a DBAPI to be available. Calls to context.execute() here emit the given string to the script output. Here is the function: def run_migrations_offline(): """Run migrations in 'offline' mode. This configures the context with just a URL and not an Engine, though an Engine is acceptable here as well. By skipping the Engine creation we don't even need a DBAPI to be available. Calls to context.execute() here emit the given string to the script output. """ url = config.get_main_option("sqlalchemy.url") context.configure( url=url, target_metadata=get_metadata(), literal_binds=True ) with context.begin_transaction(): context.run_migrations()
Run migrations in 'offline' mode. This configures the context with just a URL and not an Engine, though an Engine is acceptable here as well. By skipping the Engine creation we don't even need a DBAPI to be available. Calls to context.execute() here emit the given string to the script output.
16,979
import logging from logging.config import fileConfig from alembic import context from flask import current_app config = context.config logger = logging.getLogger('alembic.env') def get_engine(): return current_app.extensions['migrate'].db.engine config.set_main_option('sqlalchemy.url', get_engine_url()) def get_metadata(): if hasattr(target_db, 'metadatas'): return target_db.metadatas[None] return target_db.metadata def include_object(object, name, type_, reflected, compare_to): if type_ == "foreign_key_constraint": return False else: return True if context.is_offline_mode(): run_migrations_offline() else: run_migrations_online() The provided code snippet includes necessary dependencies for implementing the `run_migrations_online` function. Write a Python function `def run_migrations_online()` to solve the following problem: Run migrations in 'online' mode. In this scenario we need to create an Engine and associate a connection with the context. Here is the function: def run_migrations_online(): """Run migrations in 'online' mode. In this scenario we need to create an Engine and associate a connection with the context. """ # this callback is used to prevent an auto-migration from being generated # when there are no changes to the schema # reference: http://alembic.zzzcomputing.com/en/latest/cookbook.html def process_revision_directives(context, revision, directives): if getattr(config.cmd_opts, 'autogenerate', False): script = directives[0] if script.upgrade_ops.is_empty(): directives[:] = [] logger.info('No changes in schema detected.') connectable = get_engine() with connectable.connect() as connection: context.configure( connection=connection, target_metadata=get_metadata(), process_revision_directives=process_revision_directives, include_object=include_object, **current_app.extensions['migrate'].configure_args ) with context.begin_transaction(): context.run_migrations()
Run migrations in 'online' mode. In this scenario we need to create an Engine and associate a connection with the context.
16,980
from collections.abc import Callable from datetime import datetime from enum import Enum from functools import wraps from typing import Optional from flask import current_app, request from flask_login import user_logged_in from flask_restful import Resource from pydantic import BaseModel from werkzeug.exceptions import NotFound, Unauthorized from extensions.ext_database import db from libs.login import _get_user from models.account import Account, Tenant, TenantAccountJoin from models.model import ApiToken, App, EndUser from services.feature_service import FeatureService class WhereisUserArg(Enum): """ Enum for whereis_user_arg. """ QUERY = 'query' JSON = 'json' FORM = 'form' class FetchUserArg(BaseModel): fetch_from: WhereisUserArg required: bool = False def validate_and_get_api_token(scope=None): """ Validate and get API token. """ auth_header = request.headers.get('Authorization') if auth_header is None or ' ' not in auth_header: raise Unauthorized("Authorization header must be provided and start with 'Bearer'") auth_scheme, auth_token = auth_header.split(None, 1) auth_scheme = auth_scheme.lower() if auth_scheme != 'bearer': raise Unauthorized("Authorization scheme must be 'Bearer'") api_token = db.session.query(ApiToken).filter( ApiToken.token == auth_token, ApiToken.type == scope, ).first() if not api_token: raise Unauthorized("Access token is invalid") api_token.last_used_at = datetime.utcnow() db.session.commit() return api_token def create_or_update_end_user_for_user_id(app_model: App, user_id: Optional[str] = None) -> EndUser: """ Create or update session terminal based on user ID. """ if not user_id: user_id = 'DEFAULT-USER' end_user = db.session.query(EndUser) \ .filter( EndUser.tenant_id == app_model.tenant_id, EndUser.app_id == app_model.id, EndUser.session_id == user_id, EndUser.type == 'service_api' ).first() if end_user is None: end_user = EndUser( tenant_id=app_model.tenant_id, app_id=app_model.id, type='service_api', is_anonymous=True if user_id == 'DEFAULT-USER' else False, session_id=user_id ) db.session.add(end_user) db.session.commit() return end_user db = SQLAlchemy() class App(db.Model): __tablename__ = 'apps' __table_args__ = ( db.PrimaryKeyConstraint('id', name='app_pkey'), db.Index('app_tenant_id_idx', 'tenant_id') ) id = db.Column(UUID, server_default=db.text('uuid_generate_v4()')) tenant_id = db.Column(UUID, nullable=False) name = db.Column(db.String(255), nullable=False) mode = db.Column(db.String(255), nullable=False) icon = db.Column(db.String(255)) icon_background = db.Column(db.String(255)) app_model_config_id = db.Column(UUID, nullable=True) status = db.Column(db.String(255), nullable=False, server_default=db.text("'normal'::character varying")) enable_site = db.Column(db.Boolean, nullable=False) enable_api = db.Column(db.Boolean, nullable=False) api_rpm = db.Column(db.Integer, nullable=False) api_rph = db.Column(db.Integer, nullable=False) is_demo = db.Column(db.Boolean, nullable=False, server_default=db.text('false')) is_public = db.Column(db.Boolean, nullable=False, server_default=db.text('false')) is_universal = db.Column(db.Boolean, nullable=False, server_default=db.text('false')) created_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) updated_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) def site(self): site = db.session.query(Site).filter(Site.app_id == self.id).first() return site def app_model_config(self): app_model_config = db.session.query(AppModelConfig).filter( AppModelConfig.id == self.app_model_config_id).first() return app_model_config def api_base_url(self): return (current_app.config['SERVICE_API_URL'] if current_app.config['SERVICE_API_URL'] else request.host_url.rstrip('/')) + '/v1' def tenant(self): tenant = db.session.query(Tenant).filter(Tenant.id == self.tenant_id).first() return tenant def is_agent(self) -> bool: app_model_config = self.app_model_config if not app_model_config: return False if not app_model_config.agent_mode: return False if self.app_model_config.agent_mode_dict.get('enabled', False) \ and self.app_model_config.agent_mode_dict.get('strategy', '') in ['function_call', 'react']: return True return False def deleted_tools(self) -> list: # get agent mode tools app_model_config = self.app_model_config if not app_model_config: return [] if not app_model_config.agent_mode: return [] agent_mode = app_model_config.agent_mode_dict tools = agent_mode.get('tools', []) provider_ids = [] for tool in tools: keys = list(tool.keys()) if len(keys) >= 4: provider_type = tool.get('provider_type', '') provider_id = tool.get('provider_id', '') if provider_type == 'api': # check if provider id is a uuid string, if not, skip try: uuid.UUID(provider_id) except Exception: continue provider_ids.append(provider_id) if not provider_ids: return [] api_providers = db.session.execute( text('SELECT id FROM tool_api_providers WHERE id IN :provider_ids'), {'provider_ids': tuple(provider_ids)} ).fetchall() deleted_tools = [] current_api_provider_ids = [str(api_provider.id) for api_provider in api_providers] for tool in tools: keys = list(tool.keys()) if len(keys) >= 4: provider_type = tool.get('provider_type', '') provider_id = tool.get('provider_id', '') if provider_type == 'api' and provider_id not in current_api_provider_ids: deleted_tools.append(tool['tool_name']) return deleted_tools def validate_app_token(view: Optional[Callable] = None, *, fetch_user_arg: Optional[FetchUserArg] = None): def decorator(view_func): @wraps(view_func) def decorated_view(*args, **kwargs): api_token = validate_and_get_api_token('app') app_model = db.session.query(App).filter(App.id == api_token.app_id).first() if not app_model: raise NotFound() if app_model.status != 'normal': raise NotFound() if not app_model.enable_api: raise NotFound() kwargs['app_model'] = app_model if fetch_user_arg: if fetch_user_arg.fetch_from == WhereisUserArg.QUERY: user_id = request.args.get('user') elif fetch_user_arg.fetch_from == WhereisUserArg.JSON: user_id = request.get_json().get('user') elif fetch_user_arg.fetch_from == WhereisUserArg.FORM: user_id = request.form.get('user') else: # use default-user user_id = None if not user_id and fetch_user_arg.required: raise ValueError("Arg user must be provided.") if user_id: user_id = str(user_id) kwargs['end_user'] = create_or_update_end_user_for_user_id(app_model, user_id) return view_func(*args, **kwargs) return decorated_view if view is None: return decorator else: return decorator(view)
null
16,981
from collections.abc import Callable from datetime import datetime from enum import Enum from functools import wraps from typing import Optional from flask import current_app, request from flask_login import user_logged_in from flask_restful import Resource from pydantic import BaseModel from werkzeug.exceptions import NotFound, Unauthorized from extensions.ext_database import db from libs.login import _get_user from models.account import Account, Tenant, TenantAccountJoin from models.model import ApiToken, App, EndUser from services.feature_service import FeatureService def validate_and_get_api_token(scope=None): class FeatureService: def get_features(cls, tenant_id: str) -> FeatureModel: def _fulfill_params_from_env(cls, features: FeatureModel): def _fulfill_params_from_billing_api(cls, features: FeatureModel, tenant_id: str): def cloud_edition_billing_resource_check(resource: str, api_token_type: str, error_msg: str = "You have reached the limit of your subscription."): def interceptor(view): def decorated(*args, **kwargs): api_token = validate_and_get_api_token(api_token_type) features = FeatureService.get_features(api_token.tenant_id) if features.billing.enabled: members = features.members apps = features.apps vector_space = features.vector_space documents_upload_quota = features.documents_upload_quota if resource == 'members' and 0 < members.limit <= members.size: raise Unauthorized(error_msg) elif resource == 'apps' and 0 < apps.limit <= apps.size: raise Unauthorized(error_msg) elif resource == 'vector_space' and 0 < vector_space.limit <= vector_space.size: raise Unauthorized(error_msg) elif resource == 'documents' and 0 < documents_upload_quota.limit <= documents_upload_quota.size: raise Unauthorized(error_msg) else: return view(*args, **kwargs) return view(*args, **kwargs) return decorated return interceptor
null
16,982
from collections.abc import Callable from datetime import datetime from enum import Enum from functools import wraps from typing import Optional from flask import current_app, request from flask_login import user_logged_in from flask_restful import Resource from pydantic import BaseModel from werkzeug.exceptions import NotFound, Unauthorized from extensions.ext_database import db from libs.login import _get_user from models.account import Account, Tenant, TenantAccountJoin from models.model import ApiToken, App, EndUser from services.feature_service import FeatureService def validate_and_get_api_token(scope=None): """ Validate and get API token. """ auth_header = request.headers.get('Authorization') if auth_header is None or ' ' not in auth_header: raise Unauthorized("Authorization header must be provided and start with 'Bearer'") auth_scheme, auth_token = auth_header.split(None, 1) auth_scheme = auth_scheme.lower() if auth_scheme != 'bearer': raise Unauthorized("Authorization scheme must be 'Bearer'") api_token = db.session.query(ApiToken).filter( ApiToken.token == auth_token, ApiToken.type == scope, ).first() if not api_token: raise Unauthorized("Access token is invalid") api_token.last_used_at = datetime.utcnow() db.session.commit() return api_token db = SQLAlchemy() def _get_user(): if has_request_context(): if "_login_user" not in g: current_app.login_manager._load_user() return g._login_user return None class Account(UserMixin, db.Model): __tablename__ = 'accounts' __table_args__ = ( db.PrimaryKeyConstraint('id', name='account_pkey'), db.Index('account_email_idx', 'email') ) id = db.Column(UUID, server_default=db.text('uuid_generate_v4()')) name = db.Column(db.String(255), nullable=False) email = db.Column(db.String(255), nullable=False) password = db.Column(db.String(255), nullable=True) password_salt = db.Column(db.String(255), nullable=True) avatar = db.Column(db.String(255)) interface_language = db.Column(db.String(255)) interface_theme = db.Column(db.String(255)) timezone = db.Column(db.String(255)) last_login_at = db.Column(db.DateTime) last_login_ip = db.Column(db.String(255)) last_active_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) status = db.Column(db.String(16), nullable=False, server_default=db.text("'active'::character varying")) initialized_at = db.Column(db.DateTime) created_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) updated_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) def is_password_set(self): return self.password is not None def current_tenant(self): return self._current_tenant def current_tenant(self, value): tenant = value ta = TenantAccountJoin.query.filter_by(tenant_id=tenant.id, account_id=self.id).first() if ta: tenant.current_role = ta.role else: tenant = None self._current_tenant = tenant def current_tenant_id(self): return self._current_tenant.id def current_tenant_id(self, value): try: tenant_account_join = db.session.query(Tenant, TenantAccountJoin) \ .filter(Tenant.id == value) \ .filter(TenantAccountJoin.tenant_id == Tenant.id) \ .filter(TenantAccountJoin.account_id == self.id) \ .one_or_none() if tenant_account_join: tenant, ta = tenant_account_join tenant.current_role = ta.role else: tenant = None except: tenant = None self._current_tenant = tenant def get_status(self) -> AccountStatus: status_str = self.status return AccountStatus(status_str) def get_by_openid(cls, provider: str, open_id: str) -> db.Model: account_integrate = db.session.query(AccountIntegrate). \ filter(AccountIntegrate.provider == provider, AccountIntegrate.open_id == open_id). \ one_or_none() if account_integrate: return db.session.query(Account). \ filter(Account.id == account_integrate.account_id). \ one_or_none() return None def get_integrates(self) -> list[db.Model]: ai = db.Model return db.session.query(ai).filter( ai.account_id == self.id ).all() # check current_user.current_tenant.current_role in ['admin', 'owner'] def is_admin_or_owner(self): return self._current_tenant.current_role in ['admin', 'owner'] class Tenant(db.Model): __tablename__ = 'tenants' __table_args__ = ( db.PrimaryKeyConstraint('id', name='tenant_pkey'), ) id = db.Column(UUID, server_default=db.text('uuid_generate_v4()')) name = db.Column(db.String(255), nullable=False) encrypt_public_key = db.Column(db.Text) plan = db.Column(db.String(255), nullable=False, server_default=db.text("'basic'::character varying")) status = db.Column(db.String(255), nullable=False, server_default=db.text("'normal'::character varying")) custom_config = db.Column(db.Text) created_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) updated_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) def get_accounts(self) -> list[db.Model]: Account = db.Model return db.session.query(Account).filter( Account.id == TenantAccountJoin.account_id, TenantAccountJoin.tenant_id == self.id ).all() def custom_config_dict(self) -> dict: return json.loads(self.custom_config) if self.custom_config else {} def custom_config_dict(self, value: dict): self.custom_config = json.dumps(value) class TenantAccountJoin(db.Model): __tablename__ = 'tenant_account_joins' __table_args__ = ( db.PrimaryKeyConstraint('id', name='tenant_account_join_pkey'), db.Index('tenant_account_join_account_id_idx', 'account_id'), db.Index('tenant_account_join_tenant_id_idx', 'tenant_id'), db.UniqueConstraint('tenant_id', 'account_id', name='unique_tenant_account_join') ) id = db.Column(UUID, server_default=db.text('uuid_generate_v4()')) tenant_id = db.Column(UUID, nullable=False) account_id = db.Column(UUID, nullable=False) current = db.Column(db.Boolean, nullable=False, server_default=db.text('false')) role = db.Column(db.String(16), nullable=False, server_default='normal') invited_by = db.Column(UUID, nullable=True) created_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) updated_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) def validate_dataset_token(view=None): def decorator(view): @wraps(view) def decorated(*args, **kwargs): api_token = validate_and_get_api_token('dataset') tenant_account_join = db.session.query(Tenant, TenantAccountJoin) \ .filter(Tenant.id == api_token.tenant_id) \ .filter(TenantAccountJoin.tenant_id == Tenant.id) \ .filter(TenantAccountJoin.role.in_(['owner'])) \ .one_or_none() # TODO: only owner information is required, so only one is returned. if tenant_account_join: tenant, ta = tenant_account_join account = Account.query.filter_by(id=ta.account_id).first() # Login admin if account: account.current_tenant = tenant current_app.login_manager._update_request_context_with_user(account) user_logged_in.send(current_app._get_current_object(), user=_get_user()) else: raise Unauthorized("Tenant owner account does not exist.") else: raise Unauthorized("Tenant does not exist.") return view(api_token.tenant_id, *args, **kwargs) return decorated if view: return decorator(view) # if view is None, it means that the decorator is used without parentheses # use the decorator as a function for method_decorators return decorator
null
16,983
from flask import request from flask_restful import marshal, reqparse import services.dataset_service from controllers.service_api import api from controllers.service_api.dataset.error import DatasetNameDuplicateError from controllers.service_api.wraps import DatasetApiResource from core.model_runtime.entities.model_entities import ModelType from core.provider_manager import ProviderManager from fields.dataset_fields import dataset_detail_fields from libs.login import current_user from models.dataset import Dataset from services.dataset_service import DatasetService def _validate_name(name): if not name or len(name) < 1 or len(name) > 40: raise ValueError('Name must be between 1 to 40 characters.') return name
null
16,984
import json import logging from collections.abc import Generator from typing import Union from flask import Response, stream_with_context from flask_restful import Resource, reqparse from werkzeug.exceptions import InternalServerError, NotFound import services from controllers.service_api import api from controllers.service_api.app.error import ( AppUnavailableError, CompletionRequestError, ConversationCompletedError, NotChatAppError, ProviderModelCurrentlyNotSupportError, ProviderNotInitializeError, ProviderQuotaExceededError, ) from controllers.service_api.wraps import FetchUserArg, WhereisUserArg, validate_app_token from core.application_queue_manager import ApplicationQueueManager from core.entities.application_entities import InvokeFrom from core.errors.error import ModelCurrentlyNotSupportError, ProviderTokenNotInitError, QuotaExceededError from core.model_runtime.errors.invoke import InvokeError from libs.helper import uuid_value from models.model import App, EndUser from services.completion_service import CompletionService def compact_response(response: Union[dict, Generator]) -> Response: if isinstance(response, dict): return Response(response=json.dumps(response), status=200, mimetype='application/json') else: def generate() -> Generator: yield from response return Response(stream_with_context(generate()), status=200, mimetype='text/event-stream')
null
16,985
from functools import wraps from flask import request from flask_restful import Resource from werkzeug.exceptions import NotFound, Unauthorized from extensions.ext_database import db from libs.passport import PassportService from models.model import App, EndUser, Site def decode_jwt_token(): auth_header = request.headers.get('Authorization') if auth_header is None: raise Unauthorized('Authorization header is missing.') if ' ' not in auth_header: raise Unauthorized('Invalid Authorization header format. Expected \'Bearer <api-key>\' format.') auth_scheme, tk = auth_header.split(None, 1) auth_scheme = auth_scheme.lower() if auth_scheme != 'bearer': raise Unauthorized('Invalid Authorization header format. Expected \'Bearer <api-key>\' format.') decoded = PassportService().verify(tk) app_code = decoded.get('app_code') app_model = db.session.query(App).filter(App.id == decoded['app_id']).first() site = db.session.query(Site).filter(Site.code == app_code).first() if not app_model: raise NotFound() if not app_code or not site: raise Unauthorized('Site URL is no longer valid.') if app_model.enable_site is False: raise Unauthorized('Site is disabled.') end_user = db.session.query(EndUser).filter(EndUser.id == decoded['end_user_id']).first() if not end_user: raise NotFound() return app_model, end_user def validate_jwt_token(view=None): def decorator(view): @wraps(view) def decorated(*args, **kwargs): app_model, end_user = decode_jwt_token() return view(app_model, end_user, *args, **kwargs) return decorated if view: return decorator(view) return decorator
null
16,986
import uuid from flask import request from flask_restful import Resource from werkzeug.exceptions import NotFound, Unauthorized from controllers.web import api from extensions.ext_database import db from libs.passport import PassportService from models.model import App, EndUser, Site db = SQLAlchemy() class EndUser(UserMixin, db.Model): __tablename__ = 'end_users' __table_args__ = ( db.PrimaryKeyConstraint('id', name='end_user_pkey'), db.Index('end_user_session_id_idx', 'session_id', 'type'), db.Index('end_user_tenant_session_id_idx', 'tenant_id', 'session_id', 'type'), ) id = db.Column(UUID, server_default=db.text('uuid_generate_v4()')) tenant_id = db.Column(UUID, nullable=False) app_id = db.Column(UUID, nullable=True) type = db.Column(db.String(255), nullable=False) external_user_id = db.Column(db.String(255), nullable=True) name = db.Column(db.String(255)) is_anonymous = db.Column(db.Boolean, nullable=False, server_default=db.text('true')) session_id = db.Column(db.String(255), nullable=False) created_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) updated_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) The provided code snippet includes necessary dependencies for implementing the `generate_session_id` function. Write a Python function `def generate_session_id()` to solve the following problem: Generate a unique session ID. Here is the function: def generate_session_id(): """ Generate a unique session ID. """ while True: session_id = str(uuid.uuid4()) existing_count = db.session.query(EndUser) \ .filter(EndUser.session_id == session_id).count() if existing_count == 0: return session_id
Generate a unique session ID.
16,987
import json import logging from collections.abc import Generator from typing import Union from flask import Response, stream_with_context from flask_restful import fields, marshal_with, reqparse from flask_restful.inputs import int_range from werkzeug.exceptions import InternalServerError, NotFound import services from controllers.web import api from controllers.web.error import ( AppMoreLikeThisDisabledError, AppSuggestedQuestionsAfterAnswerDisabledError, CompletionRequestError, NotChatAppError, NotCompletionAppError, ProviderModelCurrentlyNotSupportError, ProviderNotInitializeError, ProviderQuotaExceededError, ) from controllers.web.wraps import WebApiResource from core.entities.application_entities import InvokeFrom from core.errors.error import ModelCurrentlyNotSupportError, ProviderTokenNotInitError, QuotaExceededError from core.model_runtime.errors.invoke import InvokeError from fields.conversation_fields import message_file_fields from fields.message_fields import agent_thought_fields from libs.helper import TimestampField, uuid_value from services.completion_service import CompletionService from services.errors.app import MoreLikeThisDisabledError from services.errors.conversation import ConversationNotExistsError from services.errors.message import MessageNotExistsError, SuggestedQuestionsAfterAnswerDisabledError from services.message_service import MessageService def compact_response(response: Union[dict, Generator]) -> Response: if isinstance(response, dict): return Response(response=json.dumps(response), status=200, mimetype='application/json') else: def generate() -> Generator: yield from response return Response(stream_with_context(generate()), status=200, mimetype='text/event-stream')
null
16,988
import json import logging from collections.abc import Generator from typing import Union from flask import Response, stream_with_context from flask_restful import reqparse from werkzeug.exceptions import InternalServerError, NotFound import services from controllers.web import api from controllers.web.error import ( AppUnavailableError, CompletionRequestError, ConversationCompletedError, NotChatAppError, NotCompletionAppError, ProviderModelCurrentlyNotSupportError, ProviderNotInitializeError, ProviderQuotaExceededError, ) from controllers.web.wraps import WebApiResource from core.application_queue_manager import ApplicationQueueManager from core.entities.application_entities import InvokeFrom from core.errors.error import ModelCurrentlyNotSupportError, ProviderTokenNotInitError, QuotaExceededError from core.model_runtime.errors.invoke import InvokeError from libs.helper import uuid_value from services.completion_service import CompletionService def compact_response(response: Union[dict, Generator]) -> Response: if isinstance(response, dict): return Response(response=json.dumps(response), status=200, mimetype='application/json') else: def generate() -> Generator: yield from response return Response(stream_with_context(generate()), status=200, mimetype='text/event-stream')
null
16,989
from functools import wraps from flask import current_app, request from flask_restful import Resource, reqparse from extensions.ext_database import db from libs.helper import email, str_len from libs.password import valid_password from models.model import DifySetup from services.account_service import AccountService, RegisterService, TenantService from . import api from .error import AlreadySetupError, NotInitValidateError, NotSetupError from .init_validate import get_init_validate_status from .wraps import only_edition_self_hosted db = SQLAlchemy() class DifySetup(db.Model): __tablename__ = 'dify_setups' __table_args__ = ( db.PrimaryKeyConstraint('version', name='dify_setup_pkey'), ) version = db.Column(db.String(255), nullable=False) setup_at = db.Column(db.DateTime, nullable=False, server_default=db.text('CURRENT_TIMESTAMP(0)')) def setup(): dify_setup = DifySetup( version=current_app.config['CURRENT_VERSION'] ) db.session.add(dify_setup)
null
16,990
from functools import wraps from flask import current_app, request from flask_restful import Resource, reqparse from extensions.ext_database import db from libs.helper import email, str_len from libs.password import valid_password from models.model import DifySetup from services.account_service import AccountService, RegisterService, TenantService from . import api from .error import AlreadySetupError, NotInitValidateError, NotSetupError from .init_validate import get_init_validate_status from .wraps import only_edition_self_hosted def get_setup_status(): if current_app.config['EDITION'] == 'SELF_HOSTED': return DifySetup.query.first() else: return True class NotSetupError(BaseHTTPException): error_code = 'not_setup' description = "Dify has not been initialized and installed yet. " \ "Please proceed with the initialization and installation process first." code = 401 class NotInitValidateError(BaseHTTPException): error_code = 'not_init_validated' description = "Init validation has not been completed yet. " \ "Please proceed with the init validation process first." code = 401 def get_init_validate_status(): if current_app.config['EDITION'] == 'SELF_HOSTED': if os.environ.get('INIT_PASSWORD'): return session.get('is_init_validated') or DifySetup.query.first() return True def setup_required(view): @wraps(view) def decorated(*args, **kwargs): # check setup if not get_init_validate_status(): raise NotInitValidateError() elif not get_setup_status(): raise NotSetupError() return view(*args, **kwargs) return decorated
null
16,991
from functools import wraps from flask_login import current_user from flask_restful import Resource from werkzeug.exceptions import NotFound from controllers.console.wraps import account_initialization_required from extensions.ext_database import db from libs.login import login_required from models.model import InstalledApp db = SQLAlchemy() class InstalledApp(db.Model): def app(self): def tenant(self): def is_agent(self) -> bool: def installed_app_required(view=None): def decorator(view): @wraps(view) def decorated(*args, **kwargs): if not kwargs.get('installed_app_id'): raise ValueError('missing installed_app_id in path parameters') installed_app_id = kwargs.get('installed_app_id') installed_app_id = str(installed_app_id) del kwargs['installed_app_id'] installed_app = db.session.query(InstalledApp).filter( InstalledApp.id == str(installed_app_id), InstalledApp.tenant_id == current_user.current_tenant_id ).first() if installed_app is None: raise NotFound('Installed app not found') if not installed_app.app: db.session.delete(installed_app) db.session.commit() raise NotFound('Installed app not found') return view(installed_app, *args, **kwargs) return decorated if view: return decorator(view) return decorator
null
16,992
import json import logging from collections.abc import Generator from typing import Union from flask import Response, stream_with_context from flask_login import current_user from flask_restful import marshal_with, reqparse from flask_restful.inputs import int_range from werkzeug.exceptions import InternalServerError, NotFound import services from controllers.console import api from controllers.console.app.error import ( AppMoreLikeThisDisabledError, CompletionRequestError, ProviderModelCurrentlyNotSupportError, ProviderNotInitializeError, ProviderQuotaExceededError, ) from controllers.console.explore.error import ( AppSuggestedQuestionsAfterAnswerDisabledError, NotChatAppError, NotCompletionAppError, ) from controllers.console.explore.wraps import InstalledAppResource from core.entities.application_entities import InvokeFrom from core.errors.error import ModelCurrentlyNotSupportError, ProviderTokenNotInitError, QuotaExceededError from core.model_runtime.errors.invoke import InvokeError from fields.message_fields import message_infinite_scroll_pagination_fields from libs.helper import uuid_value from services.completion_service import CompletionService from services.errors.app import MoreLikeThisDisabledError from services.errors.conversation import ConversationNotExistsError from services.errors.message import MessageNotExistsError, SuggestedQuestionsAfterAnswerDisabledError from services.message_service import MessageService def compact_response(response: Union[dict, Generator]) -> Response: if isinstance(response, dict): return Response(response=json.dumps(response), status=200, mimetype='application/json') else: def generate() -> Generator: yield from response return Response(stream_with_context(generate()), status=200, mimetype='text/event-stream')
null
16,993
import json import logging from collections.abc import Generator from datetime import datetime from typing import Union from flask import Response, stream_with_context from flask_login import current_user from flask_restful import reqparse from werkzeug.exceptions import InternalServerError, NotFound import services from controllers.console import api from controllers.console.app.error import ( AppUnavailableError, CompletionRequestError, ConversationCompletedError, ProviderModelCurrentlyNotSupportError, ProviderNotInitializeError, ProviderQuotaExceededError, ) from controllers.console.explore.error import NotChatAppError, NotCompletionAppError from controllers.console.explore.wraps import InstalledAppResource from core.application_queue_manager import ApplicationQueueManager from core.entities.application_entities import InvokeFrom from core.errors.error import ModelCurrentlyNotSupportError, ProviderTokenNotInitError, QuotaExceededError from core.model_runtime.errors.invoke import InvokeError from extensions.ext_database import db from libs.helper import uuid_value from services.completion_service import CompletionService def compact_response(response: Union[dict, Generator]) -> Response: if isinstance(response, dict): return Response(response=json.dumps(response), status=200, mimetype='application/json') else: def generate() -> Generator: yield from response return Response(stream_with_context(generate()), status=200, mimetype='text/event-stream')
null
16,994
import os from functools import wraps from flask import request from flask_restful import Resource, reqparse from werkzeug.exceptions import NotFound, Unauthorized from constants.languages import supported_language from controllers.console import api from controllers.console.wraps import only_edition_cloud from extensions.ext_database import db from models.model import App, InstalledApp, RecommendedApp def admin_required(view): @wraps(view) def decorated(*args, **kwargs): if not os.getenv('ADMIN_API_KEY'): raise Unauthorized('API key is invalid.') auth_header = request.headers.get('Authorization') if auth_header is None: raise Unauthorized('Authorization header is missing.') if ' ' not in auth_header: raise Unauthorized('Invalid Authorization header format. Expected \'Bearer <api-key>\' format.') auth_scheme, auth_token = auth_header.split(None, 1) auth_scheme = auth_scheme.lower() if auth_scheme != 'bearer': raise Unauthorized('Invalid Authorization header format. Expected \'Bearer <api-key>\' format.') if os.getenv('ADMIN_API_KEY') != auth_token: raise Unauthorized('API key is invalid.') return view(*args, **kwargs) return decorated
null
16,995
import logging import requests from flask import current_app, redirect, request from flask_login import current_user from flask_restful import Resource from werkzeug.exceptions import Forbidden from controllers.console import api from libs.login import login_required from libs.oauth_data_source import NotionOAuth from ..setup import setup_required from ..wraps import account_initialization_required class NotionOAuth(OAuthDataSource): _AUTH_URL = 'https://api.notion.com/v1/oauth/authorize' _TOKEN_URL = 'https://api.notion.com/v1/oauth/token' _NOTION_PAGE_SEARCH = "https://api.notion.com/v1/search" _NOTION_BLOCK_SEARCH = "https://api.notion.com/v1/blocks" _NOTION_BOT_USER = "https://api.notion.com/v1/users/me" def get_authorization_url(self): params = { 'client_id': self.client_id, 'response_type': 'code', 'redirect_uri': self.redirect_uri, 'owner': 'user' } return f"{self._AUTH_URL}?{urllib.parse.urlencode(params)}" def get_access_token(self, code: str): data = { 'code': code, 'grant_type': 'authorization_code', 'redirect_uri': self.redirect_uri } headers = {'Accept': 'application/json'} auth = (self.client_id, self.client_secret) response = requests.post(self._TOKEN_URL, data=data, auth=auth, headers=headers) response_json = response.json() access_token = response_json.get('access_token') if not access_token: raise ValueError(f"Error in Notion OAuth: {response_json}") workspace_name = response_json.get('workspace_name') workspace_icon = response_json.get('workspace_icon') workspace_id = response_json.get('workspace_id') # get all authorized pages pages = self.get_authorized_pages(access_token) source_info = { 'workspace_name': workspace_name, 'workspace_icon': workspace_icon, 'workspace_id': workspace_id, 'pages': pages, 'total': len(pages) } # save data source binding data_source_binding = DataSourceBinding.query.filter( db.and_( DataSourceBinding.tenant_id == current_user.current_tenant_id, DataSourceBinding.provider == 'notion', DataSourceBinding.access_token == access_token ) ).first() if data_source_binding: data_source_binding.source_info = source_info data_source_binding.disabled = False db.session.commit() else: new_data_source_binding = DataSourceBinding( tenant_id=current_user.current_tenant_id, access_token=access_token, source_info=source_info, provider='notion' ) db.session.add(new_data_source_binding) db.session.commit() def save_internal_access_token(self, access_token: str): workspace_name = self.notion_workspace_name(access_token) workspace_icon = None workspace_id = current_user.current_tenant_id # get all authorized pages pages = self.get_authorized_pages(access_token) source_info = { 'workspace_name': workspace_name, 'workspace_icon': workspace_icon, 'workspace_id': workspace_id, 'pages': pages, 'total': len(pages) } # save data source binding data_source_binding = DataSourceBinding.query.filter( db.and_( DataSourceBinding.tenant_id == current_user.current_tenant_id, DataSourceBinding.provider == 'notion', DataSourceBinding.access_token == access_token ) ).first() if data_source_binding: data_source_binding.source_info = source_info data_source_binding.disabled = False db.session.commit() else: new_data_source_binding = DataSourceBinding( tenant_id=current_user.current_tenant_id, access_token=access_token, source_info=source_info, provider='notion' ) db.session.add(new_data_source_binding) db.session.commit() def sync_data_source(self, binding_id: str): # save data source binding data_source_binding = DataSourceBinding.query.filter( db.and_( DataSourceBinding.tenant_id == current_user.current_tenant_id, DataSourceBinding.provider == 'notion', DataSourceBinding.id == binding_id, DataSourceBinding.disabled == False ) ).first() if data_source_binding: # get all authorized pages pages = self.get_authorized_pages(data_source_binding.access_token) source_info = data_source_binding.source_info new_source_info = { 'workspace_name': source_info['workspace_name'], 'workspace_icon': source_info['workspace_icon'], 'workspace_id': source_info['workspace_id'], 'pages': pages, 'total': len(pages) } data_source_binding.source_info = new_source_info data_source_binding.disabled = False db.session.commit() else: raise ValueError('Data source binding not found') def get_authorized_pages(self, access_token: str): pages = [] page_results = self.notion_page_search(access_token) database_results = self.notion_database_search(access_token) # get page detail for page_result in page_results: page_id = page_result['id'] if 'Name' in page_result['properties']: if len(page_result['properties']['Name']['title']) > 0: page_name = page_result['properties']['Name']['title'][0]['plain_text'] else: page_name = 'Untitled' elif 'title' in page_result['properties']: if len(page_result['properties']['title']['title']) > 0: page_name = page_result['properties']['title']['title'][0]['plain_text'] else: page_name = 'Untitled' elif 'Title' in page_result['properties']: if len(page_result['properties']['Title']['title']) > 0: page_name = page_result['properties']['Title']['title'][0]['plain_text'] else: page_name = 'Untitled' else: page_name = 'Untitled' page_icon = page_result['icon'] if page_icon: icon_type = page_icon['type'] if icon_type == 'external' or icon_type == 'file': url = page_icon[icon_type]['url'] icon = { 'type': 'url', 'url': url if url.startswith('http') else f'https://www.notion.so{url}' } else: icon = { 'type': 'emoji', 'emoji': page_icon[icon_type] } else: icon = None parent = page_result['parent'] parent_type = parent['type'] if parent_type == 'block_id': parent_id = self.notion_block_parent_page_id(access_token, parent[parent_type]) elif parent_type == 'workspace': parent_id = 'root' else: parent_id = parent[parent_type] page = { 'page_id': page_id, 'page_name': page_name, 'page_icon': icon, 'parent_id': parent_id, 'type': 'page' } pages.append(page) # get database detail for database_result in database_results: page_id = database_result['id'] if len(database_result['title']) > 0: page_name = database_result['title'][0]['plain_text'] else: page_name = 'Untitled' page_icon = database_result['icon'] if page_icon: icon_type = page_icon['type'] if icon_type == 'external' or icon_type == 'file': url = page_icon[icon_type]['url'] icon = { 'type': 'url', 'url': url if url.startswith('http') else f'https://www.notion.so{url}' } else: icon = { 'type': icon_type, icon_type: page_icon[icon_type] } else: icon = None parent = database_result['parent'] parent_type = parent['type'] if parent_type == 'block_id': parent_id = self.notion_block_parent_page_id(access_token, parent[parent_type]) elif parent_type == 'workspace': parent_id = 'root' else: parent_id = parent[parent_type] page = { 'page_id': page_id, 'page_name': page_name, 'page_icon': icon, 'parent_id': parent_id, 'type': 'database' } pages.append(page) return pages def notion_page_search(self, access_token: str): data = { 'filter': { "value": "page", "property": "object" } } headers = { 'Content-Type': 'application/json', 'Authorization': f"Bearer {access_token}", 'Notion-Version': '2022-06-28', } response = requests.post(url=self._NOTION_PAGE_SEARCH, json=data, headers=headers) response_json = response.json() if 'results' in response_json: results = response_json['results'] else: results = [] return results def notion_block_parent_page_id(self, access_token: str, block_id: str): headers = { 'Authorization': f"Bearer {access_token}", 'Notion-Version': '2022-06-28', } response = requests.get(url=f'{self._NOTION_BLOCK_SEARCH}/{block_id}', headers=headers) response_json = response.json() parent = response_json['parent'] parent_type = parent['type'] if parent_type == 'block_id': return self.notion_block_parent_page_id(access_token, parent[parent_type]) return parent[parent_type] def notion_workspace_name(self, access_token: str): headers = { 'Authorization': f"Bearer {access_token}", 'Notion-Version': '2022-06-28', } response = requests.get(url=self._NOTION_BOT_USER, headers=headers) response_json = response.json() if 'object' in response_json and response_json['object'] == 'user': user_type = response_json['type'] user_info = response_json[user_type] if 'workspace_name' in user_info: return user_info['workspace_name'] return 'workspace' def notion_database_search(self, access_token: str): data = { 'filter': { "value": "database", "property": "object" } } headers = { 'Content-Type': 'application/json', 'Authorization': f"Bearer {access_token}", 'Notion-Version': '2022-06-28', } response = requests.post(url=self._NOTION_PAGE_SEARCH, json=data, headers=headers) response_json = response.json() if 'results' in response_json: results = response_json['results'] else: results = [] return results def get_oauth_providers(): with current_app.app_context(): notion_oauth = NotionOAuth(client_id=current_app.config.get('NOTION_CLIENT_ID'), client_secret=current_app.config.get( 'NOTION_CLIENT_SECRET'), redirect_uri=current_app.config.get( 'CONSOLE_API_URL') + '/console/api/oauth/data-source/callback/notion') OAUTH_PROVIDERS = { 'notion': notion_oauth } return OAUTH_PROVIDERS
null
16,996
import logging from datetime import datetime from typing import Optional import requests from flask import current_app, redirect, request from flask_restful import Resource from constants.languages import languages from extensions.ext_database import db from libs.oauth import GitHubOAuth, GoogleOAuth, OAuthUserInfo from models.account import Account, AccountStatus from services.account_service import AccountService, RegisterService, TenantService from .. import api class GitHubOAuth(OAuth): _AUTH_URL = 'https://github.com/login/oauth/authorize' _TOKEN_URL = 'https://github.com/login/oauth/access_token' _USER_INFO_URL = 'https://api.github.com/user' _EMAIL_INFO_URL = 'https://api.github.com/user/emails' def get_authorization_url(self): params = { 'client_id': self.client_id, 'redirect_uri': self.redirect_uri, 'scope': 'user:email' # Request only basic user information } return f"{self._AUTH_URL}?{urllib.parse.urlencode(params)}" def get_access_token(self, code: str): data = { 'client_id': self.client_id, 'client_secret': self.client_secret, 'code': code, 'redirect_uri': self.redirect_uri } headers = {'Accept': 'application/json'} response = requests.post(self._TOKEN_URL, data=data, headers=headers) response_json = response.json() access_token = response_json.get('access_token') if not access_token: raise ValueError(f"Error in GitHub OAuth: {response_json}") return access_token def get_raw_user_info(self, token: str): headers = {'Authorization': f"token {token}"} response = requests.get(self._USER_INFO_URL, headers=headers) response.raise_for_status() user_info = response.json() email_response = requests.get(self._EMAIL_INFO_URL, headers=headers) email_info = email_response.json() primary_email = next((email for email in email_info if email['primary'] == True), None) return {**user_info, 'email': primary_email['email']} def _transform_user_info(self, raw_info: dict) -> OAuthUserInfo: email = raw_info.get('email') if not email: email = f"{raw_info['id']}+{raw_info['login']}@users.noreply.github.com" return OAuthUserInfo( id=str(raw_info['id']), name=raw_info['name'], email=email ) class GoogleOAuth(OAuth): _AUTH_URL = 'https://accounts.google.com/o/oauth2/v2/auth' _TOKEN_URL = 'https://oauth2.googleapis.com/token' _USER_INFO_URL = 'https://www.googleapis.com/oauth2/v3/userinfo' def get_authorization_url(self): params = { 'client_id': self.client_id, 'response_type': 'code', 'redirect_uri': self.redirect_uri, 'scope': 'openid email' } return f"{self._AUTH_URL}?{urllib.parse.urlencode(params)}" def get_access_token(self, code: str): data = { 'client_id': self.client_id, 'client_secret': self.client_secret, 'code': code, 'grant_type': 'authorization_code', 'redirect_uri': self.redirect_uri } headers = {'Accept': 'application/json'} response = requests.post(self._TOKEN_URL, data=data, headers=headers) response_json = response.json() access_token = response_json.get('access_token') if not access_token: raise ValueError(f"Error in Google OAuth: {response_json}") return access_token def get_raw_user_info(self, token: str): headers = {'Authorization': f"Bearer {token}"} response = requests.get(self._USER_INFO_URL, headers=headers) response.raise_for_status() return response.json() def _transform_user_info(self, raw_info: dict) -> OAuthUserInfo: return OAuthUserInfo( id=str(raw_info['sub']), name=None, email=raw_info['email'] ) def get_oauth_providers(): with current_app.app_context(): github_oauth = GitHubOAuth(client_id=current_app.config.get('GITHUB_CLIENT_ID'), client_secret=current_app.config.get( 'GITHUB_CLIENT_SECRET'), redirect_uri=current_app.config.get( 'CONSOLE_API_URL') + '/console/api/oauth/authorize/github') google_oauth = GoogleOAuth(client_id=current_app.config.get('GOOGLE_CLIENT_ID'), client_secret=current_app.config.get( 'GOOGLE_CLIENT_SECRET'), redirect_uri=current_app.config.get( 'CONSOLE_API_URL') + '/console/api/oauth/authorize/google') OAUTH_PROVIDERS = { 'github': github_oauth, 'google': google_oauth } return OAUTH_PROVIDERS
null
16,997
import logging from datetime import datetime from typing import Optional import requests from flask import current_app, redirect, request from flask_restful import Resource from constants.languages import languages from extensions.ext_database import db from libs.oauth import GitHubOAuth, GoogleOAuth, OAuthUserInfo from models.account import Account, AccountStatus from services.account_service import AccountService, RegisterService, TenantService from .. import api def _get_account_by_openid_or_email(provider: str, user_info: OAuthUserInfo) -> Optional[Account]: languages = ['en-US', 'zh-Hans', 'pt-BR', 'es-ES', 'fr-FR', 'de-DE', 'ja-JP', 'ko-KR', 'ru-RU', 'it-IT', 'uk-UA'] db = SQLAlchemy() class OAuthUserInfo: class AccountService: def load_user(user_id: str) -> Account: def get_account_jwt_token(account): def authenticate(email: str, password: str) -> Account: def update_account_password(account, password, new_password): def create_account(email: str, name: str, interface_language: str, password: str = None, interface_theme: str = 'light', timezone: str = 'America/New_York', ) -> Account: def link_account_integrate(provider: str, open_id: str, account: Account) -> None: def close_account(account: Account) -> None: def update_account(account, **kwargs): def update_last_login(account: Account, request) -> None: class RegisterService: def _get_invitation_token_key(cls, token: str) -> str: def register(cls, email, name, password: str = None, open_id: str = None, provider: str = None, language: str = None, status: AccountStatus = None) -> Account: def invite_new_member(cls, tenant: Tenant, email: str, language: str, role: str = 'normal', inviter: Account = None) -> str: def generate_invite_token(cls, tenant: Tenant, account: Account) -> str: def revoke_token(cls, workspace_id: str, email: str, token: str): def get_invitation_if_token_valid(cls, workspace_id: str, email: str, token: str) -> Optional[dict[str, Any]]: def _get_invitation_by_token(cls, token: str, workspace_id: str, email: str) -> Optional[dict[str, str]]: def _generate_account(provider: str, user_info: OAuthUserInfo): # Get account by openid or email. account = _get_account_by_openid_or_email(provider, user_info) if not account: # Create account account_name = user_info.name if user_info.name else 'Dify' account = RegisterService.register( email=user_info.email, name=account_name, password=None, open_id=user_info.id, provider=provider ) # Set interface language preferred_lang = request.accept_languages.best_match(languages) if preferred_lang and preferred_lang in languages: interface_language = preferred_lang else: interface_language = languages[0] account.interface_language = interface_language db.session.commit() # Link account AccountService.link_account_integrate(provider, user_info.id, account) return account
null
16,998
import json from functools import wraps from flask import abort, current_app, request from flask_login import current_user from controllers.console.workspace.error import AccountNotInitializedError from services.feature_service import FeatureService from services.operation_service import OperationService class AccountNotInitializedError(BaseHTTPException): error_code = 'account_not_initialized' description = "The account has not been initialized yet. Please proceed with the initialization process first." code = 400 def account_initialization_required(view): @wraps(view) def decorated(*args, **kwargs): # check account initialization account = current_user if account.status == 'uninitialized': raise AccountNotInitializedError() return view(*args, **kwargs) return decorated
null
16,999
import json from functools import wraps from flask import abort, current_app, request from flask_login import current_user from controllers.console.workspace.error import AccountNotInitializedError from services.feature_service import FeatureService from services.operation_service import OperationService def only_edition_cloud(view): @wraps(view) def decorated(*args, **kwargs): if current_app.config['EDITION'] != 'CLOUD': abort(404) return view(*args, **kwargs) return decorated
null
17,000
import json from functools import wraps from flask import abort, current_app, request from flask_login import current_user from controllers.console.workspace.error import AccountNotInitializedError from services.feature_service import FeatureService from services.operation_service import OperationService def only_edition_self_hosted(view): @wraps(view) def decorated(*args, **kwargs): if current_app.config['EDITION'] != 'SELF_HOSTED': abort(404) return view(*args, **kwargs) return decorated
null