id int64 0 190k | prompt stringlengths 21 13.4M | docstring stringlengths 1 12k ⌀ |
|---|---|---|
188,026 | import os
import os.path
import math
import torch
import torch.utils.data
import numpy as np
import librosa
from librosa.filters import mel as librosa_mel_fn
import torch.nn.functional as F
The provided code snippet includes necessary dependencies for implementing the `pitch_shifting` function. Write a Python function... | shift the audio pitch. |
188,027 | import os
import os.path
import math
import torch
import torch.utils.data
import numpy as np
import librosa
from librosa.filters import mel as librosa_mel_fn
import torch.nn.functional as F
The provided code snippet includes necessary dependencies for implementing the `speed_change` function. Write a Python function `... | change the speed of input audio. Note that we return the speed_rate to change the speed of landmarks or videos. Args: data: [n,] audio clip landmark: [m, pts, 2] aligned landmarks with audio if existed. |
188,028 | import os
import os.path
import math
import torch
import torch.utils.data
import numpy as np
import librosa
from librosa.filters import mel as librosa_mel_fn
import torch.nn.functional as F
def world_augment(wav, sr, op):
f0, sp, ap = pw.wav2world(wav.astype(np.float64), sr)
op = op if op is not None else np.r... | null |
188,029 | import os
import os.path
import math
import torch
import torch.utils.data
import numpy as np
import librosa
from librosa.filters import mel as librosa_mel_fn
import torch.nn.functional as F
def sox_augment(wav, sr, tempo_ratio=1.0, op=None):
aug_choice = op if op is not None else np.random.randint(low=1, high=8)
... | null |
188,030 | import os
import os.path
import math
import torch
import torch.utils.data
import numpy as np
import librosa
from librosa.filters import mel as librosa_mel_fn
import torch.nn.functional as F
def sox_augment_v2(wav, sr, op=None):
aug_choice = op if op is not None else np.random.randint(low=1, high=5)
hop_length... | null |
188,031 | import os
import os.path
import math
import torch
import torch.utils.data
import numpy as np
import librosa
from librosa.filters import mel as librosa_mel_fn
import torch.nn.functional as F
def audio_output_augment(wav, sr, op=None):
aug_choice = op if op is not None else np.random.randint(low=1, high=4)
tfm ... | null |
188,032 | import os
import os.path
import math
import torch
import torch.utils.data
import numpy as np
import librosa
from librosa.filters import mel as librosa_mel_fn
import torch.nn.functional as F
def audio_time_augment(wav, sr, time_scale):
tfm = sox.Transformer()
tfm.tempo(time_scale, 's', quick=False)
aug_wav... | null |
188,033 | import os
import os.path
import math
import torch
import torch.utils.data
import numpy as np
import librosa
from librosa.filters import mel as librosa_mel_fn
import torch.nn.functional as F
def prepare_noises(scp_file, root=None, sampline_rate=None, ignore_class=None):
noises = []
print('Loading augmentation n... | null |
188,034 | import os
import os.path
import math
import torch
import torch.utils.data
import numpy as np
import librosa
from librosa.filters import mel as librosa_mel_fn
import torch.nn.functional as F
def add_gauss_noise(wav, noise_std=0.03, max_wav_value=1.0):
if isinstance(wav, np.ndarray):
wav = torch.tensor(wav.co... | null |
188,035 | import torch
import torch.nn as nn
from torch.autograd import Variable
import math
import torch.nn.functional as F
from torchvision import models
The provided code snippet includes necessary dependencies for implementing the `Sample_GMM` function. Write a Python function `def Sample_GMM(gmm_params, ncenter, ndim, weig... | Sample values from a given a GMM distribution. Args: gmm_params: [b, target_length, (2 * ndim + 1) * ncenter], including the distribution weights, average and sigma ncenter: numbers of gaussian distribution ndim: dimension of each gaussian distribution weight_smooth: float, smooth the gaussian distribution weights sigm... |
188,036 | import torch
import torch.nn as nn
from torch.autograd import Variable
import math
import torch.nn.functional as F
from torchvision import models
def gram_matrix(input):
a, b, c, d = input.size() # a=batch size(=1)
# b=number of feature maps
# (c,d)=dimensions of a f. map (N=c*d)
features = input.view... | null |
188,037 | import os
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.optim import lr_scheduler
from torch.nn import init
import functools
from torch.nn.parallel import DistributedDataParallel as DDP
from torch.nn.utils.rnn import pad_packed_sequence, pack_padded_sequence
def conv2... | null |
188,038 | import os
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.optim import lr_scheduler
from torch.nn import init
import functools
from torch.nn.parallel import DistributedDataParallel as DDP
from torch.nn.utils.rnn import pad_packed_sequence, pack_padded_sequence
def init_w... | Initialize a network: 1. register CPU/GPU device (with multi-GPU support); 2. initialize the network weights Parameters: net (network) -- the network to be initialized init_type (str) -- the name of an initialization method: normal | xavier | kaiming | orthogonal gain (float) -- scaling factor for normal, xavier and or... |
188,039 | import os
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.optim import lr_scheduler
from torch.nn import init
import functools
from torch.nn.parallel import DistributedDataParallel as DDP
from torch.nn.utils.rnn import pad_packed_sequence, pack_padded_sequence
The provi... | Return a learning rate scheduler Parameters: optimizer -- the optimizer of the network opt (option class) -- stores all the experiment flags; needs to be a subclass of BaseOptions. opt.lr_policy is the name of learning rate policy: linear | step | plateau | cosine For 'linear', we keep the same learning rate for the fi... |
188,040 | import os
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.optim import lr_scheduler
from torch.nn import init
import functools
from torch.nn.parallel import DistributedDataParallel as DDP
from torch.nn.utils.rnn import pad_packed_sequence, pack_padded_sequence
def weigh... | null |
188,041 | import os
import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.optim import lr_scheduler
from torch.nn import init
import functools
from torch.nn.parallel import DistributedDataParallel as DDP
from torch.nn.utils.rnn import pad_packed_sequence, pack_padded_sequence
def print... | null |
188,042 | from __future__ import print_function
import torch
import numpy as np
from PIL import Image
import inspect, re
import numpy as np
import os
import collections
from PIL import Image
import cv2
from collections import OrderedDict
from . import flow_viz
def tensor2im(image_tensor, imtype=np.uint8, normalize=True):
if... | null |
188,043 | from __future__ import print_function
import torch
import numpy as np
from PIL import Image
import inspect, re
import numpy as np
import os
import collections
from PIL import Image
import cv2
from collections import OrderedDict
from . import flow_viz
def tensor2flow(flo, imtype=np.uint8):
flo = flo[0].permute(1,2,... | null |
188,044 | from __future__ import print_function
import torch
import numpy as np
from PIL import Image
import inspect, re
import numpy as np
import os
import collections
from PIL import Image
import cv2
from collections import OrderedDict
from . import flow_viz
def add_dummy_to_tensor(tensors, add_size=0):
if add_size == 0 o... | null |
188,045 | from __future__ import print_function
import torch
import numpy as np
from PIL import Image
import inspect, re
import numpy as np
import os
import collections
from PIL import Image
import cv2
from collections import OrderedDict
from . import flow_viz
def remove_dummy_from_tensor(tensors, remove_size=0):
if remove_... | null |
188,046 | from __future__ import print_function
import torch
import numpy as np
from PIL import Image
import inspect, re
import numpy as np
import os
import collections
from PIL import Image
import cv2
from collections import OrderedDict
from . import flow_viz
def save_image(image_numpy, image_path):
image_pil = Image.froma... | null |
188,047 | from __future__ import print_function
import torch
import numpy as np
from PIL import Image
import inspect, re
import numpy as np
import os
import collections
from PIL import Image
import cv2
from collections import OrderedDict
from . import flow_viz
def print_numpy(x, val=True, shp=False):
x = x.astype(np.float64... | null |
188,048 | from __future__ import print_function
import torch
import numpy as np
from PIL import Image
import inspect, re
import numpy as np
import os
import collections
from PIL import Image
import cv2
from collections import OrderedDict
from . import flow_viz
def mkdir(path):
def mkdirs(paths):
if isinstance(paths, list) a... | null |
188,049 |
def sort_readme():
with open("README.md", "r", encoding="utf-8") as f:
all = f.readlines()
table_start = "|Category|Company|Description|GitHub Stars|Alternative to|\n"
table_end = "<!-- END STARTUP LIST -->\n"
idx = all.index(table_start)
idx_end = all.index(table_end)
find_name = ... | null |
188,050 | import yaml
import os
from add_company import add_new_company
def parse_all_yamls():
arr = []
for filename in os.listdir("submissions"):
if filename.endswith(".yaml"):
with open(f"submissions/{filename}", "r") as file:
obj = yaml.load(file, yaml.Loader)
arr.append... | null |
188,051 | import yaml
def read_readme():
with open("README.md", "r", encoding="utf-8") as f:
all = f.readlines()
table_start = "|Category|Company|Description|GitHub Stars|Alternative to|\n"
table_end = "<!-- END STARTUP LIST -->\n"
idx = all.index(table_start)
idx_end = all.index(ta... | null |
188,052 | import yaml
def parse_line(line: str):
arr = line.split("|")
category = arr[0]
name = arr[1].split("]")[0][1:]
website = arr[1].split("]")[1][1:-1]
description = arr[2]
github = arr[3].split(">")[0].split("href=")[1]
alts = list(map(lambda x: x.strip().split("]")[0][1:], arr[4].split(",")))... | null |
188,053 |
def count_companies_in_readme():
with open('README.md', 'r', encoding='utf-8') as f:
all = f.readlines()
table_start = '|Category|Company|Description|GitHub Stars|Alternative to|\n'
idx = all.index(table_start)
return len(all[idx + 2: -1]) | null |
188,054 | import yaml
import os
appl = lambda x: SPECIAL_MAPPING[x] if x in SPECIAL_MAPPING else x
def get_all_companies():
arr = []
for filename in os.listdir("submissions"):
if filename.endswith(".yaml"):
with open(f"submissions/{filename}", "r", encoding="utf-8") as file:
obj = yam... | null |
188,055 | import yaml
import os
def get_all_categories(arr):
categories = set()
for obj in arr:
categories.add(obj["category"])
return categories | null |
188,056 | import yaml
import os
def create_website_directories(categories):
for category in categories:
if not os.path.exists(f"website/docs/{category}"):
os.mkdir(f"website/docs/{category}") | null |
188,057 | import yaml
import os
def remove_github_com(s: str):
return s.replace("https://github.com/", "")
def remove_https(s: str):
s = s.replace("https://", "")
s = s.replace("http://", "")
return s.strip("/")
markdown_template = """
# {company_name}
<a href="{link}"><img src="https://icons.duckduckgo.com/ip3/... | null |
188,058 | def add_new_company(
category, company_name, description, link, gh_link, alts_names, alts_links
):
with open("README.md", "r", encoding="utf-8") as f:
all = f.readlines()
table_start = "|Category|Company|Description|GitHub Stars|Alternative to|\n"
table_end = "<!-- END STARTUP LIST -->\n"
id... | null |
188,059 | import argparse
import copy
import glob
import os
import random
import time
import numpy as np
import torch
import wandb
from mmengine import Config
from torch.nn.parallel import DistributedDataParallel as DDP
from torch.utils.data import DataLoader, DistributedSampler
from tqdm import tqdm
from transformers import (
... | null |
188,060 | import argparse
import copy
import glob
import os
import random
import time
import numpy as np
import torch
import wandb
from mmengine import Config
from torch.nn.parallel import DistributedDataParallel as DDP
from torch.utils.data import DataLoader, DistributedSampler
from tqdm import tqdm
from transformers import (
... | null |
188,063 | import os
import torch
try:
import horovod.torch as hvd
except ImportError:
hvd = None
def is_using_distributed():
if "WORLD_SIZE" in os.environ:
return int(os.environ["WORLD_SIZE"]) > 1
if "SLURM_NTASKS" in os.environ:
return int(os.environ["SLURM_NTASKS"]) > 1
return False
def worl... | null |
188,064 | import os
import torch
def is_rank0():
if not torch.distributed.is_initialized():
return True
return torch.distributed.get_rank() == 0 | null |
188,065 | import time
from contextlib import suppress
import torch
from tqdm import tqdm
def get_cast_dtype(precision: str):
def get_autocast(precision):
class AverageMeter(object):
def __init__(self):
def reset(self):
def update(self, val, n=1):
def train_one_epoch(
args,
model,
epoch,
laion_load... | null |
188,066 | import time
from contextlib import suppress
import torch
from tqdm import tqdm
def get_checkpoint(model: torch.nn.Module):
state_dict = model.state_dict()
parameters = {k: v for k, v in model.named_parameters()}
# remove duplicate parameters
duplicate_keys = set(state_dict.keys()) - set(parameters.keys... | null |
188,067 | import copy
import json
import os
import random
from collections import defaultdict
import torch
import torch.nn.functional as F
from PIL import Image
from .vqa_dataset import VQADataset
class NLVRv1Dataset(VQADataset):
"""Visual Reasoning Dataset."""
def __init__(self, tokenizer, vis_processor, vis_root, ann_p... | null |
188,068 | import copy
import json
import os
import random
from collections import defaultdict
import torch
import torch.nn.functional as F
from PIL import Image
from .vqa_dataset import VQADataset
class NLVRv2Dataset(VQADataset):
def __init__(self, tokenizer, vis_processor, vis_root, ann_paths, **kwargs):
def parse_ann... | null |
188,069 | import json
import os
import random
from collections import defaultdict
from PIL import Image
from .vqa_dataset import VQADataset
class GQADataset(VQADataset):
"""Visual Reasoning Dataset."""
def __init__(self, tokenizer, vis_processor, vis_root, ann_paths, **kwargs):
super().__init__(tokenizer, vis_pro... | null |
188,070 | import copy
import json
import numpy as np
from torch.utils.data import Dataset
from transformers import LlamaTokenizer
class DollyDataset(Dataset):
"""Each line of the annotation file is a json object with the following fields:
{
"instruction": "What is a dispersive prism?",
"context": "In opti... | null |
188,071 | import random
from .vqa_dataset import VQADataset
class AOKVQADataset(VQADataset):
def __init__(self, tokenizer, vis_processor, vis_root, ann_paths, **kwargs):
super().__init__(tokenizer, vis_processor, vis_root, ann_paths, **kwargs)
def process_text(self, ann):
question = ann["question"]
... | null |
188,072 | import json
import os
import random
from PIL import Image
from .vqa_dataset import VQADataset, VQAPrompter
class CcSbuAlignDataset(VQADataset):
def __init__(self, tokenizer, vis_processor, vis_root, ann_paths, add_eos=True, ignore_instruction=True):
self.tokenizer = tokenizer
self.vis_root = vis_roo... | null |
188,073 | import numpy as np
import torch
from .alpaca_gpt4_dataset import AlpacaGPT4Dataset
from .aokvqa_dataset import AOKVQADataset
from .cc_sbu_align_dataset import CcSbuAlignDataset
from .clevr_dataset import CLEVRDataset
from .coco_caption_dataset import COCOCaptionDataset
from .dial_dataset import DialDataset
from .... | null |
188,074 | import json
import os
import random
from collections import defaultdict
from PIL import Image
from .vqa_dataset import VQADataset
class CLEVRDataset(VQADataset):
"""Visual Reasoning Dataset. It also contains Dialog.
Note: The image is a little bit simple. with several objects and simple background.
"""
... | null |
188,078 | import open_clip
import torch
import torch.nn as nn
from bigmodelvis import Visualization
from peft import LoraConfig, get_peft_model
from transformers import LlamaForCausalLM, LlamaTokenizer
from .flamingo import Flamingo
from .flamingo_lm import FlamingoLMMixin
from .utils import extend_instance
def _infer_decoder_la... | Initialize a Flamingo model from a pretrained vision encoder and language encoder. Appends special tokens to the tokenizer and freezes backbones. Args: clip_vision_encoder_path (str): path to pretrained clip model (e.g. "ViT-B-32") clip_vision_encoder_pretrained (str): name of pretraining dataset for clip model (e.g. "... |
188,079 | from .open_flamingo import create_model_and_transforms as create_open_flamingo_model_and_transforms
import torch.nn as nn
from transformers import LlamaTokenizer, LlamaForCausalLM
def create_model_and_transforms(
model_name: str,
clip_vision_encoder_path: str,
clip_vision_encoder_pretrained: str,
lang_... | null |
188,080 | from .open_flamingo import create_model_and_transforms as create_open_flamingo_model_and_transforms
import torch.nn as nn
from transformers import LlamaTokenizer, LlamaForCausalLM
def create_toy_model_and_transforms(
model_name: str,
clip_vision_encoder_path: str,
clip_vision_encoder_pretrained: str,
l... | null |
188,081 | import os
import gradio as gr
import torch
from PIL import Image
from mmgpt.models.builder import create_model_and_transforms
TEMPLATE = "Below is an instruction that describes a task. Write a response that appropriately completes the request."
class PromptGenerator:
def __init__(
self,
pro... | null |
188,082 | import nonebot
from nonebot import Driver
from configs.path_config import IMAGE_PATH
from services.log import logger
from utils.image_template import help_template
from utils.image_utils import BuildImage, build_sort_image, group_image, text2image
from utils.manager import group_manager, plugin_data_manager
from utils.... | null |
188,083 | import nonebot
from nonebot import Driver
from configs.path_config import IMAGE_PATH
from services.log import logger
from utils.image_template import help_template
from utils.image_utils import BuildImage, build_sort_image, group_image, text2image
from utils.manager import group_manager, plugin_data_manager
from utils.... | 创建管理员帮助图片 |
188,084 | import os
import shutil
import zipfile
from pathlib import Path
from typing import Tuple, Union
import ujson as json
from configs.path_config import DATA_PATH, TEMP_PATH
from services import logger
from utils.http_utils import AsyncHttpx
from utils.image_utils import BuildImage, text2image
from utils.manager import plu... | 安装插件 :param name: 插件名或下标 |
188,085 | import os
import shutil
import zipfile
from pathlib import Path
from typing import Tuple, Union
import ujson as json
from configs.path_config import DATA_PATH, TEMP_PATH
from services import logger
from utils.http_utils import AsyncHttpx
from utils.image_utils import BuildImage, text2image
from utils.manager import plu... | 删除插件 :param name: 插件名或下标 |
188,086 | from typing import Optional
from configs.path_config import IMAGE_PATH
from utils.image_utils import BuildImage
from utils.manager import admin_manager, plugin_data_manager, plugins2settings_manager
from ._utils import HelpImageBuild
class HelpImageBuild:
def __init__(self):
self._data: Dict[str, PluginDat... | 说明: 生成帮助图片 参数: :param group_id: 群号 |
188,087 | from typing import Optional
from configs.path_config import IMAGE_PATH
from utils.image_utils import BuildImage
from utils.manager import admin_manager, plugin_data_manager, plugins2settings_manager
from ._utils import HelpImageBuild
background = IMAGE_PATH / "background" / "0.png"
class BuildImage:
"""
快捷生成图片... | 说明: 获取功能的帮助信息 参数: :param msg: 功能cmd :param is_super: 是否为超级用户 |
188,088 | import random
from asyncio.exceptions import TimeoutError
import nonebot
from nonebot.adapters.onebot.v11 import Bot
from nonebot.drivers import Driver
from configs.path_config import TEXT_PATH
from models.bag_user import BagUser
from models.group_info import GroupInfo
from services.log import logger
from utils.http_ut... | 版本某些需要的变换 |
188,089 | import random
from asyncio.exceptions import TimeoutError
import nonebot
from nonebot.adapters.onebot.v11 import Bot
from nonebot.drivers import Driver
from configs.path_config import TEXT_PATH
from models.bag_user import BagUser
from models.group_info import GroupInfo
from services.log import logger
from utils.http_ut... | null |
188,090 | from nonebot import on_command
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from configs.config import Config
from services.log import logger
from utils.manager import (
group_manager,
plugins2block_manager,
plugins2cd_manager,
plugins2settings_manager,
)
from utils.utils impo... | null |
188,091 | from nonebot import on_command
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from configs.config import Config
from services.log import logger
from utils.manager import (
group_manager,
plugins2block_manager,
plugins2cd_manager,
plugins2settings_manager,
)
from utils.utils impo... | null |
188,092 | from nonebot import on_command
from nonebot.adapters.onebot.v11 import Bot, MessageEvent
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from models.friend_user import FriendUser
from models.group_info import GroupInfo
from services.log import logger
update_group_info = on_command(
"更新群信息", ... | null |
188,093 | from nonebot import on_command
from nonebot.adapters.onebot.v11 import Bot, MessageEvent
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from models.friend_user import FriendUser
from models.group_info import GroupInfo
from services.log import logger
update_friend_info = on_command(
"更新好友信息"... | null |
188,094 | import asyncio
import os
import time
from nonebot import on_command
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from configs.path_config import TEMP_PATH
from services.log import logger
from utils.manager import resources_manager
from utils.utils import scheduler
clear_data = on_command(
... | null |
188,095 | import asyncio
import os
import time
from nonebot import on_command
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from configs.path_config import TEMP_PATH
from services.log import logger
from utils.manager import resources_manager
from utils.utils import scheduler
logger.debug("开始清理临时文件...... | null |
188,096 | import asyncio
from nonebot import on_command
from nonebot.adapters.onebot.v11 import Bot, GroupMessageEvent, Message, MessageEvent
from nonebot.params import CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from tortoise import Tortoise
from services.db_context import TestSQL
from ser... | null |
188,097 | import asyncio
from nonebot import on_command
from nonebot.adapters.onebot.v11 import Bot, GroupMessageEvent, Message, MessageEvent
from nonebot.params import CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from tortoise import Tortoise
from services.db_context import TestSQL
from ser... | null |
188,098 | from typing import Tuple
from nonebot import on_command
from nonebot.adapters.onebot.v11 import Bot, Message, MessageEvent
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from models.group_info import GroupInfo
from services.log import logger
from u... | null |
188,099 | from typing import Tuple
from nonebot import on_command
from nonebot.adapters.onebot.v11 import Bot, Message, MessageEvent
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from models.group_info import GroupInfo
from services.log import logger
from u... | null |
188,100 | from typing import Tuple
from nonebot import on_command
from nonebot.adapters.onebot.v11 import Bot, Message, MessageEvent
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from models.group_info import GroupInfo
from services.log import logger
from u... | null |
188,101 | from typing import Tuple
from nonebot import on_command
from nonebot.adapters.onebot.v11 import Bot, Message, MessageEvent
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from models.group_info import GroupInfo
from services.log import logger
from u... | null |
188,102 | from typing import Tuple
from nonebot import on_command
from nonebot.adapters.onebot.v11 import Bot, Message, MessageEvent
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from models.group_info import GroupInfo
from services.log import logger
from u... | null |
188,103 | from typing import Tuple
from nonebot import on_command
from nonebot.adapters.onebot.v11 import Bot, Message, MessageEvent
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from models.group_info import GroupInfo
from services.log import logger
from u... | null |
188,104 | from typing import List, Tuple
from nonebot import on_command
from nonebot.adapters.onebot.v11 import Bot, GroupMessageEvent, Message, MessageEvent
from nonebot.exception import ActionFailed
from nonebot.params import CommandArg
from nonebot.permission import SUPERUSER
from models.level_user import LevelUser
from servi... | null |
188,105 | from typing import Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupMessageEvent,
Message,
MessageEvent,
)
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from configs.conf... | null |
188,106 | from typing import Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupMessageEvent,
Message,
MessageEvent,
)
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from configs.conf... | null |
188,107 | from typing import Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupMessageEvent,
Message,
MessageEvent,
)
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from configs.conf... | null |
188,108 | from typing import Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupMessageEvent,
Message,
MessageEvent,
)
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from configs.conf... | null |
188,109 | from typing import Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupMessageEvent,
Message,
MessageEvent,
)
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from configs.conf... | null |
188,110 | from typing import Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupMessageEvent,
Message,
MessageEvent,
)
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from configs.conf... | null |
188,111 | from typing import Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupMessageEvent,
Message,
MessageEvent,
)
from nonebot.params import Command, CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from configs.conf... | null |
188,112 | from typing import List
from nonebot import on_command
from nonebot.adapters.onebot.v11 import GroupMessageEvent, Message
from nonebot.adapters.onebot.v11.permission import GROUP
from nonebot.params import CommandArg
from configs.config import Config
from services.log import logger
from utils.depends import ImageList, ... | null |
188,113 | import time
from nonebot.adapters.onebot.v11 import Event
from services.log import logger
from utils.manager import group_manager, plugins2settings_manager
from utils.utils import get_message_text
cmd = []
v = time.time()
class logger:
TEMPLATE_A = "{}"
TEMPLATE_B = "[<u><c>{}</c></u>]: {}"
TEMPLATE_C = "... | 说明: 检测文本是否是关闭功能命令 参数: :param event: pass |
188,114 | from typing import Any, Tuple
from nonebot import on_command, on_message, on_regex
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupMessageEvent,
Message,
MessageEvent,
PrivateMessageEvent,
)
from nonebot.params import CommandArg, RegexGroup
from nonebot.permission import SUPERUSER
fro... | null |
188,115 | from typing import Any, Tuple
from nonebot import on_command, on_message, on_regex
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupMessageEvent,
Message,
MessageEvent,
PrivateMessageEvent,
)
from nonebot.params import CommandArg, RegexGroup
from nonebot.permission import SUPERUSER
fro... | null |
188,116 | from typing import Any, Tuple
from nonebot import on_command, on_message, on_regex
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupMessageEvent,
Message,
MessageEvent,
PrivateMessageEvent,
)
from nonebot.params import CommandArg, RegexGroup
from nonebot.permission import SUPERUSER
fro... | null |
188,117 | from typing import Any, Tuple
from nonebot import on_command, on_message, on_regex
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupMessageEvent,
Message,
MessageEvent,
PrivateMessageEvent,
)
from nonebot.params import CommandArg, RegexGroup
from nonebot.permission import SUPERUSER
fro... | null |
188,118 | from nonebot import on_notice
from nonebot.adapters.onebot.v11 import GroupAdminNoticeEvent
from configs.config import Config
from models.group_member_info import GroupInfoUser
from models.level_user import LevelUser
from services.log import logger
class GroupInfoUser(Model):
id = fields.IntField(pk=True, generat... | null |
188,119 | from nonebot import get_bots
from services.log import logger
from utils.utils import scheduler
from ._data_source import update_member_info
class logger:
TEMPLATE_A = "{}"
TEMPLATE_B = "[<u><c>{}</c></u>]: {}"
TEMPLATE_C = "用户[<u><e>{}</e></u>] 触发 [<u><c>{}</c></u>]: {}"
TEMPLATE_D = "群聊[<u><e>{}</e><... | null |
188,120 | from nonebot import on_command, on_notice
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupIncreaseNoticeEvent,
GroupMessageEvent,
)
from services.log import logger
from ._data_source import update_member_info
refresh_member_group = on_command(
"更新群组成员列表", aliases={"更新群组成员信息"}, permission=... | null |
188,121 | from nonebot import on_command, on_notice
from nonebot.adapters.onebot.v11 import (
GROUP,
Bot,
GroupIncreaseNoticeEvent,
GroupMessageEvent,
)
from services.log import logger
from ._data_source import update_member_info
class logger:
def info(
cls,
info: str,
co... | null |
188,122 | import random
from typing import Any, List, Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import Bot, GroupMessageEvent, Message, MessageEvent
from nonebot.internal.matcher import Matcher
from nonebot.internal.params import Depends
from nonebot.params import CommandArg, RegexGroup
from... | 说明: 检查名称是否合法 |
188,123 | import random
from typing import Any, List, Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import Bot, GroupMessageEvent, Message, MessageEvent
from nonebot.internal.matcher import Matcher
from nonebot.internal.params import Depends
from nonebot.params import CommandArg, RegexGroup
from... | null |
188,124 | import random
from typing import Any, List, Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import Bot, GroupMessageEvent, Message, MessageEvent
from nonebot.internal.matcher import Matcher
from nonebot.internal.params import Depends
from nonebot.params import CommandArg, RegexGroup
from... | null |
188,125 | import random
from typing import Any, List, Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import Bot, GroupMessageEvent, Message, MessageEvent
from nonebot.internal.matcher import Matcher
from nonebot.internal.params import Depends
from nonebot.params import CommandArg, RegexGroup
from... | null |
188,126 | import random
from typing import Any, List, Tuple
from nonebot import on_command, on_regex
from nonebot.adapters.onebot.v11 import Bot, GroupMessageEvent, Message, MessageEvent
from nonebot.internal.matcher import Matcher
from nonebot.internal.params import Depends
from nonebot.params import CommandArg, RegexGroup
from... | null |
188,127 | import time
from nonebot.adapters.onebot.v11 import (
Bot,
Event,
GroupMessageEvent,
Message,
MessageEvent,
PokeNotifyEvent,
PrivateMessageEvent,
)
from nonebot.exception import ActionFailed, IgnoredException
from nonebot.internal.matcher import Matcher
from configs.config import Config
from... | 说明: 发送信息 参数: :param msg: pass :param bot: pass :param event: pass |
188,128 | import asyncio
from typing import Optional
from nonebot.adapters.onebot.v11 import Bot, Event
from nonebot.matcher import Matcher
from nonebot.message import run_postprocessor
from nonebot.typing import T_State
from services.log import logger
from utils.manager import withdraw_message_manager
async def _withdraw_messag... | null |
188,129 | from nonebot.matcher import Matcher
from nonebot.message import run_preprocessor, IgnoredException
from nonebot.typing import T_State
from ._utils import status_message_manager
from nonebot.adapters.onebot.v11 import (
Bot,
MessageEvent,
PrivateMessageEvent,
GroupMessageEvent,
)
async def _(matcher: Ma... | null |
188,130 | from nonebot.matcher import Matcher
from nonebot.message import run_preprocessor, IgnoredException
from nonebot.typing import T_State
from ._utils import status_message_manager
from nonebot.adapters.onebot.v11 import (
Bot,
MessageEvent,
PrivateMessageEvent,
GroupMessageEvent,
)
status_message_manager ... | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.