repo_name stringlengths 7 71 | file_path stringlengths 5 118 | context list | import_statement stringlengths 45 12.5k | token_num int64 641 99.4k | cropped_code stringlengths 44 17k | all_code stringlengths 43 754k | next_line stringlengths 2 330 | gold_snippet_index int64 0 68 | created_at stringlengths 25 25 | level stringclasses 9 values |
|---|---|---|---|---|---|---|---|---|---|---|
google-research/semivl | model/vlm.py | [
{
"identifier": "aggregate_concept_predictions",
"path": "model/text_embeddings.py",
"snippet": "def aggregate_concept_predictions(pred, class_to_concept_idxs):\n B, _, H, W = pred.shape\n agg_pred = torch.zeros(B, len(class_to_concept_idxs), H, W, device=pred.device)\n for cls_i, conc_i in cla... | import numpy as np
import torch
import torch.nn.functional as F
from mmseg.models import builder
from mmseg.models.builder import SEGMENTORS
from mmseg.models.segmentors.encoder_decoder import EncoderDecoder
from model.text_embeddings import (aggregate_concept_predictions,
get_class_to_concept_idxs) | 1,186 | # Copyright 2023 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
@SEGMENTORS.register_module()
class VLM(EncoderDecoder):
def __init__(self,
freeze_backbone=False,
exclude_keys=None,
load_text_embedding=None,
load_mcc_text_embedding=None,
load_pl_text_embedding=None,
clip_encoder=None,
conv_encoder=None,
maskclip_class_filter=None,
maskclip_trust_head=None,
renorm_clip_img=False,
**args):
super(VLM, self).__init__(**args)
assert load_text_embedding == load_pl_text_embedding
assert maskclip_class_filter is None
assert maskclip_trust_head is None
self.local_iter = 0
self.clip_encoder = None
if clip_encoder is not None:
self.clip_encoder = builder.build_backbone(clip_encoder)
self.conv_encoder = None
if conv_encoder is not None:
self.conv_encoder = builder.build_backbone(conv_encoder)
self.load_text_embedding = load_text_embedding
self.decode_head.load_text_embedding = load_text_embedding
self.load_mcc_text_embedding = load_mcc_text_embedding
self.renorm_clip_img = renorm_clip_img
if renorm_clip_img:
print('Renormalize clip image.')
if self.load_mcc_text_embedding:
self.loaded_mcc_text_feat = np.load(self.load_mcc_text_embedding)
self.loaded_mcc_text_feat = torch.from_numpy(self.loaded_mcc_text_feat).float()
else:
raise NotImplementedError
if freeze_backbone:
self.freeze(self.backbone, exclude_keys=exclude_keys)
def renormalize_img_for_clip(self, img):
if not self.renorm_clip_img:
return img
loader_mean, loader_std = [0.485, 0.456, 0.406], [0.229, 0.224, 0.225]
clip_mean, clip_std = [0.48145466, 0.4578275, 0.40821073], [0.26862954, 0.26130258, 0.27577711]
loader_mean = torch.tensor(loader_mean, device=img.device).view(1, -1, 1, 1)
loader_std = torch.tensor(loader_std, device=img.device).view(1, -1, 1, 1)
clip_mean = torch.tensor(clip_mean, device=img.device).view(1, -1, 1, 1)
clip_std = torch.tensor(clip_std, device=img.device).view(1, -1, 1, 1)
return (img * loader_std + loader_mean - clip_mean) / clip_std
def freeze(self, model, exclude_keys=None):
for n, m in model.named_parameters():
m.requires_grad = False
if exclude_keys is not None:
assert isinstance(exclude_keys, list)
for k in exclude_keys:
if str(k) in n:
m.requires_grad = True
print(f'Finetune {n}')
def forward_maskclip(self, img, conf_tresh):
img = self.renormalize_img_for_clip(img)
self.clip_encoder.eval()
with torch.no_grad():
text_feat = self.loaded_mcc_text_feat.detach().to(img.device)
visual_feat, _ = self.clip_encoder(img)
visual_feat = visual_feat[-1]
dense_pred = F.conv2d(visual_feat, text_feat[:, :, None, None])
if dense_pred.shape[1] != self.num_classes:
| # Copyright 2023 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
@SEGMENTORS.register_module()
class VLM(EncoderDecoder):
def __init__(self,
freeze_backbone=False,
exclude_keys=None,
load_text_embedding=None,
load_mcc_text_embedding=None,
load_pl_text_embedding=None,
clip_encoder=None,
conv_encoder=None,
maskclip_class_filter=None,
maskclip_trust_head=None,
renorm_clip_img=False,
**args):
super(VLM, self).__init__(**args)
assert load_text_embedding == load_pl_text_embedding
assert maskclip_class_filter is None
assert maskclip_trust_head is None
self.local_iter = 0
self.clip_encoder = None
if clip_encoder is not None:
self.clip_encoder = builder.build_backbone(clip_encoder)
self.conv_encoder = None
if conv_encoder is not None:
self.conv_encoder = builder.build_backbone(conv_encoder)
self.load_text_embedding = load_text_embedding
self.decode_head.load_text_embedding = load_text_embedding
self.load_mcc_text_embedding = load_mcc_text_embedding
self.renorm_clip_img = renorm_clip_img
if renorm_clip_img:
print('Renormalize clip image.')
if self.load_mcc_text_embedding:
self.loaded_mcc_text_feat = np.load(self.load_mcc_text_embedding)
self.loaded_mcc_text_feat = torch.from_numpy(self.loaded_mcc_text_feat).float()
else:
raise NotImplementedError
if freeze_backbone:
self.freeze(self.backbone, exclude_keys=exclude_keys)
def renormalize_img_for_clip(self, img):
if not self.renorm_clip_img:
return img
loader_mean, loader_std = [0.485, 0.456, 0.406], [0.229, 0.224, 0.225]
clip_mean, clip_std = [0.48145466, 0.4578275, 0.40821073], [0.26862954, 0.26130258, 0.27577711]
loader_mean = torch.tensor(loader_mean, device=img.device).view(1, -1, 1, 1)
loader_std = torch.tensor(loader_std, device=img.device).view(1, -1, 1, 1)
clip_mean = torch.tensor(clip_mean, device=img.device).view(1, -1, 1, 1)
clip_std = torch.tensor(clip_std, device=img.device).view(1, -1, 1, 1)
return (img * loader_std + loader_mean - clip_mean) / clip_std
def freeze(self, model, exclude_keys=None):
for n, m in model.named_parameters():
m.requires_grad = False
if exclude_keys is not None:
assert isinstance(exclude_keys, list)
for k in exclude_keys:
if str(k) in n:
m.requires_grad = True
print(f'Finetune {n}')
def forward_maskclip(self, img, conf_tresh):
img = self.renormalize_img_for_clip(img)
self.clip_encoder.eval()
with torch.no_grad():
text_feat = self.loaded_mcc_text_feat.detach().to(img.device)
visual_feat, _ = self.clip_encoder(img)
visual_feat = visual_feat[-1]
dense_pred = F.conv2d(visual_feat, text_feat[:, :, None, None])
if dense_pred.shape[1] != self.num_classes: | cls2con = get_class_to_concept_idxs(self.load_mcc_text_embedding) | 1 | 2023-11-02 14:49:38+00:00 | 2k |
ej52/hass-ollama-conversation | custom_components/ollama_conversation/api.py | [
{
"identifier": "TIMEOUT",
"path": "custom_components/ollama_conversation/const.py",
"snippet": "TIMEOUT = 60"
},
{
"identifier": "ApiClientError",
"path": "custom_components/ollama_conversation/exceptions.py",
"snippet": "class ApiClientError(HomeAssistantError):\n \"\"\"Exception to... | import asyncio
import socket
import aiohttp
import async_timeout
from .const import TIMEOUT
from .exceptions import (
ApiClientError,
ApiCommError,
ApiJsonError,
ApiTimeoutError
) | 692 | """Ollama API Client."""
from __future__ import annotations
class OllamaApiClient:
"""Ollama API Client."""
def __init__(
self,
base_url: str,
session: aiohttp.ClientSession,
) -> None:
"""Sample API Client."""
self._base_url = base_url.rstrip("/")
self._session = session
async def async_get_heartbeat(self) -> bool:
"""Get heartbeat from the API."""
response: str = await self._api_wrapper(
method="get", url=self._base_url, decode_json=False
)
return response.strip() == "Ollama is running"
async def async_get_models(self) -> any:
"""Get models from the API."""
return await self._api_wrapper(
method="get",
url=f"{self._base_url}/api/tags",
headers={"Content-type": "application/json; charset=UTF-8"},
)
async def async_generate(self, data: dict | None = None,) -> any:
"""Generate a completion from the API."""
return await self._api_wrapper(
method="post",
url=f"{self._base_url}/api/generate",
data=data,
headers={"Content-type": "application/json; charset=UTF-8"},
)
async def _api_wrapper(
self,
method: str,
url: str,
data: dict | None = None,
headers: dict | None = None,
decode_json: bool = True,
) -> any:
"""Get information from the API."""
try:
async with async_timeout.timeout(TIMEOUT):
response = await self._session.request(
method=method,
url=url,
headers=headers,
json=data,
)
if response.status == 404 and decode_json:
json = await response.json()
raise ApiJsonError(json["error"])
response.raise_for_status()
if decode_json:
return await response.json()
return await response.text()
except ApiJsonError as e:
raise e
except asyncio.TimeoutError as e:
raise ApiTimeoutError("timeout while talking to the server") from e
except (aiohttp.ClientError, socket.gaierror) as e:
| """Ollama API Client."""
from __future__ import annotations
class OllamaApiClient:
"""Ollama API Client."""
def __init__(
self,
base_url: str,
session: aiohttp.ClientSession,
) -> None:
"""Sample API Client."""
self._base_url = base_url.rstrip("/")
self._session = session
async def async_get_heartbeat(self) -> bool:
"""Get heartbeat from the API."""
response: str = await self._api_wrapper(
method="get", url=self._base_url, decode_json=False
)
return response.strip() == "Ollama is running"
async def async_get_models(self) -> any:
"""Get models from the API."""
return await self._api_wrapper(
method="get",
url=f"{self._base_url}/api/tags",
headers={"Content-type": "application/json; charset=UTF-8"},
)
async def async_generate(self, data: dict | None = None,) -> any:
"""Generate a completion from the API."""
return await self._api_wrapper(
method="post",
url=f"{self._base_url}/api/generate",
data=data,
headers={"Content-type": "application/json; charset=UTF-8"},
)
async def _api_wrapper(
self,
method: str,
url: str,
data: dict | None = None,
headers: dict | None = None,
decode_json: bool = True,
) -> any:
"""Get information from the API."""
try:
async with async_timeout.timeout(TIMEOUT):
response = await self._session.request(
method=method,
url=url,
headers=headers,
json=data,
)
if response.status == 404 and decode_json:
json = await response.json()
raise ApiJsonError(json["error"])
response.raise_for_status()
if decode_json:
return await response.json()
return await response.text()
except ApiJsonError as e:
raise e
except asyncio.TimeoutError as e:
raise ApiTimeoutError("timeout while talking to the server") from e
except (aiohttp.ClientError, socket.gaierror) as e: | raise ApiCommError("unknown error while talking to the server") from e | 2 | 2023-11-03 14:48:45+00:00 | 2k |
Zaczero/openstreetmap-ng | src/repositories/message_repository.py | [
{
"identifier": "DB",
"path": "src/db.py",
"snippet": "DB = async_sessionmaker(\n DB_ENGINE,\n expire_on_commit=False,\n)"
},
{
"identifier": "Message",
"path": "src/models/db/message.py",
"snippet": "class Message(Base.Sequential, CreatedAtMixin, RichTextMixin):\n __tablename__... | from sqlalchemy import false, func, select
from src.db import DB
from src.models.db.message import Message | 808 |
class MessageRepository:
@staticmethod
async def count_received_by_user_id(user_id: int) -> tuple[int, int]:
"""
Count received messages by user id.
Returns a tuple of (total, unread).
"""
|
class MessageRepository:
@staticmethod
async def count_received_by_user_id(user_id: int) -> tuple[int, int]:
"""
Count received messages by user id.
Returns a tuple of (total, unread).
"""
| async with DB() as session: | 0 | 2023-11-04 01:12:13+00:00 | 2k |
codefuse-ai/Collinear-Constrained-Attention | data/multi_task_dataset.py | [
{
"identifier": "print_rank_0",
"path": "utils/common_utils.py",
"snippet": "def print_rank_0(*message):\n \"\"\"If distributed is initialized print only on rank 0.\"\"\"\n if torch.distributed.is_initialized():\n if torch.distributed.get_rank() == 0:\n print(*message, flush=True... | import os
import math
import json
import random
import time
import numpy as np
import torch
from functools import partial
from utils.common_utils import print_rank_0, TASK2ID, ID2TASK, get_local_rank
from data import helpers | 864 |
class SingleTaskDataset(torch.utils.data.Dataset):
def __init__(
self,
name,
data_prefix,
input_dataset,
# loss_mask_dataset,
# num_samples,
seq_length,
weighted_loss_mode=None,
ds_weight=1.0,
):
self.name = name
self.input_dataset = input_dataset
self.num_samples = len(self.input_dataset['input_ids'])
# self.loss_mask_dataset = loss_mask_dataset
self.seq_length = seq_length
self.weighted_loss_mode = weighted_loss_mode
self.ds_weight = ds_weight
self.task_name = data_prefix.split('/')[-1]
self.task_id = TASK2ID[self.task_name]
# Checks
def update_ds_weight(self, weight):
self.ds_weight = weight
def __len__(self):
return self.num_samples
def __getitem__(self, idx):
try:
# Get the shuffled index.
idx = idx % self.num_samples
idx_data = {key: self.input_dataset[key][idx]
for key in self.input_dataset}
if self.weighted_loss_mode:
idx_data["weight"] = np.array([self.ds_weight], dtype=np.float32)
idx_data["task_id"] = np.array([self.task_id], dtype=np.int)
return idx_data
else:
idx_data["task_id"] = np.array([self.task_id], dtype=np.int)
return idx_data
except IndexError:
new_idx = idx % len(self)
print(
f"WARNING: Got index out of bounds error with index {idx} - taking modulo of index instead ({new_idx})"
)
return self[new_idx]
class MultiTaskDataset(torch.utils.data.Dataset):
def __init__(self,
args,
dataset_type,
data_paths,
tokenizer,
max_input_length=550,
max_output_length=550,
max_length=1024,
no_append_glm_mask=False,
gpt_data=False,
world_size=1,
global_rank=0,
left_truncate=False,
shard_data=False,
**kwargs):
super().__init__()
self.args = args
self.dataset_type = dataset_type
self.mode = args.tokenize_mode
# self.seq_length = max_length
# self.max_seq_length = args.seq_length + 1
self.data_paths = data_paths
self.tokenizer = tokenizer
self.gpt_data = gpt_data
# if self.gpt_data:
# self.tokenizer.sop_token = "<|endoftext|>"
# self.tokenizer.eop_token = "<|endoftext|>"
# self.tokenizer.pad_token = "<|extratoken_1|>"
self.sop_id = self.tokenizer.convert_tokens_to_ids(
self.tokenizer.sop_token)
# self.eop_id = self.tokenizer.convert_tokens_to_ids(
# self.tokenizer.eop_token)
self.eop_id = self.tokenizer.convert_tokens_to_ids(
self.tokenizer.eos_token)
self.pad_id = self.tokenizer.convert_tokens_to_ids(
self.tokenizer.pad_token)
|
class SingleTaskDataset(torch.utils.data.Dataset):
def __init__(
self,
name,
data_prefix,
input_dataset,
# loss_mask_dataset,
# num_samples,
seq_length,
weighted_loss_mode=None,
ds_weight=1.0,
):
self.name = name
self.input_dataset = input_dataset
self.num_samples = len(self.input_dataset['input_ids'])
# self.loss_mask_dataset = loss_mask_dataset
self.seq_length = seq_length
self.weighted_loss_mode = weighted_loss_mode
self.ds_weight = ds_weight
self.task_name = data_prefix.split('/')[-1]
self.task_id = TASK2ID[self.task_name]
# Checks
def update_ds_weight(self, weight):
self.ds_weight = weight
def __len__(self):
return self.num_samples
def __getitem__(self, idx):
try:
# Get the shuffled index.
idx = idx % self.num_samples
idx_data = {key: self.input_dataset[key][idx]
for key in self.input_dataset}
if self.weighted_loss_mode:
idx_data["weight"] = np.array([self.ds_weight], dtype=np.float32)
idx_data["task_id"] = np.array([self.task_id], dtype=np.int)
return idx_data
else:
idx_data["task_id"] = np.array([self.task_id], dtype=np.int)
return idx_data
except IndexError:
new_idx = idx % len(self)
print(
f"WARNING: Got index out of bounds error with index {idx} - taking modulo of index instead ({new_idx})"
)
return self[new_idx]
class MultiTaskDataset(torch.utils.data.Dataset):
def __init__(self,
args,
dataset_type,
data_paths,
tokenizer,
max_input_length=550,
max_output_length=550,
max_length=1024,
no_append_glm_mask=False,
gpt_data=False,
world_size=1,
global_rank=0,
left_truncate=False,
shard_data=False,
**kwargs):
super().__init__()
self.args = args
self.dataset_type = dataset_type
self.mode = args.tokenize_mode
# self.seq_length = max_length
# self.max_seq_length = args.seq_length + 1
self.data_paths = data_paths
self.tokenizer = tokenizer
self.gpt_data = gpt_data
# if self.gpt_data:
# self.tokenizer.sop_token = "<|endoftext|>"
# self.tokenizer.eop_token = "<|endoftext|>"
# self.tokenizer.pad_token = "<|extratoken_1|>"
self.sop_id = self.tokenizer.convert_tokens_to_ids(
self.tokenizer.sop_token)
# self.eop_id = self.tokenizer.convert_tokens_to_ids(
# self.tokenizer.eop_token)
self.eop_id = self.tokenizer.convert_tokens_to_ids(
self.tokenizer.eos_token)
self.pad_id = self.tokenizer.convert_tokens_to_ids(
self.tokenizer.pad_token) | print_rank_0(f'self.tokenizer.sop_token {self.tokenizer.sop_token} id: {self.sop_id}') | 0 | 2023-11-02 01:37:01+00:00 | 2k |
rezaakb/pinns-tf2 | pinnstf2/models/pinn_module.py | [
{
"identifier": "gradient",
"path": "pinnstf2/utils/gradient.py",
"snippet": "def gradient(dy, dx, grad_ys=None):\n if grad_ys is None:\n dy_dx = tf.gradients(dy, dx)\n else:\n dy_dx = tf.gradients(dy, dx, grad_ys=grad_ys)\n if len(dy_dx)==1:\n dy_dx = dy_dx[0]\n return ... | from typing import List, Dict, Callable, Any, Tuple, Union
from pinnstf2.utils import fwd_gradient, gradient
from pinnstf2.utils import (
fix_extra_variables,
mse,
relative_l2_error,
sse
)
import tensorflow as tf
import sys, os, logging, time | 1,554 |
class PINNModule:
def __init__(
self,
net,
pde_fn: Callable[[Any, ...], tf.Tensor],
optimizer: tf.keras.optimizers.Adam = tf.keras.optimizers.Adam,
loss_fn: str = "sse",
extra_variables: Dict[str, Any] = None,
output_fn: Callable[[Any, ...], tf.Tensor] = None,
runge_kutta=None,
jit_compile: bool = True,
amp: bool = False,
dtype: str = 'float32'
) -> None:
"""
Initialize a `PINNModule`.
:param net: The neural network model to be used for approximating solutions.
:param pde_fn: The partial differential equation (PDE) function defining the PDE to solve.
:param optimizer: The optimizer used for training the neural network.
:param loss_fn: The name of the loss function to be used. Default is 'sse' (sum of squared errors).
:param extra_variables: Additional variables used in the model, provided as a dictionary. Default is None.
:param output_fn: A function applied to the output of the network, for post-processing or transformations.
:param runge_kutta: An optional Runge-Kutta method implementation for solving discrete problems. Default is None.
:param jit_compile: If True, TensorFlow's JIT compiler will be used for optimizing computations. Default is True.
:param amp: Automatic mixed precision (amp) for optimizing training performance. Default is False.
:param dtype: Data type to be used for the computations. Default is 'float32'.
"""
super().__init__()
self.net = net
self.tf_dtype = tf.as_dtype(dtype)
if hasattr(self.net, 'model'):
self.trainable_variables = self.net.model.trainable_variables
else:
self.trainable_variables = self.net.trainable_variables
(self.trainable_variables,
|
class PINNModule:
def __init__(
self,
net,
pde_fn: Callable[[Any, ...], tf.Tensor],
optimizer: tf.keras.optimizers.Adam = tf.keras.optimizers.Adam,
loss_fn: str = "sse",
extra_variables: Dict[str, Any] = None,
output_fn: Callable[[Any, ...], tf.Tensor] = None,
runge_kutta=None,
jit_compile: bool = True,
amp: bool = False,
dtype: str = 'float32'
) -> None:
"""
Initialize a `PINNModule`.
:param net: The neural network model to be used for approximating solutions.
:param pde_fn: The partial differential equation (PDE) function defining the PDE to solve.
:param optimizer: The optimizer used for training the neural network.
:param loss_fn: The name of the loss function to be used. Default is 'sse' (sum of squared errors).
:param extra_variables: Additional variables used in the model, provided as a dictionary. Default is None.
:param output_fn: A function applied to the output of the network, for post-processing or transformations.
:param runge_kutta: An optional Runge-Kutta method implementation for solving discrete problems. Default is None.
:param jit_compile: If True, TensorFlow's JIT compiler will be used for optimizing computations. Default is True.
:param amp: Automatic mixed precision (amp) for optimizing training performance. Default is False.
:param dtype: Data type to be used for the computations. Default is 'float32'.
"""
super().__init__()
self.net = net
self.tf_dtype = tf.as_dtype(dtype)
if hasattr(self.net, 'model'):
self.trainable_variables = self.net.model.trainable_variables
else:
self.trainable_variables = self.net.trainable_variables
(self.trainable_variables, | self.extra_variables) = fix_extra_variables(self.trainable_variables, extra_variables, self.tf_dtype) | 2 | 2023-11-01 03:25:51+00:00 | 2k |
djinni-co/djinni-inbox-test | app/sandbox/views.py | [
{
"identifier": "Recruiter",
"path": "app/sandbox/models.py",
"snippet": "class Recruiter(models.Model):\n USERTYPE = \"recruiter\"\n\n name = models.CharField(max_length=255, blank=True, default='')\n email = models.EmailField(blank=False, db_index=True, unique=True)\n picture_url = models.... | from django.http import HttpResponse
from django.db.models import Count, Q
from django.shortcuts import render
from .models import Recruiter, MessageThread | 736 |
# Hardcode for logged in as recruiter
RECRUITER_ID = 125528
def inbox(request):
recruiter = Recruiter.objects.get(id = RECRUITER_ID)
|
# Hardcode for logged in as recruiter
RECRUITER_ID = 125528
def inbox(request):
recruiter = Recruiter.objects.get(id = RECRUITER_ID) | threads = MessageThread.objects.filter(recruiter = recruiter).select_related('candidate', 'job') | 1 | 2023-11-02 15:12:54+00:00 | 2k |
XinyuanWangCS/PromptAgent | src/prompt_optim_agent/world_model/beam_world_model.py | [
{
"identifier": "eval_instruction_with_loader",
"path": "src/prompt_optim_agent/test_helper.py",
"snippet": "def eval_instruction_with_loader(task, eval_prompt, dataloader, model='gpt-3.5-turbo', temperature=0, record_outputs=True):\n '''\n evaluate cur_prompt on task testing dataset\n '''\... | from .gradient_descent import *
from typing import NamedTuple
from ..test_helper import eval_instruction_with_loader
from typing import Generic
from ..search_algo.base_algo import State, Action
from ..search_algo.beam_search import BeamNode
from ..utils import gpt_chat_completion | 1,462 |
class BeamSearchWorldModel(Generic[State, Action]):
def __init__(
self,
task,
logger,
# model
pred_model: str,
optim_model: str,
pred_temperature: float,
optim_temperature: float,
prompt_length_limit:int,
num_new_prompts = 3,
train_shuffle = True,
train_batch_size: int = 5,
test_batch_size: int = 1,
eval_batch_size: int = 1,
**kwargs
) -> None:
self.task = task
self.logger = logger
self.pred_model = pred_model
self.optim_model = optim_model
self.pred_temperature=pred_temperature
self.optim_temperature = optim_temperature
self.train_dataloader = self.task.get_dataloader('train',
batch_size=train_batch_size,
shuffle=train_shuffle)
self.train_data_iterator = self._infinite_data_loader(self.train_dataloader)
self.test_dataloader = self.task.get_dataloader('test',
batch_size=test_batch_size,
shuffle=False)
self.eval_dataloader = self.task.get_dataloader('eval',
batch_size=eval_batch_size,
shuffle=False)
self.gradient_descent = GradientDescent(task=self.task,
logger=self.logger,
pred_model=pred_model,
optim_model=optim_model,
num_new_prompts = num_new_prompts,
forward_temperature=pred_temperature,
optim_temperature = optim_temperature,
prompt_length_limit=prompt_length_limit)
def _infinite_data_loader(self, data_loader):
while True:
for batch in data_loader:
yield batch
def get_train_batch(self):
return next(self.train_data_iterator)
|
class BeamSearchWorldModel(Generic[State, Action]):
def __init__(
self,
task,
logger,
# model
pred_model: str,
optim_model: str,
pred_temperature: float,
optim_temperature: float,
prompt_length_limit:int,
num_new_prompts = 3,
train_shuffle = True,
train_batch_size: int = 5,
test_batch_size: int = 1,
eval_batch_size: int = 1,
**kwargs
) -> None:
self.task = task
self.logger = logger
self.pred_model = pred_model
self.optim_model = optim_model
self.pred_temperature=pred_temperature
self.optim_temperature = optim_temperature
self.train_dataloader = self.task.get_dataloader('train',
batch_size=train_batch_size,
shuffle=train_shuffle)
self.train_data_iterator = self._infinite_data_loader(self.train_dataloader)
self.test_dataloader = self.task.get_dataloader('test',
batch_size=test_batch_size,
shuffle=False)
self.eval_dataloader = self.task.get_dataloader('eval',
batch_size=eval_batch_size,
shuffle=False)
self.gradient_descent = GradientDescent(task=self.task,
logger=self.logger,
pred_model=pred_model,
optim_model=optim_model,
num_new_prompts = num_new_prompts,
forward_temperature=pred_temperature,
optim_temperature = optim_temperature,
prompt_length_limit=prompt_length_limit)
def _infinite_data_loader(self, data_loader):
while True:
for batch in data_loader:
yield batch
def get_train_batch(self):
return next(self.train_data_iterator)
| def _get_trajectory_prompts(self, node: BeamNode): | 2 | 2023-11-03 19:14:00+00:00 | 2k |
evaluable-ai/auto-eval | evaluableai/models/candidate_models/null_model.py | [
{
"identifier": "InputRow",
"path": "evaluableai/data_model/input_row_object.py",
"snippet": "class InputRow:\n def __init__(self, input_text, context, input_id=None):\n self._input_id = input_id if input_id is not None else uuid.uuid4()\n self._input_text = input_text\n self._co... | import json
import logging
import uuid
from evaluableai.data_model.input_row_object import InputRow
from evaluableai.data_model.model_response_object import ModelResponseObject | 1,108 |
# Make sure to import InputRow if it's a separate class
class NullModel:
def __init__(self, model_name, model_version):
self._model_name = model_name
self._model_version = model_version
@property
def model_name(self):
return self._model_name
@property
def model_version(self):
return self._model_version
@classmethod
def from_json_array(cls, json_array):
data_array = [cls.parse_json(item) for item in json_array]
response_objects = []
for data in data_array:
input_text = data.get('input', '')
context = data.get('context', '')
|
# Make sure to import InputRow if it's a separate class
class NullModel:
def __init__(self, model_name, model_version):
self._model_name = model_name
self._model_version = model_version
@property
def model_name(self):
return self._model_name
@property
def model_version(self):
return self._model_version
@classmethod
def from_json_array(cls, json_array):
data_array = [cls.parse_json(item) for item in json_array]
response_objects = []
for data in data_array:
input_text = data.get('input', '')
context = data.get('context', '') | input_row = InputRow(input_text, context) # Assuming InputRow is imported | 0 | 2023-11-06 01:26:17+00:00 | 2k |
allenai/wimbd | wimbd/contamination/promptsource_parse.py | [
{
"identifier": "INCLUDED_USERS",
"path": "wimbd/contamination/templates.py",
"snippet": "INCLUDED_USERS = {\"Zaid\", \"craffel\"}"
},
{
"identifier": "TemplateCollection",
"path": "wimbd/contamination/templates.py",
"snippet": "class TemplateCollection:\n \"\"\"\n This helper clas... | import argparse
import csv
import re
from glob import glob
from wimbd.contamination.templates import INCLUDED_USERS, TemplateCollection
from wimbd.contamination.utils import get_dataset | 1,194 |
def main():
parse = argparse.ArgumentParser("")
parse.add_argument("--path", type=str)
parse.add_argument("--out_file", type=str)
args = parse.parse_args()
datasets = []
for path in glob(args.path + '/**/templates.yaml', recursive=True):
datasets.append(path)
with open(args.out_file, 'w', newline='') as f:
writer = csv.writer(f, delimiter='\t', lineterminator='\n')
for dataset in datasets:
path = dataset.split('/')
dataset_name = path[8]
subset_name = path[9] if len(path) == 11 else ''
|
def main():
parse = argparse.ArgumentParser("")
parse.add_argument("--path", type=str)
parse.add_argument("--out_file", type=str)
args = parse.parse_args()
datasets = []
for path in glob(args.path + '/**/templates.yaml', recursive=True):
datasets.append(path)
with open(args.out_file, 'w', newline='') as f:
writer = csv.writer(f, delimiter='\t', lineterminator='\n')
for dataset in datasets:
path = dataset.split('/')
dataset_name = path[8]
subset_name = path[9] if len(path) == 11 else ''
| template_collection = TemplateCollection() | 1 | 2023-11-08 18:18:41+00:00 | 2k |
kakaobrain/cxr-clip | cxrclip/data/datasets/imagetext_eval.py | [
{
"identifier": "load_transform",
"path": "cxrclip/data/data_utils.py",
"snippet": "def load_transform(split: str = \"train\", transform_config: Dict = None):\n assert split in {\"train\", \"valid\", \"test\", \"aug\"}\n\n config = []\n if transform_config:\n if split in transform_config... | import ast
import pandas as pd
from typing import Dict, List
from PIL import Image
from torch.utils.data import default_collate
from torch.utils.data.dataset import Dataset
from cxrclip.data.data_utils import load_transform, transform_image
from cxrclip.prompt.constants import CHEXPERT_CLASS_PROMPTS | 1,417 |
class ImageTextEvalDataset(Dataset):
def __init__(
self,
name: str,
data_path: str,
split: str,
data_frac: float = 1.0,
tokenizer=None,
text_max_length: int = 256,
transform_config: Dict = None,
normalize: str = "huggingface",
**kwargs
):
super().__init__()
self.name = name
self.split = split
self.tokenizer = tokenizer
self.text_max_length = text_max_length
self.data_frac = data_frac
self.normalize = normalize
if self.name == "chexpert5x200":
|
class ImageTextEvalDataset(Dataset):
def __init__(
self,
name: str,
data_path: str,
split: str,
data_frac: float = 1.0,
tokenizer=None,
text_max_length: int = 256,
transform_config: Dict = None,
normalize: str = "huggingface",
**kwargs
):
super().__init__()
self.name = name
self.split = split
self.tokenizer = tokenizer
self.text_max_length = text_max_length
self.data_frac = data_frac
self.normalize = normalize
if self.name == "chexpert5x200": | self.label_list = list(CHEXPERT_CLASS_PROMPTS.keys()) | 2 | 2023-11-01 07:24:52+00:00 | 2k |
mihirp1998/Diffusion-TTA | diff_tta/models/build.py | [
{
"identifier": "get_obj_from_str",
"path": "diff_tta/utils.py",
"snippet": "def get_obj_from_str(string, reload=False):\n \"\"\"A helper function to instantiate a class from a config object.\n See https://github.com/CompVis/stable-diffusion/blob/main/ldm/util.py\n \"\"\"\n module, cls = str... | import torch
import torch.nn as nn
import torchvision
from diffusers import (
AutoencoderKL,
UNet2DConditionModel,
DDPMScheduler,
StableDiffusionPipeline,
EulerDiscreteScheduler
)
from transformers import CLIPTextModel, CLIPTokenizer
from diff_tta.utils import get_obj_from_str
from diff_tta.models.DiT.models import DiT_XL_2
from diff_tta.models.DiT.download import find_model
from diff_tta.models.clip_classifier import ClipClassifier
from diff_tta.models.DiT.diffusion import create_diffusion
from diff_tta import utils | 893 |
def load_dit_model(config, device):
"""Load DiT model"""
#@param ["stabilityai/sd-vae-ft-mse", "stabilityai/sd-vae-ft-ema"]
vae_model = "stabilityai/sd-vae-ft-ema"
image_size = config.input.sd_img_res
latent_size = int(image_size) // 8
model = DiT_XL_2(input_size=latent_size).to(device)
state_dict = find_model(f"DiT-XL-2-{image_size}x{image_size}.pt")
model.load_state_dict(state_dict)
model.eval() # important!
vae = AutoencoderKL.from_pretrained(vae_model).to(device)
vae.eval()
# default: 1000 steps, linear noise schedule
diffusion = create_diffusion(timestep_respacing="")
|
def load_dit_model(config, device):
"""Load DiT model"""
#@param ["stabilityai/sd-vae-ft-mse", "stabilityai/sd-vae-ft-ema"]
vae_model = "stabilityai/sd-vae-ft-ema"
image_size = config.input.sd_img_res
latent_size = int(image_size) // 8
model = DiT_XL_2(input_size=latent_size).to(device)
state_dict = find_model(f"DiT-XL-2-{image_size}x{image_size}.pt")
model.load_state_dict(state_dict)
model.eval() # important!
vae = AutoencoderKL.from_pretrained(vae_model).to(device)
vae.eval()
# default: 1000 steps, linear noise schedule
diffusion = create_diffusion(timestep_respacing="") | image_renormalizer = utils.VQVAEUnNormalize( | 2 | 2023-11-07 21:09:50+00:00 | 2k |
pofey/MemAI-Flow | memflow/main.py | [
{
"identifier": "CuboxErrorException",
"path": "memflow/exceptions.py",
"snippet": "class CuboxErrorException(RuntimeError):\n def __init__(self, message):\n self.message = message"
},
{
"identifier": "LOGGING_CONFIG",
"path": "memflow/common/logging.py",
"snippet": "LOGGING_CO... | import os
import logging.config
import inject
import httpx
import uvicorn
from memflow.exceptions import CuboxErrorException
from apscheduler.schedulers.background import BackgroundScheduler
from fastapi.exceptions import RequestValidationError
from memflow.common.logging import LOGGING_CONFIG
from memflow.memapi import MemApi
from starlette.exceptions import HTTPException
from fastapi import FastAPI
from memflow.databases import create_all
from memflow.common.response import json_200, json_500, json_with_status
from memflow.models import *
from memflow.tasks.cuboxsynctask import CuboxSyncTask | 1,424 | """
程序启动入口类
"""
if not os.environ.get("WORKDIR"):
workdir = os.path.join(os.path.dirname(os.path.dirname(os.path.abspath(__file__))), 'data')
else:
workdir = os.environ.get("WORKDIR")
if not os.path.exists(workdir):
os.makedirs(workdir)
log_dir = os.path.join(workdir, 'logs')
if not os.path.exists(log_dir):
os.makedirs(log_dir)
os.environ["WORKDIR"] = workdir
| """
程序启动入口类
"""
if not os.environ.get("WORKDIR"):
workdir = os.path.join(os.path.dirname(os.path.dirname(os.path.abspath(__file__))), 'data')
else:
workdir = os.environ.get("WORKDIR")
if not os.path.exists(workdir):
os.makedirs(workdir)
log_dir = os.path.join(workdir, 'logs')
if not os.path.exists(log_dir):
os.makedirs(log_dir)
os.environ["WORKDIR"] = workdir
| logging.config.dictConfig(LOGGING_CONFIG) | 1 | 2023-11-08 10:02:00+00:00 | 2k |
sdebruyn/dbt-timescaledb | dbt/adapters/timescaledb/timescaledb_adapter.py | [
{
"identifier": "NO_TRANSACTION_MARKER",
"path": "dbt/adapters/timescaledb/timescaledb_connection_manager.py",
"snippet": "NO_TRANSACTION_MARKER = \"/* MARKER SHOULD RUN OUTSIDE TRANSACTION */\""
},
{
"identifier": "TimescaleDBConnectionManager",
"path": "dbt/adapters/timescaledb/timescaledb... | from typing import Any, Optional
from dbt.adapters.base.meta import available
from dbt.adapters.postgres import PostgresAdapter
from dbt.adapters.timescaledb.timescaledb_connection_manager import (
NO_TRANSACTION_MARKER,
TimescaleDBConnectionManager,
)
from dbt.adapters.timescaledb.timescaledb_index_config import TimescaleDBIndexConfig | 699 |
class TimescaleDBAdapter(PostgresAdapter):
ConnectionManager = TimescaleDBConnectionManager
@available
def parse_index(self, raw_index: Any) -> Optional[TimescaleDBIndexConfig]:
return TimescaleDBIndexConfig.parse(raw_index)
@available
def marker_run_outside_transaction(self) -> str:
|
class TimescaleDBAdapter(PostgresAdapter):
ConnectionManager = TimescaleDBConnectionManager
@available
def parse_index(self, raw_index: Any) -> Optional[TimescaleDBIndexConfig]:
return TimescaleDBIndexConfig.parse(raw_index)
@available
def marker_run_outside_transaction(self) -> str: | return NO_TRANSACTION_MARKER | 0 | 2023-11-07 21:54:46+00:00 | 2k |
jax-ml/bayeux | bayeux/_src/shared.py | [
{
"identifier": "debug",
"path": "bayeux/_src/debug.py",
"snippet": "def debug(model, seed, verbosity, printer, kwargs, catch_exceptions: bool):\n \"\"\"Debugger that includes the inverse log det jacobian.\"\"\"\n checkers = [\n check_shapes,\n check_test_point_log_density,\n check_kwar... | import dataclasses
import functools
import inspect
import jax
import jax.numpy as jnp
import oryx
from typing import Callable, Optional
from bayeux._src import debug
from bayeux._src import initialization
from bayeux._src import types | 1,108 | # Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Shared functionality for MCMC sampling."""
def map_fn(chain_method, fn):
if chain_method == "parallel":
return jax.pmap(fn)
elif chain_method == "vectorized":
return jax.vmap(fn)
elif chain_method == "sequential":
return functools.partial(jax.tree_map, fn)
raise ValueError(f"Chain method {chain_method} not supported.")
def _default_init(
*,
initial_state,
test_point,
inverse_transform_fn,
transform_fn,
num_points,
key
):
"""Initialization in case there is no explicit init provided."""
if initial_state is None:
return initialization.uniform(
test_point=test_point,
inverse_transform_fn=inverse_transform_fn,
transform_fn=transform_fn,
num_points=num_points,
key=key)
else:
return initial_state
def constrain(
transform_fn: types.JAXFn,
inverse_log_det_jacobian: Optional[types.JAXFn] = None,
) -> Callable[[types.LogDensityFn], types.LogDensityFn]:
"""Returns a log density function that operates in an unconstrained space.
Adapted from oryx (https://github.com/jax-ml/oryx)
Args:
transform_fn: Constraining bijector, mapping from R^n to the support of the
target log density.
inverse_log_det_jacobian: Optional inverse log det jacobian, if known.
"""
if inverse_log_det_jacobian is None:
inverse_log_det_jacobian = oryx.core.ildj(transform_fn)
def wrap_log_density(target_log_density):
def wrapped(args):
mapped_args = transform_fn(args)
ildjs = inverse_log_det_jacobian(mapped_args)
return target_log_density(mapped_args) - jnp.sum(
jnp.array(jax.tree_util.tree_leaves(ildjs)))
return wrapped
return wrap_log_density
def get_default_signature(fn):
defaults = {}
required = set()
for key, val in inspect.signature(fn).parameters.items():
if val.default is inspect.Signature.empty:
required.add(key)
else:
defaults[key] = val.default
return defaults, required
def _nothing(x: types.Point) -> types.Point:
return x
@dataclasses.dataclass
class Base:
"""Base class for MCMC sampling."""
log_density: types.LogDensityFn
test_point: types.Point
transform_fn: types.JAXFn = _nothing
inverse_transform_fn: Optional[types.JAXFn] = None
inverse_log_det_jacobian: Optional[types.JAXFn] = None
initial_state: Optional[types.Point] = None
def get_initial_state(self, key, num_chains=8):
return _default_init(
initial_state=self.initial_state,
test_point=self.test_point,
inverse_transform_fn=self.inverse_transform_fn,
transform_fn=self.transform_fn,
num_points=num_chains,
key=key)
def get_kwargs(self, **kwargs):
raise NotImplementedError()
def constrained_log_density(self):
return constrain(
self.transform_fn,
inverse_log_det_jacobian=self.inverse_log_det_jacobian,
)(self.log_density)
| # Copyright 2023 The bayeux Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Shared functionality for MCMC sampling."""
def map_fn(chain_method, fn):
if chain_method == "parallel":
return jax.pmap(fn)
elif chain_method == "vectorized":
return jax.vmap(fn)
elif chain_method == "sequential":
return functools.partial(jax.tree_map, fn)
raise ValueError(f"Chain method {chain_method} not supported.")
def _default_init(
*,
initial_state,
test_point,
inverse_transform_fn,
transform_fn,
num_points,
key
):
"""Initialization in case there is no explicit init provided."""
if initial_state is None:
return initialization.uniform(
test_point=test_point,
inverse_transform_fn=inverse_transform_fn,
transform_fn=transform_fn,
num_points=num_points,
key=key)
else:
return initial_state
def constrain(
transform_fn: types.JAXFn,
inverse_log_det_jacobian: Optional[types.JAXFn] = None,
) -> Callable[[types.LogDensityFn], types.LogDensityFn]:
"""Returns a log density function that operates in an unconstrained space.
Adapted from oryx (https://github.com/jax-ml/oryx)
Args:
transform_fn: Constraining bijector, mapping from R^n to the support of the
target log density.
inverse_log_det_jacobian: Optional inverse log det jacobian, if known.
"""
if inverse_log_det_jacobian is None:
inverse_log_det_jacobian = oryx.core.ildj(transform_fn)
def wrap_log_density(target_log_density):
def wrapped(args):
mapped_args = transform_fn(args)
ildjs = inverse_log_det_jacobian(mapped_args)
return target_log_density(mapped_args) - jnp.sum(
jnp.array(jax.tree_util.tree_leaves(ildjs)))
return wrapped
return wrap_log_density
def get_default_signature(fn):
defaults = {}
required = set()
for key, val in inspect.signature(fn).parameters.items():
if val.default is inspect.Signature.empty:
required.add(key)
else:
defaults[key] = val.default
return defaults, required
def _nothing(x: types.Point) -> types.Point:
return x
@dataclasses.dataclass
class Base:
"""Base class for MCMC sampling."""
log_density: types.LogDensityFn
test_point: types.Point
transform_fn: types.JAXFn = _nothing
inverse_transform_fn: Optional[types.JAXFn] = None
inverse_log_det_jacobian: Optional[types.JAXFn] = None
initial_state: Optional[types.Point] = None
def get_initial_state(self, key, num_chains=8):
return _default_init(
initial_state=self.initial_state,
test_point=self.test_point,
inverse_transform_fn=self.inverse_transform_fn,
transform_fn=self.transform_fn,
num_points=num_chains,
key=key)
def get_kwargs(self, **kwargs):
raise NotImplementedError()
def constrained_log_density(self):
return constrain(
self.transform_fn,
inverse_log_det_jacobian=self.inverse_log_det_jacobian,
)(self.log_density)
| def debug( | 0 | 2023-11-02 16:52:57+00:00 | 2k |
zamaniamin/fastapi-shop | apps/main.py | [
{
"identifier": "DatabaseManager",
"path": "config/database.py",
"snippet": "class DatabaseManager:\n \"\"\"\n A utility class for managing database operations using SQLAlchemy.\n\n The DatabaseManager simplifies the process of initializing and managing database connections, creating database\n... | from fastapi import FastAPI
from fastapi.middleware.cors import CORSMiddleware
from fastapi.staticfiles import StaticFiles
from config.database import DatabaseManager
from config.routers import RouterManager
from config.settings import MEDIA_DIR | 1,496 |
# -------------------
# --- Init Models ---
# -------------------
DatabaseManager().create_database_tables()
# --------------------
# --- Init FastAPI ---
# --------------------
app = FastAPI()
# ------------------
# --- Middleware ---
# ------------------
app.add_middleware(
CORSMiddleware,
allow_origins=["http://localhost:3000"],
allow_credentials=True,
allow_methods=["*"],
allow_headers=["*"])
# -------------------
# --- Static File ---
# -------------------
# add static-file support, for see images by URL
app.mount("/media", StaticFiles(directory=MEDIA_DIR), name="media")
# --------------------
# --- Init Routers ---
# --------------------
|
# -------------------
# --- Init Models ---
# -------------------
DatabaseManager().create_database_tables()
# --------------------
# --- Init FastAPI ---
# --------------------
app = FastAPI()
# ------------------
# --- Middleware ---
# ------------------
app.add_middleware(
CORSMiddleware,
allow_origins=["http://localhost:3000"],
allow_credentials=True,
allow_methods=["*"],
allow_headers=["*"])
# -------------------
# --- Static File ---
# -------------------
# add static-file support, for see images by URL
app.mount("/media", StaticFiles(directory=MEDIA_DIR), name="media")
# --------------------
# --- Init Routers ---
# --------------------
| RouterManager(app).import_routers() | 1 | 2023-11-06 04:46:03+00:00 | 2k |
jkulhanek/nerfbaselines | tests/test_utils.py | [
{
"identifier": "Indices",
"path": "nerfbaselines/utils.py",
"snippet": "class Indices:\n def __init__(self, steps):\n self._steps = steps\n self.total: Optional[int] = None\n\n def __contains__(self, x):\n if isinstance(self._steps, list):\n steps = self._steps\n ... | import pytest
from time import sleep, perf_counter
from nerfbaselines.utils import Indices
from nerfbaselines.utils import cancellable, CancellationToken, CancelledException
from nerfbaselines.utils import get_resources_utilization_info | 1,160 |
def test_indices_last():
indices = Indices([-1])
indices.total = 12
for i in range(12):
if i == indices.total - 1:
assert i in indices
else:
assert i not in indices
class TimeLimitCancellationToken(CancellationToken):
def __init__(self, timeout=0.003):
super().__init__()
self.timeout = timeout
self.start = perf_counter()
@property
def cancelled(self):
return super().cancelled or perf_counter() - self.start > self.timeout
def test_cancellable():
was_called = False
|
def test_indices_last():
indices = Indices([-1])
indices.total = 12
for i in range(12):
if i == indices.total - 1:
assert i in indices
else:
assert i not in indices
class TimeLimitCancellationToken(CancellationToken):
def __init__(self, timeout=0.003):
super().__init__()
self.timeout = timeout
self.start = perf_counter()
@property
def cancelled(self):
return super().cancelled or perf_counter() - self.start > self.timeout
def test_cancellable():
was_called = False
| @cancellable | 1 | 2023-11-07 20:22:35+00:00 | 2k |
microsoft/Everything-of-Thoughts-XoT | xot_all_in_one/xot/prompter/prompter_cube.py | [
{
"identifier": "doAlgStr",
"path": "xot_all_in_one/xot/prompter/utils/py222.py",
"snippet": "def doAlgStr(s, alg):\n # print('',alg)\n moves = alg.split(\" \")\n # print('moves',moves)\n for m in moves:\n if m in moveInds:\n s = doMove(s, moveInds[m])\n return s"
},
{
"identifier":... | import re
import os
import sympy
import numpy as np
import pandas as pd
from .prompts.prompts_cube import *
from .utils.py222 import doAlgStr, getCube | 755 | # Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.
class CubePrompter():
"""
CubePrompter provides the generation of prompts specific to the cube
example for the language models.
"""
def __init__(self, last_step=True):
self.last_step = int(last_step)
self.value_cache = {}
def get_current_state(self, x, y: str) -> str:
moves = y.strip().replace('\n', ' ')
s = doAlgStr(np.array(x), moves)
return s.tolist()
def count_inconsistencies(self, input_):
color_key = ['Upper', 'Right', 'Front', 'Down', 'Left', 'Back']
color_dict = dict()
for i in range(6):
order = i*4
color = input_[order:order+4]
color_dict[color_key[i]] = len(set(color))
return color_dict
def format_thoughts(self, x: str, y: str) -> str:
current_s = x
moves = ''
thoughts = y.strip().split('\n')
for i, m in enumerate(thoughts):
s2 = doAlgStr(np.array(current_s), m)
| # Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.
class CubePrompter():
"""
CubePrompter provides the generation of prompts specific to the cube
example for the language models.
"""
def __init__(self, last_step=True):
self.last_step = int(last_step)
self.value_cache = {}
def get_current_state(self, x, y: str) -> str:
moves = y.strip().replace('\n', ' ')
s = doAlgStr(np.array(x), moves)
return s.tolist()
def count_inconsistencies(self, input_):
color_key = ['Upper', 'Right', 'Front', 'Down', 'Left', 'Back']
color_dict = dict()
for i in range(6):
order = i*4
color = input_[order:order+4]
color_dict[color_key[i]] = len(set(color))
return color_dict
def format_thoughts(self, x: str, y: str) -> str:
current_s = x
moves = ''
thoughts = y.strip().split('\n')
for i, m in enumerate(thoughts):
s2 = doAlgStr(np.array(current_s), m) | state2 = getCube(s2) | 1 | 2023-11-08 09:48:34+00:00 | 2k |
ultraleap/leapc-python-bindings | leapc-python-api/src/leap/device.py | [
{
"identifier": "LeapCStruct",
"path": "leapc-python-api/src/leap/datatypes.py",
"snippet": "class FrameData:\nclass FrameHeader(LeapCStruct):\nclass Vector(LeapCStruct):\nclass Quaternion(LeapCStruct):\nclass Palm(LeapCStruct):\nclass Bone(LeapCStruct):\nclass Digit(LeapCStruct):\nclass Hand(LeapCStruc... | from contextlib import contextmanager
from leapc_cffi import ffi, libleapc
from .datatypes import LeapCStruct
from .enums import get_enum_entries, DevicePID, DeviceStatus
from .exceptions import success_or_raise, LeapError, LeapCannotOpenDeviceError | 863 |
class DeviceNotOpenException(LeapError):
pass
class DeviceStatusInfo:
def __init__(self, status: ffi.CData):
"""Create the DeviceStatusInfo
:param status: The CData defining the status
"""
|
class DeviceNotOpenException(LeapError):
pass
class DeviceStatusInfo:
def __init__(self, status: ffi.CData):
"""Create the DeviceStatusInfo
:param status: The CData defining the status
""" | self._status_flags = get_enum_entries(DeviceStatus, status) | 3 | 2023-11-08 13:35:40+00:00 | 2k |
UMass-Foundation-Model/CoVLM | open_flamingo/src/flamingo_lm.py | [
{
"identifier": "getattr_recursive",
"path": "open_flamingo/src/utils.py",
"snippet": "def getattr_recursive(obj, att):\n \"\"\"\n Return nested attribute of obj\n Example: getattr_recursive(obj, 'a.b.c') is equivalent to obj.a.b.c\n \"\"\"\n if att == \"\":\n return obj\n i = a... | import random
import torch
import torch.nn as nn
import numpy as np
from .utils import getattr_recursive, setattr_recursive | 1,155 |
class FlamingoLayer(nn.Module):
def __init__(self, decoder_layer):
super().__init__()
self.decoder_layer = decoder_layer
self.vis_x = None
self.image_nums = None
self.image_start_index_list = None
def is_conditioned(self) -> bool:
"""Check whether the layer is conditioned."""
return self.vis_x is not None
# Used this great idea from this implementation of Flamingo (https://github.com/dhansmair/flamingo-mini/)
def condition_vis_x(self, vis_x, image_nums=None, image_start_index_list=None, num_beams=None, visual_tokens=None, data_list=None):
self.vis_x = vis_x
self.image_nums = image_nums
self.image_start_index_list = image_start_index_list
self.num_beams = num_beams
self.visual_tokens = visual_tokens
self.data_list = data_list
def forward(
self,
hidden_states, # alignment with hugging face name
attention_mask=None,
**decoder_layer_kwargs,
):
if self.vis_x is not None:
if self.training:
single_length = self.vis_x.shape[-2]
image_nums = self.image_nums
image_start_index_list = self.image_start_index_list
image_nums = [0] + np.cumsum(image_nums).tolist()
for i, (image_num_begin, image_num_end, start_indices) in enumerate(zip(image_nums[:-1], image_nums[1:], image_start_index_list)):
for index in start_indices:
if image_num_begin < image_num_end:
hidden_states[i, index:index+single_length] = self.vis_x[image_num_begin]
image_num_begin += 1
if self.visual_tokens is not None and len(self.visual_tokens) != 0:
for i, (x, y) in enumerate(self.data_list):
if len(self.visual_tokens[i].shape) > 1:
# print(self.visual_tokens[i].shape[0], "embedding")
hidden_states[x, y+1-self.visual_tokens[i].shape[0]:y+1] = self.visual_tokens[i]
else:
# print(self.visual_tokens[i].shape[0], "embedding")
hidden_states[x, y] = self.visual_tokens[i]
elif not self.training:
if (
("past_key_value" in decoder_layer_kwargs and decoder_layer_kwargs["past_key_value"] is None) or
("layer_past" in decoder_layer_kwargs and decoder_layer_kwargs["layer_past"] is None)
):
single_length = self.vis_x.shape[-2]
image_nums = self.image_nums
image_start_index_list = self.image_start_index_list
image_nums = [0] + np.cumsum(image_nums).tolist()
for i, (image_num_begin, image_num_end, start_indices) in enumerate(zip(image_nums[:-1], image_nums[1:], image_start_index_list)):
for index in start_indices:
if image_num_begin < image_num_end:
hidden_states[i, index:index+single_length] = self.vis_x[image_num_begin]
image_num_begin += 1
if self.visual_tokens is not None and len(self.visual_tokens) != 0:
for i, (x, y) in enumerate(self.data_list):
# import pdb; pdb.set_trace()
# print(x, y, self.visual_tokens[i].shape)
if len(self.visual_tokens[i].shape) > 1:
# print(self.visual_tokens[i].shape[0], "embedding")
hidden_states[x, y+1-self.visual_tokens[i].shape[0]:y+1] = self.visual_tokens[i]
else:
# print(self.visual_tokens[i].shape[0], "embedding")
hidden_states[x, y] = self.visual_tokens[i]
hidden_states = self.decoder_layer(
hidden_states, attention_mask=attention_mask, **decoder_layer_kwargs
)
return hidden_states
class FlamingoLMMixin(nn.Module):
"""
Mixin to add cross-attention layers to a language model.
"""
def set_decoder_layers_attr_name(self, decoder_layers_attr_name):
self.decoder_layers_attr_name = decoder_layers_attr_name
def _get_decoder_layers(self):
|
class FlamingoLayer(nn.Module):
def __init__(self, decoder_layer):
super().__init__()
self.decoder_layer = decoder_layer
self.vis_x = None
self.image_nums = None
self.image_start_index_list = None
def is_conditioned(self) -> bool:
"""Check whether the layer is conditioned."""
return self.vis_x is not None
# Used this great idea from this implementation of Flamingo (https://github.com/dhansmair/flamingo-mini/)
def condition_vis_x(self, vis_x, image_nums=None, image_start_index_list=None, num_beams=None, visual_tokens=None, data_list=None):
self.vis_x = vis_x
self.image_nums = image_nums
self.image_start_index_list = image_start_index_list
self.num_beams = num_beams
self.visual_tokens = visual_tokens
self.data_list = data_list
def forward(
self,
hidden_states, # alignment with hugging face name
attention_mask=None,
**decoder_layer_kwargs,
):
if self.vis_x is not None:
if self.training:
single_length = self.vis_x.shape[-2]
image_nums = self.image_nums
image_start_index_list = self.image_start_index_list
image_nums = [0] + np.cumsum(image_nums).tolist()
for i, (image_num_begin, image_num_end, start_indices) in enumerate(zip(image_nums[:-1], image_nums[1:], image_start_index_list)):
for index in start_indices:
if image_num_begin < image_num_end:
hidden_states[i, index:index+single_length] = self.vis_x[image_num_begin]
image_num_begin += 1
if self.visual_tokens is not None and len(self.visual_tokens) != 0:
for i, (x, y) in enumerate(self.data_list):
if len(self.visual_tokens[i].shape) > 1:
# print(self.visual_tokens[i].shape[0], "embedding")
hidden_states[x, y+1-self.visual_tokens[i].shape[0]:y+1] = self.visual_tokens[i]
else:
# print(self.visual_tokens[i].shape[0], "embedding")
hidden_states[x, y] = self.visual_tokens[i]
elif not self.training:
if (
("past_key_value" in decoder_layer_kwargs and decoder_layer_kwargs["past_key_value"] is None) or
("layer_past" in decoder_layer_kwargs and decoder_layer_kwargs["layer_past"] is None)
):
single_length = self.vis_x.shape[-2]
image_nums = self.image_nums
image_start_index_list = self.image_start_index_list
image_nums = [0] + np.cumsum(image_nums).tolist()
for i, (image_num_begin, image_num_end, start_indices) in enumerate(zip(image_nums[:-1], image_nums[1:], image_start_index_list)):
for index in start_indices:
if image_num_begin < image_num_end:
hidden_states[i, index:index+single_length] = self.vis_x[image_num_begin]
image_num_begin += 1
if self.visual_tokens is not None and len(self.visual_tokens) != 0:
for i, (x, y) in enumerate(self.data_list):
# import pdb; pdb.set_trace()
# print(x, y, self.visual_tokens[i].shape)
if len(self.visual_tokens[i].shape) > 1:
# print(self.visual_tokens[i].shape[0], "embedding")
hidden_states[x, y+1-self.visual_tokens[i].shape[0]:y+1] = self.visual_tokens[i]
else:
# print(self.visual_tokens[i].shape[0], "embedding")
hidden_states[x, y] = self.visual_tokens[i]
hidden_states = self.decoder_layer(
hidden_states, attention_mask=attention_mask, **decoder_layer_kwargs
)
return hidden_states
class FlamingoLMMixin(nn.Module):
"""
Mixin to add cross-attention layers to a language model.
"""
def set_decoder_layers_attr_name(self, decoder_layers_attr_name):
self.decoder_layers_attr_name = decoder_layers_attr_name
def _get_decoder_layers(self): | return getattr_recursive(self, self.decoder_layers_attr_name) | 0 | 2023-11-07 04:23:57+00:00 | 2k |
nouu-me/document_vector_search_benchmark | tools/run_benchmark.py | [
{
"identifier": "DATASET_REGISTRY",
"path": "dvsb/data/dataset.py",
"snippet": "DATASET_REGISTRY = Registry[Dataset]()"
},
{
"identifier": "Dataset",
"path": "dvsb/data/dataset.py",
"snippet": "class Dataset(ABC):\n @abstractmethod\n def get_name(self) -> str:\n \"\"\"Return... | import argparse
import json
import numpy as np
import numpy.typing as npt
import pandas as pd
import yaml
from pathlib import Path
from typing import Iterable
from dvsb.data import DATASET_REGISTRY, Dataset
from dvsb.embedding import EMBEDDING_REGISTRY, Embedding
from dvsb.metric import METRIC_REGISTRY, Metric
from dvsb.relevance import RELEVANCE_REGISTRY, Relevance
from loguru import logger
from tqdm import tqdm | 776 |
def get_parser() -> argparse.ArgumentParser:
parser = argparse.ArgumentParser("run_benchmark")
parser.add_argument("-n", "--name", help="config name", required=False, default="default")
parser.add_argument("--no-cache", action="store_true")
return parser
|
def get_parser() -> argparse.ArgumentParser:
parser = argparse.ArgumentParser("run_benchmark")
parser.add_argument("-n", "--name", help="config name", required=False, default="default")
parser.add_argument("--no-cache", action="store_true")
return parser
| def load_dataset(dataset_config: dict, cache: bool) -> Dataset: | 1 | 2023-11-09 00:04:51+00:00 | 2k |
HKU-BAL/ClairS-TO | src/nonsomatic_tagging.py | [
{
"identifier": "VcfReader",
"path": "shared/vcf.py",
"snippet": "class TruthStdout(object):\nclass VcfWriter(object):\nclass VcfReader(object):\n def __init__(self, handle):\n def __del__(self):\n def __init__(self,\n vcf_fn,\n ctg_name=None,\n r... | import os
import shlex
from argparse import ArgumentParser, SUPPRESS
from collections import defaultdict
from shared.vcf import VcfReader, VcfWriter, Position
from shared.utils import str2bool, str_none, reference_sequence_from, subprocess_popen | 1,163 |
major_contigs_order = ["chr" + str(a) for a in list(range(1, 23)) + ["X", "Y"]] + [str(a) for a in
list(range(1, 23)) + ["X", "Y"]]
class VcfReader_Database(object):
def __init__(self, vcf_fn,
ctg_name=None,
direct_open=False,
keep_row_str=False,
save_header=False):
self.vcf_fn = vcf_fn
self.ctg_name = ctg_name
|
major_contigs_order = ["chr" + str(a) for a in list(range(1, 23)) + ["X", "Y"]] + [str(a) for a in
list(range(1, 23)) + ["X", "Y"]]
class VcfReader_Database(object):
def __init__(self, vcf_fn,
ctg_name=None,
direct_open=False,
keep_row_str=False,
save_header=False):
self.vcf_fn = vcf_fn
self.ctg_name = ctg_name | self.variant_dict = defaultdict(Position) | 0 | 2023-11-07 04:39:16+00:00 | 2k |
the-siesta-group/edfio | tests/test_utils.py | [
{
"identifier": "decode_edfplus_date",
"path": "edfio/_utils.py",
"snippet": "def decode_edfplus_date(date: str) -> datetime.date:\n day, month, year = date.split(\"-\")\n try:\n month_int = _MONTH_NAMES.index(month.upper()) + 1\n except ValueError:\n raise ValueError(f\"Invalid m... | import datetime
import math
import pytest
from edfio._utils import (
decode_edfplus_date,
encode_annotation_duration,
encode_annotation_onset,
encode_edfplus_date,
round_float_to_8_characters,
) | 1,114 |
VALID_EDFPLUS_DATE_PAIRS = (
("02-MAY-1951", datetime.date(1951, 5, 2)),
("02-DEC-1951", datetime.date(1951, 12, 2)),
("02-AUG-1951", datetime.date(1951, 8, 2)),
("02-MAY-2051", datetime.date(2051, 5, 2)),
)
@pytest.mark.parametrize(("string", "datetime_"), VALID_EDFPLUS_DATE_PAIRS)
def test_decode_edfplus_date(string: str, datetime_: datetime.date):
assert decode_edfplus_date(string) == datetime_
@pytest.mark.parametrize(("string", "datetime_"), VALID_EDFPLUS_DATE_PAIRS)
def test_encode_edfplus_date(string: str, datetime_: datetime.date):
assert encode_edfplus_date(datetime_) == string
def test_decode_edfplus_date_invalid_month_name():
with pytest.raises(ValueError, match="Invalid month"):
decode_edfplus_date("02-MAI-1951")
@pytest.mark.parametrize(
("onset", "expected"),
[
(0, "+0"),
(0.0, "+0"),
(0.1, "+0.1"),
(0.01, "+0.01"),
(0.001, "+0.001"),
(0.0001, "+0.0001"),
(0.00001, "+0.00001"),
(0.000001, "+0.000001"),
(0.0000001, "+0.0000001"),
(0.00000001, "+0.00000001"),
(0.00000000001, "+0.00000000001"),
(100000000000.0, "+100000000000"),
(-0.1, "-0.1"),
(-0.0000001, "-0.0000001"),
(-0.0000000001, "-0.0000000001"),
(-100000000000.0, "-100000000000"),
],
)
def test_encode_annotation_onset(onset: float, expected: str):
assert encode_annotation_onset(onset) == expected
@pytest.mark.parametrize(
("duration", "expected"),
[
(0, "0"),
(0.0, "0"),
(0.1, "0.1"),
(0.01, "0.01"),
(0.001, "0.001"),
(0.0001, "0.0001"),
(0.00001, "0.00001"),
(0.000001, "0.000001"),
(0.0000001, "0.0000001"),
(0.00000000001, "0.00000000001"),
(100000000000.0, "100000000000"),
],
)
def test_encode_annotation_duration(duration: float, expected: str):
|
VALID_EDFPLUS_DATE_PAIRS = (
("02-MAY-1951", datetime.date(1951, 5, 2)),
("02-DEC-1951", datetime.date(1951, 12, 2)),
("02-AUG-1951", datetime.date(1951, 8, 2)),
("02-MAY-2051", datetime.date(2051, 5, 2)),
)
@pytest.mark.parametrize(("string", "datetime_"), VALID_EDFPLUS_DATE_PAIRS)
def test_decode_edfplus_date(string: str, datetime_: datetime.date):
assert decode_edfplus_date(string) == datetime_
@pytest.mark.parametrize(("string", "datetime_"), VALID_EDFPLUS_DATE_PAIRS)
def test_encode_edfplus_date(string: str, datetime_: datetime.date):
assert encode_edfplus_date(datetime_) == string
def test_decode_edfplus_date_invalid_month_name():
with pytest.raises(ValueError, match="Invalid month"):
decode_edfplus_date("02-MAI-1951")
@pytest.mark.parametrize(
("onset", "expected"),
[
(0, "+0"),
(0.0, "+0"),
(0.1, "+0.1"),
(0.01, "+0.01"),
(0.001, "+0.001"),
(0.0001, "+0.0001"),
(0.00001, "+0.00001"),
(0.000001, "+0.000001"),
(0.0000001, "+0.0000001"),
(0.00000001, "+0.00000001"),
(0.00000000001, "+0.00000000001"),
(100000000000.0, "+100000000000"),
(-0.1, "-0.1"),
(-0.0000001, "-0.0000001"),
(-0.0000000001, "-0.0000000001"),
(-100000000000.0, "-100000000000"),
],
)
def test_encode_annotation_onset(onset: float, expected: str):
assert encode_annotation_onset(onset) == expected
@pytest.mark.parametrize(
("duration", "expected"),
[
(0, "0"),
(0.0, "0"),
(0.1, "0.1"),
(0.01, "0.01"),
(0.001, "0.001"),
(0.0001, "0.0001"),
(0.00001, "0.00001"),
(0.000001, "0.000001"),
(0.0000001, "0.0000001"),
(0.00000000001, "0.00000000001"),
(100000000000.0, "100000000000"),
],
)
def test_encode_annotation_duration(duration: float, expected: str): | assert encode_annotation_duration(duration) == expected | 1 | 2023-11-09 09:53:27+00:00 | 2k |
microsoft/folx | folx/operators.py | [
{
"identifier": "Array",
"path": "folx/api.py",
"snippet": "T = TypeVar(\"T\", bound=PyTree[Array])\nR = TypeVar(\"R\", bound=PyTree[Array])\nJAC_DIM = 0 # should be either 0 or -1. TODO: switching is not support.\n GENERAL = 0\n LINEAR_IN_FIRST = 1\n LINEAR_IN_ONE = 2 | LINEAR_IN_FIRST\n L... | from dataclasses import dataclass
from typing import Callable, Protocol
from .api import Array
from .interpreter import forward_laplacian
import jax
import jax.numpy as jnp | 1,332 |
__all__ = [
"Laplacian",
"LaplacianOperator",
"ForwardLaplacianOperator",
"LoopLaplacianOperator",
"ParallelLaplacianOperator",
]
class Laplacian(Protocol):
|
__all__ = [
"Laplacian",
"LaplacianOperator",
"ForwardLaplacianOperator",
"LoopLaplacianOperator",
"ParallelLaplacianOperator",
]
class Laplacian(Protocol): | def __call__(self, x: Array) -> tuple[Array, Array]: | 0 | 2023-11-07 16:32:46+00:00 | 2k |
shuttworth/NICE-SLAM-Easyread | visualizer.py | [
{
"identifier": "config",
"path": "src/config.py",
"snippet": "def load_config(path, default_path=None):\ndef update_recursive(dict1, dict2):\ndef get_model(cfg, nice=True):"
},
{
"identifier": "SLAMFrontend",
"path": "src/tools/viz.py",
"snippet": "class SLAMFrontend:\n def __init__(... | import argparse
import os
import time
import numpy as np
import torch
import cv2
from tqdm import tqdm
from torch.utils.data import DataLoader
from src import config
from src.tools.viz import SLAMFrontend
from src.utils.datasets import get_dataset | 985 |
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Arguments to visualize the SLAM process.'
)
parser.add_argument('config', type=str, help='Path to config file.')
parser.add_argument('--input_folder', type=str,
help='input folder, this have higher priority, can overwrite the one in config file')
parser.add_argument('--output', type=str,
help='output folder, this have higher priority, can overwrite the one inconfig file')
nice_parser = parser.add_mutually_exclusive_group(required=False)
nice_parser.add_argument('--nice', dest='nice', action='store_true')
nice_parser.add_argument('--imap', dest='nice', action='store_false')
parser.set_defaults(nice=True)
parser.add_argument('--save_rendering',
action='store_true', help='save rendering video to `vis.mp4` in output folder ')
parser.add_argument('--vis_input_frame',
action='store_true', help='visualize input frames')
parser.add_argument('--no_gt_traj',
action='store_true', help='not visualize gt trajectory')
args = parser.parse_args()
cfg = config.load_config(
args.config, 'configs/nice_slam.yaml' if args.nice else 'configs/imap.yaml')
scale = cfg['scale']
output = cfg['data']['output'] if args.output is None else args.output
if args.vis_input_frame:
frame_reader = get_dataset(cfg, args, scale, device='cpu')
frame_loader = DataLoader(
frame_reader, batch_size=1, shuffle=False, num_workers=4)
ckptsdir = f'{output}/ckpts'
if os.path.exists(ckptsdir):
ckpts = [os.path.join(ckptsdir, f)
for f in sorted(os.listdir(ckptsdir)) if 'tar' in f]
if len(ckpts) > 0:
ckpt_path = ckpts[-1]
print('Get ckpt :', ckpt_path)
ckpt = torch.load(ckpt_path, map_location=torch.device('cpu'))
estimate_c2w_list = ckpt['estimate_c2w_list']
gt_c2w_list = ckpt['gt_c2w_list']
N = ckpt['idx']
estimate_c2w_list[:, :3, 3] /= scale
gt_c2w_list[:, :3, 3] /= scale
estimate_c2w_list = estimate_c2w_list.cpu().numpy()
gt_c2w_list = gt_c2w_list.cpu().numpy()
|
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Arguments to visualize the SLAM process.'
)
parser.add_argument('config', type=str, help='Path to config file.')
parser.add_argument('--input_folder', type=str,
help='input folder, this have higher priority, can overwrite the one in config file')
parser.add_argument('--output', type=str,
help='output folder, this have higher priority, can overwrite the one inconfig file')
nice_parser = parser.add_mutually_exclusive_group(required=False)
nice_parser.add_argument('--nice', dest='nice', action='store_true')
nice_parser.add_argument('--imap', dest='nice', action='store_false')
parser.set_defaults(nice=True)
parser.add_argument('--save_rendering',
action='store_true', help='save rendering video to `vis.mp4` in output folder ')
parser.add_argument('--vis_input_frame',
action='store_true', help='visualize input frames')
parser.add_argument('--no_gt_traj',
action='store_true', help='not visualize gt trajectory')
args = parser.parse_args()
cfg = config.load_config(
args.config, 'configs/nice_slam.yaml' if args.nice else 'configs/imap.yaml')
scale = cfg['scale']
output = cfg['data']['output'] if args.output is None else args.output
if args.vis_input_frame:
frame_reader = get_dataset(cfg, args, scale, device='cpu')
frame_loader = DataLoader(
frame_reader, batch_size=1, shuffle=False, num_workers=4)
ckptsdir = f'{output}/ckpts'
if os.path.exists(ckptsdir):
ckpts = [os.path.join(ckptsdir, f)
for f in sorted(os.listdir(ckptsdir)) if 'tar' in f]
if len(ckpts) > 0:
ckpt_path = ckpts[-1]
print('Get ckpt :', ckpt_path)
ckpt = torch.load(ckpt_path, map_location=torch.device('cpu'))
estimate_c2w_list = ckpt['estimate_c2w_list']
gt_c2w_list = ckpt['gt_c2w_list']
N = ckpt['idx']
estimate_c2w_list[:, :3, 3] /= scale
gt_c2w_list[:, :3, 3] /= scale
estimate_c2w_list = estimate_c2w_list.cpu().numpy()
gt_c2w_list = gt_c2w_list.cpu().numpy()
| frontend = SLAMFrontend(output, init_pose=estimate_c2w_list[0], cam_scale=0.3, | 1 | 2023-11-07 05:09:36+00:00 | 2k |
mileswyn/SAMIHS | models/segment_anything/modeling/image_encoder.py | [
{
"identifier": "LayerNorm2d",
"path": "models/segment_anything/modeling/common.py",
"snippet": "class LayerNorm2d(nn.Module):\n def __init__(self, num_channels: int, eps: float = 1e-6) -> None:\n super().__init__()\n self.weight = nn.Parameter(torch.ones(num_channels))\n self.bi... | import torch
import torch.nn as nn
import torch.nn.functional as F
from typing import Optional, Tuple, Type
from .common import LayerNorm2d, MLPBlock | 1,138 | # Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
# This class and its supporting functions below lightly adapted from the ViTDet backbone available at: https://github.com/facebookresearch/detectron2/blob/main/detectron2/modeling/backbone/vit.py # noqa
class ImageEncoderViT(nn.Module):
def __init__(
self,
img_size: int = 1024,
patch_size: int = 16,
in_chans: int = 3,
embed_dim: int = 768,
depth: int = 12,
num_heads: int = 12,
mlp_ratio: float = 4.0,
out_chans: int = 256,
qkv_bias: bool = True,
norm_layer: Type[nn.Module] = nn.LayerNorm,
act_layer: Type[nn.Module] = nn.GELU,
use_abs_pos: bool = True,
use_rel_pos: bool = False,
rel_pos_zero_init: bool = True,
window_size: int = 0,
global_attn_indexes: Tuple[int, ...] = (),
) -> None:
"""
Args:
img_size (int): Input image size.
patch_size (int): Patch size.
in_chans (int): Number of input image channels.
embed_dim (int): Patch embedding dimension.
depth (int): Depth of ViT.
num_heads (int): Number of attention heads in each ViT block.
mlp_ratio (float): Ratio of mlp hidden dim to embedding dim.
qkv_bias (bool): If True, add a learnable bias to query, key, value.
norm_layer (nn.Module): Normalization layer.
act_layer (nn.Module): Activation layer.
use_abs_pos (bool): If True, use absolute positional embeddings.
use_rel_pos (bool): If True, add relative positional embeddings to the attention map.
rel_pos_zero_init (bool): If True, zero initialize relative positional parameters.
window_size (int): Window size for window attention blocks.
global_attn_indexes (list): Indexes for blocks using global attention.
"""
super().__init__()
self.img_size = img_size
self.patch_embed = PatchEmbed(
kernel_size=(patch_size, patch_size),
stride=(patch_size, patch_size),
in_chans=in_chans,
embed_dim=embed_dim,
)
self.pos_embed: Optional[nn.Parameter] = None
if use_abs_pos:
# Initialize absolute positional embedding with pretrain image size.
self.pos_embed = nn.Parameter(
torch.zeros(1, img_size // patch_size, img_size // patch_size, embed_dim)
)
self.blocks = nn.ModuleList()
for i in range(depth):
block = Block(
dim=embed_dim,
num_heads=num_heads,
mlp_ratio=mlp_ratio,
qkv_bias=qkv_bias,
norm_layer=norm_layer,
act_layer=act_layer,
use_rel_pos=use_rel_pos,
rel_pos_zero_init=rel_pos_zero_init,
window_size=window_size if i not in global_attn_indexes else 0,
input_size=(img_size // patch_size, img_size // patch_size),
)
self.blocks.append(block)
self.neck = nn.Sequential(
nn.Conv2d(
embed_dim,
out_chans,
kernel_size=1,
bias=False,
),
| # Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
# This class and its supporting functions below lightly adapted from the ViTDet backbone available at: https://github.com/facebookresearch/detectron2/blob/main/detectron2/modeling/backbone/vit.py # noqa
class ImageEncoderViT(nn.Module):
def __init__(
self,
img_size: int = 1024,
patch_size: int = 16,
in_chans: int = 3,
embed_dim: int = 768,
depth: int = 12,
num_heads: int = 12,
mlp_ratio: float = 4.0,
out_chans: int = 256,
qkv_bias: bool = True,
norm_layer: Type[nn.Module] = nn.LayerNorm,
act_layer: Type[nn.Module] = nn.GELU,
use_abs_pos: bool = True,
use_rel_pos: bool = False,
rel_pos_zero_init: bool = True,
window_size: int = 0,
global_attn_indexes: Tuple[int, ...] = (),
) -> None:
"""
Args:
img_size (int): Input image size.
patch_size (int): Patch size.
in_chans (int): Number of input image channels.
embed_dim (int): Patch embedding dimension.
depth (int): Depth of ViT.
num_heads (int): Number of attention heads in each ViT block.
mlp_ratio (float): Ratio of mlp hidden dim to embedding dim.
qkv_bias (bool): If True, add a learnable bias to query, key, value.
norm_layer (nn.Module): Normalization layer.
act_layer (nn.Module): Activation layer.
use_abs_pos (bool): If True, use absolute positional embeddings.
use_rel_pos (bool): If True, add relative positional embeddings to the attention map.
rel_pos_zero_init (bool): If True, zero initialize relative positional parameters.
window_size (int): Window size for window attention blocks.
global_attn_indexes (list): Indexes for blocks using global attention.
"""
super().__init__()
self.img_size = img_size
self.patch_embed = PatchEmbed(
kernel_size=(patch_size, patch_size),
stride=(patch_size, patch_size),
in_chans=in_chans,
embed_dim=embed_dim,
)
self.pos_embed: Optional[nn.Parameter] = None
if use_abs_pos:
# Initialize absolute positional embedding with pretrain image size.
self.pos_embed = nn.Parameter(
torch.zeros(1, img_size // patch_size, img_size // patch_size, embed_dim)
)
self.blocks = nn.ModuleList()
for i in range(depth):
block = Block(
dim=embed_dim,
num_heads=num_heads,
mlp_ratio=mlp_ratio,
qkv_bias=qkv_bias,
norm_layer=norm_layer,
act_layer=act_layer,
use_rel_pos=use_rel_pos,
rel_pos_zero_init=rel_pos_zero_init,
window_size=window_size if i not in global_attn_indexes else 0,
input_size=(img_size // patch_size, img_size // patch_size),
)
self.blocks.append(block)
self.neck = nn.Sequential(
nn.Conv2d(
embed_dim,
out_chans,
kernel_size=1,
bias=False,
), | LayerNorm2d(out_chans), | 0 | 2023-11-09 07:26:33+00:00 | 2k |
AlexandrErohin/home-assistant-tplink-router | custom_components/tplink_router/switch.py | [
{
"identifier": "DOMAIN",
"path": "custom_components/tplink_router/const.py",
"snippet": "DOMAIN = \"tplink_router\""
},
{
"identifier": "TPLinkRouterCoordinator",
"path": "custom_components/tplink_router/coordinator.py",
"snippet": "class TPLinkRouterCoordinator(DataUpdateCoordinator):\... | from collections.abc import Callable
from dataclasses import dataclass
from typing import Any
from homeassistant.components.switch import SwitchEntity, SwitchEntityDescription
from homeassistant.config_entries import ConfigEntry
from homeassistant.const import EntityCategory
from homeassistant.core import HomeAssistant
from homeassistant.helpers.entity_platform import AddEntitiesCallback
from .const import DOMAIN
from homeassistant.helpers.update_coordinator import CoordinatorEntity
from .coordinator import TPLinkRouterCoordinator
from tplinkrouterc6u import Wifi, Status
from homeassistant.helpers.device_registry import DeviceInfo | 1,050 | from __future__ import annotations
@dataclass
class TPLinkRouterSwitchEntityDescriptionMixin:
method: Callable[[TPLinkRouterCoordinator, bool], Any]
property: str
@dataclass
class TPLinkRouterSwitchEntityDescription(SwitchEntityDescription, TPLinkRouterSwitchEntityDescriptionMixin):
"""A class that describes sensor entities."""
SWITCH_TYPES = (
TPLinkRouterSwitchEntityDescription(
key="wifi_guest_24g",
name="Guest WIFI 2.4G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='guest_2g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_GUEST_2G, value),
),
TPLinkRouterSwitchEntityDescription(
key="wifi_guest_5g",
name="Guest WIFI 5G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='guest_5g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_GUEST_5G, value),
),
TPLinkRouterSwitchEntityDescription(
key="wifi_24g",
name="WIFI 2.4G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='wifi_2g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_2G, value),
),
TPLinkRouterSwitchEntityDescription(
key="wifi_5g",
name="WIFI 5G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='wifi_5g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_5G, value),
),
TPLinkRouterSwitchEntityDescription(
key="iot_24g",
name="IoT WIFI 2.4G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='iot_2g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_IOT_2G, value),
),
TPLinkRouterSwitchEntityDescription(
key="iot_5g",
name="IoT WIFI 5G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='iot_5g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_IOT_5G, value),
),
)
async def async_setup_entry(
hass: HomeAssistant,
entry: ConfigEntry,
async_add_entities: AddEntitiesCallback,
) -> None:
| from __future__ import annotations
@dataclass
class TPLinkRouterSwitchEntityDescriptionMixin:
method: Callable[[TPLinkRouterCoordinator, bool], Any]
property: str
@dataclass
class TPLinkRouterSwitchEntityDescription(SwitchEntityDescription, TPLinkRouterSwitchEntityDescriptionMixin):
"""A class that describes sensor entities."""
SWITCH_TYPES = (
TPLinkRouterSwitchEntityDescription(
key="wifi_guest_24g",
name="Guest WIFI 2.4G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='guest_2g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_GUEST_2G, value),
),
TPLinkRouterSwitchEntityDescription(
key="wifi_guest_5g",
name="Guest WIFI 5G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='guest_5g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_GUEST_5G, value),
),
TPLinkRouterSwitchEntityDescription(
key="wifi_24g",
name="WIFI 2.4G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='wifi_2g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_2G, value),
),
TPLinkRouterSwitchEntityDescription(
key="wifi_5g",
name="WIFI 5G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='wifi_5g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_5G, value),
),
TPLinkRouterSwitchEntityDescription(
key="iot_24g",
name="IoT WIFI 2.4G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='iot_2g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_IOT_2G, value),
),
TPLinkRouterSwitchEntityDescription(
key="iot_5g",
name="IoT WIFI 5G",
icon="mdi:wifi",
entity_category=EntityCategory.CONFIG,
property='iot_5g_enable',
method=lambda coordinator, value: coordinator.set_wifi(Wifi.WIFI_IOT_5G, value),
),
)
async def async_setup_entry(
hass: HomeAssistant,
entry: ConfigEntry,
async_add_entities: AddEntitiesCallback,
) -> None: | coordinator = hass.data[DOMAIN][entry.entry_id] | 0 | 2023-11-09 17:38:33+00:00 | 2k |
DaveParr/starpilot | tests/test_utils.py | [
{
"identifier": "get_repo_contents",
"path": "starpilot/utils/utils.py",
"snippet": "def get_repo_contents(\n repos: List[Repository], g: Github, include_readmes: bool = False\n) -> List[Dict]:\n repo_infos = []\n for repo in track(repos, description=\"Reading the stars...\"):\n repo_inf... | from unittest.mock import Mock
from starpilot.utils.utils import get_repo_contents, get_user_starred_repos
import pytest
import os
import github | 1,103 |
def test_get_user_starred_repos_mocked():
# Mock the necessary objects
class MockRepo:
def __init__(self, stargazers_count):
self.stargazers_count = stargazers_count
class MockUser:
def get_starred(self):
return [MockRepo(10), MockRepo(5), MockRepo(8), MockRepo(3), MockRepo(7)]
class MockGithub:
def get_user(self, user):
return MockUser()
# Call the function under test
|
def test_get_user_starred_repos_mocked():
# Mock the necessary objects
class MockRepo:
def __init__(self, stargazers_count):
self.stargazers_count = stargazers_count
class MockUser:
def get_starred(self):
return [MockRepo(10), MockRepo(5), MockRepo(8), MockRepo(3), MockRepo(7)]
class MockGithub:
def get_user(self, user):
return MockUser()
# Call the function under test | result = get_user_starred_repos("testuser", MockGithub(), num_repos=3) | 1 | 2023-11-07 20:03:08+00:00 | 2k |
xarray-contrib/xdggs | xdggs/h3.py | [
{
"identifier": "DGGSIndex",
"path": "xdggs/index.py",
"snippet": "class DGGSIndex(Index):\n _dim: str\n _pd_index: PandasIndex\n\n def __init__(self, cell_ids: Any | PandasIndex, dim: str):\n self._dim = dim\n\n if isinstance(cell_ids, PandasIndex):\n self._pd_index = ... | from collections.abc import Mapping
from typing import Any
from h3ronpy.arrow.vector import cells_to_coordinates, coordinates_to_cells
from xarray.indexes import PandasIndex
from xdggs.index import DGGSIndex
from xdggs.utils import _extract_cell_id_variable, register_dggs
import numpy as np
import xarray as xr | 883 |
@register_dggs("h3")
class H3Index(DGGSIndex):
_resolution: int
def __init__(
self,
cell_ids: Any | PandasIndex,
dim: str,
resolution: int,
):
super().__init__(cell_ids, dim)
self._resolution = int(resolution)
@classmethod
def from_variables(
cls: type["H3Index"],
variables: Mapping[Any, xr.Variable],
*,
options: Mapping[str, Any],
) -> "H3Index":
|
@register_dggs("h3")
class H3Index(DGGSIndex):
_resolution: int
def __init__(
self,
cell_ids: Any | PandasIndex,
dim: str,
resolution: int,
):
super().__init__(cell_ids, dim)
self._resolution = int(resolution)
@classmethod
def from_variables(
cls: type["H3Index"],
variables: Mapping[Any, xr.Variable],
*,
options: Mapping[str, Any],
) -> "H3Index": | _, var, dim = _extract_cell_id_variable(variables) | 1 | 2023-11-06 16:11:15+00:00 | 2k |
ApolloAuto/apollo-model-centerpoint | paddle3d/utils/checkpoint.py | [
{
"identifier": "PRETRAINED_HOME",
"path": "paddle3d/env.py",
"snippet": "PRETRAINED_HOME = get_sub_home('pretrained')"
},
{
"identifier": "TMP_HOME",
"path": "paddle3d/env.py",
"snippet": "TMP_HOME = get_sub_home('tmp')"
},
{
"identifier": "download_with_progress",
"path": "... | import os
import filelock
import paddle
from typing import Union
from urllib.parse import unquote, urlparse
from paddle3d.env import PRETRAINED_HOME, TMP_HOME
from paddle3d.utils.download import download_with_progress
from paddle3d.utils.logger import logger
from paddle3d.utils.xarfile import unarchive_with_progress | 1,120 | # Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
def load_pretrained_model_from_url(model: paddle.nn.Layer,
url: str,
overwrite: bool = False):
"""
"""
pretrained_model = unquote(url)
savename = pretrained_model.split('/')[-1]
savedir = os.path.join(PRETRAINED_HOME, savename.split('.')[0])
os.makedirs(savedir, exist_ok=True)
savepath = os.path.join(savedir, savename)
if os.path.exists(savepath) and not overwrite:
| # Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
def load_pretrained_model_from_url(model: paddle.nn.Layer,
url: str,
overwrite: bool = False):
"""
"""
pretrained_model = unquote(url)
savename = pretrained_model.split('/')[-1]
savedir = os.path.join(PRETRAINED_HOME, savename.split('.')[0])
os.makedirs(savedir, exist_ok=True)
savepath = os.path.join(savedir, savename)
if os.path.exists(savepath) and not overwrite: | logger.warning( | 3 | 2023-11-08 07:08:03+00:00 | 2k |
camlsys/fl-project-template | project/task/default/train_test.py | [
{
"identifier": "ClientConfig",
"path": "project/client/client.py",
"snippet": "class ClientConfig(BaseModel):\n \"\"\"Fit/eval config, allows '.' member acces and static checking.\n\n Used to check weather each component has its own independent config present. Each\n component should then use ... | from collections.abc import Sized
from pathlib import Path
from typing import cast
from flwr.common import NDArrays
from pydantic import BaseModel
from torch import nn
from torch.utils.data import DataLoader
from project.client.client import ClientConfig
from project.fed.utils.utils import generic_set_parameters
from project.types.common import (
FedDataloaderGen,
FedEvalFN,
NetGen,
OnFitConfigFN,
TestFunc,
)
from project.utils.utils import obtain_device
import torch | 1,246 |
class TrainConfig(BaseModel):
"""Training configuration, allows '.' member acces and static checking.
Guarantees that all necessary components are present, fails early if config is
mismatched to client.
"""
device: torch.device
# epochs: int
# learning_rate: float
class Config:
"""Setting to allow any types, including library ones like torch.device."""
arbitrary_types_allowed = True
def train(
net: nn.Module,
trainloader: DataLoader,
_config: dict,
_working_dir: Path,
) -> tuple[int, dict]:
"""Train the network on the training set.
Parameters
----------
net : nn.Module
The neural network to train.
trainloader : DataLoader
The DataLoader containing the data to train the network on.
_config : Dict
The configuration for the training.
Contains the device, number of epochs and learning rate.
Static type checking is done by the TrainConfig class.
Returns
-------
Tuple[int, Dict]
The number of samples used for training,
the loss, and the accuracy of the input model on the given data.
"""
if len(cast(Sized, trainloader.dataset)) == 0:
raise ValueError(
"Trainloader can't be 0, exiting...",
)
config: TrainConfig = TrainConfig(**_config)
del _config
net.to(config.device)
net.train()
return len(cast(Sized, trainloader.dataset)), {}
class TestConfig(BaseModel):
"""Testing configuration, allows '.' member acces and static checking.
Guarantees that all necessary components are present, fails early if config is
mismatched to client.
"""
device: torch.device
class Config:
"""Setting to allow any types, including library ones like torch.device."""
arbitrary_types_allowed = True
def test(
net: nn.Module,
testloader: DataLoader,
_config: dict,
_working_dir: Path,
) -> tuple[float, int, dict]:
"""Evaluate the network on the test set.
Parameters
----------
net : nn.Module
The neural network to test.
testloader : DataLoader
The DataLoader containing the data to test the network on.
config : Dict
The configuration for the testing.
Contains the device.
Static type checking is done by the TestConfig class.
Returns
-------
Tuple[float, int, float]
The loss, number of test samples,
and the accuracy of the input model on the given data.
"""
if len(cast(Sized, testloader.dataset)) == 0:
raise ValueError(
"Testloader can't be 0, exiting...",
)
config: TestConfig = TestConfig(**_config)
del _config
net.to(config.device)
net.eval()
return (
0.0,
len(cast(Sized, testloader.dataset)),
{},
)
def get_fed_eval_fn(
net_generator: NetGen,
| """Default training and testing functions, local and federated."""
class TrainConfig(BaseModel):
"""Training configuration, allows '.' member acces and static checking.
Guarantees that all necessary components are present, fails early if config is
mismatched to client.
"""
device: torch.device
# epochs: int
# learning_rate: float
class Config:
"""Setting to allow any types, including library ones like torch.device."""
arbitrary_types_allowed = True
def train(
net: nn.Module,
trainloader: DataLoader,
_config: dict,
_working_dir: Path,
) -> tuple[int, dict]:
"""Train the network on the training set.
Parameters
----------
net : nn.Module
The neural network to train.
trainloader : DataLoader
The DataLoader containing the data to train the network on.
_config : Dict
The configuration for the training.
Contains the device, number of epochs and learning rate.
Static type checking is done by the TrainConfig class.
Returns
-------
Tuple[int, Dict]
The number of samples used for training,
the loss, and the accuracy of the input model on the given data.
"""
if len(cast(Sized, trainloader.dataset)) == 0:
raise ValueError(
"Trainloader can't be 0, exiting...",
)
config: TrainConfig = TrainConfig(**_config)
del _config
net.to(config.device)
net.train()
return len(cast(Sized, trainloader.dataset)), {}
class TestConfig(BaseModel):
"""Testing configuration, allows '.' member acces and static checking.
Guarantees that all necessary components are present, fails early if config is
mismatched to client.
"""
device: torch.device
class Config:
"""Setting to allow any types, including library ones like torch.device."""
arbitrary_types_allowed = True
def test(
net: nn.Module,
testloader: DataLoader,
_config: dict,
_working_dir: Path,
) -> tuple[float, int, dict]:
"""Evaluate the network on the test set.
Parameters
----------
net : nn.Module
The neural network to test.
testloader : DataLoader
The DataLoader containing the data to test the network on.
config : Dict
The configuration for the testing.
Contains the device.
Static type checking is done by the TestConfig class.
Returns
-------
Tuple[float, int, float]
The loss, number of test samples,
and the accuracy of the input model on the given data.
"""
if len(cast(Sized, testloader.dataset)) == 0:
raise ValueError(
"Testloader can't be 0, exiting...",
)
config: TestConfig = TestConfig(**_config)
del _config
net.to(config.device)
net.eval()
return (
0.0,
len(cast(Sized, testloader.dataset)),
{},
)
def get_fed_eval_fn(
net_generator: NetGen, | fed_dataloater_generator: FedDataloaderGen, | 2 | 2023-11-08 15:31:44+00:00 | 2k |
alibaba/CloudEval-YAML | evaluate.py | [
{
"identifier": "bleu",
"path": "metrics/bleu.py",
"snippet": "def test(result_str=\"\", reference_str=\"\"):"
},
{
"identifier": "edit_distance",
"path": "metrics/edit_distance.py",
"snippet": "def test(result_str=\"\", reference_str=\"\"):"
},
{
"identifier": "exact_match",
... | import loader
import prompt
import query
import json
import ray
import os
import openai
import time
import importlib
import sys
import random
from tqdm import tqdm
from metrics import bleu, edit_distance, exact_match, kv_match
from metrics import kv_wildcard, unit_test, unit_test_pred | 663 |
metric_map = {
'bleu': bleu,
'edit_distance': edit_distance,
'exact_match': exact_match,
'kv_match': kv_match,
}
def import_module_from_string(module_name, module_code):
module_spec = importlib.util.spec_from_loader(module_name, loader=None)
module = importlib.util.module_from_spec(module_spec)
sys.modules[module_name] = module
exec(module_code, module.__dict__)
return module
def evaluate(config):
problem_keys = json.loads(loader.read_memory(config, 'problem_keys'))
if 'Kubernetes' in config['libs'] and config['metrics']['unit_test']:
# restart minikube
print("Preparing minikube for unit tests ...")
unit_test.run_bash('minikube delete && minikube start && sleep 20')
for problem_key in tqdm(problem_keys, desc='Evaluating'):
reference_code = loader.read_memory_problem(config, problem_key, f'reference_code')
unit_test_code = loader.read_memory_problem(config, problem_key, f'unit_test_code')
if unit_test_code is not None:
unit_test_code = unit_test_code
scores_dict = {}
for metric_name in config['metrics'].keys():
scores_dict[metric_name] = []
for sample_id in range(config['num_samples']):
generated_code = loader.read_memory_problem(config, problem_key, f'generated_code_{sample_id}')
for metric_name in config['metrics'].keys():
if config['metrics'][metric_name]:
# if the metric is enabled
if metric_name in metric_map.keys():
# general scoring
score = metric_map[metric_name].test(generated_code, reference_code)
# problem-specific scoring
elif metric_name == 'kv_wildcard':
try:
|
metric_map = {
'bleu': bleu,
'edit_distance': edit_distance,
'exact_match': exact_match,
'kv_match': kv_match,
}
def import_module_from_string(module_name, module_code):
module_spec = importlib.util.spec_from_loader(module_name, loader=None)
module = importlib.util.module_from_spec(module_spec)
sys.modules[module_name] = module
exec(module_code, module.__dict__)
return module
def evaluate(config):
problem_keys = json.loads(loader.read_memory(config, 'problem_keys'))
if 'Kubernetes' in config['libs'] and config['metrics']['unit_test']:
# restart minikube
print("Preparing minikube for unit tests ...")
unit_test.run_bash('minikube delete && minikube start && sleep 20')
for problem_key in tqdm(problem_keys, desc='Evaluating'):
reference_code = loader.read_memory_problem(config, problem_key, f'reference_code')
unit_test_code = loader.read_memory_problem(config, problem_key, f'unit_test_code')
if unit_test_code is not None:
unit_test_code = unit_test_code
scores_dict = {}
for metric_name in config['metrics'].keys():
scores_dict[metric_name] = []
for sample_id in range(config['num_samples']):
generated_code = loader.read_memory_problem(config, problem_key, f'generated_code_{sample_id}')
for metric_name in config['metrics'].keys():
if config['metrics'][metric_name]:
# if the metric is enabled
if metric_name in metric_map.keys():
# general scoring
score = metric_map[metric_name].test(generated_code, reference_code)
# problem-specific scoring
elif metric_name == 'kv_wildcard':
try: | score = kv_wildcard.test(generated_code, reference_code) | 4 | 2023-11-08 08:13:39+00:00 | 2k |
KAIST-AILab/palr | rlkit/torch/sac/sac.py | [
{
"identifier": "LossFunction",
"path": "rlkit/core/loss.py",
"snippet": "class LossFunction(object, metaclass=abc.ABCMeta):\n def compute_loss(self, batch, skip_statistics=False, **kwargs):"
},
{
"identifier": "create_stats_ordered_dict",
"path": "rlkit/core/eval_util.py",
"snippet":... | from collections import OrderedDict, namedtuple
from typing import Tuple
from rlkit.core.loss import LossFunction, LossStatistics
from torch import nn as nn
from rlkit.core.eval_util import create_stats_ordered_dict
from rlkit.torch.torch_rl_algorithm import TorchTrainer
from rlkit.core.logging import add_prefix
from mine.mine import MINE_DV
import numpy as np
import torch
import torch.optim as optim
import rlkit.torch.pytorch_util as ptu
import gtimer as gt | 1,192 |
SACLosses = namedtuple(
'SACLosses',
'policy_loss qf1_loss qf2_loss alpha_loss',
)
|
SACLosses = namedtuple(
'SACLosses',
'policy_loss qf1_loss qf2_loss alpha_loss',
)
| class SACTrainer(TorchTrainer, LossFunction): | 0 | 2023-11-06 08:35:34+00:00 | 2k |
JustlfC03/SCUNet-plusplus | trainer.py | [
{
"identifier": "DiceLoss",
"path": "utils.py",
"snippet": "class DiceLoss(nn.Module):\n def __init__(self, n_classes):\n super(DiceLoss, self).__init__()\n self.n_classes = n_classes\n\n def _one_hot_encoder(self, input_tensor):\n tensor_list = []\n for i in range(self... | import argparse
import logging
import os
import random
import sys
import time
import numpy as np
import torch
import torch.nn as nn
import torch.optim as optim
from tensorboardX import SummaryWriter
from torch.nn.modules.loss import CrossEntropyLoss
from torch.utils.data import DataLoader
from tqdm import tqdm
from utils import DiceLoss
from torchvision import transforms
from utils import test_single_volume
from datasets.dataset_synapse import Synapse_dataset, RandomGenerator | 1,250 |
def trainer_synapse(args, model, snapshot_path):
logging.basicConfig(filename=snapshot_path + "/log.txt", level=logging.INFO,
format='[%(asctime)s.%(msecs)03d] %(message)s', datefmt='%H:%M:%S')
logging.getLogger().addHandler(logging.StreamHandler(sys.stdout))
logging.info(str(args))
base_lr = args.base_lr
num_classes = args.num_classes
batch_size = args.batch_size * args.n_gpu
max_iterations = args.max_iterations
db_train = Synapse_dataset(base_dir=args.root_path, list_dir=args.list_dir, split="train",
transform=transforms.Compose(
[RandomGenerator(output_size=[args.img_size, args.img_size])]))
print("The length of train set is: {}".format(len(db_train)))
def worker_init_fn(worker_id):
random.seed(args.seed + worker_id)
trainloader = DataLoader(db_train, batch_size=batch_size, shuffle=True, num_workers=0, pin_memory=True,
worker_init_fn=worker_init_fn)
if args.n_gpu > 1:
model = nn.DataParallel(model)
model.train()
ce_loss = CrossEntropyLoss()
|
def trainer_synapse(args, model, snapshot_path):
logging.basicConfig(filename=snapshot_path + "/log.txt", level=logging.INFO,
format='[%(asctime)s.%(msecs)03d] %(message)s', datefmt='%H:%M:%S')
logging.getLogger().addHandler(logging.StreamHandler(sys.stdout))
logging.info(str(args))
base_lr = args.base_lr
num_classes = args.num_classes
batch_size = args.batch_size * args.n_gpu
max_iterations = args.max_iterations
db_train = Synapse_dataset(base_dir=args.root_path, list_dir=args.list_dir, split="train",
transform=transforms.Compose(
[RandomGenerator(output_size=[args.img_size, args.img_size])]))
print("The length of train set is: {}".format(len(db_train)))
def worker_init_fn(worker_id):
random.seed(args.seed + worker_id)
trainloader = DataLoader(db_train, batch_size=batch_size, shuffle=True, num_workers=0, pin_memory=True,
worker_init_fn=worker_init_fn)
if args.n_gpu > 1:
model = nn.DataParallel(model)
model.train()
ce_loss = CrossEntropyLoss() | dice_loss = DiceLoss(num_classes) | 0 | 2023-11-04 11:42:02+00:00 | 2k |
corcel-api/cortex.t | validators/image_validator.py | [
{
"identifier": "get_question",
"path": "template/utils.py",
"snippet": "async def get_question(category, num_questions_needed):\n if category not in [\"text\", \"images\"]:\n raise ValueError(\"Invalid category. Must be 'text' or 'images'.\")\n\n question = await update_counters_and_get_ne... | import io
import torch
import wandb
import random
import asyncio
import aiohttp
import base64
import traceback
import template.reward
import bittensor as bt
from PIL import Image
from io import BytesIO
from template.utils import get_question
from base_validator import BaseValidator
from template.protocol import ImageResponse | 1,332 |
class ImageValidator(BaseValidator):
def __init__(self, dendrite, config, subtensor, wallet):
super().__init__(dendrite, config, subtensor, wallet, timeout=25)
self.streaming = False
self.query_type = "images"
self.model = "dall-e-2"
self.weight = .5
self.provider = "OpenAI"
self.size = "1024x1024"
self.width = 1024
self.height = 1024
self.quality = "standard"
self.style = "vivid"
self.steps = 30
self.seed = 123456
self.wandb_data = {
"modality": "images",
"prompts": {},
"responses": {},
"images": {},
"scores": {},
"timestamps": {},
}
async def start_query(self, available_uids, metagraph):
try:
query_tasks = []
uid_to_question = {}
# Randomly choose the provider based on specified probabilities
providers = ["OpenAI"] * 6 + ["Stability"] * 4
self.provider = random.choice(providers)
if self.provider == "Stability":
self.seed = random.randint(1000, 1000000)
self.model = "stable-diffusion-xl-1024-v1-0"
elif self.provider == "OpenAI":
self.model = "dall-e-2"
# Query all images concurrently
for uid in available_uids:
messages = await get_question("images", len(available_uids))
uid_to_question[uid] = messages # Store messages for each UID
|
class ImageValidator(BaseValidator):
def __init__(self, dendrite, config, subtensor, wallet):
super().__init__(dendrite, config, subtensor, wallet, timeout=25)
self.streaming = False
self.query_type = "images"
self.model = "dall-e-2"
self.weight = .5
self.provider = "OpenAI"
self.size = "1024x1024"
self.width = 1024
self.height = 1024
self.quality = "standard"
self.style = "vivid"
self.steps = 30
self.seed = 123456
self.wandb_data = {
"modality": "images",
"prompts": {},
"responses": {},
"images": {},
"scores": {},
"timestamps": {},
}
async def start_query(self, available_uids, metagraph):
try:
query_tasks = []
uid_to_question = {}
# Randomly choose the provider based on specified probabilities
providers = ["OpenAI"] * 6 + ["Stability"] * 4
self.provider = random.choice(providers)
if self.provider == "Stability":
self.seed = random.randint(1000, 1000000)
self.model = "stable-diffusion-xl-1024-v1-0"
elif self.provider == "OpenAI":
self.model = "dall-e-2"
# Query all images concurrently
for uid in available_uids:
messages = await get_question("images", len(available_uids))
uid_to_question[uid] = messages # Store messages for each UID
| syn = ImageResponse(messages=messages, model=self.model, size=self.size, quality=self.quality, style=self.style, provider=self.provider, seed=self.seed, steps=self.steps) | 1 | 2023-11-06 10:35:34+00:00 | 2k |
flatypus/flowchat | flowchat/chain.py | [
{
"identifier": "autodedent",
"path": "flowchat/autodedent.py",
"snippet": "def autodedent(*text_lines) -> str:\n \"\"\"Format multiline strings, including with multiple levels of indentation, to align with the first line.\n\n Example:\n\n code = '''\n def add(a, b):\n return a + b\n ... | from .autodedent import autodedent
from .private._private_helpers import _encode_image
from retry import retry
from typing import List, TypedDict, Union, Callable, Dict, Literal, Any
from wrapt_timeout_decorator import timeout
import json
import openai
import os
import logging | 1,205 |
logging.basicConfig(level=logging.WARNING,
format='[%(asctime)s] %(levelname)s: %(message)s')
Message = TypedDict('Message', {'role': str, 'content': str | List[Any]})
ResponseFormat = TypedDict(
'ResponseFormat', {'type': Literal['text', 'json_object']})
ImageFormat = TypedDict('ImageFormat', {
'url': str,
'format_type': str,
'detail': Literal['low', 'high']
})
class Chain:
def __init__(self, model: str, api_key: str = None, environ_key="OPENAI_API_KEY"):
super().__init__()
if type(model) is not str:
raise TypeError(
f"Model argument must be a string, not {type(model)}"
)
if api_key is not None and type(api_key) is not str:
raise TypeError(
f"API key argument must be a string, not {type(api_key)}"
)
if type(environ_key) is not str:
raise TypeError(
f"Environment key argument must be a string, not {type(environ_key)}"
)
if api_key is None:
api_key = os.environ.get(environ_key)
if not api_key:
raise ValueError(
"OpenAI API key not found. Please set the OPENAI_API_KEY environment variable, "
"pass in an api_key parameter, or set the environ_key parameter to the environment "
"variable that contains your API key."
)
openai.api_key = api_key
self.model = model
self.system = None
self.user_prompt = []
self.model_response = None
self.prompt_tokens = 0
self.completion_tokens = 0
def _query_api(self, function: callable, *args, max_query_time=None, **kwargs):
"""Call the API for max_query_time seconds, and if it times out, it will retry."""
timeouted_function = timeout(
dec_timeout=max_query_time, use_signals=False)(function)
return timeouted_function(*args, **kwargs)
def _try_query_and_parse(self, function: callable, json_schema, *args, max_query_time=None, **kwargs):
"""Query and try to parse the response, and if it fails, it will retry."""
completion = self._query_api(
function, *args, max_query_time=max_query_time, **kwargs)
if completion is None:
return None
if kwargs.get('stream', False):
return completion
message = completion.choices[0].message.content
if not json_schema is None:
open_bracket = message.find('{')
close_bracket = message.rfind('}')
message = message[open_bracket:close_bracket+1]
try:
message = json.loads(message)
except json.JSONDecodeError:
raise Exception(
"Response was not in the expected JSON format. Please try again. Check that you haven't accidentally lowered the max_tokens parameter so that the response is truncated."
)
self.prompt_tokens += completion.usage.prompt_tokens
self.completion_tokens += completion.usage.completion_tokens
return message
def _ask(
self,
system: Message,
user_messages: List[Message],
json_schema: Any = None,
max_query_time=None,
tries=-1,
**params
):
"""Ask a question to the chatbot with a system prompt and return the response."""
if not user_messages:
return None
messages = [
system,
*user_messages
] if system else user_messages
message = retry(delay=1, logger=logging, tries=tries)(self._try_query_and_parse)(
openai.chat.completions.create,
json_schema=json_schema,
messages=messages,
max_query_time=max_query_time,
**params
)
return message
def _format_images(self, image: str | ImageFormat | Any):
"""Format whatever image format we receive into the specific format that OpenAI's API expects."""
if isinstance(image, str):
return {"url": image}
elif not isinstance(image, dict):
# not string or dict so assume PIL image
# no specific file format, so default to PNG
|
logging.basicConfig(level=logging.WARNING,
format='[%(asctime)s] %(levelname)s: %(message)s')
Message = TypedDict('Message', {'role': str, 'content': str | List[Any]})
ResponseFormat = TypedDict(
'ResponseFormat', {'type': Literal['text', 'json_object']})
ImageFormat = TypedDict('ImageFormat', {
'url': str,
'format_type': str,
'detail': Literal['low', 'high']
})
class Chain:
def __init__(self, model: str, api_key: str = None, environ_key="OPENAI_API_KEY"):
super().__init__()
if type(model) is not str:
raise TypeError(
f"Model argument must be a string, not {type(model)}"
)
if api_key is not None and type(api_key) is not str:
raise TypeError(
f"API key argument must be a string, not {type(api_key)}"
)
if type(environ_key) is not str:
raise TypeError(
f"Environment key argument must be a string, not {type(environ_key)}"
)
if api_key is None:
api_key = os.environ.get(environ_key)
if not api_key:
raise ValueError(
"OpenAI API key not found. Please set the OPENAI_API_KEY environment variable, "
"pass in an api_key parameter, or set the environ_key parameter to the environment "
"variable that contains your API key."
)
openai.api_key = api_key
self.model = model
self.system = None
self.user_prompt = []
self.model_response = None
self.prompt_tokens = 0
self.completion_tokens = 0
def _query_api(self, function: callable, *args, max_query_time=None, **kwargs):
"""Call the API for max_query_time seconds, and if it times out, it will retry."""
timeouted_function = timeout(
dec_timeout=max_query_time, use_signals=False)(function)
return timeouted_function(*args, **kwargs)
def _try_query_and_parse(self, function: callable, json_schema, *args, max_query_time=None, **kwargs):
"""Query and try to parse the response, and if it fails, it will retry."""
completion = self._query_api(
function, *args, max_query_time=max_query_time, **kwargs)
if completion is None:
return None
if kwargs.get('stream', False):
return completion
message = completion.choices[0].message.content
if not json_schema is None:
open_bracket = message.find('{')
close_bracket = message.rfind('}')
message = message[open_bracket:close_bracket+1]
try:
message = json.loads(message)
except json.JSONDecodeError:
raise Exception(
"Response was not in the expected JSON format. Please try again. Check that you haven't accidentally lowered the max_tokens parameter so that the response is truncated."
)
self.prompt_tokens += completion.usage.prompt_tokens
self.completion_tokens += completion.usage.completion_tokens
return message
def _ask(
self,
system: Message,
user_messages: List[Message],
json_schema: Any = None,
max_query_time=None,
tries=-1,
**params
):
"""Ask a question to the chatbot with a system prompt and return the response."""
if not user_messages:
return None
messages = [
system,
*user_messages
] if system else user_messages
message = retry(delay=1, logger=logging, tries=tries)(self._try_query_and_parse)(
openai.chat.completions.create,
json_schema=json_schema,
messages=messages,
max_query_time=max_query_time,
**params
)
return message
def _format_images(self, image: str | ImageFormat | Any):
"""Format whatever image format we receive into the specific format that OpenAI's API expects."""
if isinstance(image, str):
return {"url": image}
elif not isinstance(image, dict):
# not string or dict so assume PIL image
# no specific file format, so default to PNG | return {"url": _encode_image(image, "PNG")} | 1 | 2023-11-08 00:45:21+00:00 | 2k |
WHU-USI3DV/PatchAugNet | place_recognition/Minkloc3D_V2/misc/utils.py | [
{
"identifier": "PolarQuantizer",
"path": "place_recognition/Minkloc3D_V2/misc/quantization.py",
"snippet": "class PolarQuantizer(Quantizer):\n def __init__(self, quant_step: List[float]):\n assert len(quant_step) == 3, '3 quantization steps expected: for sector (in degrees), ring and z-coordi... | import os
import configparser
import time
import numpy as np
from place_recognition.Minkloc3D_V2.misc.quantization import PolarQuantizer, CartesianQuantizer | 838 | # Warsaw University of Technology
class ModelParams:
def __init__(self, model_params_path):
config = configparser.ConfigParser()
config.read(model_params_path)
params = config['MODEL']
self.model_params_path = model_params_path
self.model = params.get('model')
self.output_dim = params.getint('output_dim', 256) # Size of the final descriptor
#######################################################################
# Model dependent
#######################################################################
self.coordinates = params.get('coordinates', 'polar')
assert self.coordinates in ['polar', 'cartesian'], f'Unsupported coordinates: {self.coordinates}'
if 'polar' in self.coordinates:
# 3 quantization steps for polar coordinates: for sectors (in degrees), rings (in meters) and z coordinate (in meters)
self.quantization_step = tuple([float(e) for e in params['quantization_step'].split(',')])
assert len(self.quantization_step) == 3, f'Expected 3 quantization steps: for sectors (degrees), rings (meters) and z coordinate (meters)'
| # Warsaw University of Technology
class ModelParams:
def __init__(self, model_params_path):
config = configparser.ConfigParser()
config.read(model_params_path)
params = config['MODEL']
self.model_params_path = model_params_path
self.model = params.get('model')
self.output_dim = params.getint('output_dim', 256) # Size of the final descriptor
#######################################################################
# Model dependent
#######################################################################
self.coordinates = params.get('coordinates', 'polar')
assert self.coordinates in ['polar', 'cartesian'], f'Unsupported coordinates: {self.coordinates}'
if 'polar' in self.coordinates:
# 3 quantization steps for polar coordinates: for sectors (in degrees), rings (in meters) and z coordinate (in meters)
self.quantization_step = tuple([float(e) for e in params['quantization_step'].split(',')])
assert len(self.quantization_step) == 3, f'Expected 3 quantization steps: for sectors (degrees), rings (meters) and z coordinate (meters)' | self.quantizer = PolarQuantizer(quant_step=self.quantization_step) | 0 | 2023-11-02 13:52:20+00:00 | 2k |
WeiLab-Biology/DeepProSite | DeepProSite-main/edge_features.py | [
{
"identifier": "gather_edges",
"path": "self_attention.py",
"snippet": "def gather_edges(edges, neighbor_idx):\n # Features [B,N,N,C] at Neighbor indices [B,N,K] => Neighbor features [B,N,K,C]\n neighbors = neighbor_idx.unsqueeze(-1).expand(-1, -1, -1, edges.size(-1))\n edge_features = torch.g... | import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
from self_attention import gather_edges, gather_nodes, Normalize | 850 |
class PositionalEncodings(nn.Module):
def __init__(self, num_embeddings):
super(PositionalEncodings, self).__init__()
self.num_embeddings = num_embeddings
def forward(self, E_idx):
# i-j
N_batch = E_idx.size(0)
N_nodes = E_idx.size(1)
N_neighbors = E_idx.size(2)
ii = torch.arange(N_nodes, dtype=torch.float32).view((1, -1, 1)).cuda()
d = (E_idx.float() - ii).unsqueeze(-1)
# Original Transformer frequencies
frequency = torch.exp(
torch.arange(0, self.num_embeddings, 2, dtype=torch.float32)
* -(np.log(10000.0) / self.num_embeddings)).cuda()
angles = d * frequency.view((1,1,1,-1))
E = torch.cat((torch.cos(angles), torch.sin(angles)), -1)
return E # [N_batch, N_nodes, N_neighbors, num_embeddings]
class EdgeFeatures(nn.Module):
def __init__(self, edge_features, num_positional_embeddings=16,
num_rbf=16, top_k=30, augment_eps=0.):
super(EdgeFeatures, self).__init__()
self.top_k = top_k
self.augment_eps = augment_eps
self.num_rbf = num_rbf
# Positional encoding
self.PE = PositionalEncodings(num_positional_embeddings)
# Embedding and normalization
self.edge_embedding = nn.Linear(num_positional_embeddings + num_rbf + 7, edge_features, bias=True)
|
class PositionalEncodings(nn.Module):
def __init__(self, num_embeddings):
super(PositionalEncodings, self).__init__()
self.num_embeddings = num_embeddings
def forward(self, E_idx):
# i-j
N_batch = E_idx.size(0)
N_nodes = E_idx.size(1)
N_neighbors = E_idx.size(2)
ii = torch.arange(N_nodes, dtype=torch.float32).view((1, -1, 1)).cuda()
d = (E_idx.float() - ii).unsqueeze(-1)
# Original Transformer frequencies
frequency = torch.exp(
torch.arange(0, self.num_embeddings, 2, dtype=torch.float32)
* -(np.log(10000.0) / self.num_embeddings)).cuda()
angles = d * frequency.view((1,1,1,-1))
E = torch.cat((torch.cos(angles), torch.sin(angles)), -1)
return E # [N_batch, N_nodes, N_neighbors, num_embeddings]
class EdgeFeatures(nn.Module):
def __init__(self, edge_features, num_positional_embeddings=16,
num_rbf=16, top_k=30, augment_eps=0.):
super(EdgeFeatures, self).__init__()
self.top_k = top_k
self.augment_eps = augment_eps
self.num_rbf = num_rbf
# Positional encoding
self.PE = PositionalEncodings(num_positional_embeddings)
# Embedding and normalization
self.edge_embedding = nn.Linear(num_positional_embeddings + num_rbf + 7, edge_features, bias=True) | self.norm_edges = Normalize(edge_features) | 2 | 2023-11-04 15:32:31+00:00 | 2k |
gchada/ROAM | real/rail_real_walker/robots/go1_remote.py | [
{
"identifier": "Go1RemoteActionMsg",
"path": "real/rail_real_walker/robots/go1_remote_runner.py",
"snippet": "class Go1RemoteActionMsg:\n target_action : np.ndarray"
},
{
"identifier": "Go1RemoteObservation",
"path": "real/rail_real_walker/robots/go1_remote_runner.py",
"snippet": "cl... | from .go1_remote_runner import Go1RemoteActionMsg, Go1RemoteObservation, Go1RemoteConfigMsg, empty_obs, DataPack, REAL_CONTROL_TIMESTEP
from typing import Optional
from rail_walker_interface import BaseWalker, BaseWalkerWithFootContact, BaseWalkerWithJoystick, BaseWalkerWithJointTemperatureSensor, Walker3DVelocityEstimator
from serde.msgpack import from_msgpack, to_msgpack
from serde import SerdeError
from functools import cached_property
import numpy as np
import time
import socket
import msgpack.exceptions
import unitree_go1_wrapper.go1_constants as go1_constants
import transforms3d as tr3d
import errno
import traceback | 1,273 |
class Go1RealWalkerRemote(BaseWalker[Go1RemoteObservation], BaseWalkerWithFootContact, BaseWalkerWithJoystick, BaseWalkerWithJointTemperatureSensor):
def __init__(
self,
velocity_estimator: Walker3DVelocityEstimator,
power_protect_factor : float = 0.5,
foot_contact_threshold: np.ndarray = np.array([80, 170, 170, 170]),
action_interpolation: bool = True,
name: Optional[str] = "robot",
Kp: float = 40,
Kd: float = 5,
control_timestep : float = 0.05,
force_real_control_timestep : bool = True,
limit_action_range : float = 1.0,
server_addr = ("192.168.123.161",6001)
):
assert control_timestep >= 0.02
assert power_protect_factor > 0.0 and power_protect_factor <= 1.0
# Init communication with the robot
self.socket_client = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
self.socket_client.setblocking(True)
self.socket_client.settimeout(0.5)
self.socket_client.connect(server_addr)
self.socket_client.setblocking(False)
self._server_addr = server_addr
|
class Go1RealWalkerRemote(BaseWalker[Go1RemoteObservation], BaseWalkerWithFootContact, BaseWalkerWithJoystick, BaseWalkerWithJointTemperatureSensor):
def __init__(
self,
velocity_estimator: Walker3DVelocityEstimator,
power_protect_factor : float = 0.5,
foot_contact_threshold: np.ndarray = np.array([80, 170, 170, 170]),
action_interpolation: bool = True,
name: Optional[str] = "robot",
Kp: float = 40,
Kd: float = 5,
control_timestep : float = 0.05,
force_real_control_timestep : bool = True,
limit_action_range : float = 1.0,
server_addr = ("192.168.123.161",6001)
):
assert control_timestep >= 0.02
assert power_protect_factor > 0.0 and power_protect_factor <= 1.0
# Init communication with the robot
self.socket_client = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
self.socket_client.setblocking(True)
self.socket_client.settimeout(0.5)
self.socket_client.connect(server_addr)
self.socket_client.setblocking(False)
self._server_addr = server_addr | self.data_pack = DataPack(self.deal_with_data) | 4 | 2023-11-02 23:21:38+00:00 | 2k |
NUCCASJNR/PaystackPyAPI | paystackpyAPI/transaction.py | [
{
"identifier": "PaystackAPI",
"path": "paystackpyAPI/base.py",
"snippet": "class PaystackAPI:\n \n def __init__(self, api_key: str) -> None:\n self.api_key = api_key"
},
{
"identifier": "APIError",
"path": "errors.py",
"snippet": "class APIError(PaystackError):\n \"\"\"E... | import requests
import datetime
import webbrowser
from .base import PaystackAPI
from typing import Dict, Union
from errors import APIError
from decimal import Decimal | 681 | #!/usr/bin/env python3
"""Handles All Paystack related tasks"""
class Transaction(PaystackAPI):
INITIALIZATION_OPTIONAL_PARAMS = [
"currency",
"reference",
"callback_url",
"plan",
"invoice_limit",
"metadata",
"channels",
"split_code",
"subaccount",
"transaction_charge",
"bearer"
]
TRANSACTION_LIST_OPTIONAL_PARAMS = [
"customer",
"terminalid",
"status",
"from",
"to",
"amount"
]
CHARGE_AUTHORIZATION_OPTIONAL_PARAMS = [
"reference",
"currency",
"metadata",
"channels",
"subaccount",
"transaction_charge",
"bearer",
"queue"
]
EXPORT_OPTIONAL_PARAMS = [
'from',
'to',
'customer',
'status',
'currency',
'amount',
'settled',
'settlement',
'payment_page'
]
def __init__(self, api_key: str):
super().__init__(api_key)
self.paystack_initialization_url = "https://api.paystack.co/transaction/initialize"
self.paystack_verification_url = "https://api.paystack.co/transaction/verify"
self.list_transaction_url = "https://api.paystack.co/transaction"
self.fetch_transaction_url = "https://api.paystack.co/transaction"
self.charge_authorization_url = "https://api.paystack.co/transaction/charge_authorization"
self.transaction_timeline_url = "https://api.paystack.co/transaction/timeline"
self.transaction_totals_url = "https://api.paystack.co/transaction/totals"
self.export_transactions_url = "https://api.paystack.co/transaction/export"
def initialize_transaction(self, email: str, amount: int, **kwargs):
"""
Initialize a Paystack transaction.
:param email: Customer's email address.
:param amount: Transaction amount.
:param kwargs: Optional parameters for the transaction.
Example: `currency`, `callback_url`, etc.
:return: JSON response from Paystack API.
:raises APIError: If required parameters are missing or the API key is invalid.
"""
if not email or not amount:
| #!/usr/bin/env python3
"""Handles All Paystack related tasks"""
class Transaction(PaystackAPI):
INITIALIZATION_OPTIONAL_PARAMS = [
"currency",
"reference",
"callback_url",
"plan",
"invoice_limit",
"metadata",
"channels",
"split_code",
"subaccount",
"transaction_charge",
"bearer"
]
TRANSACTION_LIST_OPTIONAL_PARAMS = [
"customer",
"terminalid",
"status",
"from",
"to",
"amount"
]
CHARGE_AUTHORIZATION_OPTIONAL_PARAMS = [
"reference",
"currency",
"metadata",
"channels",
"subaccount",
"transaction_charge",
"bearer",
"queue"
]
EXPORT_OPTIONAL_PARAMS = [
'from',
'to',
'customer',
'status',
'currency',
'amount',
'settled',
'settlement',
'payment_page'
]
def __init__(self, api_key: str):
super().__init__(api_key)
self.paystack_initialization_url = "https://api.paystack.co/transaction/initialize"
self.paystack_verification_url = "https://api.paystack.co/transaction/verify"
self.list_transaction_url = "https://api.paystack.co/transaction"
self.fetch_transaction_url = "https://api.paystack.co/transaction"
self.charge_authorization_url = "https://api.paystack.co/transaction/charge_authorization"
self.transaction_timeline_url = "https://api.paystack.co/transaction/timeline"
self.transaction_totals_url = "https://api.paystack.co/transaction/totals"
self.export_transactions_url = "https://api.paystack.co/transaction/export"
def initialize_transaction(self, email: str, amount: int, **kwargs):
"""
Initialize a Paystack transaction.
:param email: Customer's email address.
:param amount: Transaction amount.
:param kwargs: Optional parameters for the transaction.
Example: `currency`, `callback_url`, etc.
:return: JSON response from Paystack API.
:raises APIError: If required parameters are missing or the API key is invalid.
"""
if not email or not amount: | raise APIError(400, "Missing required parameters: email and/or amount") | 1 | 2023-11-07 18:00:39+00:00 | 2k |
Dataherald/Assistant | assistant.py | [
{
"identifier": "Function",
"path": "function.py",
"snippet": "class Function(BaseModel, ABC):\n name: str\n description: Optional[str] = None\n parameters: Optional[List[Property]] = None\n\n def to_dict(self):\n if self.parameters is None:\n return {\n \"na... | from openai import OpenAI
from openai import Client
from function import Function, FunctionCall
from openai.types.beta import Thread, Assistant
from openai.types.beta.threads import Run, ThreadMessage
from yaspin import yaspin
import json
import random
import time | 1,288 |
PRINT_COLORS = [
'\033[31m',
'\033[32m',
'\033[33m',
'\033[34m',
'\033[35m',
'\033[36m',
]
class Message:
thread_id: str
role: str
content: str
file_ids: list[str]
def __init__(
self, thread_id: str, role: str, content: str, file_ids: list[str] = None
):
self.thread_id = thread_id
self.role = role
self.content = content
self.file_ids = file_ids
class Conversation:
messages: list[Message]
def __init__(self, messages: list[Message]):
self.messages = messages
def print_conversation(self):
for message in self.messages:
print(f"{message.role}: {message.content}")
class AIAssistant:
assistant: Assistant
client: OpenAI
assistant_name: str
assistant_description: str
instruction: str
model: str
use_retrieval: bool
use_code_interpreter: bool
functions: list[Function]
threads: list[Thread]
tools: list[dict]
file_ids: list[str]
conversation: Conversation
verbose: bool
auto_delete: bool = True
def __init__(
self,
instruction: str,
model: str,
use_retrieval: bool = False,
use_code_interpreter: bool = False,
file_ids: list[str] = None,
functions: list[Function] = None,
assistant_name: str = "AI Assistant",
assistant_description: str = "An AI Assistant",
verbose: bool = False,
auto_delete: bool = True,
):
self.client = Client()
self.instruction = instruction
self.model = model
self.use_retrieval = use_retrieval
self.use_code_interpreter = use_code_interpreter
self.file_ids = file_ids
self.functions = functions
self.assistant_name = assistant_name
self.assistant_description = assistant_description
self.tools = [
{"type": "function", "function": f.to_dict()} for f in self.functions
] if self.functions else []
if self.use_retrieval:
self.tools.append({"type": "retrieval"})
if self.use_code_interpreter:
self.tools.append({"type": "code_interpreter"})
self.assistant = self.client.beta.assistants.create(
name=self.assistant_name,
description=self.assistant_description,
instructions=self.instruction,
model=self.model,
tools=self.tools,
file_ids=self.file_ids if self.file_ids else [],
)
self.threads = []
self.conversation = Conversation(messages=[])
self.verbose = verbose
self.auto_delete = auto_delete
def delete_assistant_file_by_id(self, file_id: str):
file_deletion_status = self.client.beta.assistants.files.delete(
assistant_id=self.assistant.id, file_id=file_id
)
return file_deletion_status
def create_thread(self) -> Thread:
thread = self.client.beta.threads.create()
self.threads.append(thread)
return thread
def create_tool_outputs(self, run: Run) -> list[dict]:
tool_outputs = []
for tool in run.required_action.submit_tool_outputs.tool_calls:
tool_found = False
function_name = tool.function.name
if tool.function.arguments:
function_arguments = json.loads(tool.function.arguments)
else:
function_arguments = {}
call_id = tool.id
|
PRINT_COLORS = [
'\033[31m',
'\033[32m',
'\033[33m',
'\033[34m',
'\033[35m',
'\033[36m',
]
class Message:
thread_id: str
role: str
content: str
file_ids: list[str]
def __init__(
self, thread_id: str, role: str, content: str, file_ids: list[str] = None
):
self.thread_id = thread_id
self.role = role
self.content = content
self.file_ids = file_ids
class Conversation:
messages: list[Message]
def __init__(self, messages: list[Message]):
self.messages = messages
def print_conversation(self):
for message in self.messages:
print(f"{message.role}: {message.content}")
class AIAssistant:
assistant: Assistant
client: OpenAI
assistant_name: str
assistant_description: str
instruction: str
model: str
use_retrieval: bool
use_code_interpreter: bool
functions: list[Function]
threads: list[Thread]
tools: list[dict]
file_ids: list[str]
conversation: Conversation
verbose: bool
auto_delete: bool = True
def __init__(
self,
instruction: str,
model: str,
use_retrieval: bool = False,
use_code_interpreter: bool = False,
file_ids: list[str] = None,
functions: list[Function] = None,
assistant_name: str = "AI Assistant",
assistant_description: str = "An AI Assistant",
verbose: bool = False,
auto_delete: bool = True,
):
self.client = Client()
self.instruction = instruction
self.model = model
self.use_retrieval = use_retrieval
self.use_code_interpreter = use_code_interpreter
self.file_ids = file_ids
self.functions = functions
self.assistant_name = assistant_name
self.assistant_description = assistant_description
self.tools = [
{"type": "function", "function": f.to_dict()} for f in self.functions
] if self.functions else []
if self.use_retrieval:
self.tools.append({"type": "retrieval"})
if self.use_code_interpreter:
self.tools.append({"type": "code_interpreter"})
self.assistant = self.client.beta.assistants.create(
name=self.assistant_name,
description=self.assistant_description,
instructions=self.instruction,
model=self.model,
tools=self.tools,
file_ids=self.file_ids if self.file_ids else [],
)
self.threads = []
self.conversation = Conversation(messages=[])
self.verbose = verbose
self.auto_delete = auto_delete
def delete_assistant_file_by_id(self, file_id: str):
file_deletion_status = self.client.beta.assistants.files.delete(
assistant_id=self.assistant.id, file_id=file_id
)
return file_deletion_status
def create_thread(self) -> Thread:
thread = self.client.beta.threads.create()
self.threads.append(thread)
return thread
def create_tool_outputs(self, run: Run) -> list[dict]:
tool_outputs = []
for tool in run.required_action.submit_tool_outputs.tool_calls:
tool_found = False
function_name = tool.function.name
if tool.function.arguments:
function_arguments = json.loads(tool.function.arguments)
else:
function_arguments = {}
call_id = tool.id | function_call = FunctionCall( | 1 | 2023-11-09 01:58:07+00:00 | 2k |
Skytliang/SpyGame | utils/agent.py | [
{
"identifier": "OutOfQuotaException",
"path": "utils/openai_utils.py",
"snippet": "class OutOfQuotaException(Exception):\n \"Raised when the key exceeded the current quota\"\n def __init__(self, key, cause=None):\n super().__init__(f\"No quota for key: {key}\")\n self.key = key\n ... | import os
import openai
import backoff
import time
import random
import json
import copy
import numpy as np
from datetime import datetime
from openai.error import RateLimitError, APIError, ServiceUnavailableError, APIConnectionError, AuthenticationError
from utils.openai_utils import OutOfQuotaException, AccessTerminatedException, TimeOutException
from utils.openai_utils import num_tokens_from_string, model2max_context | 1,371 |
# from bardapi import Bard
# import requests
# import torch
# from transformers import AutoTokenizer, AutoModelForCausalLM
# from FastChat.fastchat.model.model_adapter import load_model, get_conversation_template, add_model_args
cycle_all_keys = True
current_path = os.path.abspath(__file__).rsplit('/', 1)[0]
gpt3_api_keys = json.load(open(f'{current_path}/gpt3_apikeys.json'))
gpt4_api_keys = json.load(open(f'{current_path}/gpt4_apikeys.json'))
random.shuffle(gpt3_api_keys)
random.shuffle(gpt4_api_keys)
def cycle_keys(openai_api_keys):
while True:
for key in openai_api_keys:
yield key
gpt3_key_generator = cycle_keys(gpt3_api_keys)
gpt4_key_generator = cycle_keys(gpt4_api_keys)
def key_generator(model_name):
if model_name in ["gpt-3.5-turbo", "gpt-3.5-turbo-0301", "gpt-3.5-turbo-0613", "text-davinci-003", "text-davinci-002", "curie", "text-curie-001", "text-babbage-001", "text-ada-001"]:
return gpt3_key_generator
elif model_name in ["gpt-4", "gpt-4-0314", "gpt-4-0613"]:
return gpt4_key_generator
class Agent:
def __init__(self, model_name: str, temperature: float, sleep_time: float = 0) -> None:
"""Create an agent (gpt series by default)
Args:
model_name(str): model name
temperature (float): higher values make the output more random, while lower values make it more focused and deterministic
sleep_time (float): sleep against rate limiting
"""
self.model_name = model_name
self.temperature = temperature
self.sleep_time = sleep_time
self.memory_lst = []
self.memory_lst_idx = np.array([])
@backoff.on_exception(backoff.expo, (RateLimitError, APIError, ServiceUnavailableError, APIConnectionError, AuthenticationError), max_tries=5)
def query(self, messages: "list[dict]", max_tokens: int, api_key: str, temperature: float) -> str:
"""make a query
Args:
messages (list[dict]): chat history in turbo format
max_tokens (int): max token in api call
api_key (str): openai api key
temperature (float): sampling temperature
Raises:
OutOfQuotaException: the apikey has out of quota
AccessTerminatedException: the apikey has been ban
Returns:
str: the return msg
"""
# assert self.model_name in support_models, f"Not support {self.model_name}. Choices: {support_models}"
try:
openai.api_base = self.api_base
response = openai.ChatCompletion.create(
model=self.model_name,
messages=messages,
temperature=temperature,
max_tokens=max_tokens,
api_key=api_key,
presence_penalty=0.75
)
gen = response['choices'][0]['message']['content']
return gen
except RateLimitError as e:
if "You exceeded your current quota, please check your plan and billing details" in e.user_message:
self.openai_api_keys.remove(api_key)
print(f'Out Of Quota: {api_key}')
raise OutOfQuotaException(api_key)
elif "Your access was terminated due to violation of our policies" in e.user_message:
self.openai_api_keys.remove(api_key)
print(f'Access Terminated: {api_key}')
|
# from bardapi import Bard
# import requests
# import torch
# from transformers import AutoTokenizer, AutoModelForCausalLM
# from FastChat.fastchat.model.model_adapter import load_model, get_conversation_template, add_model_args
cycle_all_keys = True
current_path = os.path.abspath(__file__).rsplit('/', 1)[0]
gpt3_api_keys = json.load(open(f'{current_path}/gpt3_apikeys.json'))
gpt4_api_keys = json.load(open(f'{current_path}/gpt4_apikeys.json'))
random.shuffle(gpt3_api_keys)
random.shuffle(gpt4_api_keys)
def cycle_keys(openai_api_keys):
while True:
for key in openai_api_keys:
yield key
gpt3_key_generator = cycle_keys(gpt3_api_keys)
gpt4_key_generator = cycle_keys(gpt4_api_keys)
def key_generator(model_name):
if model_name in ["gpt-3.5-turbo", "gpt-3.5-turbo-0301", "gpt-3.5-turbo-0613", "text-davinci-003", "text-davinci-002", "curie", "text-curie-001", "text-babbage-001", "text-ada-001"]:
return gpt3_key_generator
elif model_name in ["gpt-4", "gpt-4-0314", "gpt-4-0613"]:
return gpt4_key_generator
class Agent:
def __init__(self, model_name: str, temperature: float, sleep_time: float = 0) -> None:
"""Create an agent (gpt series by default)
Args:
model_name(str): model name
temperature (float): higher values make the output more random, while lower values make it more focused and deterministic
sleep_time (float): sleep against rate limiting
"""
self.model_name = model_name
self.temperature = temperature
self.sleep_time = sleep_time
self.memory_lst = []
self.memory_lst_idx = np.array([])
@backoff.on_exception(backoff.expo, (RateLimitError, APIError, ServiceUnavailableError, APIConnectionError, AuthenticationError), max_tries=5)
def query(self, messages: "list[dict]", max_tokens: int, api_key: str, temperature: float) -> str:
"""make a query
Args:
messages (list[dict]): chat history in turbo format
max_tokens (int): max token in api call
api_key (str): openai api key
temperature (float): sampling temperature
Raises:
OutOfQuotaException: the apikey has out of quota
AccessTerminatedException: the apikey has been ban
Returns:
str: the return msg
"""
# assert self.model_name in support_models, f"Not support {self.model_name}. Choices: {support_models}"
try:
openai.api_base = self.api_base
response = openai.ChatCompletion.create(
model=self.model_name,
messages=messages,
temperature=temperature,
max_tokens=max_tokens,
api_key=api_key,
presence_penalty=0.75
)
gen = response['choices'][0]['message']['content']
return gen
except RateLimitError as e:
if "You exceeded your current quota, please check your plan and billing details" in e.user_message:
self.openai_api_keys.remove(api_key)
print(f'Out Of Quota: {api_key}')
raise OutOfQuotaException(api_key)
elif "Your access was terminated due to violation of our policies" in e.user_message:
self.openai_api_keys.remove(api_key)
print(f'Access Terminated: {api_key}') | raise AccessTerminatedException(api_key) | 1 | 2023-11-01 03:42:10+00:00 | 2k |
SpectacularAI/point-cloud-tools | formats/auto.py | [
{
"identifier": "load_ply_to_dataframe",
"path": "formats/ply.py",
"snippet": "def load_ply_to_dataframe(ply_file):\n with open(ply_file, 'rb') as f:\n return load_ply_stream_to_dataframe(f)\n load_ply_to_dataframe"
},
{
"identifier": "dataframe_to_ply",
"path": "formats/ply.py"... | import pandas as pd
from .ply import load_ply_to_dataframe, dataframe_to_ply
from .splat import splat_file_to_data_frame, dataframe_to_splat_file
from .pcd import dataframe_to_pcd
from .html import dataframe_to_gsplat_html | 673 |
def load_to_dataframe(fn):
ext = fn.split('.')[-1]
if ext == 'ply':
return load_ply_to_dataframe(fn)
elif ext == 'csv':
return pd.read_csv(fn)
elif ext == 'txt':
# assuming COLMAP CSV format
return pd.read_csv(fn, sep=' ', header=None, usecols=list(range(7)), names=['id'] + list('xyzrgb')).set_index('id')
elif ext == 'parquet':
return pd.read_parquet(fn)
elif ext == 'splat':
return splat_file_to_data_frame(fn)
elif ext == 'pcd':
raise RuntimeError("PCD import not implemented")
else:
raise RuntimeError("unrecognized extension ." + ext)
def save_to_dataframe(df, fn, args):
ext = fn.split('.')[-1]
if ext == 'ply':
|
def load_to_dataframe(fn):
ext = fn.split('.')[-1]
if ext == 'ply':
return load_ply_to_dataframe(fn)
elif ext == 'csv':
return pd.read_csv(fn)
elif ext == 'txt':
# assuming COLMAP CSV format
return pd.read_csv(fn, sep=' ', header=None, usecols=list(range(7)), names=['id'] + list('xyzrgb')).set_index('id')
elif ext == 'parquet':
return pd.read_parquet(fn)
elif ext == 'splat':
return splat_file_to_data_frame(fn)
elif ext == 'pcd':
raise RuntimeError("PCD import not implemented")
else:
raise RuntimeError("unrecognized extension ." + ext)
def save_to_dataframe(df, fn, args):
ext = fn.split('.')[-1]
if ext == 'ply': | dataframe_to_ply(df, fn) | 1 | 2023-11-02 14:16:49+00:00 | 2k |
jdelahayes/ha-voltalis | custom_components/voltalis/climate.py | [
{
"identifier": "DEFAULT_MAX_TEMP",
"path": "custom_components/voltalis/const.py",
"snippet": "DEFAULT_MAX_TEMP = 24"
},
{
"identifier": "DEFAULT_MIN_TEMP",
"path": "custom_components/voltalis/const.py",
"snippet": "DEFAULT_MIN_TEMP = 7"
},
{
"identifier": "DOMAIN",
"path": "... | import logging
from typing import Any
from homeassistant.components.climate import (
ClimateEntity,
ClimateEntityFeature,
HVACAction,
HVACMode,
)
from homeassistant.config_entries import ConfigEntry
from homeassistant.const import ATTR_TEMPERATURE, UnitOfTemperature
from homeassistant.core import HomeAssistant
from homeassistant.helpers.entity_platform import AddEntitiesCallback
from homeassistant.util.unit_conversion import TemperatureConverter
from .const import (
DEFAULT_MAX_TEMP,
DEFAULT_MIN_TEMP,
DOMAIN,
HA_PRESET_MODES,
VOLTALIS_CONTROLLER,
VOLTALIS_PRESET_MODES,
VOLTALIS_HEATER_TYPE,
)
from .entity import VoltalisEntity | 670 | """Platform for climate integration."""
from __future__ import annotations
_LOGGER = logging.getLogger(__name__)
async def async_setup_entry(
hass: HomeAssistant, entry: ConfigEntry, async_add_entities: AddEntitiesCallback
) -> None:
"""Set up climate entity for Voltalis Appliance."""
| """Platform for climate integration."""
from __future__ import annotations
_LOGGER = logging.getLogger(__name__)
async def async_setup_entry(
hass: HomeAssistant, entry: ConfigEntry, async_add_entities: AddEntitiesCallback
) -> None:
"""Set up climate entity for Voltalis Appliance.""" | controller = hass.data[DOMAIN][entry.entry_id][VOLTALIS_CONTROLLER] | 4 | 2023-11-01 09:05:17+00:00 | 2k |
r-three/licensed-pile | ubuntu/to-dolma.py | [
{
"identifier": "PermissiveLicenses",
"path": "licensed_pile/licenses.py",
"snippet": "class PermissiveLicenses(StringEnum):\n PD = \"Public Domain\"\n CC0 = \"Creative Commons Zero - Public Domain - https://creativecommons.org/publicdomain/zero/1.0/\"\n CC_BY = (\n \"Creative Commons - ... | import argparse
import datetime
import glob
import os
import urllib.parse
from charset_normalizer import from_bytes
from licensed_pile.licenses import PermissiveLicenses
from licensed_pile.write import to_dolma | 1,425 | """Convert the raw ubuntu data to the dolma format."""
SOURCE_NAME = "ubuntu-chat"
BASE_URL = "https://irclogs.ubuntu.com"
parser = argparse.ArgumentParser(description="Convert data to dolma.")
parser.add_argument(
"--data",
default="data/irclogs.ubuntu.com/",
help="Path to the directory containing ubuntu chat data.",
)
parser.add_argument(
"--output_dir",
default=f"data/{SOURCE_NAME}/raw/documents/",
help="Where the dolma formatted data goes.",
)
parser.add_argument(
"--filename", default="ubuntu.jsonl.gz", help="The base filename for our chat data."
)
parser.add_argument(
"--shard_size", type=int, default=1, help="Size, in GB, for each shard."
)
def format_dolma(chat: str, source_name: str = SOURCE_NAME, base_url: str = BASE_URL):
# Manually split because os.path.split only give (head, tail)
*_, year, month, day, channel = os.path.splitext(chat)[0].split(os.path.sep)
created = datetime.date(int(year), int(month), int(day))
with open(chat, "rb") as f:
# There is some encoding weirdness that this seems to fix.
text = str(from_bytes(f.read()).best())
return {
# We don't want each channel to be it own data source so add the date
# to the channel to make a unique string id.
"id": f"{created.isoformat()}-{channel}",
"text": text,
"source": source_name,
"added": datetime.datetime.utcnow().isoformat(),
"created": created.isoformat(),
"metadata": {
"license": str(PermissiveLicenses.PD),
# This will be added in the next phase of preprocessing.
"authors": [],
"url": urllib.parse.quote(
f"{base_url}/{year}/{month}/{day}/{channel}.txt", safe=":/"
),
"channel": channel,
},
}
def main(args):
# Use iterators so we don't have to load the whole dataset in memory.
# year month day channel
chats = map(
format_dolma, glob.iglob(os.path.join(args.data, "**", "**", "**", "*.txt"))
)
| """Convert the raw ubuntu data to the dolma format."""
SOURCE_NAME = "ubuntu-chat"
BASE_URL = "https://irclogs.ubuntu.com"
parser = argparse.ArgumentParser(description="Convert data to dolma.")
parser.add_argument(
"--data",
default="data/irclogs.ubuntu.com/",
help="Path to the directory containing ubuntu chat data.",
)
parser.add_argument(
"--output_dir",
default=f"data/{SOURCE_NAME}/raw/documents/",
help="Where the dolma formatted data goes.",
)
parser.add_argument(
"--filename", default="ubuntu.jsonl.gz", help="The base filename for our chat data."
)
parser.add_argument(
"--shard_size", type=int, default=1, help="Size, in GB, for each shard."
)
def format_dolma(chat: str, source_name: str = SOURCE_NAME, base_url: str = BASE_URL):
# Manually split because os.path.split only give (head, tail)
*_, year, month, day, channel = os.path.splitext(chat)[0].split(os.path.sep)
created = datetime.date(int(year), int(month), int(day))
with open(chat, "rb") as f:
# There is some encoding weirdness that this seems to fix.
text = str(from_bytes(f.read()).best())
return {
# We don't want each channel to be it own data source so add the date
# to the channel to make a unique string id.
"id": f"{created.isoformat()}-{channel}",
"text": text,
"source": source_name,
"added": datetime.datetime.utcnow().isoformat(),
"created": created.isoformat(),
"metadata": {
"license": str(PermissiveLicenses.PD),
# This will be added in the next phase of preprocessing.
"authors": [],
"url": urllib.parse.quote(
f"{base_url}/{year}/{month}/{day}/{channel}.txt", safe=":/"
),
"channel": channel,
},
}
def main(args):
# Use iterators so we don't have to load the whole dataset in memory.
# year month day channel
chats = map(
format_dolma, glob.iglob(os.path.join(args.data, "**", "**", "**", "*.txt"))
) | to_dolma(chats, args.output_dir, args.filename, args.shard_size) | 1 | 2023-11-06 16:04:10+00:00 | 2k |
UMass-Foundation-Model/genome | engine/util.py | [
{
"identifier": "Wizardlm",
"path": "engine/llm.py",
"snippet": "class Wizardlm():\n @classmethod\n def init(cls, base_model=\"WizardLM/WizardCoder-Python-34B-V1.0\", n_gpus=4, max_input_tokens=16384):\n cls.llm = LLM(model=base_model, tensor_parallel_size=n_gpus, max_num_batched_tokens=max... | import os
import json
import openai
import pdb
from engine.llm import Wizardlm
from engine.llm import Codellama
from engine.datasets import get_dataset | 1,203 |
def strip_dict(dict):
for k, v in dict.items():
if isinstance(v, str):
dict[k] = v.strip()
return dict
def get_module_list(args):
if not args.use_new_module:
return []
module_save_dir = args.module_save_dir
if os.path.isdir(module_save_dir):
file_list = os.listdir(module_save_dir)
module_name_dict = {}
for filename in file_list:
# relieve the name constraint
if 'MODULE' in filename and 'json' in filename or 'json' in filename:
file_path = os.path.join(module_save_dir, filename)
try:
module_dict = json.load(open(file_path))
module_dict = strip_dict(module_dict)
name_key = "module_name" if "module_name" in module_dict else "name"
module_name = module_dict[name_key]
module_dict['module_name'] = module_name
if 'module_program' not in module_dict:
module_dict['module_program'] = module_dict['module']
#if 'annotations' not in module_dict:
# module_dict['annotations'] = module_dict['program']
if module_name not in module_name_dict or \
module_dict['test_accuracy'] > module_name_dict[module_name]['test_accuracy']:
module_name_dict[module_name] = module_dict
except:
pdb.set_trace()
module_list = []
for module_dict in module_name_dict.values():
if 'test_accuracy' not in module_dict:
module_list.append(module_dict)
elif module_dict['test_accuracy'] >= args.threshold:
module_list.append(module_dict)
else:
print("There is no available module directory: %s"%(module_save_dir))
module_list = []
return module_list
def save_output(args, output_dict, filename=None):
output_dir = args.output_dir
if args.stage == 1:
output_path = os.path.join(output_dir, output_dict['annotations'][0]['id'] + '.json')
json.dump(output_dict, open(output_path, 'w'), indent=2)
elif args.stage == 1.5:
if args.split_cases:
module_head_list = output_dict.pop('module_head_list')
for index, module_head in enumerate(module_head_list):
output_dict['module_head'] = module_head
output_path = os.path.join(output_dir, output_dict['annotations'][index]['id'] + '.json')
json.dump(output_dict, open(output_path, 'w'), indent=2)
else:
output_path = os.path.join(output_dir, output_dict['annotations'][0]['id'] + '.json')
json.dump(output_dict, open(output_path, 'w'), indent=2)
elif args.stage == 2:
if filename is None:
filename = 'MODULE_' + output_dict['module_name'] + '.json'
output_path = os.path.join(output_dir, filename)
json.dump(output_dict, open(output_path, 'w'), indent=2)
elif args.stage == 3:
output_path = os.path.join(output_dir, 'result_' + output_dict['id'] + '.json')
json.dump(output_dict, open(output_path, 'w'), indent=2)
pass
def init_gpt(args):
if args.model == 'wizardlm':
|
def strip_dict(dict):
for k, v in dict.items():
if isinstance(v, str):
dict[k] = v.strip()
return dict
def get_module_list(args):
if not args.use_new_module:
return []
module_save_dir = args.module_save_dir
if os.path.isdir(module_save_dir):
file_list = os.listdir(module_save_dir)
module_name_dict = {}
for filename in file_list:
# relieve the name constraint
if 'MODULE' in filename and 'json' in filename or 'json' in filename:
file_path = os.path.join(module_save_dir, filename)
try:
module_dict = json.load(open(file_path))
module_dict = strip_dict(module_dict)
name_key = "module_name" if "module_name" in module_dict else "name"
module_name = module_dict[name_key]
module_dict['module_name'] = module_name
if 'module_program' not in module_dict:
module_dict['module_program'] = module_dict['module']
#if 'annotations' not in module_dict:
# module_dict['annotations'] = module_dict['program']
if module_name not in module_name_dict or \
module_dict['test_accuracy'] > module_name_dict[module_name]['test_accuracy']:
module_name_dict[module_name] = module_dict
except:
pdb.set_trace()
module_list = []
for module_dict in module_name_dict.values():
if 'test_accuracy' not in module_dict:
module_list.append(module_dict)
elif module_dict['test_accuracy'] >= args.threshold:
module_list.append(module_dict)
else:
print("There is no available module directory: %s"%(module_save_dir))
module_list = []
return module_list
def save_output(args, output_dict, filename=None):
output_dir = args.output_dir
if args.stage == 1:
output_path = os.path.join(output_dir, output_dict['annotations'][0]['id'] + '.json')
json.dump(output_dict, open(output_path, 'w'), indent=2)
elif args.stage == 1.5:
if args.split_cases:
module_head_list = output_dict.pop('module_head_list')
for index, module_head in enumerate(module_head_list):
output_dict['module_head'] = module_head
output_path = os.path.join(output_dir, output_dict['annotations'][index]['id'] + '.json')
json.dump(output_dict, open(output_path, 'w'), indent=2)
else:
output_path = os.path.join(output_dir, output_dict['annotations'][0]['id'] + '.json')
json.dump(output_dict, open(output_path, 'w'), indent=2)
elif args.stage == 2:
if filename is None:
filename = 'MODULE_' + output_dict['module_name'] + '.json'
output_path = os.path.join(output_dir, filename)
json.dump(output_dict, open(output_path, 'w'), indent=2)
elif args.stage == 3:
output_path = os.path.join(output_dir, 'result_' + output_dict['id'] + '.json')
json.dump(output_dict, open(output_path, 'w'), indent=2)
pass
def init_gpt(args):
if args.model == 'wizardlm': | Wizardlm.init() | 0 | 2023-11-01 16:39:33+00:00 | 2k |
ml4bio/RhoFold | rhofold/model/primitives.py | [
{
"identifier": "permute_final_dims",
"path": "rhofold/utils/tensor_utils.py",
"snippet": "def permute_final_dims(tensor: torch.Tensor, inds: List[int]):\n zero_index = -1 * len(inds)\n first_inds = list(range(len(tensor.shape[:zero_index])))\n return tensor.permute(first_inds + [zero_index + i... | import math
import torch
import torch.nn as nn
import matplotlib.pyplot as plt
import numpy as np
from typing import Optional, List, Tuple
from rhofold.utils.tensor_utils import (
permute_final_dims,
flatten_final_dims,
) | 757 | # Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
def _prod(nums):
out = 1
for n in nums:
out = out * n
return out
class Linear(nn.Linear):
"""
A Linear layer with built-in nonstandard initializations. Called just
like torch.nn.Linear.
Implements the initializers in 1.11.4, plus some additional ones found
in the code.
"""
def __init__(
self,
in_dim: int,
out_dim: int,
bias: bool = True,
):
"""
Args:
in_dim:
The final dimension of inputs to the layer
out_dim:
The final dimension of layer outputs
bias:
Whether to learn an additive bias. True by default
"""
super(Linear, self).__init__(in_dim, out_dim, bias=bias)
if bias:
with torch.no_grad():
self.bias.fill_(0)
class LayerNorm(nn.Module):
def __init__(self, c_in, eps=1e-5):
super(LayerNorm, self).__init__()
self.c_in = (c_in,)
self.eps = eps
self.weight = nn.Parameter(torch.ones(c_in))
self.bias = nn.Parameter(torch.zeros(c_in))
def forward(self, x):
out = nn.functional.layer_norm(
x,
self.c_in,
self.weight,
self.bias,
self.eps,
)
return out
@torch.jit.ignore
def softmax_no_cast(t: torch.Tensor, dim: int = -1) -> torch.Tensor:
"""
Softmax, but without automatic casting to fp32 when the input is of
type bfloat16
"""
s = torch.nn.functional.softmax(t, dim=dim)
return s
a_c = 0
attn_maps = []
attn_map_dir = None
seq_len = None
save_png = False
msa_depth = None
#@torch.jit.script
def _attention(query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, biases: List[torch.Tensor]) -> torch.Tensor:
# [*, H, C_hidden, K]
| # Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
def _prod(nums):
out = 1
for n in nums:
out = out * n
return out
class Linear(nn.Linear):
"""
A Linear layer with built-in nonstandard initializations. Called just
like torch.nn.Linear.
Implements the initializers in 1.11.4, plus some additional ones found
in the code.
"""
def __init__(
self,
in_dim: int,
out_dim: int,
bias: bool = True,
):
"""
Args:
in_dim:
The final dimension of inputs to the layer
out_dim:
The final dimension of layer outputs
bias:
Whether to learn an additive bias. True by default
"""
super(Linear, self).__init__(in_dim, out_dim, bias=bias)
if bias:
with torch.no_grad():
self.bias.fill_(0)
class LayerNorm(nn.Module):
def __init__(self, c_in, eps=1e-5):
super(LayerNorm, self).__init__()
self.c_in = (c_in,)
self.eps = eps
self.weight = nn.Parameter(torch.ones(c_in))
self.bias = nn.Parameter(torch.zeros(c_in))
def forward(self, x):
out = nn.functional.layer_norm(
x,
self.c_in,
self.weight,
self.bias,
self.eps,
)
return out
@torch.jit.ignore
def softmax_no_cast(t: torch.Tensor, dim: int = -1) -> torch.Tensor:
"""
Softmax, but without automatic casting to fp32 when the input is of
type bfloat16
"""
s = torch.nn.functional.softmax(t, dim=dim)
return s
a_c = 0
attn_maps = []
attn_map_dir = None
seq_len = None
save_png = False
msa_depth = None
#@torch.jit.script
def _attention(query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, biases: List[torch.Tensor]) -> torch.Tensor:
# [*, H, C_hidden, K] | key = permute_final_dims(key, (1, 0)) | 0 | 2023-11-01 10:29:08+00:00 | 2k |
ziqi-zhang/TAOISM | python/layers/flatten.py | [
{
"identifier": "SecretNonlinearLayer",
"path": "python/layers/nonlinear.py",
"snippet": "class SecretNonlinearLayer(SecretLayerBase):\n def __init__(\n self, sid, LayerName, EnclaveMode, link_prev=True, link_next=True,\n manually_register_prev=False, manually_register_next=False\n )... | from python.layers.nonlinear import SecretNonlinearLayer
from python.utils.timer_utils import NamedTimerInstance, VerboseLevel
from python.utils.torch_utils import compare_expected_actual
from python.utils.basic_utils import ExecutionModeOptions | 1,305 |
# Assume the prev. layer is of 4d. It outputs a 2d mat
# This layer doesnt pull the input in enclave if it is not so reduce duplicated action
class SecretFlattenLayer(SecretNonlinearLayer):
batch_size = None
n_features = None
input_shape = None
output_shape = None
def __init__(
self, sid, LayerName, EnclaveMode, link_prev=True, link_next=True,
manually_register_prev=False, manually_register_next=False
):
super().__init__(sid, LayerName, EnclaveMode, link_prev, link_next, manually_register_prev, manually_register_next)
self.StoreInEnclave = False
self.ForwardFuncName = "Flatten"
self.BackwardFuncName = "DerFlatten"
def init(self, start_enclave=True):
super().init(start_enclave)
self.ForwardFunc = lambda x: x.view(-1, self.n_features)
self.PlainFunc = lambda x: x.view(-1, self.n_features)
def init_shape(self):
self.input_shape = self.PrevLayer.get_output_shape()
if len(self.input_shape) != 4:
return ValueError("The dimension of the tensor form prev. layer has to be 4D.")
self.batch_size = self.input_shape[0]
self.n_features = self.input_shape[1] * self.input_shape[2] * self.input_shape[3]
self.output_shape = [self.batch_size, self.n_features]
def get_output_shape(self):
return self.output_shape
def generate_tensor_name_list(self, force=False):
if not force and self.tensor_name_list:
return
if self.sid == 2:
self.tensor_name_list = {}
return
NeededTensorNames = [("output", self.output_shape, None),
("input", self.input_shape, None),
("DerInput", self.input_shape, None),
("DerOutput", self.output_shape, None)
]
self.tensor_name_list = NeededTensorNames
def forward(self):
with NamedTimerInstance(f"S{self.sid}: {self.LayerName} Forward", verbose_level=VerboseLevel.LAYER):
|
# Assume the prev. layer is of 4d. It outputs a 2d mat
# This layer doesnt pull the input in enclave if it is not so reduce duplicated action
class SecretFlattenLayer(SecretNonlinearLayer):
batch_size = None
n_features = None
input_shape = None
output_shape = None
def __init__(
self, sid, LayerName, EnclaveMode, link_prev=True, link_next=True,
manually_register_prev=False, manually_register_next=False
):
super().__init__(sid, LayerName, EnclaveMode, link_prev, link_next, manually_register_prev, manually_register_next)
self.StoreInEnclave = False
self.ForwardFuncName = "Flatten"
self.BackwardFuncName = "DerFlatten"
def init(self, start_enclave=True):
super().init(start_enclave)
self.ForwardFunc = lambda x: x.view(-1, self.n_features)
self.PlainFunc = lambda x: x.view(-1, self.n_features)
def init_shape(self):
self.input_shape = self.PrevLayer.get_output_shape()
if len(self.input_shape) != 4:
return ValueError("The dimension of the tensor form prev. layer has to be 4D.")
self.batch_size = self.input_shape[0]
self.n_features = self.input_shape[1] * self.input_shape[2] * self.input_shape[3]
self.output_shape = [self.batch_size, self.n_features]
def get_output_shape(self):
return self.output_shape
def generate_tensor_name_list(self, force=False):
if not force and self.tensor_name_list:
return
if self.sid == 2:
self.tensor_name_list = {}
return
NeededTensorNames = [("output", self.output_shape, None),
("input", self.input_shape, None),
("DerInput", self.input_shape, None),
("DerOutput", self.output_shape, None)
]
self.tensor_name_list = NeededTensorNames
def forward(self):
with NamedTimerInstance(f"S{self.sid}: {self.LayerName} Forward", verbose_level=VerboseLevel.LAYER): | if self.EnclaveMode == ExecutionModeOptions.Enclave: | 4 | 2023-11-01 10:37:37+00:00 | 2k |
rafaelleinio/biar | biar/model.py | [
{
"identifier": "ContentCallbackError",
"path": "biar/errors.py",
"snippet": "class ContentCallbackError(Exception):\n \"\"\"Base Exception for content callback errors.\"\"\""
},
{
"identifier": "ResponseEvaluationError",
"path": "biar/errors.py",
"snippet": "class ResponseEvaluationE... | import asyncio
import aiohttp
import tenacity
from functools import cached_property
from typing import Any, Callable, Dict, List, Literal, Optional, Tuple, Type
from aiohttp import ClientResponseError
from loguru import logger
from pydantic import BaseModel, ConfigDict, Field, JsonValue, computed_field
from pyrate_limiter import Duration, InMemoryBucket, Limiter, Rate
from yarl import URL
from biar.errors import ContentCallbackError, ResponseEvaluationError | 724 |
class ProxyConfig(BaseModel):
"""Proxy configuration.
Attributes:
host: proxy address.
headers: additional configuration required by the proxy.
ssl_cadata: certificate as a string required by some proxies to use SSL.
"""
host: str
headers: Optional[Dict[str, Any]] = None
ssl_cadata: Optional[str] = None
class Response(BaseModel):
"""Attributes from the http request response.
Attributes:
url: final url after (possible) redirects.
status_code: HTTP status code.
headers: headers in the response.
json_content: response content as json dict.
text_content: raw response content as a string.
"""
model_config = ConfigDict(arbitrary_types_allowed=True)
url: URL
status_code: int
headers: Dict[str, Any] = Field(default_factory=dict)
json_content: Dict[str, JsonValue] = Field(default_factory=dict)
text_content: str = ""
class StructuredResponse(Response):
"""Attributes from the http request response.
Attributes:
url: final url after (possible) redirects.
status_code: HTTP status code.
headers: headers in the response.
json_content: response content as json dict.
text_content: raw response content as a string.
structured_content: response content as a pydantic model.
"""
structured_content: Any
class Retryer(BaseModel):
"""Retry logic with exponential backoff strategy.
Attributes:
attempts: number of attempts.
`attempts=1` means only one try and no subsequent retry attempts.
min_delay: number of seconds as the starting delay.
max_delay: number of seconds as the maximum achieving delay.
retry_if_exception_in: retry if exception found in this tuple.
A ResponseEvaluationError is always added dynamically to be retried.
"""
attempts: int = 1
min_delay: int = 0
max_delay: int = 10
retry_if_exception_in: Tuple[Type[BaseException], ...] = (
ClientResponseError,
asyncio.TimeoutError,
ContentCallbackError,
)
retry_based_on_content_callback: Optional[Callable[[BaseModel], bool]] = None
@property
def retrying_config(self) -> Dict[str, Any]:
"""Configuration for retrying logic.
Changing arguments at run time reference:
https://github.com/jd/tenacity#changing-arguments-at-run-time
Returns:
kwargs dictionary for tenacity.BaseRetrying.
"""
return dict(
stop=tenacity.stop_after_attempt(self.attempts),
retry=tenacity.retry_if_exception_type(
|
class ProxyConfig(BaseModel):
"""Proxy configuration.
Attributes:
host: proxy address.
headers: additional configuration required by the proxy.
ssl_cadata: certificate as a string required by some proxies to use SSL.
"""
host: str
headers: Optional[Dict[str, Any]] = None
ssl_cadata: Optional[str] = None
class Response(BaseModel):
"""Attributes from the http request response.
Attributes:
url: final url after (possible) redirects.
status_code: HTTP status code.
headers: headers in the response.
json_content: response content as json dict.
text_content: raw response content as a string.
"""
model_config = ConfigDict(arbitrary_types_allowed=True)
url: URL
status_code: int
headers: Dict[str, Any] = Field(default_factory=dict)
json_content: Dict[str, JsonValue] = Field(default_factory=dict)
text_content: str = ""
class StructuredResponse(Response):
"""Attributes from the http request response.
Attributes:
url: final url after (possible) redirects.
status_code: HTTP status code.
headers: headers in the response.
json_content: response content as json dict.
text_content: raw response content as a string.
structured_content: response content as a pydantic model.
"""
structured_content: Any
class Retryer(BaseModel):
"""Retry logic with exponential backoff strategy.
Attributes:
attempts: number of attempts.
`attempts=1` means only one try and no subsequent retry attempts.
min_delay: number of seconds as the starting delay.
max_delay: number of seconds as the maximum achieving delay.
retry_if_exception_in: retry if exception found in this tuple.
A ResponseEvaluationError is always added dynamically to be retried.
"""
attempts: int = 1
min_delay: int = 0
max_delay: int = 10
retry_if_exception_in: Tuple[Type[BaseException], ...] = (
ClientResponseError,
asyncio.TimeoutError,
ContentCallbackError,
)
retry_based_on_content_callback: Optional[Callable[[BaseModel], bool]] = None
@property
def retrying_config(self) -> Dict[str, Any]:
"""Configuration for retrying logic.
Changing arguments at run time reference:
https://github.com/jd/tenacity#changing-arguments-at-run-time
Returns:
kwargs dictionary for tenacity.BaseRetrying.
"""
return dict(
stop=tenacity.stop_after_attempt(self.attempts),
retry=tenacity.retry_if_exception_type( | exception_types=self.retry_if_exception_in + (ResponseEvaluationError,) | 1 | 2023-11-03 00:03:59+00:00 | 2k |
NVlabs/M2T2 | m2t2/action_decoder.py | [
{
"identifier": "MLP",
"path": "m2t2/model_utils.py",
"snippet": "class MLP(nn.Module):\n def __init__(\n self, input_dim, hidden_dim, output_dim, num_layers,\n activation=\"ReLU\", dropout=0.\n ):\n super().__init__()\n h = [hidden_dim] * (num_layers - 1)\n laye... | import numpy as np
import torch
import torch.nn.functional as F
import trimesh.transformations as tra
from m2t2.model_utils import MLP, repeat_new_axis | 1,520 | # distribution of this software and related documentation without an express
# license agreement from NVIDIA CORPORATION is strictly prohibited.
#
# Author: Wentao Yuan
"""
Modules to compute gripper poses from contact masks and parameters.
"""
def double_split(tensor, chunks):
tensor = list(tensor.split([sum(chunk) for chunk in chunks]))
tensor = [
list(elem.split([n for n in chunk]))
for elem, chunk in zip(tensor, chunks)
]
return tensor
def build_6d_grasp(
contact_pt, contact_dir, approach_dir, offset, gripper_depth=0.1034
):
grasp_tr = torch.stack([
contact_dir,
torch.cross(approach_dir, contact_dir),
approach_dir,
contact_pt + contact_dir * offset.unsqueeze(-1) / 2
- gripper_depth * approach_dir
], axis=-1)
last_row = torch.tensor([[0, 0, 0, 1]]).to(grasp_tr.device)
if len(grasp_tr.shape) > 2:
last_row = last_row * torch.ones(
*grasp_tr.shape[:-2], 1, 4, device=grasp_tr.device
)
grasp_tr = torch.cat([grasp_tr, last_row], dim=-2)
return grasp_tr
def build_6d_place(contact_pts, rot, offset, ee_pose):
# Transformation order: first rotate gripper to grasp pose,
# then add offset between gripper center and reference point,
# then rotate around object center, finally translate to contact point.
rot = rot @ ee_pose[..., :3, :3]
trans = (contact_pts + offset).unsqueeze(-1)
place_tr = torch.cat([rot, trans], axis=-1)
last_row = torch.tensor([[0, 0, 0, 1]]).to(place_tr.device)
if len(place_tr.shape) > 2:
last_row = last_row * torch.ones(
*place_tr.shape[:-2], 1, 4, device=place_tr.device
)
place_tr = torch.cat([place_tr, last_row], dim=-2)
return place_tr
def compute_offset(obj_pts, ee_pose, rot, grid_res=0, cam_pose=None):
# rot R is about object center o
# offset is ee position e - target position t
# R(e - o) - R(t - o) = -R(t - e)
if cam_pose is not None:
rot = cam_pose[:3, :3] @ rot
obj_pts_stable = (obj_pts - ee_pose[:3, 3]) @ rot.transpose(-1, -2)
if grid_res > 0:
obj_pts_grid = (obj_pts_stable[..., :2] / grid_res).round()
offset = obj_pts_stable.min(dim=0)[0]
offset[:2] = obj_pts_grid.unique(dim=0).mean(dim=0) * grid_res
else:
offset = obj_pts_stable.mean(dim=0)
offset[..., 2] = obj_pts_stable[..., 2].min(dim=1)[0]
offset = -offset
if cam_pose is not None:
offset = offset @ cam_pose[:3, :3]
return offset
def infer_placements(
xyz, logits, bottom_center, ee_poses, cam_poses, conf_thresh, height
):
rot_prompts = torch.stack([torch.from_numpy(
tra.euler_matrix(0, 0, 2 * np.pi / logits.shape[1] * i)
)[:3, :3].float() for i in range(logits.shape[1])]).to(xyz.device)
rot_prompts = repeat_new_axis(rot_prompts, xyz.shape[1], dim=1)
placements, confidence, contact_points = [], [], []
for i, (pts, bc, ee_pose, logit) in enumerate(zip(
xyz, bottom_center, ee_poses, logits
)):
conf = logit.sigmoid()
mask = conf > conf_thresh
num = list(mask.sum(dim=1))
rot = rot_prompts[mask]
offsets = (ee_pose[:3, 3] - bc) @ rot.transpose(1, 2)
if cam_poses is not None:
pts = pts @ cam_poses[i, :3, :3].T + cam_poses[i, :3, 3]
contacts = repeat_new_axis(pts, mask.shape[0], dim=0)[mask]
place = build_6d_place(contacts, rot, offsets, ee_pose)
place[:, 2, 3] = place[:, 2, 3] + height
if cam_poses is not None:
place = cam_poses[i].inverse() @ place
placements.append(list(place.split(num)))
confidence.append(list(conf[mask].split(num)))
contact_points.append(list(contacts.split(num)))
outputs = {
'placements': placements,
'placement_confidence': confidence,
'placement_contacts': contact_points
}
return outputs
class ActionDecoder(torch.nn.Module):
def __init__(
self, mask_dim, use_embed, embed_dim, max_num_pred, num_params,
hidden_dim, num_layers, activation, offset_bins
):
super(ActionDecoder, self).__init__()
feat_dim = mask_dim
if use_embed:
feat_dim += embed_dim
self.feat_dim = feat_dim
self.use_embed = use_embed
| # Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved.
#
# NVIDIA CORPORATION and its licensors retain all intellectual property
# and proprietary rights in and to this software, related documentation
# and any modifications thereto. Any use, reproduction, disclosure or
# distribution of this software and related documentation without an express
# license agreement from NVIDIA CORPORATION is strictly prohibited.
#
# Author: Wentao Yuan
"""
Modules to compute gripper poses from contact masks and parameters.
"""
def double_split(tensor, chunks):
tensor = list(tensor.split([sum(chunk) for chunk in chunks]))
tensor = [
list(elem.split([n for n in chunk]))
for elem, chunk in zip(tensor, chunks)
]
return tensor
def build_6d_grasp(
contact_pt, contact_dir, approach_dir, offset, gripper_depth=0.1034
):
grasp_tr = torch.stack([
contact_dir,
torch.cross(approach_dir, contact_dir),
approach_dir,
contact_pt + contact_dir * offset.unsqueeze(-1) / 2
- gripper_depth * approach_dir
], axis=-1)
last_row = torch.tensor([[0, 0, 0, 1]]).to(grasp_tr.device)
if len(grasp_tr.shape) > 2:
last_row = last_row * torch.ones(
*grasp_tr.shape[:-2], 1, 4, device=grasp_tr.device
)
grasp_tr = torch.cat([grasp_tr, last_row], dim=-2)
return grasp_tr
def build_6d_place(contact_pts, rot, offset, ee_pose):
# Transformation order: first rotate gripper to grasp pose,
# then add offset between gripper center and reference point,
# then rotate around object center, finally translate to contact point.
rot = rot @ ee_pose[..., :3, :3]
trans = (contact_pts + offset).unsqueeze(-1)
place_tr = torch.cat([rot, trans], axis=-1)
last_row = torch.tensor([[0, 0, 0, 1]]).to(place_tr.device)
if len(place_tr.shape) > 2:
last_row = last_row * torch.ones(
*place_tr.shape[:-2], 1, 4, device=place_tr.device
)
place_tr = torch.cat([place_tr, last_row], dim=-2)
return place_tr
def compute_offset(obj_pts, ee_pose, rot, grid_res=0, cam_pose=None):
# rot R is about object center o
# offset is ee position e - target position t
# R(e - o) - R(t - o) = -R(t - e)
if cam_pose is not None:
rot = cam_pose[:3, :3] @ rot
obj_pts_stable = (obj_pts - ee_pose[:3, 3]) @ rot.transpose(-1, -2)
if grid_res > 0:
obj_pts_grid = (obj_pts_stable[..., :2] / grid_res).round()
offset = obj_pts_stable.min(dim=0)[0]
offset[:2] = obj_pts_grid.unique(dim=0).mean(dim=0) * grid_res
else:
offset = obj_pts_stable.mean(dim=0)
offset[..., 2] = obj_pts_stable[..., 2].min(dim=1)[0]
offset = -offset
if cam_pose is not None:
offset = offset @ cam_pose[:3, :3]
return offset
def infer_placements(
xyz, logits, bottom_center, ee_poses, cam_poses, conf_thresh, height
):
rot_prompts = torch.stack([torch.from_numpy(
tra.euler_matrix(0, 0, 2 * np.pi / logits.shape[1] * i)
)[:3, :3].float() for i in range(logits.shape[1])]).to(xyz.device)
rot_prompts = repeat_new_axis(rot_prompts, xyz.shape[1], dim=1)
placements, confidence, contact_points = [], [], []
for i, (pts, bc, ee_pose, logit) in enumerate(zip(
xyz, bottom_center, ee_poses, logits
)):
conf = logit.sigmoid()
mask = conf > conf_thresh
num = list(mask.sum(dim=1))
rot = rot_prompts[mask]
offsets = (ee_pose[:3, 3] - bc) @ rot.transpose(1, 2)
if cam_poses is not None:
pts = pts @ cam_poses[i, :3, :3].T + cam_poses[i, :3, 3]
contacts = repeat_new_axis(pts, mask.shape[0], dim=0)[mask]
place = build_6d_place(contacts, rot, offsets, ee_pose)
place[:, 2, 3] = place[:, 2, 3] + height
if cam_poses is not None:
place = cam_poses[i].inverse() @ place
placements.append(list(place.split(num)))
confidence.append(list(conf[mask].split(num)))
contact_points.append(list(contacts.split(num)))
outputs = {
'placements': placements,
'placement_confidence': confidence,
'placement_contacts': contact_points
}
return outputs
class ActionDecoder(torch.nn.Module):
def __init__(
self, mask_dim, use_embed, embed_dim, max_num_pred, num_params,
hidden_dim, num_layers, activation, offset_bins
):
super(ActionDecoder, self).__init__()
feat_dim = mask_dim
if use_embed:
feat_dim += embed_dim
self.feat_dim = feat_dim
self.use_embed = use_embed | self.contact_dir_head = MLP( | 0 | 2023-11-03 22:32:05+00:00 | 2k |
Codra-Ingenierie-Informatique/DataLab | cdl/tests/features/common/newobject_unit.py | [
{
"identifier": "execenv",
"path": "cdl/env.py",
"snippet": "DEBUG = os.environ.get(\"DEBUG\", \"\").lower() in (\"1\", \"true\")\n QUIET = \"quiet\"\n NORMAL = \"normal\"\n DEBUG = \"debug\"\n UNATTENDED_ARG = \"unattended\"\n VERBOSE_ARG = \"verbose\"\n SCREENSHOT_ARG = \"screenshot\... | from collections.abc import Generator
from guidata.qthelpers import qt_app_context
from cdl.env import execenv
from cdl.obj import (
Gauss2DParam,
ImageDatatypes,
ImageObj,
ImageTypes,
NormalRandomParam,
SignalObj,
SignalTypes,
UniformRandomParam,
create_image_from_param,
create_signal_from_param,
new_image_param,
new_signal_param,
)
from cdl.utils.vistools import view_curves, view_images | 1,082 | # -*- coding: utf-8 -*-
#
# Licensed under the terms of the BSD 3-Clause
# (see cdl/LICENSE for details)
"""
New signal/image test
Testing functions related to signal/image creation.
"""
# pylint: disable=invalid-name # Allows short reference names like x, y, ...
# pylint: disable=duplicate-code
# guitest: show
from __future__ import annotations
def iterate_signal_creation(
data_size: int = 500, non_zero: bool = False, verbose: bool = True
) -> Generator[SignalObj, None, None]:
"""Iterate over all possible signals created from parameters"""
if verbose:
| # -*- coding: utf-8 -*-
#
# Licensed under the terms of the BSD 3-Clause
# (see cdl/LICENSE for details)
"""
New signal/image test
Testing functions related to signal/image creation.
"""
# pylint: disable=invalid-name # Allows short reference names like x, y, ...
# pylint: disable=duplicate-code
# guitest: show
from __future__ import annotations
def iterate_signal_creation(
data_size: int = 500, non_zero: bool = False, verbose: bool = True
) -> Generator[SignalObj, None, None]:
"""Iterate over all possible signals created from parameters"""
if verbose: | execenv.print( | 0 | 2023-11-09 16:56:03+00:00 | 2k |
sxwyh/pytradecn | src/pytradecn/control/wrappersa.py | [
{
"identifier": "BaseUIAWrapper",
"path": "src/pytradecn/control/baseuiawrapper.py",
"snippet": "class BaseUIAWrapper(UIAWrapper):\n\n _control_types = ['BaseUIA']\n\n def __init__(self, element_info):\n super(BaseUIAWrapper, self).__init__(element_info)\n self._client = get_client(p... | from os import remove
from csv import DictReader
from decimal import Decimal
from tempfile import NamedTemporaryFile
from os.path import exists
from .baseuiawrapper import BaseUIAWrapper
from ..error import RecordNotFoundError, RecordAmbiguousError, ItemKeyError, TimeoutError | 1,392 | #
# 券商客户端自动化测试库
# Copyright (C) 2023 谁的谁(41715399@qq.com) All rights reserved.
#
# 模块功能:各种自定义控件
# 建立日期:2023.07.20
# 联系方式:谁的谁(41715399@qq.com)
#
# 开源软件声明:
# 本软件遵守“MIT License”开源协议开源,仅供学习和参考。您可以自由使用或修改源代码或二进制文件,但必须保留上述版权声明。
# 该软件旨在深度学习和挖掘python pywinauto库的功能和潜力,由于环境的不确定性和该软件的不可靠性,请不要将该软件应用于
# 实盘交易。如您确需量化交易实盘功能,请使用券商提供的量化交易平台,否则由于您使用该软件实盘交易所造成的账户损失或政策风
# 险,开源软件提供者或插件提供者均不承担任何责任。同时,无论是直接的、间接的、偶然的、潜在的因使用该软件所造成的账号安全
# 损失、数据安全损失、账户资产损失或其他任何责任事故,开源软件提供者或插件提供者均不承担任何责任。请不要将该软件应用于商
# 业活动,否则由于把该软件应用于商业活动所造成的一切损失或法律责任,开源软件提供者或插件提供者均不承担任何责任。
#
# 修改日志:
# 2022-07-20 第一次编写
#
class PromptWrapper(BaseUIAWrapper):
_control_types = ['Prompt']
def __init__(self, elem):
super(PromptWrapper, self).__init__(elem)
def __wait_prompt_close(self):
try:
# NOTE 使用_get_control从顶层窗口查找
self._get_control({'handle': self.handle}).wait_not('exists')
| #
# 券商客户端自动化测试库
# Copyright (C) 2023 谁的谁(41715399@qq.com) All rights reserved.
#
# 模块功能:各种自定义控件
# 建立日期:2023.07.20
# 联系方式:谁的谁(41715399@qq.com)
#
# 开源软件声明:
# 本软件遵守“MIT License”开源协议开源,仅供学习和参考。您可以自由使用或修改源代码或二进制文件,但必须保留上述版权声明。
# 该软件旨在深度学习和挖掘python pywinauto库的功能和潜力,由于环境的不确定性和该软件的不可靠性,请不要将该软件应用于
# 实盘交易。如您确需量化交易实盘功能,请使用券商提供的量化交易平台,否则由于您使用该软件实盘交易所造成的账户损失或政策风
# 险,开源软件提供者或插件提供者均不承担任何责任。同时,无论是直接的、间接的、偶然的、潜在的因使用该软件所造成的账号安全
# 损失、数据安全损失、账户资产损失或其他任何责任事故,开源软件提供者或插件提供者均不承担任何责任。请不要将该软件应用于商
# 业活动,否则由于把该软件应用于商业活动所造成的一切损失或法律责任,开源软件提供者或插件提供者均不承担任何责任。
#
# 修改日志:
# 2022-07-20 第一次编写
#
class PromptWrapper(BaseUIAWrapper):
_control_types = ['Prompt']
def __init__(self, elem):
super(PromptWrapper, self).__init__(elem)
def __wait_prompt_close(self):
try:
# NOTE 使用_get_control从顶层窗口查找
self._get_control({'handle': self.handle}).wait_not('exists') | except TimeoutError: | 1 | 2023-11-03 02:22:34+00:00 | 2k |
ingra14m/Tensor4D-DNeRF | models/fields.py | [
{
"identifier": "get_embedder",
"path": "models/embedder.py",
"snippet": "def get_embedder(multires, input_dims=3):\n embed_kwargs = {\n 'include_input': True,\n 'input_dims': input_dims,\n 'max_freq_log2': multires-1,\n 'num_freqs': multires,\n 'log_sampling': True... | from configparser import NoOptionError
from models.embedder import get_embedder
from models.mip_utils import integrated_pos_enc
from tqdm import tqdm
import torch
import torch.nn as nn
import torch.nn.functional as F
import numpy as np | 978 |
class FieldNetwork(nn.Module):
def __init__(self,
d_in,
d_out,
d_hidden,
d_t4d,
min_emb,
max_emb,
n_layers,
t_emb=-1,
skip_in=(4,),
bias=0.5,
geometric_init=True,
weight_norm=True):
super(FieldNetwork, self).__init__()
dims = [d_in] + [d_hidden for _ in range(n_layers)] + [d_out]
dims[0] = d_in + (max_emb - min_emb)*3*2
self.num_layers = len(dims)
self.skip_in = skip_in
self.min_emb = min_emb
self.max_emb = max_emb
self.t_emb = t_emb
if t_emb > 0:
|
class FieldNetwork(nn.Module):
def __init__(self,
d_in,
d_out,
d_hidden,
d_t4d,
min_emb,
max_emb,
n_layers,
t_emb=-1,
skip_in=(4,),
bias=0.5,
geometric_init=True,
weight_norm=True):
super(FieldNetwork, self).__init__()
dims = [d_in] + [d_hidden for _ in range(n_layers)] + [d_out]
dims[0] = d_in + (max_emb - min_emb)*3*2
self.num_layers = len(dims)
self.skip_in = skip_in
self.min_emb = min_emb
self.max_emb = max_emb
self.t_emb = t_emb
if t_emb > 0: | embed_fn, time_input_ch = get_embedder(t_emb, input_dims=1) | 0 | 2023-11-07 10:16:33+00:00 | 2k |
865charlesw/homeassistant-kidde | custom_components/kidde/binary_sensor.py | [
{
"identifier": "DOMAIN",
"path": "custom_components/kidde/const.py",
"snippet": "DOMAIN = \"kidde\""
},
{
"identifier": "KiddeCoordinator",
"path": "custom_components/kidde/coordinator.py",
"snippet": "class KiddeCoordinator(DataUpdateCoordinator):\n \"\"\"Coordinator for Kidde HomeS... | from homeassistant.components.binary_sensor import BinarySensorEntity
from homeassistant.components.binary_sensor import BinarySensorEntityDescription
from homeassistant.config_entries import ConfigEntry
from homeassistant.core import HomeAssistant
from homeassistant.helpers.entity_platform import AddEntitiesCallback
from .const import DOMAIN
from .coordinator import KiddeCoordinator
from .entity import KiddeEntity | 825 | """Binary sensor platform for Kidde Homesafe integration."""
from __future__ import annotations
_BINARY_SENSOR_DESCRIPTIONS = (
BinarySensorEntityDescription(
"smoke_alarm", icon="mdi:smoke-detector-variant-alert", name="Smoke Alarm"
),
BinarySensorEntityDescription(
"smoke_hushed", icon="mdi:smoke-detector-variant-off", name="Smoke Hushed"
),
BinarySensorEntityDescription("co_alarm", icon="mdi:molecule-co", name="CO Alarm"),
)
async def async_setup_entry(
hass: HomeAssistant, entry: ConfigEntry, async_add_devices: AddEntitiesCallback
) -> None:
"""Set up the binary sensor platform."""
| """Binary sensor platform for Kidde Homesafe integration."""
from __future__ import annotations
_BINARY_SENSOR_DESCRIPTIONS = (
BinarySensorEntityDescription(
"smoke_alarm", icon="mdi:smoke-detector-variant-alert", name="Smoke Alarm"
),
BinarySensorEntityDescription(
"smoke_hushed", icon="mdi:smoke-detector-variant-off", name="Smoke Hushed"
),
BinarySensorEntityDescription("co_alarm", icon="mdi:molecule-co", name="CO Alarm"),
)
async def async_setup_entry(
hass: HomeAssistant, entry: ConfigEntry, async_add_devices: AddEntitiesCallback
) -> None:
"""Set up the binary sensor platform.""" | coordinator: KiddeCoordinator = hass.data[DOMAIN][entry.entry_id] | 0 | 2023-11-09 23:25:02+00:00 | 2k |
humemarx/CPG-LCF | models/backbone2d/resnet.py | [
{
"identifier": "constant_init",
"path": "models/utils/weight_init.py",
"snippet": "def constant_init(module, val, bias=0):\n if hasattr(module, 'weight') and module.weight is not None:\n nn.init.constant_(module.weight, val)\n if hasattr(module, 'bias') and module.bias is not None:\n ... | import warnings
import torch
import torch.nn as nn
import torch.utils.checkpoint as cp
from collections import OrderedDict
from torch.nn.modules.batchnorm import _BatchNorm
from torch.nn.modules.instancenorm import _InstanceNorm
from torch.nn.modules.conv import _ConvNd
from models.utils.weight_init import (constant_init, kaiming_init) | 1,521 |
downsample = None
if stride != 1 or inplanes != planes * block.expansion:
downsample = []
conv_stride = stride
if avg_down:
conv_stride = 1
downsample.append(
nn.AvgPool2d(
kernel_size=stride,
stride=stride,
ceil_mode=True,
count_include_pad=False))
downsample.extend([
conv_type(
inplanes,
planes * block.expansion,
kernel_size=1,
stride=conv_stride,
bias=False),
norm_type(planes * block.expansion)
])
downsample = nn.Sequential(*downsample)
layers = []
if multi_grid is None:
if dilation > 1 and contract_dilation:
first_dilation = dilation // 2
else:
first_dilation = dilation
else:
first_dilation = multi_grid[0]
layers.append(
block(
inplanes=inplanes,
planes=planes,
stride=stride,
dilation=first_dilation,
downsample=downsample,
conv_type=conv_type,
norm_type=norm_type,
**kwargs))
inplanes = planes * block.expansion
for i in range(1, num_blocks):
layers.append(
block(
inplanes=inplanes,
planes=planes,
stride=1,
dilation=dilation if multi_grid is None else multi_grid[i],
conv_type=conv_type,
norm_type=norm_type,
**kwargs))
super().__init__(*layers)
class BasicBlock(nn.Module):
"""Basic block for ResNet."""
expansion = 1
def __init__(self,
inplanes,
planes,
stride=1,
dilation=1,
downsample=None,
style='pytorch',
with_cp=False,
conv_type=nn.Conv2d,
norm_type=nn.BatchNorm2d,
dcn=None,
plugins=None,
zero_init_residual=True):
super().__init__()
assert dcn is None, 'Not implemented yet.'
assert plugins is None, 'Not implemented yet.'
self.zero_init_residual = zero_init_residual
norm1 = norm_type(planes)
self.norm1_name = get_norm_name(norm_type, postfix=1)
norm2 = norm_type(planes)
self.norm2_name = get_norm_name(norm_type, postfix=2)
self.conv1 = conv_type(
inplanes,
planes,
3,
stride=stride,
padding=dilation,
dilation=dilation,
bias=False)
self.add_module(self.norm1_name, norm1)
self.conv2 = conv_type(
planes,
planes,
3,
padding=1,
bias=False)
self.add_module(self.norm2_name, norm2)
self.relu = nn.ReLU(inplace=True)
self.downsample = downsample
self.stride = stride
self.dilation = dilation
self.with_cp = with_cp
@property
def norm1(self):
"""nn.Module: normalization layer after the first convolution layer"""
return getattr(self, self.norm1_name)
@property
def norm2(self):
"""nn.Module: normalization layer after the second convolution layer"""
return getattr(self, self.norm2_name)
def _init_weights(self, m):
if isinstance(m, nn.Conv2d):
kaiming_init(m)
elif isinstance(m, nn.GroupNorm) or isinstance(m, _BatchNorm):
| # coding=utf-8
'''
Author: husserl
License: Apache Licence
Software: VSCode
Date: 2023-07-17 06:36:26
LastEditors: husserl
LastEditTime: 2023-11-02 15:36:30
'''
def get_norm_name(norm_type, postfix=1):
if issubclass(norm_type, _InstanceNorm): # IN is a subclass of BN
return 'in{}'.format(postfix)
elif issubclass(norm_type, _BatchNorm):
return 'bn{}'.format(postfix)
elif issubclass(norm_type, nn.GroupNorm):
return 'gn{}'.format(postfix)
elif issubclass(norm_type, nn.LayerNorm):
return 'ln{}'.format(postfix)
class ResLayer(nn.Sequential):
"""ResLayer to build ResNet style backbone.
Args:
block (nn.Module): block used to build ResLayer.
inplanes (int): inplanes of block.
planes (int): planes of block.
num_blocks (int): number of blocks.
stride (int): stride of the first block. Default: 1
avg_down (bool): Use AvgPool instead of stride conv when
downsampling in the bottleneck. Default: False
conv_cfg (dict): dictionary to construct and config conv layer.
Default: None
norm_cfg (dict): dictionary to construct and config norm layer.
Default: dict(type='BN')
multi_grid (int | None): Multi grid dilation rates of last
stage. Default: None
contract_dilation (bool): Whether contract first dilation of each layer
Default: False
"""
def __init__(self,
block,
inplanes,
planes,
num_blocks,
stride=1,
dilation=1,
avg_down=False,
conv_type=nn.Conv2d,
norm_type=nn.BatchNorm2d,
multi_grid=None,
contract_dilation=False,
**kwargs):
self.block = block
downsample = None
if stride != 1 or inplanes != planes * block.expansion:
downsample = []
conv_stride = stride
if avg_down:
conv_stride = 1
downsample.append(
nn.AvgPool2d(
kernel_size=stride,
stride=stride,
ceil_mode=True,
count_include_pad=False))
downsample.extend([
conv_type(
inplanes,
planes * block.expansion,
kernel_size=1,
stride=conv_stride,
bias=False),
norm_type(planes * block.expansion)
])
downsample = nn.Sequential(*downsample)
layers = []
if multi_grid is None:
if dilation > 1 and contract_dilation:
first_dilation = dilation // 2
else:
first_dilation = dilation
else:
first_dilation = multi_grid[0]
layers.append(
block(
inplanes=inplanes,
planes=planes,
stride=stride,
dilation=first_dilation,
downsample=downsample,
conv_type=conv_type,
norm_type=norm_type,
**kwargs))
inplanes = planes * block.expansion
for i in range(1, num_blocks):
layers.append(
block(
inplanes=inplanes,
planes=planes,
stride=1,
dilation=dilation if multi_grid is None else multi_grid[i],
conv_type=conv_type,
norm_type=norm_type,
**kwargs))
super().__init__(*layers)
class BasicBlock(nn.Module):
"""Basic block for ResNet."""
expansion = 1
def __init__(self,
inplanes,
planes,
stride=1,
dilation=1,
downsample=None,
style='pytorch',
with_cp=False,
conv_type=nn.Conv2d,
norm_type=nn.BatchNorm2d,
dcn=None,
plugins=None,
zero_init_residual=True):
super().__init__()
assert dcn is None, 'Not implemented yet.'
assert plugins is None, 'Not implemented yet.'
self.zero_init_residual = zero_init_residual
norm1 = norm_type(planes)
self.norm1_name = get_norm_name(norm_type, postfix=1)
norm2 = norm_type(planes)
self.norm2_name = get_norm_name(norm_type, postfix=2)
self.conv1 = conv_type(
inplanes,
planes,
3,
stride=stride,
padding=dilation,
dilation=dilation,
bias=False)
self.add_module(self.norm1_name, norm1)
self.conv2 = conv_type(
planes,
planes,
3,
padding=1,
bias=False)
self.add_module(self.norm2_name, norm2)
self.relu = nn.ReLU(inplace=True)
self.downsample = downsample
self.stride = stride
self.dilation = dilation
self.with_cp = with_cp
@property
def norm1(self):
"""nn.Module: normalization layer after the first convolution layer"""
return getattr(self, self.norm1_name)
@property
def norm2(self):
"""nn.Module: normalization layer after the second convolution layer"""
return getattr(self, self.norm2_name)
def _init_weights(self, m):
if isinstance(m, nn.Conv2d):
kaiming_init(m)
elif isinstance(m, nn.GroupNorm) or isinstance(m, _BatchNorm): | constant_init(m, val=1.0, bias=0.) | 0 | 2023-11-02 09:50:13+00:00 | 2k |
intelheropuck/steam.com-scraping | listingcollector.py | [
{
"identifier": "parse_price",
"path": "helpers.py",
"snippet": "def parse_price(string):\n return int(string[1:].replace(\",\", \"\").split('.')[0])"
},
{
"identifier": "Database",
"path": "helpers.py",
"snippet": "class Database:\n _instance = None\n\n @classmethod\n def ge... | import json
import threading
import time
import argparse
import requests
import requests
import re
from queue import Queue
from urllib.parse import unquote
from helpers import parse_price
from helpers import Database, Listing
from urllib import parse
from bs4 import BeautifulSoup | 1,282 |
ACTIVITY_URL = "https://steamcommunity.com/market/itemordersactivity"\
"?item_nameid={item_id}&country=RU&language=english¤cy=1&&two_factor=0&norender=1"
db = Database().get_instance()
def get_activities(item_id):
return requests.get(ACTIVITY_URL.format(item_id=item_id)).json()["activity"]
STEAM_SESSION = "cafea932c9929bccba6f1583"
def getSteamidByAvatarNameAndUsername(user_name, avatar_name):
timeout=3
username = user_name
avatarname = avatar_name.split('/')[3].split(".")[0]
resp = requests.get("https://steamcommunity.com/search/SearchCommunityAjax?text=" + parse.quote(username) + "&filter=users&sessionid=" + STEAM_SESSION, headers={"Cookie": "sessionid=" + STEAM_SESSION}, timeout=timeout)
data = resp.json()
soup = BeautifulSoup(data["html"], "html.parser")
for div in soup.find_all('div', class_='avatarMedium'):
img = div.find('img')
imgFileName = img['src'].rsplit('/', 1)
imgName = imgFileName[1].split('_', 1)[0]
if imgName == avatarname:
return div.find('a')['href'].rsplit("/", 1)[1]
return ""
def get_profile_link(steamIdOrUsername):
if re.match(r'^\d+$', steamIdOrUsername):
url = "https://steamcommunity.com/profiles/" + steamIdOrUsername
else:
url = "https://steamcommunity.com/id/" + steamIdOrUsername
return url
def worker(queue):
while True:
listing_id, listing_link = queue.get()
for activity in get_activities(listing_id):
if activity["type"] == "BuyOrderCancel" or activity["type"] == "BuyOrderMulti":
continue
item_name=unquote(listing_link.split('/')[6])
price=parse_price(activity["price"])
owner_name=activity["persona_seller"] or activity["persona_buyer"]
owner_avatar=activity["avatar_seller"] or activity["avatar_buyer"]
steamIdOrUsername=getSteamidByAvatarNameAndUsername(owner_name, owner_avatar)
url=get_profile_link(steamIdOrUsername)
if url is None:
continue
else:
rrr = requests.get(url)
soup1 = BeautifulSoup(rrr.content, 'html.parser')
situation = soup1.find('div', class_ = 'profile_in_game_header')
if situation is None:
continue
else:
online = situation.text.split(' ')[1]
if online == 'Offline':
continue
else:
|
ACTIVITY_URL = "https://steamcommunity.com/market/itemordersactivity"\
"?item_nameid={item_id}&country=RU&language=english¤cy=1&&two_factor=0&norender=1"
db = Database().get_instance()
def get_activities(item_id):
return requests.get(ACTIVITY_URL.format(item_id=item_id)).json()["activity"]
STEAM_SESSION = "cafea932c9929bccba6f1583"
def getSteamidByAvatarNameAndUsername(user_name, avatar_name):
timeout=3
username = user_name
avatarname = avatar_name.split('/')[3].split(".")[0]
resp = requests.get("https://steamcommunity.com/search/SearchCommunityAjax?text=" + parse.quote(username) + "&filter=users&sessionid=" + STEAM_SESSION, headers={"Cookie": "sessionid=" + STEAM_SESSION}, timeout=timeout)
data = resp.json()
soup = BeautifulSoup(data["html"], "html.parser")
for div in soup.find_all('div', class_='avatarMedium'):
img = div.find('img')
imgFileName = img['src'].rsplit('/', 1)
imgName = imgFileName[1].split('_', 1)[0]
if imgName == avatarname:
return div.find('a')['href'].rsplit("/", 1)[1]
return ""
def get_profile_link(steamIdOrUsername):
if re.match(r'^\d+$', steamIdOrUsername):
url = "https://steamcommunity.com/profiles/" + steamIdOrUsername
else:
url = "https://steamcommunity.com/id/" + steamIdOrUsername
return url
def worker(queue):
while True:
listing_id, listing_link = queue.get()
for activity in get_activities(listing_id):
if activity["type"] == "BuyOrderCancel" or activity["type"] == "BuyOrderMulti":
continue
item_name=unquote(listing_link.split('/')[6])
price=parse_price(activity["price"])
owner_name=activity["persona_seller"] or activity["persona_buyer"]
owner_avatar=activity["avatar_seller"] or activity["avatar_buyer"]
steamIdOrUsername=getSteamidByAvatarNameAndUsername(owner_name, owner_avatar)
url=get_profile_link(steamIdOrUsername)
if url is None:
continue
else:
rrr = requests.get(url)
soup1 = BeautifulSoup(rrr.content, 'html.parser')
situation = soup1.find('div', class_ = 'profile_in_game_header')
if situation is None:
continue
else:
online = situation.text.split(' ')[1]
if online == 'Offline':
continue
else: | listing = Listing( | 2 | 2023-11-05 04:47:16+00:00 | 2k |
JaeBinCHA7/DEMUCS-for-Speech-Enhancement | models/HDDEMUCS_TF.py | [
{
"identifier": "capture_init",
"path": "models/tools.py",
"snippet": "def capture_init(init):\n \"\"\"capture_init.\n Decorate `__init__` with this, and you can then\n recover the *args and **kwargs passed to it in `self._init_args_kwargs`\n \"\"\"\n @functools.wraps(init)\n def __ini... | import math
import torch
import torch as th
import typing as tp
from torch import nn
from torch.nn import functional as F
from einops import rearrange
from .tools import capture_init, spectro, ispectro | 1,343 | """
Reference: https://github.com/facebookresearch/denoiser/blob/main/denoiser/demucs.py
Copyright (c) Facebook, Inc. and its affiliates.
All rights reserved.
This source code is licensed under the license found in the
LICENSE file in the root directory of this source tree.
author: adefossez
"""
class BLSTM(nn.Module):
def __init__(self, dim, layers=2, bi=True):
super().__init__()
klass = nn.LSTM
self.lstm = klass(bidirectional=bi, num_layers=layers, hidden_size=dim, input_size=dim, batch_first=True)
self.linear = None
if bi:
self.linear = nn.Linear(2 * dim, dim)
def forward(self, x, hidden=None):
x, hidden = self.lstm(x, hidden)
if self.linear:
x = self.linear(x)
return x, hidden
def pad1d(x: torch.Tensor, paddings: tp.Tuple[int, int], mode: str = 'constant', value: float = 0.):
"""Tiny wrapper around F.pad, just to allow for reflect padding on small input.
If this is the case, we insert extra 0 padding to the right before the reflection happen."""
x0 = x
length = x.shape[-1]
padding_left, padding_right = paddings
if mode == 'reflect':
max_pad = max(padding_left, padding_right)
if length <= max_pad:
extra_pad = max_pad - length + 1
extra_pad_right = min(padding_right, extra_pad)
extra_pad_left = extra_pad - extra_pad_right
paddings = (padding_left - extra_pad_left, padding_right - extra_pad_right)
x = F.pad(x, (extra_pad_left, extra_pad_right))
out = F.pad(x, paddings, mode, value)
assert out.shape[-1] == length + padding_left + padding_right
assert (out[..., padding_left: padding_left + length] == x0).all()
return out
def rescale_conv(conv, reference):
std = conv.weight.std().detach()
scale = (std / reference) ** 0.5
conv.weight.data /= scale
if conv.bias is not None:
conv.bias.data /= scale
def rescale_module(module, reference):
for sub in module.modules():
if isinstance(sub, (nn.Conv1d, nn.ConvTranspose1d)):
rescale_conv(sub, reference)
class HDDEMUCS_TF(nn.Module):
"""
Demucs speech enhancement model.
Args:
- chin (int): number of input channels.
- chout (int): number of output channels.
- hidden (int): number of initial hidden channels.
- depth (int): number of layers.
- kernel_size (int): kernel size for each layer.
- stride (int): stride for each layer.
- causal (bool): if false, uses BiLSTM instead of LSTM.
- resample (int): amount of resampling to apply to the input/output.
Can be one of 1, 2 or 4.
- growth (float): number of channels is multiplied by this for every layer.
- max_hidden (int): maximum number of channels. Can be useful to
control the size/speed of the model.
- normalize (bool): if true, normalize the input.
- glu (bool): if true uses GLU instead of ReLU in 1x1 convolutions.
- rescale (float): controls custom weight initialization.
See https://arxiv.org/abs/1911.13254.
- floor (float): stability flooring when normalizing.
- sample_rate (float): sample_rate used for training the model.
"""
| """
Reference: https://github.com/facebookresearch/denoiser/blob/main/denoiser/demucs.py
Copyright (c) Facebook, Inc. and its affiliates.
All rights reserved.
This source code is licensed under the license found in the
LICENSE file in the root directory of this source tree.
author: adefossez
"""
class BLSTM(nn.Module):
def __init__(self, dim, layers=2, bi=True):
super().__init__()
klass = nn.LSTM
self.lstm = klass(bidirectional=bi, num_layers=layers, hidden_size=dim, input_size=dim, batch_first=True)
self.linear = None
if bi:
self.linear = nn.Linear(2 * dim, dim)
def forward(self, x, hidden=None):
x, hidden = self.lstm(x, hidden)
if self.linear:
x = self.linear(x)
return x, hidden
def pad1d(x: torch.Tensor, paddings: tp.Tuple[int, int], mode: str = 'constant', value: float = 0.):
"""Tiny wrapper around F.pad, just to allow for reflect padding on small input.
If this is the case, we insert extra 0 padding to the right before the reflection happen."""
x0 = x
length = x.shape[-1]
padding_left, padding_right = paddings
if mode == 'reflect':
max_pad = max(padding_left, padding_right)
if length <= max_pad:
extra_pad = max_pad - length + 1
extra_pad_right = min(padding_right, extra_pad)
extra_pad_left = extra_pad - extra_pad_right
paddings = (padding_left - extra_pad_left, padding_right - extra_pad_right)
x = F.pad(x, (extra_pad_left, extra_pad_right))
out = F.pad(x, paddings, mode, value)
assert out.shape[-1] == length + padding_left + padding_right
assert (out[..., padding_left: padding_left + length] == x0).all()
return out
def rescale_conv(conv, reference):
std = conv.weight.std().detach()
scale = (std / reference) ** 0.5
conv.weight.data /= scale
if conv.bias is not None:
conv.bias.data /= scale
def rescale_module(module, reference):
for sub in module.modules():
if isinstance(sub, (nn.Conv1d, nn.ConvTranspose1d)):
rescale_conv(sub, reference)
class HDDEMUCS_TF(nn.Module):
"""
Demucs speech enhancement model.
Args:
- chin (int): number of input channels.
- chout (int): number of output channels.
- hidden (int): number of initial hidden channels.
- depth (int): number of layers.
- kernel_size (int): kernel size for each layer.
- stride (int): stride for each layer.
- causal (bool): if false, uses BiLSTM instead of LSTM.
- resample (int): amount of resampling to apply to the input/output.
Can be one of 1, 2 or 4.
- growth (float): number of channels is multiplied by this for every layer.
- max_hidden (int): maximum number of channels. Can be useful to
control the size/speed of the model.
- normalize (bool): if true, normalize the input.
- glu (bool): if true uses GLU instead of ReLU in 1x1 convolutions.
- rescale (float): controls custom weight initialization.
See https://arxiv.org/abs/1911.13254.
- floor (float): stability flooring when normalizing.
- sample_rate (float): sample_rate used for training the model.
"""
| @capture_init | 0 | 2023-11-06 08:16:24+00:00 | 2k |
pauloxnet/generatedfields | samples/tests.py | [
{
"identifier": "Circle",
"path": "samples/models.py",
"snippet": "class Circle(models.Model):\n radius = models.FloatField()\n area = models.GeneratedField(\n expression=Round(\n Power(\"radius\", 2) * Pi(),\n precision=2,\n ),\n output_field=models.Floa... | from django.test import TestCase
from samples.models import (
Circle,
Event,
Item,
Order,
Package,
Rectangle,
RightTriangle,
Square,
User,
) | 1,227 |
class RectangleTestCase(TestCase):
@classmethod
def setUpTestData(cls):
cls.rectangle = Rectangle.objects.create(base=6, height=7)
def test_str(self):
self.assertEqual(str(self.rectangle), "6×7=42.0")
class SquareTestCase(TestCase):
@classmethod
def setUpTestData(cls):
cls.square = Square.objects.create(side=3)
def test_str(self):
self.assertEqual(str(self.square), "3²=9.0")
class CircleTestCase(TestCase):
@classmethod
def setUpTestData(cls):
cls.circle = Circle.objects.create(radius=3.1415)
def test_str(self):
self.assertEqual(str(self.circle), "3.1415²×π=31.0")
class RightTriangleTestCase(TestCase):
@classmethod
def setUpTestData(cls):
|
class RectangleTestCase(TestCase):
@classmethod
def setUpTestData(cls):
cls.rectangle = Rectangle.objects.create(base=6, height=7)
def test_str(self):
self.assertEqual(str(self.rectangle), "6×7=42.0")
class SquareTestCase(TestCase):
@classmethod
def setUpTestData(cls):
cls.square = Square.objects.create(side=3)
def test_str(self):
self.assertEqual(str(self.square), "3²=9.0")
class CircleTestCase(TestCase):
@classmethod
def setUpTestData(cls):
cls.circle = Circle.objects.create(radius=3.1415)
def test_str(self):
self.assertEqual(str(self.circle), "3.1415²×π=31.0")
class RightTriangleTestCase(TestCase):
@classmethod
def setUpTestData(cls): | cls.righttriangle = RightTriangle.objects.create(hypotenuse=5, angle=45) | 6 | 2023-11-07 17:06:11+00:00 | 2k |
akhilravidas/stack-sparrow | sparrow/assistant/run.py | [
{
"identifier": "actions",
"path": "sparrow/assistant/actions.py",
"snippet": "class FileReviewComments(BaseModel):\nclass FileReviewResult(BaseModel):\n def new(cls, json_input: str) -> FileReviewResult:"
},
{
"identifier": "BaseReview",
"path": "sparrow/assistant/review.py",
"snippe... | import json
import logging
import os
import time
import pydantic
from functools import lru_cache
from typing import List, Optional, Tuple
from openai import OpenAI
from openai.types.beta.threads import Run
from rich import print # pylint: disable=redefined-builtin
from rich.progress import Progress, SpinnerColumn, TextColumn
from sparrow.assistant import actions
from sparrow.assistant.review import BaseReview, ReviewFile, ReviewPlan
from sparrow.libs import config, constants, llm, scm, strings | 1,248 | """
OpenAI assistant
"""
ASSISTANT_INSTRUCTIONS = """
You an an assistant that helps with DevOps tasks. You review code, help with adding documentation etc..
""".strip()
REVIEW_THREAD_INSTRUCTIONS = """
Each message in this thread represents changes made to a file in the patch set.
The first line is the file path. The subsequent lines contains the file contents annotated with line numbers.
Only the lines that start with an asterisk were updated.
IMPORTANT:
- Review code and flag substantive issues for updated code (lines marked with an asterisk).
- Only reject if you are sure that there is an underlying issue with the code.
- Do not flag formatting or style issues.
""".strip()
@lru_cache(maxsize=None)
def _client() -> OpenAI:
| """
OpenAI assistant
"""
ASSISTANT_INSTRUCTIONS = """
You an an assistant that helps with DevOps tasks. You review code, help with adding documentation etc..
""".strip()
REVIEW_THREAD_INSTRUCTIONS = """
Each message in this thread represents changes made to a file in the patch set.
The first line is the file path. The subsequent lines contains the file contents annotated with line numbers.
Only the lines that start with an asterisk were updated.
IMPORTANT:
- Review code and flag substantive issues for updated code (lines marked with an asterisk).
- Only reject if you are sure that there is an underlying issue with the code.
- Do not flag formatting or style issues.
""".strip()
@lru_cache(maxsize=None)
def _client() -> OpenAI: | return OpenAI(api_key=config.AppConfig.instance().openai_token) | 4 | 2023-11-07 00:55:26+00:00 | 2k |
nimamahmoudi/LLMStreamlitDemoBasic | app-agent.py | [
{
"identifier": "get_agent_chain",
"path": "llm_helper.py",
"snippet": "def get_agent_chain(file_name=\"Mahmoudi_Nima_202202_PhD.pdf\", index_folder=\"index\", callbacks=None, st_cb: Optional[StreamlitCallbackHandler] = None, ):\n if callbacks is None:\n callbacks = []\n\n from langchain.ag... | import streamlit as st
from langchain.agents import initialize_agent, AgentType
from langchain.callbacks import StreamlitCallbackHandler
from langchain.chat_models import ChatOpenAI
from langchain.prompts import ChatPromptTemplate, MessagesPlaceholder
from llm_helper import get_agent_chain, get_lc_oai_tools | 1,017 |
with st.sidebar:
openai_api_key = st.secrets["OPENAI_API_KEY"]
"[Get an OpenAI API key](https://platform.openai.com/account/api-keys)"
"[View the source code](https://github.com/streamlit/llm-examples/blob/main/pages/2_Chat_with_search.py)"
"[](https://codespaces.new/streamlit/llm-examples?quickstart=1)"
st.title("🔎 LangChain - Chat with search")
"""
In this example, we're using `StreamlitCallbackHandler` to display the thoughts and actions of an agent in an interactive Streamlit app.
Try more LangChain 🤝 Streamlit Agent examples at [github.com/langchain-ai/streamlit-agent](https://github.com/langchain-ai/streamlit-agent).
"""
if "messages" not in st.session_state:
st.session_state["messages"] = [
{"role": "assistant", "content": "Hi, I'm a chatbot who can search the web. How can I help you?"}
]
for msg in st.session_state.messages:
st.chat_message(msg["role"]).write(msg["content"])
if prompt := st.chat_input(placeholder="Who won the Women's U.S. Open in 2018?"):
st.session_state.messages.append({"role": "user", "content": prompt})
st.chat_message("user").write(prompt)
if not openai_api_key:
st.info("Please add your OpenAI API key to continue.")
st.stop()
llm = ChatOpenAI(model_name="gpt-3.5-turbo-1106", openai_api_key=openai_api_key, streaming=True)
|
with st.sidebar:
openai_api_key = st.secrets["OPENAI_API_KEY"]
"[Get an OpenAI API key](https://platform.openai.com/account/api-keys)"
"[View the source code](https://github.com/streamlit/llm-examples/blob/main/pages/2_Chat_with_search.py)"
"[](https://codespaces.new/streamlit/llm-examples?quickstart=1)"
st.title("🔎 LangChain - Chat with search")
"""
In this example, we're using `StreamlitCallbackHandler` to display the thoughts and actions of an agent in an interactive Streamlit app.
Try more LangChain 🤝 Streamlit Agent examples at [github.com/langchain-ai/streamlit-agent](https://github.com/langchain-ai/streamlit-agent).
"""
if "messages" not in st.session_state:
st.session_state["messages"] = [
{"role": "assistant", "content": "Hi, I'm a chatbot who can search the web. How can I help you?"}
]
for msg in st.session_state.messages:
st.chat_message(msg["role"]).write(msg["content"])
if prompt := st.chat_input(placeholder="Who won the Women's U.S. Open in 2018?"):
st.session_state.messages.append({"role": "user", "content": prompt})
st.chat_message("user").write(prompt)
if not openai_api_key:
st.info("Please add your OpenAI API key to continue.")
st.stop()
llm = ChatOpenAI(model_name="gpt-3.5-turbo-1106", openai_api_key=openai_api_key, streaming=True) | lc_tools, _ = get_lc_oai_tools() | 1 | 2023-11-05 13:19:04+00:00 | 2k |
JakubPluta/gymhero | gymhero/api/routes/user.py | [
{
"identifier": "get_current_superuser",
"path": "gymhero/api/dependencies.py",
"snippet": "def get_current_superuser(\n current_user: User = Depends(get_current_user),\n) -> User:\n \"\"\"Returns the current superuser.\n\n Parameters:\n current_user (User, optional): The current user.\n... | from typing import List, Optional
from fastapi import APIRouter, Depends, HTTPException, status
from sqlalchemy.orm import Session
from gymhero.api.dependencies import get_current_superuser, get_pagination_params
from gymhero.crud import user_crud
from gymhero.database.db import get_db
from gymhero.log import get_logger
from gymhero.models import User
from gymhero.schemas.user import UserCreate, UserInDB, UserOut, UserUpdate
from gymhero.security import get_password_hash | 1,318 |
log = get_logger(__name__)
router = APIRouter(dependencies=[Depends(get_current_superuser)])
@router.get(
"/all", response_model=List[Optional[UserOut]], status_code=status.HTTP_200_OK
)
def fetch_all_users(
|
log = get_logger(__name__)
router = APIRouter(dependencies=[Depends(get_current_superuser)])
@router.get(
"/all", response_model=List[Optional[UserOut]], status_code=status.HTTP_200_OK
)
def fetch_all_users( | db: Session = Depends(get_db), pagination_params=Depends(get_pagination_params) | 3 | 2023-11-05 14:37:46+00:00 | 2k |
IIMunchII/restllm | src/restllm/models/chat.py | [
{
"identifier": "MetaModel",
"path": "src/restllm/models/base.py",
"snippet": "class MetaModel(BaseModel):\n id: int = Field(gt=0, examples=[1, 2, 3])\n class_name: str\n owner: int\n object: Any\n created_at: Datetime = Field(default_factory=Datetime)\n updated_at: Datetime = Field(de... | from enum import auto, UNIQUE, verify, StrEnum
from typing import Optional
from pydantic import BaseModel, Field
from .base import MetaModel
from .functions import FunctionCall
from .completion import CompletionParameters
from ..models.functions import get_function_schemas | 1,232 |
@verify(UNIQUE)
class RoleTypes(StrEnum):
USER = auto()
SYSTEM = auto()
ASSISTANT = auto()
FUNCTION = auto()
@verify(UNIQUE)
class ModelTypes(StrEnum):
GPT3_TURBO = "gpt-3.5-turbo"
GPT3_TURBO_16K = "gpt-3.5-turbo-16k"
GPT4 = "gpt-4"
GPT4_32K = "gpt-4-32k"
class ChatMessage(BaseModel):
role: RoleTypes = Field(
description="The role of the message's author. Roles can be: system, user, assistant, or function.",
examples=[RoleTypes.USER, RoleTypes.SYSTEM],
)
content: str = Field(
description="The contents of the message. It is required for all messages, but may be null for assistant messages with function calls.",
examples=["Can you write a function in Python that adds two numbers together?"],
)
name: Optional[str] = Field(
default=None,
max_length=64,
pattern="^[a-zA-Z0-9_]*$",
description="The name of the author of the message. It is required if the role is 'function'. The name should match the name of the function represented in the content. It can contain characters (a-z, A-Z, 0-9), and underscores, with a maximum length of 64 characters.",
examples=["Alice", "AI Assistant"],
)
|
@verify(UNIQUE)
class RoleTypes(StrEnum):
USER = auto()
SYSTEM = auto()
ASSISTANT = auto()
FUNCTION = auto()
@verify(UNIQUE)
class ModelTypes(StrEnum):
GPT3_TURBO = "gpt-3.5-turbo"
GPT3_TURBO_16K = "gpt-3.5-turbo-16k"
GPT4 = "gpt-4"
GPT4_32K = "gpt-4-32k"
class ChatMessage(BaseModel):
role: RoleTypes = Field(
description="The role of the message's author. Roles can be: system, user, assistant, or function.",
examples=[RoleTypes.USER, RoleTypes.SYSTEM],
)
content: str = Field(
description="The contents of the message. It is required for all messages, but may be null for assistant messages with function calls.",
examples=["Can you write a function in Python that adds two numbers together?"],
)
name: Optional[str] = Field(
default=None,
max_length=64,
pattern="^[a-zA-Z0-9_]*$",
description="The name of the author of the message. It is required if the role is 'function'. The name should match the name of the function represented in the content. It can contain characters (a-z, A-Z, 0-9), and underscores, with a maximum length of 64 characters.",
examples=["Alice", "AI Assistant"],
) | function_call: Optional[FunctionCall] = Field( | 1 | 2023-11-05 19:16:00+00:00 | 2k |
rabilrbl/deepseek-api | deepseek_api/deepseek_api.py | [
{
"identifier": "API_URL",
"path": "deepseek_api/constants.py",
"snippet": "class API_URL:\n \"\"\"Deepseek API URL constants\"\"\"\n\n BASE_URL = \"https://coder.deepseek.com/api/v0\"\n LOGIN = BASE_URL + \"/users/login\"\n CLEAR_CONTEXT = BASE_URL + \"/chat/clear_context\"\n CHAT = BASE... | import requests
import aiohttp
import aiofiles
import threading
import json
import jwt
import datetime
from abc import ABC, abstractmethod
from deepseek_api.constants import API_URL, DeepseekConstants
from deepseek_api.errors import EmptyEmailOrPasswordError, NotLoggedInError | 743 |
class DeepseekBase(ABC):
"""
A base class to create DeepseekAPI instances.
"""
def __init__(
self,
email: str,
password: str,
model_class: str = "deepseek_code",
save_login: bool = False,
):
"""
Constructor method for DeepseekAPI class.
Initializes a DeepseekAPI instance with provided credentials and settings.
Parameters:
email (str): User's email for Deepseek account
password (str): Password for user's Deepseek account
model_class (str): Deepseek model to use, either 'deepseek_chat' or 'deepseek_code'
save_login (bool): Whether to save credentials to login.json to avoid re-login
"""
self.email = email
self.password = password
self.model_class = model_class
self.save_login = save_login
|
class DeepseekBase(ABC):
"""
A base class to create DeepseekAPI instances.
"""
def __init__(
self,
email: str,
password: str,
model_class: str = "deepseek_code",
save_login: bool = False,
):
"""
Constructor method for DeepseekAPI class.
Initializes a DeepseekAPI instance with provided credentials and settings.
Parameters:
email (str): User's email for Deepseek account
password (str): Password for user's Deepseek account
model_class (str): Deepseek model to use, either 'deepseek_chat' or 'deepseek_code'
save_login (bool): Whether to save credentials to login.json to avoid re-login
"""
self.email = email
self.password = password
self.model_class = model_class
self.save_login = save_login | self.headers = DeepseekConstants.BASE_HEADERS | 1 | 2023-11-09 18:42:43+00:00 | 2k |
HealthSciTech/E2E-PPG | example.py | [
{
"identifier": "e2e_hrv_extraction",
"path": "e2e_ppg_pipeline.py",
"snippet": "def e2e_hrv_extraction(\n input_sig: np.ndarray,\n sampling_rate: int,\n window_length_sec: int = 60,\n peak_detection_method: str = 'kazemi'\n) -> pd.DataFrame:\n '''\n End-to-end HR and H... | import os
import warnings
from e2e_ppg_pipeline import e2e_hrv_extraction
from utils import get_data | 1,085 | # -*- coding: utf-8 -*-
warnings.filterwarnings("ignore")
# Import a sample data
file_name = "201902020222_Data.csv"
sampling_frequency = 20
input_sig = get_data(file_name=file_name)
# Set the window length for HR and HRV extraction in seconds
window_length_sec = 90
# Extract HRV parameters from the input PPG signal
| # -*- coding: utf-8 -*-
warnings.filterwarnings("ignore")
# Import a sample data
file_name = "201902020222_Data.csv"
sampling_frequency = 20
input_sig = get_data(file_name=file_name)
# Set the window length for HR and HRV extraction in seconds
window_length_sec = 90
# Extract HRV parameters from the input PPG signal | hrv_data = e2e_hrv_extraction( | 0 | 2023-11-07 22:52:14+00:00 | 2k |
Antelcat/ida_copilot | ida_copilot/copilot.py | [
{
"identifier": "core",
"path": "ida_copilot/core.py",
"snippet": "def push_async_call_result(result):\ndef pop_async_call_result(index):\ndef preprocess_prompt(template: str) -> str:\ndef escape_agent_input(query: str, tool_name: str) -> str:\ndef get_screen_func():\ndef get_safe_new_name(new_func_name... | import asyncio
import concurrent.futures
import re
import idaapi
from typing import Any, Optional
from langchain.agents import tool, initialize_agent, AgentType
from langchain.callbacks import FileCallbackHandler
from langchain.callbacks.base import BaseCallbackManager
from langchain.callbacks.manager import CallbackManagerForToolRun
from langchain.chat_models import ChatOpenAI
from langchain.tools import BaseTool
from ida_copilot import core, prompts | 1,570 |
class Copilot:
def run(self, temperature=0.2, model='gpt-3.5-turbo-0613'):
ea = idaapi.get_screen_ea()
func_name = idaapi.get_func_name(ea)
tools = [
self.__GetAddressInfoTool(),
self.__GetDefinitionTool(),
self.__GetPseudocodeTool(),
self.__SetFunctionCommentTool(),
self.__SetFunctionDefinitionTool(),
self.__SetFunctionNameTool(),
self.__GetIsMyWorkDoneTool(ea)
]
agent = initialize_agent(
agent_type=AgentType.OPENAI_MULTI_FUNCTIONS,
llm=ChatOpenAI(temperature=temperature, model=model),
tools=tools,
# callback_manager=BaseCallbackManager(handlers=[
# CopilotPanelCallbackManager()]),
verbose=True,
)
prompt = prompts.default_prompt_zh.format(
binary_description=f'name: {func_name}, address 0x{ea:x}'
# pseudocode=pseudocode
)
# 开启新线程运行agent
t = concurrent.futures.ThreadPoolExecutor()
loop = asyncio.get_event_loop()
loop.run_in_executor(t, agent.run, prompt)
class __GetAddressInfoTool(BaseTool):
name = 'get_address_info'
description = ('Given a hex address or function name, show its information. '
'**Input Format**: `<hex_address_or_function_name>`. '
'**Input Example1**: `sub_140007080`. '
'**Input Example2**: `0x140007080`.')
@staticmethod
def __get_address_info(name_or_hex_address: str):
try:
if name_or_hex_address.lower().startswith('0x'):
ea = int(name_or_hex_address, 16)
else:
ea = idaapi.get_name_ea(idaapi.BADADDR, name_or_hex_address)
if ea == idaapi.BADADDR:
raise Exception
except Exception:
return f'{name_or_hex_address} is not a valid address or name.'
flags = idaapi.get_flags(ea)
result = ''
# 检查地址是否位于函数内部
func = idaapi.get_func(ea)
if func:
result += "Address 0x%X is inside a function.\n" % ea
result += "Function start: 0x%X\n" % func.start_ea
result += "Function end: 0x%X\n" % func.end_ea
func_name = idaapi.get_func_name(func.start_ea)
if func_name:
result += "Function name: %s\n" % func_name
elif idaapi.is_code(flags):
result += "Address 0x%X is code.\n" % ea
elif idaapi.is_data(flags):
result += "Address 0x%X is data.\n" % ea
if idaapi.is_byte(flags):
result += "Data type: Byte\n"
result += "Value: %d\n" % idaapi.get_wide_byte(ea)
elif idaapi.is_word(flags):
result += "Data type: Word\n"
result += "Value: %d\n" % idaapi.get_wide_word(ea)
elif idaapi.is_dword(flags):
result += "Data type: Dword\n"
result += "Value: %d\n" % idaapi.get_wide_dword(ea)
elif idaapi.is_qword(flags):
result += "Data type: Qword\n"
result += "Value: %d\n" % idaapi.get_qword(ea)
elif idaapi.is_float(flags):
result += "Data type: Float\n"
# result += "Value: %f\n" % idaapi.get_wide_float(address)
elif idaapi.is_double(flags):
result += "Data type: Double\n"
# result += "Value: %f\n" % idaapi.get_wide_double(address)
elif idaapi.is_strlit(flags):
result += "Data type: String\n"
result += "Value: %s\n" % idaapi.get_strlit_contents(ea)
elif idaapi.is_struct(flags):
result += "Data type: Struct\n"
# ... 其他数据类型检查
elif idaapi.is_unknown(flags):
result += "Address 0x%X is unknown.\n" % ea
# 名称和注释
if idaapi.has_name(flags):
result += "Name: %s\n" % idaapi.get_name(ea)
elif idaapi.has_dummy_name(flags):
result += "Dummy name: %s\n" % idaapi.get_name(ea)
if idaapi.has_cmt(flags):
result += "Comment: %s\n" % idaapi.get_cmt(ea, 0)
if result == '':
result = 'Address not found.'
elif result[-1] == '\n':
result = result[:-1]
return result
def _run(self, query: str, run_manager: Optional[CallbackManagerForToolRun] = None) -> Any:
|
class Copilot:
def run(self, temperature=0.2, model='gpt-3.5-turbo-0613'):
ea = idaapi.get_screen_ea()
func_name = idaapi.get_func_name(ea)
tools = [
self.__GetAddressInfoTool(),
self.__GetDefinitionTool(),
self.__GetPseudocodeTool(),
self.__SetFunctionCommentTool(),
self.__SetFunctionDefinitionTool(),
self.__SetFunctionNameTool(),
self.__GetIsMyWorkDoneTool(ea)
]
agent = initialize_agent(
agent_type=AgentType.OPENAI_MULTI_FUNCTIONS,
llm=ChatOpenAI(temperature=temperature, model=model),
tools=tools,
# callback_manager=BaseCallbackManager(handlers=[
# CopilotPanelCallbackManager()]),
verbose=True,
)
prompt = prompts.default_prompt_zh.format(
binary_description=f'name: {func_name}, address 0x{ea:x}'
# pseudocode=pseudocode
)
# 开启新线程运行agent
t = concurrent.futures.ThreadPoolExecutor()
loop = asyncio.get_event_loop()
loop.run_in_executor(t, agent.run, prompt)
class __GetAddressInfoTool(BaseTool):
name = 'get_address_info'
description = ('Given a hex address or function name, show its information. '
'**Input Format**: `<hex_address_or_function_name>`. '
'**Input Example1**: `sub_140007080`. '
'**Input Example2**: `0x140007080`.')
@staticmethod
def __get_address_info(name_or_hex_address: str):
try:
if name_or_hex_address.lower().startswith('0x'):
ea = int(name_or_hex_address, 16)
else:
ea = idaapi.get_name_ea(idaapi.BADADDR, name_or_hex_address)
if ea == idaapi.BADADDR:
raise Exception
except Exception:
return f'{name_or_hex_address} is not a valid address or name.'
flags = idaapi.get_flags(ea)
result = ''
# 检查地址是否位于函数内部
func = idaapi.get_func(ea)
if func:
result += "Address 0x%X is inside a function.\n" % ea
result += "Function start: 0x%X\n" % func.start_ea
result += "Function end: 0x%X\n" % func.end_ea
func_name = idaapi.get_func_name(func.start_ea)
if func_name:
result += "Function name: %s\n" % func_name
elif idaapi.is_code(flags):
result += "Address 0x%X is code.\n" % ea
elif idaapi.is_data(flags):
result += "Address 0x%X is data.\n" % ea
if idaapi.is_byte(flags):
result += "Data type: Byte\n"
result += "Value: %d\n" % idaapi.get_wide_byte(ea)
elif idaapi.is_word(flags):
result += "Data type: Word\n"
result += "Value: %d\n" % idaapi.get_wide_word(ea)
elif idaapi.is_dword(flags):
result += "Data type: Dword\n"
result += "Value: %d\n" % idaapi.get_wide_dword(ea)
elif idaapi.is_qword(flags):
result += "Data type: Qword\n"
result += "Value: %d\n" % idaapi.get_qword(ea)
elif idaapi.is_float(flags):
result += "Data type: Float\n"
# result += "Value: %f\n" % idaapi.get_wide_float(address)
elif idaapi.is_double(flags):
result += "Data type: Double\n"
# result += "Value: %f\n" % idaapi.get_wide_double(address)
elif idaapi.is_strlit(flags):
result += "Data type: String\n"
result += "Value: %s\n" % idaapi.get_strlit_contents(ea)
elif idaapi.is_struct(flags):
result += "Data type: Struct\n"
# ... 其他数据类型检查
elif idaapi.is_unknown(flags):
result += "Address 0x%X is unknown.\n" % ea
# 名称和注释
if idaapi.has_name(flags):
result += "Name: %s\n" % idaapi.get_name(ea)
elif idaapi.has_dummy_name(flags):
result += "Dummy name: %s\n" % idaapi.get_name(ea)
if idaapi.has_cmt(flags):
result += "Comment: %s\n" % idaapi.get_cmt(ea, 0)
if result == '':
result = 'Address not found.'
elif result[-1] == '\n':
result = result[:-1]
return result
def _run(self, query: str, run_manager: Optional[CallbackManagerForToolRun] = None) -> Any: | query = core.escape_agent_input( | 0 | 2023-11-02 14:23:11+00:00 | 2k |
WSH032/fastapi-proxy-lib | tests/test_http.py | [
{
"identifier": "AppFactoryFixture",
"path": "tests/conftest.py",
"snippet": "_P = ParamSpec(\"_P\")\nclass LifeAppDataclass4Test(AppDataclass4Test):\nclass UvicornServerFixture(Protocol): # noqa: D101\n def __call__( # noqa: D102\n self, config: uvicorn.Config, contx_exit_timeout: Union[int... | import httpx
import pytest
from fastapi_proxy_lib.core.tool import default_proxy_filter
from typing_extensions import override
from .conftest import AppFactoryFixture, LifeAppDataclass4Test
from .tool import (
DEFAULT_URL,
PRIVATE_IP_URL,
WRONG_PROTO_URL,
AbstractTestProxy,
Tool4TestFixture,
check_if_err_resp_is_from_px_serv,
) | 1,070 | # noqa: D100
DEFAULT_TARGET_SERVER_BASE_URL = "http://www.echo.com/"
DEFAULT_PROXY_SERVER_BASE_URL = "http://www.proxy.com/"
class TestReverseHttpProxy(AbstractTestProxy):
"""For testing reverse http proxy."""
@override
@pytest.fixture()
async def tool_4_test_fixture( # pyright: ignore[reportIncompatibleMethodOverride]
self,
echo_http_test_model: LifeAppDataclass4Test,
| # noqa: D100
DEFAULT_TARGET_SERVER_BASE_URL = "http://www.echo.com/"
DEFAULT_PROXY_SERVER_BASE_URL = "http://www.proxy.com/"
class TestReverseHttpProxy(AbstractTestProxy):
"""For testing reverse http proxy."""
@override
@pytest.fixture()
async def tool_4_test_fixture( # pyright: ignore[reportIncompatibleMethodOverride]
self,
echo_http_test_model: LifeAppDataclass4Test, | reverse_http_app_fct: AppFactoryFixture, | 0 | 2023-11-08 04:38:36+00:00 | 2k |
simorxb/PID-Controller-Python | main.py | [
{
"identifier": "Car",
"path": "lib.py",
"snippet": "class Car:\n\n \"\"\" This class represents a car moving in 1D, subject to a throttle force F, with mass m, \n aerodynamic drag coefficient b, F_max/F_min forces, and time step T. \n \"\"\"\n\n def __init__(self, m, b, F_max_0, F_max_m... | import numpy as np
import matplotlib.pyplot as plt
from lib import Car, PID | 1,348 |
def main():
# -------- Configuration --------
# Simulation parameters
time_step = 0.1
end_time = 25
length = round(end_time/time_step)
t = np.zeros(length)
stp = np.zeros(length)
v = np.zeros(length)
command = np.zeros(length)
# Car parameters
m = 2140
b = 0.33
F_max_0 = 22000
F_max_max = 1710
v_max = 72
# PID parameters
Kp = 800.0
Ki = 70.0
Kaw = 1.0
Kd = 20.0
T_C = 1.0
# Initialize PID controller
|
def main():
# -------- Configuration --------
# Simulation parameters
time_step = 0.1
end_time = 25
length = round(end_time/time_step)
t = np.zeros(length)
stp = np.zeros(length)
v = np.zeros(length)
command = np.zeros(length)
# Car parameters
m = 2140
b = 0.33
F_max_0 = 22000
F_max_max = 1710
v_max = 72
# PID parameters
Kp = 800.0
Ki = 70.0
Kaw = 1.0
Kd = 20.0
T_C = 1.0
# Initialize PID controller | pid = PID(Kp, Ki, Kd, Kaw, T_C, time_step, F_max_0, 0, 30000) | 1 | 2023-11-03 19:38:34+00:00 | 2k |
aws-samples/amazon-location-geospatial-agent | geospatial_agent/agent/geospatial/planner/planner.py | [
{
"identifier": "_graph_generation_instructions",
"path": "geospatial_agent/agent/geospatial/planner/prompts.py",
"snippet": ""
},
{
"identifier": "GIS_AGENT_ROLE_INTRO",
"path": "geospatial_agent/shared/prompts.py",
"snippet": "GIS_AGENT_ROLE_INTRO = r'You are a geospatial data scienti... | import time
from langchain import PromptTemplate, LLMChain
from langchain.llms.base import LLM
from geospatial_agent.agent.geospatial.planner.prompts import _graph_generation_instructions, \
_graph_reply_example, _task_name_generation_prompt, _graph_requirement_list, \
_planning_graph_task_prompt_template
from geospatial_agent.shared.prompts import GIS_AGENT_ROLE_INTRO, HUMAN_STOP_SEQUENCE
from geospatial_agent.shared.utils import extract_code | 730 |
class PlannerException(Exception):
def __init__(self, message: str):
self.message = message
super().__init__(self.message)
def gen_task_name(llm: LLM, task: str) -> str:
"""Returns a task name for creating unix folders from task description using LLM"""
task_name_gen_prompt_template: PromptTemplate = PromptTemplate.from_template(_task_name_generation_prompt)
task_name_gen_prompt = task_name_gen_prompt_template.format(human_role="Human",
assistant_role="Assistant",
task_definition=task)
task_name = llm.predict(text=task_name_gen_prompt, stop=[HUMAN_STOP_SEQUENCE]).strip()
task_name = f'{int(time.time())}_{task_name}'
return task_name
def gen_plan_graph(llm: LLM, task_definition: str, data_locations_instructions: str) -> str:
"""Returns a plan graph in the form of python code from a task definition."""
try:
graph_plan_code = _gen_plan_graph_code(llm, task_definition, data_locations_instructions)
return graph_plan_code
except Exception as e:
raise PlannerException(f"Failed to generate graph plan code for task") from e
def _gen_plan_graph_code(llm: LLM, task_definition: str, data_locations_instructions: str):
# Generating a graph plan python code using the LLM.
graph_requirements = _get_graph_requirements()
graph_gen_prompt_template: PromptTemplate = PromptTemplate.from_template(_planning_graph_task_prompt_template)
chain = LLMChain(llm=llm, prompt=graph_gen_prompt_template)
graph_plan_response = chain.run(human_role="Human",
planner_role_intro=GIS_AGENT_ROLE_INTRO,
graph_generation_instructions=_graph_generation_instructions,
task_definition=task_definition.strip("\n").strip(),
graph_requirements=graph_requirements,
graph_reply_example=_graph_reply_example,
data_locations_instructions=data_locations_instructions,
assistant_role="Assistant",
stop=[HUMAN_STOP_SEQUENCE])
# Use the LLM to generate a plan graph code
|
class PlannerException(Exception):
def __init__(self, message: str):
self.message = message
super().__init__(self.message)
def gen_task_name(llm: LLM, task: str) -> str:
"""Returns a task name for creating unix folders from task description using LLM"""
task_name_gen_prompt_template: PromptTemplate = PromptTemplate.from_template(_task_name_generation_prompt)
task_name_gen_prompt = task_name_gen_prompt_template.format(human_role="Human",
assistant_role="Assistant",
task_definition=task)
task_name = llm.predict(text=task_name_gen_prompt, stop=[HUMAN_STOP_SEQUENCE]).strip()
task_name = f'{int(time.time())}_{task_name}'
return task_name
def gen_plan_graph(llm: LLM, task_definition: str, data_locations_instructions: str) -> str:
"""Returns a plan graph in the form of python code from a task definition."""
try:
graph_plan_code = _gen_plan_graph_code(llm, task_definition, data_locations_instructions)
return graph_plan_code
except Exception as e:
raise PlannerException(f"Failed to generate graph plan code for task") from e
def _gen_plan_graph_code(llm: LLM, task_definition: str, data_locations_instructions: str):
# Generating a graph plan python code using the LLM.
graph_requirements = _get_graph_requirements()
graph_gen_prompt_template: PromptTemplate = PromptTemplate.from_template(_planning_graph_task_prompt_template)
chain = LLMChain(llm=llm, prompt=graph_gen_prompt_template)
graph_plan_response = chain.run(human_role="Human",
planner_role_intro=GIS_AGENT_ROLE_INTRO,
graph_generation_instructions=_graph_generation_instructions,
task_definition=task_definition.strip("\n").strip(),
graph_requirements=graph_requirements,
graph_reply_example=_graph_reply_example,
data_locations_instructions=data_locations_instructions,
assistant_role="Assistant",
stop=[HUMAN_STOP_SEQUENCE])
# Use the LLM to generate a plan graph code | graph_plan_code = extract_code(graph_plan_response) | 3 | 2023-11-09 18:29:25+00:00 | 2k |
Hojagulyyev/rp2 | apps/diaries/interactors.py | [
{
"identifier": "COMMIT_MIN_LENGTH",
"path": "rp2/business_logic.py",
"snippet": "COMMIT_MIN_LENGTH = 10"
},
{
"identifier": "Diary",
"path": "apps/diaries/models.py",
"snippet": "class Diary(models.Model):\n\n account = models.ForeignKey(Account, on_delete=models.CASCADE, related_nam... | import datetime
from django.contrib.auth.decorators import login_required
from django.shortcuts import redirect
from django.contrib import messages
from django.urls import reverse
from rp2.business_logic import COMMIT_MIN_LENGTH
from .models import Diary, DiaryCommit, DiaryComment
from .signals import diary_commit_created | 728 |
@login_required
def create_commit(request, diary_id: int):
# ===== DTO
message = request.POST.get("message", "").strip()
diary = Diary.objects.get(id=diary_id)
# ===== VALIDATION
if diary.account != request.user.account:
messages.error(request, f"others' diaries are readonly")
return redirect(
f"{reverse('diaries:detail_view', kwargs={'id': diary.id})}"
)
if len(message) < COMMIT_MIN_LENGTH:
messages.error(request, f"message length is less than {COMMIT_MIN_LENGTH} character")
return redirect(
f"{reverse('diaries:detail_view', kwargs={'id': diary.id})}"
f"?message={message}"
)
if (
DiaryCommit.objects
.filter(
diary=diary,
message=message,
created_datetime__date=datetime.date.today()
)
.exists()
):
messages.error(request, message="this message already exists for today")
return redirect(
f"{reverse(viewname='diaries:detail_view', kwargs={'id': diary.id})}"
f"?message={message}"
)
# ===== PROCESS
diary_commit = DiaryCommit()
diary_commit.diary = diary
diary_commit.message = message
diary_commit.save()
|
@login_required
def create_commit(request, diary_id: int):
# ===== DTO
message = request.POST.get("message", "").strip()
diary = Diary.objects.get(id=diary_id)
# ===== VALIDATION
if diary.account != request.user.account:
messages.error(request, f"others' diaries are readonly")
return redirect(
f"{reverse('diaries:detail_view', kwargs={'id': diary.id})}"
)
if len(message) < COMMIT_MIN_LENGTH:
messages.error(request, f"message length is less than {COMMIT_MIN_LENGTH} character")
return redirect(
f"{reverse('diaries:detail_view', kwargs={'id': diary.id})}"
f"?message={message}"
)
if (
DiaryCommit.objects
.filter(
diary=diary,
message=message,
created_datetime__date=datetime.date.today()
)
.exists()
):
messages.error(request, message="this message already exists for today")
return redirect(
f"{reverse(viewname='diaries:detail_view', kwargs={'id': diary.id})}"
f"?message={message}"
)
# ===== PROCESS
diary_commit = DiaryCommit()
diary_commit.diary = diary
diary_commit.message = message
diary_commit.save()
| diary_commit_created.send(sender=DiaryCommit, instance=diary_commit) | 4 | 2023-11-05 07:57:17+00:00 | 2k |
soobin419/DWT | basicsr/utils/logger.py | [
{
"identifier": "get_dist_info",
"path": "basicsr/utils/dist_util.py",
"snippet": "def get_dist_info():\n if dist.is_available():\n initialized = dist.is_initialized()\n else:\n initialized = False\n if initialized:\n rank = dist.get_rank()\n world_size = dist.get_wo... | import datetime
import logging
import time
import wandb
import torch
import torchvision
from .dist_util import get_dist_info, master_only
from torch.utils.tensorboard import SummaryWriter
from version import __version__ | 1,528 | train (dict): Contains 'total_iter' (int) for total iters.
use_tb_logger (bool): Use tensorboard logger.
start_iter (int): Start iter. Default: 1.
tb_logger (obj:`tb_logger`): Tensorboard logger. Default: None.
"""
def __init__(self, opt, start_iter=1, tb_logger=None):
self.exp_name = opt['name']
self.interval = opt['logger']['print_freq']
self.start_iter = start_iter
self.max_iters = opt['train']['total_iter']
self.use_tb_logger = opt['logger']['use_tb_logger']
self.tb_logger = tb_logger
self.start_time = time.time()
self.logger = get_root_logger()
def reset_start_time(self):
self.start_time = time.time()
@master_only
def __call__(self, log_vars):
"""Format logging message.
Args:
log_vars (dict): It contains the following keys:
epoch (int): Epoch number.
iter (int): Current iter.
lrs (list): List for learning rates.
time (float): Iter time.
data_time (float): Data time for each iter.
"""
# epoch, iter, learning rates
epoch = log_vars.pop('epoch')
current_iter = log_vars.pop('iter')
lrs = log_vars.pop('lrs')
message = (f'[{self.exp_name[:5]}..][epoch:{epoch:3d}, iter:{current_iter:8,d}, lr:(')
for v in lrs:
message += f'{v:.3e},'
message += ')] '
# time and estimated time
if 'time' in log_vars.keys():
iter_time = log_vars.pop('time')
data_time = log_vars.pop('data_time')
total_time = time.time() - self.start_time
time_sec_avg = total_time / (current_iter - self.start_iter + 1)
eta_sec = time_sec_avg * (self.max_iters - current_iter - 1)
eta_str = str(datetime.timedelta(seconds=int(eta_sec)))
message += f'[eta: {eta_str}, '
message += f'time (data): {iter_time:.3f} ({data_time:.3f})] '
# other items, especially losses
for k, v in log_vars.items():
message += f'{k}: {v:.4e} '
# tensorboard logger
if self.use_tb_logger and 'debug' not in self.exp_name:
if k.startswith('l_'):
self.tb_logger.add_scalar(f'losses/{k}', v, current_iter)
else:
self.tb_logger.add_scalar(k, v, current_iter)
self.logger.info(message)
@master_only
def init_tb_logger(log_dir):
tb_logger = SummaryWriter(log_dir=log_dir)
return tb_logger
@master_only
def init_wandb_logger(opt):
"""We now only use wandb to sync tensorboard log."""
logger = get_root_logger()
project = opt['logger']['wandb']['project']
resume_id = opt['logger']['wandb'].get('resume_id')
if resume_id:
wandb_id = resume_id
resume = 'allow'
logger.warning(f'Resume wandb logger with id={wandb_id}.')
else:
wandb_id = wandb.util.generate_id()
resume = 'never'
wandb.init(id=wandb_id, resume=resume, name=opt['name'], config=opt, project=project, sync_tensorboard=True)
logger.info(f'Use wandb logger with id={wandb_id}; project={project}.')
def get_root_logger(logger_name='basicsr', log_level=logging.INFO, log_file=None):
"""Get the root logger.
The logger will be initialized if it has not been initialized. By default a
StreamHandler will be added. If `log_file` is specified, a FileHandler will
also be added.
Args:
logger_name (str): root logger name. Default: 'basicsr'.
log_file (str | None): The log filename. If specified, a FileHandler
will be added to the root logger.
log_level (int): The root logger level. Note that only the process of
rank 0 is affected, while other processes will set the level to
"Error" and be silent most of the time.
Returns:
logging.Logger: The root logger.
"""
logger = logging.getLogger(logger_name)
# if the logger has been initialized, just return it
if logger_name in initialized_logger:
return logger
format_str = '%(asctime)s %(levelname)s: %(message)s'
stream_handler = logging.StreamHandler()
stream_handler.setFormatter(logging.Formatter(format_str))
logger.addHandler(stream_handler)
logger.propagate = False
|
initialized_logger = {}
class AvgTimer():
def __init__(self, window=200):
self.window = window # average window
self.current_time = 0
self.total_time = 0
self.count = 0
self.avg_time = 0
self.start()
def start(self):
self.start_time = self.tic = time.time()
def record(self):
self.count += 1
self.toc = time.time()
self.current_time = self.toc - self.tic
self.total_time += self.current_time
# calculate average time
self.avg_time = self.total_time / self.count
# reset
if self.count > self.window:
self.count = 0
self.total_time = 0
self.tic = time.time()
def get_current_time(self):
return self.current_time
def get_avg_time(self):
return self.avg_time
class MessageLogger():
"""Message logger for printing.
Args:
opt (dict): Config. It contains the following keys:
name (str): Exp name.
logger (dict): Contains 'print_freq' (str) for logger interval.
train (dict): Contains 'total_iter' (int) for total iters.
use_tb_logger (bool): Use tensorboard logger.
start_iter (int): Start iter. Default: 1.
tb_logger (obj:`tb_logger`): Tensorboard logger. Default: None.
"""
def __init__(self, opt, start_iter=1, tb_logger=None):
self.exp_name = opt['name']
self.interval = opt['logger']['print_freq']
self.start_iter = start_iter
self.max_iters = opt['train']['total_iter']
self.use_tb_logger = opt['logger']['use_tb_logger']
self.tb_logger = tb_logger
self.start_time = time.time()
self.logger = get_root_logger()
def reset_start_time(self):
self.start_time = time.time()
@master_only
def __call__(self, log_vars):
"""Format logging message.
Args:
log_vars (dict): It contains the following keys:
epoch (int): Epoch number.
iter (int): Current iter.
lrs (list): List for learning rates.
time (float): Iter time.
data_time (float): Data time for each iter.
"""
# epoch, iter, learning rates
epoch = log_vars.pop('epoch')
current_iter = log_vars.pop('iter')
lrs = log_vars.pop('lrs')
message = (f'[{self.exp_name[:5]}..][epoch:{epoch:3d}, iter:{current_iter:8,d}, lr:(')
for v in lrs:
message += f'{v:.3e},'
message += ')] '
# time and estimated time
if 'time' in log_vars.keys():
iter_time = log_vars.pop('time')
data_time = log_vars.pop('data_time')
total_time = time.time() - self.start_time
time_sec_avg = total_time / (current_iter - self.start_iter + 1)
eta_sec = time_sec_avg * (self.max_iters - current_iter - 1)
eta_str = str(datetime.timedelta(seconds=int(eta_sec)))
message += f'[eta: {eta_str}, '
message += f'time (data): {iter_time:.3f} ({data_time:.3f})] '
# other items, especially losses
for k, v in log_vars.items():
message += f'{k}: {v:.4e} '
# tensorboard logger
if self.use_tb_logger and 'debug' not in self.exp_name:
if k.startswith('l_'):
self.tb_logger.add_scalar(f'losses/{k}', v, current_iter)
else:
self.tb_logger.add_scalar(k, v, current_iter)
self.logger.info(message)
@master_only
def init_tb_logger(log_dir):
tb_logger = SummaryWriter(log_dir=log_dir)
return tb_logger
@master_only
def init_wandb_logger(opt):
"""We now only use wandb to sync tensorboard log."""
logger = get_root_logger()
project = opt['logger']['wandb']['project']
resume_id = opt['logger']['wandb'].get('resume_id')
if resume_id:
wandb_id = resume_id
resume = 'allow'
logger.warning(f'Resume wandb logger with id={wandb_id}.')
else:
wandb_id = wandb.util.generate_id()
resume = 'never'
wandb.init(id=wandb_id, resume=resume, name=opt['name'], config=opt, project=project, sync_tensorboard=True)
logger.info(f'Use wandb logger with id={wandb_id}; project={project}.')
def get_root_logger(logger_name='basicsr', log_level=logging.INFO, log_file=None):
"""Get the root logger.
The logger will be initialized if it has not been initialized. By default a
StreamHandler will be added. If `log_file` is specified, a FileHandler will
also be added.
Args:
logger_name (str): root logger name. Default: 'basicsr'.
log_file (str | None): The log filename. If specified, a FileHandler
will be added to the root logger.
log_level (int): The root logger level. Note that only the process of
rank 0 is affected, while other processes will set the level to
"Error" and be silent most of the time.
Returns:
logging.Logger: The root logger.
"""
logger = logging.getLogger(logger_name)
# if the logger has been initialized, just return it
if logger_name in initialized_logger:
return logger
format_str = '%(asctime)s %(levelname)s: %(message)s'
stream_handler = logging.StreamHandler()
stream_handler.setFormatter(logging.Formatter(format_str))
logger.addHandler(stream_handler)
logger.propagate = False | rank, _ = get_dist_info() | 0 | 2023-11-09 08:08:09+00:00 | 2k |
Rishit-dagli/Astroformer | pytorch-image-models/timm/layers/create_norm.py | [
{
"identifier": "GroupNorm",
"path": "pytorch-image-models/timm/layers/norm.py",
"snippet": "class GroupNorm(nn.GroupNorm):\n def __init__(self, num_channels, num_groups=32, eps=1e-5, affine=True):\n # NOTE num_channels is swapped to first arg for consistency in swapping norm layers with BN\n ... | import functools
import types
import torch.nn as nn
from typing import Type
from .norm import GroupNorm, GroupNorm1, LayerNorm, LayerNorm2d, RmsNorm
from torchvision.ops.misc import FrozenBatchNorm2d | 1,331 | """ Norm Layer Factory
Create norm modules by string (to mirror create_act and creat_norm-act fns)
Copyright 2022 Ross Wightman
"""
_NORM_MAP = dict(
batchnorm=nn.BatchNorm2d,
batchnorm2d=nn.BatchNorm2d,
batchnorm1d=nn.BatchNorm1d,
groupnorm=GroupNorm,
groupnorm1=GroupNorm1,
layernorm=LayerNorm,
layernorm2d=LayerNorm2d,
| """ Norm Layer Factory
Create norm modules by string (to mirror create_act and creat_norm-act fns)
Copyright 2022 Ross Wightman
"""
_NORM_MAP = dict(
batchnorm=nn.BatchNorm2d,
batchnorm2d=nn.BatchNorm2d,
batchnorm1d=nn.BatchNorm1d,
groupnorm=GroupNorm,
groupnorm1=GroupNorm1,
layernorm=LayerNorm,
layernorm2d=LayerNorm2d, | rmsnorm=RmsNorm, | 4 | 2023-11-05 01:25:14+00:00 | 2k |
dewgenenny/rtl_433_discoverandsubmit | rtl_433_discoverandsubmit/modules/mqtt_client.py | [
{
"identifier": "config",
"path": "rtl_433_discoverandsubmit/config.py",
"snippet": ""
},
{
"identifier": "save_devices_to_file",
"path": "rtl_433_discoverandsubmit/modules/device_manager.py",
"snippet": "def save_devices_to_file(devices):\n \"\"\"Save the list of devices to a JSON fi... | import paho.mqtt.client as mqtt
import json
import logging
from datetime import datetime
from rtl_433_discoverandsubmit import config
from rtl_433_discoverandsubmit.modules.device_manager import save_devices_to_file | 704 | log_level = getattr(logging, config.configuration['log_level'])
logging.basicConfig(filename=config.configuration['log_filename'], level=log_level)
# List to store detected devices
detected_devices = []
def reset_message_counters():
global detected_devices
for device in detected_devices:
if 'message_count' in device:
device['message_count'] = 0
def sort_detected_devices():
global detected_devices
criteria = config.configuration['current_sort_criteria']
reverse = True if criteria in ["last_detected_time", "message_count"] else False
if criteria in ["last_detected_time", "message_count"]:
detected_devices.sort(key=lambda x: x[criteria], reverse=reverse)
else:
detected_devices.sort(key=lambda x: x.get('model', '').lower())
return detected_devices
def on_connect(client, userdata, flags, rc):
"""Callback for when the client receives a CONNACK response from the server."""
client.subscribe(config.configuration['topic'])
def on_message(client, userdata, msg):
global detected_devices
payload = json.loads(msg.payload.decode())
# Safely get 'id' from payload or default to 'unknown' (fixes https://github.com/dewgenenny/rtl_433_discoverandsubmit/issues/1)
device_id_value = payload.get('id', 'unknown')
topicprefix = "/".join(msg.topic.split("/", 2)[:2])
# Construct a unique device identifier from model and id
device_id = f"{payload['model']}_{device_id_value}"
# Check if the device is already in the detected devices list
existing_device = next((device for device in detected_devices if device['id'] == device_id), None)
# If the device is new, add it to the list
if not existing_device:
device_data = {
'first_detected_time': datetime.now().strftime('%Y-%m-%d %H:%M:%S'),
'last_detected_time': datetime.now().strftime('%Y-%m-%d %H:%M:%S'),
'message_count': 1 # Initialize message count
}
device_data.update(payload) # Add all the payload attributes to the dictionary
device_data['topicprefix']= topicprefix
old_id = payload.get('id')
logging.debug("Old ID = "+ str(old_id))
device_data['original_id'] = old_id
device_data['id'] = device_id # Set the id field after updating from payload
detected_devices.append(device_data)
#save new device to file, so that it is remembered on startup
| log_level = getattr(logging, config.configuration['log_level'])
logging.basicConfig(filename=config.configuration['log_filename'], level=log_level)
# List to store detected devices
detected_devices = []
def reset_message_counters():
global detected_devices
for device in detected_devices:
if 'message_count' in device:
device['message_count'] = 0
def sort_detected_devices():
global detected_devices
criteria = config.configuration['current_sort_criteria']
reverse = True if criteria in ["last_detected_time", "message_count"] else False
if criteria in ["last_detected_time", "message_count"]:
detected_devices.sort(key=lambda x: x[criteria], reverse=reverse)
else:
detected_devices.sort(key=lambda x: x.get('model', '').lower())
return detected_devices
def on_connect(client, userdata, flags, rc):
"""Callback for when the client receives a CONNACK response from the server."""
client.subscribe(config.configuration['topic'])
def on_message(client, userdata, msg):
global detected_devices
payload = json.loads(msg.payload.decode())
# Safely get 'id' from payload or default to 'unknown' (fixes https://github.com/dewgenenny/rtl_433_discoverandsubmit/issues/1)
device_id_value = payload.get('id', 'unknown')
topicprefix = "/".join(msg.topic.split("/", 2)[:2])
# Construct a unique device identifier from model and id
device_id = f"{payload['model']}_{device_id_value}"
# Check if the device is already in the detected devices list
existing_device = next((device for device in detected_devices if device['id'] == device_id), None)
# If the device is new, add it to the list
if not existing_device:
device_data = {
'first_detected_time': datetime.now().strftime('%Y-%m-%d %H:%M:%S'),
'last_detected_time': datetime.now().strftime('%Y-%m-%d %H:%M:%S'),
'message_count': 1 # Initialize message count
}
device_data.update(payload) # Add all the payload attributes to the dictionary
device_data['topicprefix']= topicprefix
old_id = payload.get('id')
logging.debug("Old ID = "+ str(old_id))
device_data['original_id'] = old_id
device_data['id'] = device_id # Set the id field after updating from payload
detected_devices.append(device_data)
#save new device to file, so that it is remembered on startup | save_devices_to_file(detected_devices) | 1 | 2023-11-03 19:34:56+00:00 | 2k |
dvruette/pygba | src/pygba/gym_env.py | [
{
"identifier": "KEY_MAP",
"path": "src/pygba/utils.py",
"snippet": "KEY_MAP = {\n \"up\": GBA.KEY_UP,\n \"down\": GBA.KEY_DOWN,\n \"left\": GBA.KEY_LEFT,\n \"right\": GBA.KEY_RIGHT,\n \"A\": GBA.KEY_A,\n \"B\": GBA.KEY_B,\n \"L\": GBA.KEY_L,\n \"R\": GBA.KEY_R,\n \"start\": G... | import sys
import gymnasium as gym
import mgba.core
import mgba.image
import numpy as np
import pygame
from typing import Any, Literal
from .utils import KEY_MAP
from .pygba import PyGBA
from .game_wrappers.base import GameWrapper
from pygame import gfxdraw | 1,415 |
try:
except ImportError as e:
pass
def _pil_image_to_pygame(img):
return pygame.image.fromstring(img.tobytes(), img.size, img.mode).convert()
class PyGBAEnv(gym.Env):
metadata = {
"render_modes": ["human", "rgb_array"],
"render_fps": 60,
}
def __init__(
self,
gba: PyGBA,
|
try:
except ImportError as e:
pass
def _pil_image_to_pygame(img):
return pygame.image.fromstring(img.tobytes(), img.size, img.mode).convert()
class PyGBAEnv(gym.Env):
metadata = {
"render_modes": ["human", "rgb_array"],
"render_fps": 60,
}
def __init__(
self,
gba: PyGBA, | game_wrapper: GameWrapper | None = None, | 2 | 2023-11-08 20:51:13+00:00 | 2k |
BouncyKoishi/ChuCaoQi-Bot | dbConnection/draw_item.py | [
{
"identifier": "DrawItemList",
"path": "dbConnection/models.py",
"snippet": "class DrawItemList(Model):\n id = IntField(pk=True)\n name = CharField(max_length=64)\n pool = CharField(max_length=32)\n rareRank = IntField()\n detail = CharField(max_length=1024)\n author = CharField(max_l... | from random import randint
from .models import DrawItemList, DrawItemStorage
from tortoise import Tortoise
from tortoise.query_utils import Prefetch
from tortoise.functions import Sum | 646 |
async def getItem(itemId):
return await DrawItemList.filter(id=itemId).first()
async def getItemByName(itemName):
return await DrawItemList.filter(name=itemName).first()
async def getItemListByAuthor(qqNum, rareRank=None, poolName=None):
filterQuery = getRareRankAndPoolFilter(rareRank, poolName)
return await filterQuery.filter(author=qqNum).order_by("-rareRank")
async def getRandomItem(rareRank, poolName=None):
if poolName:
rareItemList = await DrawItemList.filter(rareRank=rareRank, pool=poolName)
else:
rareItemList = await DrawItemList.filter(rareRank=rareRank)
if not rareItemList:
raise Exception(f"DrawItem Error: 抽奖物品列表为空!")
return rareItemList[randint(0, len(rareItemList) - 1)]
async def searchItem(keyword, limit, offset=0):
conn = Tortoise.get_connection('default')
count = (await conn.execute_query_dict('''
SELECT
count(*) AS count
FROM
DrawItemList
WHERE
name LIKE ('%' || ? || '%')
''', [keyword,]))[0]['count']
rows = await conn.execute_query_dict('''
SELECT
name,
rareRank
FROM
DrawItemList
WHERE
name LIKE ('%' || ? || '%')
ORDER BY
rareRank DESC
LIMIT ? OFFSET ?
''', [keyword, limit, offset])
return count, rows
async def addItem(itemName, itemRare, poolName, itemDetail, author):
await DrawItemList.create(name=itemName, rareRank=itemRare, pool=poolName, detail=itemDetail, author=author)
async def deleteItem(item: DrawItemList):
await item.delete()
async def setItemDetail(item: DrawItemList, newItemDetail):
item.detail = newItemDetail
await item.save()
async def getItemsWithStorage(qqNum, rareRank=None, poolName=None):
filterQuery = getRareRankAndPoolFilter(rareRank, poolName)
return await filterQuery.order_by("-rareRank").prefetch_related(
|
async def getItem(itemId):
return await DrawItemList.filter(id=itemId).first()
async def getItemByName(itemName):
return await DrawItemList.filter(name=itemName).first()
async def getItemListByAuthor(qqNum, rareRank=None, poolName=None):
filterQuery = getRareRankAndPoolFilter(rareRank, poolName)
return await filterQuery.filter(author=qqNum).order_by("-rareRank")
async def getRandomItem(rareRank, poolName=None):
if poolName:
rareItemList = await DrawItemList.filter(rareRank=rareRank, pool=poolName)
else:
rareItemList = await DrawItemList.filter(rareRank=rareRank)
if not rareItemList:
raise Exception(f"DrawItem Error: 抽奖物品列表为空!")
return rareItemList[randint(0, len(rareItemList) - 1)]
async def searchItem(keyword, limit, offset=0):
conn = Tortoise.get_connection('default')
count = (await conn.execute_query_dict('''
SELECT
count(*) AS count
FROM
DrawItemList
WHERE
name LIKE ('%' || ? || '%')
''', [keyword,]))[0]['count']
rows = await conn.execute_query_dict('''
SELECT
name,
rareRank
FROM
DrawItemList
WHERE
name LIKE ('%' || ? || '%')
ORDER BY
rareRank DESC
LIMIT ? OFFSET ?
''', [keyword, limit, offset])
return count, rows
async def addItem(itemName, itemRare, poolName, itemDetail, author):
await DrawItemList.create(name=itemName, rareRank=itemRare, pool=poolName, detail=itemDetail, author=author)
async def deleteItem(item: DrawItemList):
await item.delete()
async def setItemDetail(item: DrawItemList, newItemDetail):
item.detail = newItemDetail
await item.save()
async def getItemsWithStorage(qqNum, rareRank=None, poolName=None):
filterQuery = getRareRankAndPoolFilter(rareRank, poolName)
return await filterQuery.order_by("-rareRank").prefetch_related( | Prefetch("draw_item_storage", queryset=DrawItemStorage.filter(qq=qqNum), to_attr="storage") | 1 | 2023-11-02 04:06:31+00:00 | 2k |
ilur98/DGQ | dgq/utils/evalutils.py | [
{
"identifier": "get_blocks",
"path": "dgq/utils/modelutils.py",
"snippet": "def get_blocks(model):\n if isinstance(model, LlamaForCausalLM):\n layers = model.model.layers\n elif isinstance(model, OPTForCausalLM):\n layers = model.model.decoder.layers\n elif isinstance(model, Bloo... | import torch
import torch.nn as nn
import numpy as np
from dgq.utils.modelutils import get_blocks, move_embed, move_norm_head
from datasets import load_dataset, load_from_disk
from tqdm import tqdm
from dgq.utils.datautils import IGNORE_INDEX, DEFAULT_PAD_TOKEN | 866 |
@torch.no_grad()
def model_eval(model, testenc, dev, local_args=None):
testenc = testenc.input_ids
nsamples = testenc.numel() // model.seqlen
# model = model.to(dev)
model.eval()
model.config.use_cache = False
# testenc = testenc.to(dev)
|
@torch.no_grad()
def model_eval(model, testenc, dev, local_args=None):
testenc = testenc.input_ids
nsamples = testenc.numel() // model.seqlen
# model = model.to(dev)
model.eval()
model.config.use_cache = False
# testenc = testenc.to(dev) | layers = get_blocks(model) | 0 | 2023-11-01 13:45:16+00:00 | 2k |
JeasunLok/ResNet-pytorch | test.py | [
{
"identifier": "AverageMeter",
"path": "utils/utils.py",
"snippet": "class AverageMeter(object):\n\n def __init__(self):\n self.reset()\n\n def reset(self):\n self.average = 0 \n self.sum = 0\n self.count = 0\n\n def update(self, val, n=1):\n self.sum += val * n\n self.count += n\n... | import numpy as np
import torch
from tqdm import tqdm
from utils.utils import AverageMeter
from utils.accuracy import accuracy | 695 |
# test model
def test_epoch(model, test_loader, device):
acc1 = AverageMeter()
acc3 = AverageMeter()
prediction = np.array([])
label = np.array([])
loop = tqdm(enumerate(test_loader), total = len(test_loader))
with torch.no_grad():
for batch_idx, (batch_data, batch_label) in loop:
batch_data = batch_data.to(device).float()
batch_label = batch_label.to(device)
batch_prediction = model(batch_data)
# calculate the accuracy
|
# test model
def test_epoch(model, test_loader, device):
acc1 = AverageMeter()
acc3 = AverageMeter()
prediction = np.array([])
label = np.array([])
loop = tqdm(enumerate(test_loader), total = len(test_loader))
with torch.no_grad():
for batch_idx, (batch_data, batch_label) in loop:
batch_data = batch_data.to(device).float()
batch_label = batch_label.to(device)
batch_prediction = model(batch_data)
# calculate the accuracy | acc_batch = accuracy(batch_prediction, batch_label, topk=(1,3)) | 1 | 2023-11-02 06:26:47+00:00 | 2k |
soobin419/EDAT | basicsr/utils/logger.py | [
{
"identifier": "get_dist_info",
"path": "basicsr/utils/dist_util.py",
"snippet": "def get_dist_info():\n if dist.is_available():\n initialized = dist.is_initialized()\n else:\n initialized = False\n if initialized:\n rank = dist.get_rank()\n world_size = dist.get_wo... | import datetime
import logging
import time
import wandb
import torch
import torchvision
from .dist_util import get_dist_info, master_only
from torch.utils.tensorboard import SummaryWriter
from version import __version__ | 1,528 | train (dict): Contains 'total_iter' (int) for total iters.
use_tb_logger (bool): Use tensorboard logger.
start_iter (int): Start iter. Default: 1.
tb_logger (obj:`tb_logger`): Tensorboard logger. Default: None.
"""
def __init__(self, opt, start_iter=1, tb_logger=None):
self.exp_name = opt['name']
self.interval = opt['logger']['print_freq']
self.start_iter = start_iter
self.max_iters = opt['train']['total_iter']
self.use_tb_logger = opt['logger']['use_tb_logger']
self.tb_logger = tb_logger
self.start_time = time.time()
self.logger = get_root_logger()
def reset_start_time(self):
self.start_time = time.time()
@master_only
def __call__(self, log_vars):
"""Format logging message.
Args:
log_vars (dict): It contains the following keys:
epoch (int): Epoch number.
iter (int): Current iter.
lrs (list): List for learning rates.
time (float): Iter time.
data_time (float): Data time for each iter.
"""
# epoch, iter, learning rates
epoch = log_vars.pop('epoch')
current_iter = log_vars.pop('iter')
lrs = log_vars.pop('lrs')
message = (f'[{self.exp_name[:5]}..][epoch:{epoch:3d}, iter:{current_iter:8,d}, lr:(')
for v in lrs:
message += f'{v:.3e},'
message += ')] '
# time and estimated time
if 'time' in log_vars.keys():
iter_time = log_vars.pop('time')
data_time = log_vars.pop('data_time')
total_time = time.time() - self.start_time
time_sec_avg = total_time / (current_iter - self.start_iter + 1)
eta_sec = time_sec_avg * (self.max_iters - current_iter - 1)
eta_str = str(datetime.timedelta(seconds=int(eta_sec)))
message += f'[eta: {eta_str}, '
message += f'time (data): {iter_time:.3f} ({data_time:.3f})] '
# other items, especially losses
for k, v in log_vars.items():
message += f'{k}: {v:.4e} '
# tensorboard logger
if self.use_tb_logger and 'debug' not in self.exp_name:
if k.startswith('l_'):
self.tb_logger.add_scalar(f'losses/{k}', v, current_iter)
else:
self.tb_logger.add_scalar(k, v, current_iter)
self.logger.info(message)
@master_only
def init_tb_logger(log_dir):
tb_logger = SummaryWriter(log_dir=log_dir)
return tb_logger
@master_only
def init_wandb_logger(opt):
"""We now only use wandb to sync tensorboard log."""
logger = get_root_logger()
project = opt['logger']['wandb']['project']
resume_id = opt['logger']['wandb'].get('resume_id')
if resume_id:
wandb_id = resume_id
resume = 'allow'
logger.warning(f'Resume wandb logger with id={wandb_id}.')
else:
wandb_id = wandb.util.generate_id()
resume = 'never'
wandb.init(id=wandb_id, resume=resume, name=opt['name'], config=opt, project=project, sync_tensorboard=True)
logger.info(f'Use wandb logger with id={wandb_id}; project={project}.')
def get_root_logger(logger_name='basicsr', log_level=logging.INFO, log_file=None):
"""Get the root logger.
The logger will be initialized if it has not been initialized. By default a
StreamHandler will be added. If `log_file` is specified, a FileHandler will
also be added.
Args:
logger_name (str): root logger name. Default: 'basicsr'.
log_file (str | None): The log filename. If specified, a FileHandler
will be added to the root logger.
log_level (int): The root logger level. Note that only the process of
rank 0 is affected, while other processes will set the level to
"Error" and be silent most of the time.
Returns:
logging.Logger: The root logger.
"""
logger = logging.getLogger(logger_name)
# if the logger has been initialized, just return it
if logger_name in initialized_logger:
return logger
format_str = '%(asctime)s %(levelname)s: %(message)s'
stream_handler = logging.StreamHandler()
stream_handler.setFormatter(logging.Formatter(format_str))
logger.addHandler(stream_handler)
logger.propagate = False
|
initialized_logger = {}
class AvgTimer():
def __init__(self, window=200):
self.window = window # average window
self.current_time = 0
self.total_time = 0
self.count = 0
self.avg_time = 0
self.start()
def start(self):
self.start_time = self.tic = time.time()
def record(self):
self.count += 1
self.toc = time.time()
self.current_time = self.toc - self.tic
self.total_time += self.current_time
# calculate average time
self.avg_time = self.total_time / self.count
# reset
if self.count > self.window:
self.count = 0
self.total_time = 0
self.tic = time.time()
def get_current_time(self):
return self.current_time
def get_avg_time(self):
return self.avg_time
class MessageLogger():
"""Message logger for printing.
Args:
opt (dict): Config. It contains the following keys:
name (str): Exp name.
logger (dict): Contains 'print_freq' (str) for logger interval.
train (dict): Contains 'total_iter' (int) for total iters.
use_tb_logger (bool): Use tensorboard logger.
start_iter (int): Start iter. Default: 1.
tb_logger (obj:`tb_logger`): Tensorboard logger. Default: None.
"""
def __init__(self, opt, start_iter=1, tb_logger=None):
self.exp_name = opt['name']
self.interval = opt['logger']['print_freq']
self.start_iter = start_iter
self.max_iters = opt['train']['total_iter']
self.use_tb_logger = opt['logger']['use_tb_logger']
self.tb_logger = tb_logger
self.start_time = time.time()
self.logger = get_root_logger()
def reset_start_time(self):
self.start_time = time.time()
@master_only
def __call__(self, log_vars):
"""Format logging message.
Args:
log_vars (dict): It contains the following keys:
epoch (int): Epoch number.
iter (int): Current iter.
lrs (list): List for learning rates.
time (float): Iter time.
data_time (float): Data time for each iter.
"""
# epoch, iter, learning rates
epoch = log_vars.pop('epoch')
current_iter = log_vars.pop('iter')
lrs = log_vars.pop('lrs')
message = (f'[{self.exp_name[:5]}..][epoch:{epoch:3d}, iter:{current_iter:8,d}, lr:(')
for v in lrs:
message += f'{v:.3e},'
message += ')] '
# time and estimated time
if 'time' in log_vars.keys():
iter_time = log_vars.pop('time')
data_time = log_vars.pop('data_time')
total_time = time.time() - self.start_time
time_sec_avg = total_time / (current_iter - self.start_iter + 1)
eta_sec = time_sec_avg * (self.max_iters - current_iter - 1)
eta_str = str(datetime.timedelta(seconds=int(eta_sec)))
message += f'[eta: {eta_str}, '
message += f'time (data): {iter_time:.3f} ({data_time:.3f})] '
# other items, especially losses
for k, v in log_vars.items():
message += f'{k}: {v:.4e} '
# tensorboard logger
if self.use_tb_logger and 'debug' not in self.exp_name:
if k.startswith('l_'):
self.tb_logger.add_scalar(f'losses/{k}', v, current_iter)
else:
self.tb_logger.add_scalar(k, v, current_iter)
self.logger.info(message)
@master_only
def init_tb_logger(log_dir):
tb_logger = SummaryWriter(log_dir=log_dir)
return tb_logger
@master_only
def init_wandb_logger(opt):
"""We now only use wandb to sync tensorboard log."""
logger = get_root_logger()
project = opt['logger']['wandb']['project']
resume_id = opt['logger']['wandb'].get('resume_id')
if resume_id:
wandb_id = resume_id
resume = 'allow'
logger.warning(f'Resume wandb logger with id={wandb_id}.')
else:
wandb_id = wandb.util.generate_id()
resume = 'never'
wandb.init(id=wandb_id, resume=resume, name=opt['name'], config=opt, project=project, sync_tensorboard=True)
logger.info(f'Use wandb logger with id={wandb_id}; project={project}.')
def get_root_logger(logger_name='basicsr', log_level=logging.INFO, log_file=None):
"""Get the root logger.
The logger will be initialized if it has not been initialized. By default a
StreamHandler will be added. If `log_file` is specified, a FileHandler will
also be added.
Args:
logger_name (str): root logger name. Default: 'basicsr'.
log_file (str | None): The log filename. If specified, a FileHandler
will be added to the root logger.
log_level (int): The root logger level. Note that only the process of
rank 0 is affected, while other processes will set the level to
"Error" and be silent most of the time.
Returns:
logging.Logger: The root logger.
"""
logger = logging.getLogger(logger_name)
# if the logger has been initialized, just return it
if logger_name in initialized_logger:
return logger
format_str = '%(asctime)s %(levelname)s: %(message)s'
stream_handler = logging.StreamHandler()
stream_handler.setFormatter(logging.Formatter(format_str))
logger.addHandler(stream_handler)
logger.propagate = False | rank, _ = get_dist_info() | 0 | 2023-11-09 08:53:40+00:00 | 2k |
noco-ai/elemental-golem | modules/noco-ai/bark-tts/handler.py | [
{
"identifier": "BaseHandler",
"path": "application/base_handler.py",
"snippet": "class BaseHandler:\n\n def __init__(self):\n self.cached_schemas = {}\n\n def execute(self, model, request) -> dict:\n raise NotImplementedError(\"The `execute` method should be implemented in the deriv... | from application.base_handler import BaseHandler
from transformers import AutoProcessor, AutoModel
from io import BytesIO
from application.progress_streamer import ProgressStreamer
import torch
import base64
import scipy
import copy
import logging | 1,247 |
logger = logging.getLogger(__name__)
class BarkHandler(BaseHandler):
def __init__(self):
|
logger = logging.getLogger(__name__)
class BarkHandler(BaseHandler):
def __init__(self): | self.progress_streamer = ProgressStreamer() | 1 | 2023-11-06 19:03:07+00:00 | 2k |
anilaltuner/personalized-news-agent | pages/chatbot.py | [
{
"identifier": "CUSTOM_ALGO_ID",
"path": "news.py",
"snippet": "CUSTOM_ALGO_ID = st.secrets[\"custom_algo_id\"]"
},
{
"identifier": "initialize_session",
"path": "news.py",
"snippet": "def initialize_session(user_input=\"\"):\n \"\"\"Initialize or restart the session.\"\"\"\n if u... | import streamlit as st
from firstbatch import AlgorithmLabel
from pydantic import BaseModel
from news import CUSTOM_ALGO_ID, initialize_session, fetch_content
from chat_tools.kernel import chat, setup_chat_with_memory
from markdowns.markdowns_chat import css_, sidebar | 1,313 |
# Pydantic models
class SessionData(BaseModel):
username: str
class PersonalizeData(BaseModel):
message: str
class SignalData(BaseModel):
sessionID: dict
id: str
def get_user_input():
return st.sidebar.text_input("Username/Session Name", st.session_state.get("username", ""))
def update_session_state(user_input):
st.session_state.session = st.session_state.personalized.session(
AlgorithmLabel.CUSTOM, vdbid="rss_db", custom_id=CUSTOM_ALGO_ID
)
st.session_state.batches = []
st.session_state.ids = []
st.session_state.likes = []
ids, batch = st.session_state.personalized.batch(st.session_state.session)
st.session_state.batches += batch
st.session_state.ids += ids
st.session_state.username = user_input
st.session_state.html_content = """
<div class="chat-container">
<div class="chat-box">
<div class="chat-output" id="chat-output"></div>
</div>
</div>
"""
st.session_state.chat_history = ""
st.session_state.chat_loader = 3
def display_sidebar():
user_input = get_user_input()
if user_input and st.session_state.get("username") != user_input:
update_session_state(user_input)
initialize_session(user_input)
|
# Pydantic models
class SessionData(BaseModel):
username: str
class PersonalizeData(BaseModel):
message: str
class SignalData(BaseModel):
sessionID: dict
id: str
def get_user_input():
return st.sidebar.text_input("Username/Session Name", st.session_state.get("username", ""))
def update_session_state(user_input):
st.session_state.session = st.session_state.personalized.session(
AlgorithmLabel.CUSTOM, vdbid="rss_db", custom_id=CUSTOM_ALGO_ID
)
st.session_state.batches = []
st.session_state.ids = []
st.session_state.likes = []
ids, batch = st.session_state.personalized.batch(st.session_state.session)
st.session_state.batches += batch
st.session_state.ids += ids
st.session_state.username = user_input
st.session_state.html_content = """
<div class="chat-container">
<div class="chat-box">
<div class="chat-output" id="chat-output"></div>
</div>
</div>
"""
st.session_state.chat_history = ""
st.session_state.chat_loader = 3
def display_sidebar():
user_input = get_user_input()
if user_input and st.session_state.get("username") != user_input:
update_session_state(user_input)
initialize_session(user_input) | fetch_content() | 2 | 2023-11-07 12:51:01+00:00 | 2k |
m4rkw/monzo-utils | monzo_utils/model/payment.py | [
{
"identifier": "Config",
"path": "monzo_utils/lib/config.py",
"snippet": "class Config(metaclass=Singleton):\n\n def __init__(self, config=None, config_path=None):\n if config_path is None:\n homedir = pwd.getpwuid(os.getuid()).pw_dir\n config_path = f\"{homedir}/.monzo\... | import re
import datetime
from monzo_utils.lib.config import Config
from monzo_utils.model.transaction import Transaction
from monzo_utils.lib.transactions import Transactions | 1,356 | self.today = datetime.datetime.now()
self.cache = {}
def data(self, abbreviate=False):
if self.num_paid is not None:
suffix = '%d/%d' % (
self.num_paid,
self.num_total
)
else:
suffix = ''
if self.remaining is not None:
remaining = self.remaining
else:
remaining = None
return {
'status': self.status,
'payment_type': self.payment_type if abbreviate is False else self.abbreviate(self.payment_type),
'name': self.name,
'suffix': suffix,
'amount': self.display_amount,
'remaining': remaining,
'last_date': self.short_date(self.last_date) if abbreviate else self.last_date,
'due_date': self.short_date(self.due_date) if abbreviate else self.due_date
}
def abbreviate(self, string):
abbreviated = ''
for i in range(0, len(string)):
if string[i].isupper():
abbreviated += string[i]
return abbreviated
def short_date(self, date):
if not date:
return None
return date.strftime('%d/%m/%y')
def display(self):
data = self.data()
print("%s: %s %s %s %s %s %s %s" % (
data['status'].rjust(7),
data['payment_type'].ljust(15),
data['name'].ljust(25),
data['suffix'].ljust(5),
('£%.2f' % (data['amount'])).ljust(8),
('£%.2f' % (data['remaining'])).ljust(8) if data['remaining'] else ''.ljust(8),
data['last_date'].strftime('%Y-%m-%d').ljust(12) if data['last_date'] else ''.ljust(12),
data['due_date'].strftime('%Y-%m-%d').ljust(10) if data['due_date'] else ''
))
@property
def name(self):
return self.payment_config['name']
@property
def status(self):
if 'start_date' in self.payment_config and self.payment_config['start_date'] >= self.next_salary_date:
return 'SKIPPED'
if 'yearly_month' in self.payment_config:
if self.yearly_payment_due_this_month(self.payment_config, self.last_salary_date) is False:
return 'SKIPPED'
if 'renew_date' in self.payment_config and self.payment_config['renew_date'] >= self.next_salary_date:
return 'SKIPPED'
if 'exclude_months' in self.payment_config and self.today.month in self.payment_config['exclude_months']:
return 'SKIPPED'
if self.last_date and self.last_date >= self.last_salary_date:
return 'PAID'
if self.due_date and self.due_date >= self.next_salary_date:
return 'SKIPPED'
return 'DUE'
@property
def payment_type(self):
return re.sub(r'(?<!^)(?=[A-Z])', '_', type(self).__name__).replace('_',' ')
@property
def num_paid(self):
return None
@property
def num_total(self):
if 'months' in self.payment_config:
return self.payment_config['months']
return None
@property
def remaining(self):
pass
@property
def display_amount(self):
today = datetime.datetime.now()
today = datetime.date(today.year, today.month, today.day)
|
class Payment:
transaction_type = 'money_out'
always_fixed = False
def __init__(self, config, payment_list_config, payment_config, last_salary_date, next_salary_date, following_salary_date):
self.config = config
self.payment_list_config = payment_list_config
self.payment_config = payment_config
self.last_salary_date = last_salary_date
self.next_salary_date = next_salary_date
self.following_salary_date = following_salary_date
self.today = datetime.datetime.now()
self.cache = {}
def data(self, abbreviate=False):
if self.num_paid is not None:
suffix = '%d/%d' % (
self.num_paid,
self.num_total
)
else:
suffix = ''
if self.remaining is not None:
remaining = self.remaining
else:
remaining = None
return {
'status': self.status,
'payment_type': self.payment_type if abbreviate is False else self.abbreviate(self.payment_type),
'name': self.name,
'suffix': suffix,
'amount': self.display_amount,
'remaining': remaining,
'last_date': self.short_date(self.last_date) if abbreviate else self.last_date,
'due_date': self.short_date(self.due_date) if abbreviate else self.due_date
}
def abbreviate(self, string):
abbreviated = ''
for i in range(0, len(string)):
if string[i].isupper():
abbreviated += string[i]
return abbreviated
def short_date(self, date):
if not date:
return None
return date.strftime('%d/%m/%y')
def display(self):
data = self.data()
print("%s: %s %s %s %s %s %s %s" % (
data['status'].rjust(7),
data['payment_type'].ljust(15),
data['name'].ljust(25),
data['suffix'].ljust(5),
('£%.2f' % (data['amount'])).ljust(8),
('£%.2f' % (data['remaining'])).ljust(8) if data['remaining'] else ''.ljust(8),
data['last_date'].strftime('%Y-%m-%d').ljust(12) if data['last_date'] else ''.ljust(12),
data['due_date'].strftime('%Y-%m-%d').ljust(10) if data['due_date'] else ''
))
@property
def name(self):
return self.payment_config['name']
@property
def status(self):
if 'start_date' in self.payment_config and self.payment_config['start_date'] >= self.next_salary_date:
return 'SKIPPED'
if 'yearly_month' in self.payment_config:
if self.yearly_payment_due_this_month(self.payment_config, self.last_salary_date) is False:
return 'SKIPPED'
if 'renew_date' in self.payment_config and self.payment_config['renew_date'] >= self.next_salary_date:
return 'SKIPPED'
if 'exclude_months' in self.payment_config and self.today.month in self.payment_config['exclude_months']:
return 'SKIPPED'
if self.last_date and self.last_date >= self.last_salary_date:
return 'PAID'
if self.due_date and self.due_date >= self.next_salary_date:
return 'SKIPPED'
return 'DUE'
@property
def payment_type(self):
return re.sub(r'(?<!^)(?=[A-Z])', '_', type(self).__name__).replace('_',' ')
@property
def num_paid(self):
return None
@property
def num_total(self):
if 'months' in self.payment_config:
return self.payment_config['months']
return None
@property
def remaining(self):
pass
@property
def display_amount(self):
today = datetime.datetime.now()
today = datetime.date(today.year, today.month, today.day)
| if 'last_amount_overrides' in Config().keys and \ | 0 | 2023-11-05 12:48:18+00:00 | 2k |
rossiyareich/inknhue | src/conditional/conditional_encoder.py | [
{
"identifier": "DownSample",
"path": "src/downsample.py",
"snippet": "class DownSample(nn.Module):\n \"\"\"\n ## Down-sampling layer\n \"\"\"\n\n def __init__(self, channels: int):\n \"\"\"\n :param channels: is the number of channels\n \"\"\"\n super().__init__(... | from typing import List
from torch import nn
from ..downsample import DownSample
from ..resnet_block import ResnetBlock
from ..utils import zero_module
import torch | 1,084 |
class ConditionalEncoder(nn.Module):
def __init__(
self,
*,
channels: int,
channel_multipliers: List[int],
n_resnet_blocks: int,
in_channels: int,
) -> None:
super().__init__()
# Number of blocks of different resolutions.
# The resolution is halved at the end each top level block
n_resolutions = len(channel_multipliers)
# Initial $3 \times 3$ convolution layer that maps the image to `channels`
self.conv_in = nn.Conv2d(in_channels, channels, 3, stride=1, padding=1)
# Number of channels in each top level block
channels_list = [m * channels for m in [1] + channel_multipliers]
# List of top-level blocks
self.down = nn.ModuleList()
self.proj = nn.ModuleList()
# Create top-level blocks
for i in range(n_resolutions):
# Each top level block consists of multiple ResNet Blocks and down-sampling
resnet_blocks = nn.ModuleList()
# Add ResNet Blocks
for _ in range(n_resnet_blocks):
resnet_blocks.append(ResnetBlock(channels, channels_list[i + 1]))
channels = channels_list[i + 1]
# Top-level block
down = nn.Module()
down.block = resnet_blocks
# Down-sampling at the end of each top level block except the last
if i != n_resolutions - 1:
down.downsample = DownSample(channels)
else:
down.downsample = nn.Identity()
self.down.append(down)
# Projection
proj = nn.Conv2d(channels, channels, 1, 1, 0)
|
class ConditionalEncoder(nn.Module):
def __init__(
self,
*,
channels: int,
channel_multipliers: List[int],
n_resnet_blocks: int,
in_channels: int,
) -> None:
super().__init__()
# Number of blocks of different resolutions.
# The resolution is halved at the end each top level block
n_resolutions = len(channel_multipliers)
# Initial $3 \times 3$ convolution layer that maps the image to `channels`
self.conv_in = nn.Conv2d(in_channels, channels, 3, stride=1, padding=1)
# Number of channels in each top level block
channels_list = [m * channels for m in [1] + channel_multipliers]
# List of top-level blocks
self.down = nn.ModuleList()
self.proj = nn.ModuleList()
# Create top-level blocks
for i in range(n_resolutions):
# Each top level block consists of multiple ResNet Blocks and down-sampling
resnet_blocks = nn.ModuleList()
# Add ResNet Blocks
for _ in range(n_resnet_blocks):
resnet_blocks.append(ResnetBlock(channels, channels_list[i + 1]))
channels = channels_list[i + 1]
# Top-level block
down = nn.Module()
down.block = resnet_blocks
# Down-sampling at the end of each top level block except the last
if i != n_resolutions - 1:
down.downsample = DownSample(channels)
else:
down.downsample = nn.Identity()
self.down.append(down)
# Projection
proj = nn.Conv2d(channels, channels, 1, 1, 0) | proj = zero_module(proj) | 2 | 2023-11-03 09:35:30+00:00 | 2k |
drakoRRR/chatSynthia | users/views.py | [
{
"identifier": "ProfileForm",
"path": "users/forms.py",
"snippet": "class ProfileForm(UserChangeForm):\n first_name = forms.CharField(widget=forms.TextInput(attrs={\n 'class': 'form-control py-4'\n }))\n last_name = forms.CharField(widget=forms.TextInput(attrs={\n 'class': 'form-... | from django.contrib import auth, messages
from django.contrib.auth.views import LoginView
from django.urls import reverse_lazy
from django.views.generic.edit import CreateView, UpdateView
from users.forms import ProfileForm, UserLoginForm, UserRegisterForm
from users.models import User | 864 |
# Create your views here.
class LoginUserView(LoginView):
template_name = 'users/login.html'
form_class = UserLoginForm
def form_invalid(self, form):
messages.error(self.request, 'There was an error with username or password, check again !')
return super().form_invalid(form)
class RegistrationView(CreateView):
model = User
form_class = UserRegisterForm
template_name = 'users/register.html'
success_url = reverse_lazy('')
def form_valid(self, form):
response = super().form_valid(form)
auth.login(self.request, self.object, backend='django.contrib.auth.backends.ModelBackend')
return response
def form_invalid(self, form):
messages.error(self.request, 'There was an error with username or password, check again !')
return super().form_invalid(form)
class ProfileView(UpdateView):
model = User
|
# Create your views here.
class LoginUserView(LoginView):
template_name = 'users/login.html'
form_class = UserLoginForm
def form_invalid(self, form):
messages.error(self.request, 'There was an error with username or password, check again !')
return super().form_invalid(form)
class RegistrationView(CreateView):
model = User
form_class = UserRegisterForm
template_name = 'users/register.html'
success_url = reverse_lazy('')
def form_valid(self, form):
response = super().form_valid(form)
auth.login(self.request, self.object, backend='django.contrib.auth.backends.ModelBackend')
return response
def form_invalid(self, form):
messages.error(self.request, 'There was an error with username or password, check again !')
return super().form_invalid(form)
class ProfileView(UpdateView):
model = User | form_class = ProfileForm | 0 | 2023-11-08 12:21:53+00:00 | 2k |
TencentBlueKing/bkflow-feel | bkflow_feel/parsers.py | [
{
"identifier": "RangeGroupData",
"path": "bkflow_feel/data_models.py",
"snippet": "class RangeGroupData(BaseModel):\n left_val: Any\n right_val: Any\n left_operator: RangeGroupOperator\n right_operator: RangeGroupOperator"
},
{
"identifier": "RangeGroupOperator",
"path": "bkflow... | import abc
import datetime
import logging
import re
import pytz
from dateutil.parser import parse as date_parse
from .data_models import RangeGroupData, RangeGroupOperator
from .utils import FEELFunctionsManager
from .validators import BinaryOperationValidator, DummyValidator, ListsLengthValidator | 1,209 | # -*- coding: utf-8 -*-
logger = logging.getLogger(__name__)
class Expression(metaclass=abc.ABCMeta):
validator_cls = DummyValidator
@abc.abstractmethod
def evaluate(self, context):
pass
class CommonExpression(Expression):
def __init__(self, value):
self.value = value
def evaluate(self, context):
return self.value
class Expr(CommonExpression):
def evaluate(self, context):
return self.value.evaluate(context)
class Number(CommonExpression):
pass
class String(CommonExpression):
pass
class Boolean(CommonExpression):
pass
class Null(Expression):
def evaluate(self, context):
return None
class List(Expression):
def __init__(self, *items):
self.items = items
def evaluate(self, context):
return [item.evaluate(context) for item in self.items]
class ListItem(Expression):
def __init__(self, list_expr, index):
self.list_expr = list_expr
self.index = index
def evaluate(self, context):
items = self.list_expr.evaluate(context)
if not isinstance(items, list) or self.index == 0 or len(items) < abs(self.index):
return None
items = items[self.index - 1] if self.index > 0 else items[self.index]
return items
class ListMatch(Expression):
| # -*- coding: utf-8 -*-
logger = logging.getLogger(__name__)
class Expression(metaclass=abc.ABCMeta):
validator_cls = DummyValidator
@abc.abstractmethod
def evaluate(self, context):
pass
class CommonExpression(Expression):
def __init__(self, value):
self.value = value
def evaluate(self, context):
return self.value
class Expr(CommonExpression):
def evaluate(self, context):
return self.value.evaluate(context)
class Number(CommonExpression):
pass
class String(CommonExpression):
pass
class Boolean(CommonExpression):
pass
class Null(Expression):
def evaluate(self, context):
return None
class List(Expression):
def __init__(self, *items):
self.items = items
def evaluate(self, context):
return [item.evaluate(context) for item in self.items]
class ListItem(Expression):
def __init__(self, list_expr, index):
self.list_expr = list_expr
self.index = index
def evaluate(self, context):
items = self.list_expr.evaluate(context)
if not isinstance(items, list) or self.index == 0 or len(items) < abs(self.index):
return None
items = items[self.index - 1] if self.index > 0 else items[self.index]
return items
class ListMatch(Expression): | validator_cls = ListsLengthValidator | 5 | 2023-11-09 13:47:26+00:00 | 2k |
namedgraph/oxijen | oxijen/model_impl/impl.py | [
{
"identifier": "Resource",
"path": "oxijen/rdf_model.py",
"snippet": "class Resource(ABC):\n\n @property\n def node(self):\n return self._node\n\n @property\n def graph(self):\n return self._graph\n\n @property\n def is_anon(self):\n if isinstance(self.node, Named... | from oxijen.rdf_model import Resource, Property, Graph, Dataset
from oxijen.model_impl.xsd import XSD
from pyoxigraph import Store, Triple, BlankNode, NamedNode, Literal, Quad, DefaultGraph
from typing import Iterator, Union, Optional, Any | 1,453 |
class ResourceImpl(Resource):
def __init__(self, node: Union[BlankNode, NamedNode], graph: Graph):
self._node = node
self._graph = graph
def __hash__(self):
return hash(self.node.value)
def __eq__(self, other):
if isinstance(other, self.__class__):
return self.node.value == other.node.value
else:
return False
def __ne__(self, other):
return not self.__eq__(other)
def __str__(self) -> str:
return self.node.__str__()
def __repr__(self) -> str:
return self.__str__()
def add_property(self, property: 'Property', value: Union[Resource, Literal]) -> 'Resource':
if isinstance(value, Resource):
value = value.node
self.graph.store.add(Quad(self.node, property.node, value, self.graph.name)) # assumes GraphStoreImpl!
return self
def list_properties(self, property: Optional[Property] = None) -> Iterator[Triple]:
if isinstance(property, Property):
property_node = property.node
else:
property_node = None
quads = self.graph.store.quads_for_pattern(self.node, property_node, None, None)
return map(lambda quad: quad.triple, quads)
def remove_all(self, property: Optional[Property] = None) -> Resource:
if isinstance(property, Property):
property_node = property.node
else:
property_node = None
quad_iter = self.graph.store.quads_for_pattern(self.node, property_node, None, None)
for quad in quad_iter:
self.graph.store.remove(quad)
return self
class PropertyImpl(ResourceImpl, Property):
pass
class GraphImpl(Graph):
def create_resource(self, uri: Optional[str] = None) -> Resource:
if uri is not None:
return ResourceImpl(NamedNode(uri), self)
else:
return ResourceImpl(BlankNode(), self)
def create_property(self, uri: str) -> Property:
return ResourceImpl(NamedNode(uri), self)
def create_literal(self, value: str, language: Optional[str] = None) -> Literal:
return Literal(value, language=language) # should it be xsd:string-typed by default as per RDF 1.1?
def create_typed_literal(self, value: Any, datatype: Optional[Union[str, NamedNode]] = None) -> Literal:
if datatype is None:
match value:
case int():
|
class ResourceImpl(Resource):
def __init__(self, node: Union[BlankNode, NamedNode], graph: Graph):
self._node = node
self._graph = graph
def __hash__(self):
return hash(self.node.value)
def __eq__(self, other):
if isinstance(other, self.__class__):
return self.node.value == other.node.value
else:
return False
def __ne__(self, other):
return not self.__eq__(other)
def __str__(self) -> str:
return self.node.__str__()
def __repr__(self) -> str:
return self.__str__()
def add_property(self, property: 'Property', value: Union[Resource, Literal]) -> 'Resource':
if isinstance(value, Resource):
value = value.node
self.graph.store.add(Quad(self.node, property.node, value, self.graph.name)) # assumes GraphStoreImpl!
return self
def list_properties(self, property: Optional[Property] = None) -> Iterator[Triple]:
if isinstance(property, Property):
property_node = property.node
else:
property_node = None
quads = self.graph.store.quads_for_pattern(self.node, property_node, None, None)
return map(lambda quad: quad.triple, quads)
def remove_all(self, property: Optional[Property] = None) -> Resource:
if isinstance(property, Property):
property_node = property.node
else:
property_node = None
quad_iter = self.graph.store.quads_for_pattern(self.node, property_node, None, None)
for quad in quad_iter:
self.graph.store.remove(quad)
return self
class PropertyImpl(ResourceImpl, Property):
pass
class GraphImpl(Graph):
def create_resource(self, uri: Optional[str] = None) -> Resource:
if uri is not None:
return ResourceImpl(NamedNode(uri), self)
else:
return ResourceImpl(BlankNode(), self)
def create_property(self, uri: str) -> Property:
return ResourceImpl(NamedNode(uri), self)
def create_literal(self, value: str, language: Optional[str] = None) -> Literal:
return Literal(value, language=language) # should it be xsd:string-typed by default as per RDF 1.1?
def create_typed_literal(self, value: Any, datatype: Optional[Union[str, NamedNode]] = None) -> Literal:
if datatype is None:
match value:
case int(): | datatype = NamedNode(XSD.INTEGER.value) | 4 | 2023-11-03 19:50:51+00:00 | 2k |
sivasurend/lyzr | build/lib/lyzr/utils/document_reading.py | [
{
"identifier": "LyzrDocxReader",
"path": "lyzr/utils/docx_reader.py",
"snippet": "class LyzrDocxReader(BaseReader):\n def __init__(self) -> None:\n try:\n import docx2txt\n except ImportError:\n raise ImportError(\n \"`docx2txt` package not found, p... | import logging
from typing import List, Sequence, Optional
from llama_index.readers.file.base import SimpleDirectoryReader
from llama_index.schema import Document
from lyzr.utils.docx_reader import LyzrDocxReader
from lyzr.utils.pdf_reader import LyzrPDFReader
from lyzr.utils.txt_reader import LyzrTxtReader
from lyzr.utils.webpage_reader import LyzrWebPageReader
from lyzr.utils.website_reader import LyzrWebsiteReader
from lyzr.utils.youtube_reader import LyzrYoutubeReader | 1,230 |
logger = logging.getLogger(__name__)
def read_pdf_as_documents(
input_dir: Optional[str] = None,
input_files: Optional[List] = None,
exclude_hidden: bool = True,
filename_as_id: bool = True,
recursive: bool = True,
required_exts: Optional[List[str]] = None,
**kwargs,
) -> Sequence[Document]:
|
logger = logging.getLogger(__name__)
def read_pdf_as_documents(
input_dir: Optional[str] = None,
input_files: Optional[List] = None,
exclude_hidden: bool = True,
filename_as_id: bool = True,
recursive: bool = True,
required_exts: Optional[List[str]] = None,
**kwargs,
) -> Sequence[Document]: | file_extractor = {".pdf": LyzrPDFReader()} | 1 | 2023-11-07 14:52:08+00:00 | 2k |
focused-labs/ai-custom-chatbot-data-pipeline | main.py | [
{
"identifier": "import_web_scrape_data",
"path": "import_service.py",
"snippet": "def import_web_scrape_data(urls: list):\n BeautifulSoupWebReader = download_loader(\"BeautifulSoupWebReader\")\n\n loader = BeautifulSoupWebReader()\n documents = loader.load_data(urls=urls)\n\n for document i... | import logging
import os
import sys
import openai
import uvicorn
from contextlib import asynccontextmanager
from dotenv import load_dotenv
from fastapi import FastAPI
from fastapi.middleware.cors import CORSMiddleware
from import_service import import_web_scrape_data, import_notion_data
from models.imported_pages import ImportedPages
from models.imported_urls import ImportedUrls
from models.question import Question
from query_service import QueryService | 689 |
load_dotenv()
openai.api_key = os.getenv('OPENAI_API_KEY')
allowed_origins = [
"http://localhost:3000",
]
|
load_dotenv()
openai.api_key = os.getenv('OPENAI_API_KEY')
allowed_origins = [
"http://localhost:3000",
]
| query_service = QueryService() | 5 | 2023-11-01 20:47:07+00:00 | 2k |
pradyunsg/pip-resolver-benchmarks | src/common/creation.py | [
{
"identifier": "DistributionInfo",
"path": "src/common/model.py",
"snippet": "class DistributionInfo(BaseModel):\n depends_by_extra: dict[str, list[str]]\n requires_python: str | None = None\n\n @field_validator(\"depends_by_extra\", mode=\"after\")\n @classmethod\n def ensure_no_empty_e... | import base64
import hashlib
import zipfile
from pathlib import Path
from rich.progress import BarColumn, MofNCompleteColumn, Progress
from .model import DistributionInfo, Scenario | 725 | """Creates the actual wheel files in a directory to pass to the resolver.
"""
from __future__ import annotations
WHEEL = """\
Wheel-Version: 1.0
Generator: pip-resolver-benchmark
Root-Is-Purelib: true
Tag: py2-none-any
Tag: py3-none-any
"""
def _make_wheel(
| """Creates the actual wheel files in a directory to pass to the resolver.
"""
from __future__ import annotations
WHEEL = """\
Wheel-Version: 1.0
Generator: pip-resolver-benchmark
Root-Is-Purelib: true
Tag: py2-none-any
Tag: py3-none-any
"""
def _make_wheel( | name: str, version: str, wheel: DistributionInfo, output_dir: Path | 0 | 2023-11-05 17:59:32+00:00 | 2k |
allmonday/pydantic2-resolve | pydantic2_resolve/resolver.py | [
{
"identifier": "ResolverTargetAttrNotFound",
"path": "pydantic2_resolve/exceptions.py",
"snippet": "class ResolverTargetAttrNotFound(Exception):\n pass"
},
{
"identifier": "LoaderFieldNotProvidedError",
"path": "pydantic2_resolve/exceptions.py",
"snippet": "class LoaderFieldNotProvid... | import asyncio
import contextvars
import inspect
import pydantic2_resolve.constant as const
import pydantic2_resolve.util as util
from collections import defaultdict
from inspect import iscoroutine
from typing import TypeVar, Dict
from .exceptions import ResolverTargetAttrNotFound, LoaderFieldNotProvidedError, MissingAnnotationError
from typing import Any, Callable, Optional
from pydantic2_resolve import core
from aiodataloader import DataLoader
from inspect import isclass
from types import MappingProxyType | 1,477 |
# for dataloader which has class attributes, you can assign the value at here
self.loader_filters = loader_filters or {}
# now you can pass your loader instance, Resolver will check `isinstance``
if loader_instances and self._validate_loader_instance(loader_instances):
self.loader_instances = loader_instances
else:
self.loader_instances = None
self.ensure_type = ensure_type
self.context = MappingProxyType(context) if context else None
self.scan_data = {}
def _add_expose_fields(self, target):
"""
1. check whether expose to descendant existed
2. add fields into contextvars (ancestor_vars_checker)
2.1 check overwrite by another class(which is forbidden)
2.2 check field exists
"""
dct: Optional[dict] = getattr(target, const.EXPOSE_TO_DESCENDANT, None)
# 1
if dct:
if type(dct) is not dict:
raise AttributeError(f'{const.EXPOSE_TO_DESCENDANT} is not dict')
# 2
for field, alias in dct.items(): # eg: name, bar_name
# 2.1
self.ancestor_vars_checker[alias].add(util.get_kls_full_path(target.__class__))
if len(self.ancestor_vars_checker[alias]) > 1:
conflict_modules = ', '.join(list(self.ancestor_vars_checker[alias]))
raise AttributeError(f'alias name conflicts, please check: {conflict_modules}')
if not self.ancestor_vars.get(alias):
self.ancestor_vars[alias] = contextvars.ContextVar(alias)
try:
val = getattr(target, field)
except AttributeError:
raise AttributeError(f'{field} does not existed')
self.ancestor_vars[alias].set(val)
def _build_ancestor_context(self):
"""get values from contextvars and put into a dict"""
return { k: v.get() for k, v in self.ancestor_vars.items()}
def _validate_loader_instance(self, loader_instances: Dict[Any, Any]):
for cls, loader in loader_instances.items():
if not issubclass(cls, DataLoader):
raise AttributeError(f'{cls.__name__} must be subclass of DataLoader')
if not isinstance(loader, cls):
raise AttributeError(f'{loader.__name__} is not instance of {cls.__name__}')
return True
def _execute_resolver_method(self, method):
"""
1. inspect method, atttach context if declared in method
2. if params includes LoaderDepend, create instance and cache it.
2.1 create from DataLoader class
2.1.1 apply loader_filters into dataloader instance
2.2 ceate from batch_load_fn
3. execute method
"""
# >>> 1
signature = inspect.signature(method)
params = {}
if signature.parameters.get('context'):
if self.context is None:
raise AttributeError('Resolver.context is missing')
params['context'] = self.context
if signature.parameters.get('ancestor_context'):
if self.ancestor_vars is None:
raise AttributeError(f'there is not class has {const.EXPOSE_TO_DESCENDANT} configed')
params['ancestor_context'] = self._build_ancestor_context()
# manage the creation of loader instances
for k, v in signature.parameters.items():
# >>> 2
if isinstance(v.default, Depends):
# Base: DataLoader or batch_load_fn
Loader = v.default.dependency
# check loader_instance first, if already defined in Resolver param, just take it.
if self.loader_instances and self.loader_instances.get(Loader):
loader = self.loader_instances.get(Loader)
params[k] = loader
continue
# module.kls to avoid same kls name from different module
cache_key = util.get_kls_full_path(v.default.dependency)
hit = self.loader_instance_cache.get(cache_key)
if hit:
loader = hit
else:
# >>> 2.1
# create loader instance
if isclass(Loader):
# if extra transform provides
loader = Loader()
filter_config = self.loader_filters.get(Loader, {})
for field in util.get_class_field_annotations(Loader):
# >>> 2.1.1
# class ExampleLoader(DataLoader):
# filtar_x: bool <--------------- set this field
try:
value = filter_config[field]
setattr(loader, field, value)
except KeyError:
|
def LoaderDepend( # noqa: N802
dependency: Optional[Callable[..., Any]] = None,
) -> Any:
return Depends(dependency=dependency)
class Depends:
def __init__(
self,
dependency: Optional[Callable[..., Any]] = None,
):
self.dependency = dependency
T = TypeVar("T")
class Resolver:
"""
Entrypoint of a resolve action
"""
def __init__(
self,
loader_filters: Optional[Dict[Any, Dict[str, Any]]] = None,
loader_instances: Optional[Dict[Any, Any]] = None,
ensure_type=False,
context: Optional[Dict[str, Any]] = None
):
self.loader_instance_cache = {}
self.ancestor_vars = {}
self.ancestor_vars_checker = defaultdict(set) # expose_field_name: set(kls fullpath) if len > 1, raise error
# for dataloader which has class attributes, you can assign the value at here
self.loader_filters = loader_filters or {}
# now you can pass your loader instance, Resolver will check `isinstance``
if loader_instances and self._validate_loader_instance(loader_instances):
self.loader_instances = loader_instances
else:
self.loader_instances = None
self.ensure_type = ensure_type
self.context = MappingProxyType(context) if context else None
self.scan_data = {}
def _add_expose_fields(self, target):
"""
1. check whether expose to descendant existed
2. add fields into contextvars (ancestor_vars_checker)
2.1 check overwrite by another class(which is forbidden)
2.2 check field exists
"""
dct: Optional[dict] = getattr(target, const.EXPOSE_TO_DESCENDANT, None)
# 1
if dct:
if type(dct) is not dict:
raise AttributeError(f'{const.EXPOSE_TO_DESCENDANT} is not dict')
# 2
for field, alias in dct.items(): # eg: name, bar_name
# 2.1
self.ancestor_vars_checker[alias].add(util.get_kls_full_path(target.__class__))
if len(self.ancestor_vars_checker[alias]) > 1:
conflict_modules = ', '.join(list(self.ancestor_vars_checker[alias]))
raise AttributeError(f'alias name conflicts, please check: {conflict_modules}')
if not self.ancestor_vars.get(alias):
self.ancestor_vars[alias] = contextvars.ContextVar(alias)
try:
val = getattr(target, field)
except AttributeError:
raise AttributeError(f'{field} does not existed')
self.ancestor_vars[alias].set(val)
def _build_ancestor_context(self):
"""get values from contextvars and put into a dict"""
return { k: v.get() for k, v in self.ancestor_vars.items()}
def _validate_loader_instance(self, loader_instances: Dict[Any, Any]):
for cls, loader in loader_instances.items():
if not issubclass(cls, DataLoader):
raise AttributeError(f'{cls.__name__} must be subclass of DataLoader')
if not isinstance(loader, cls):
raise AttributeError(f'{loader.__name__} is not instance of {cls.__name__}')
return True
def _execute_resolver_method(self, method):
"""
1. inspect method, atttach context if declared in method
2. if params includes LoaderDepend, create instance and cache it.
2.1 create from DataLoader class
2.1.1 apply loader_filters into dataloader instance
2.2 ceate from batch_load_fn
3. execute method
"""
# >>> 1
signature = inspect.signature(method)
params = {}
if signature.parameters.get('context'):
if self.context is None:
raise AttributeError('Resolver.context is missing')
params['context'] = self.context
if signature.parameters.get('ancestor_context'):
if self.ancestor_vars is None:
raise AttributeError(f'there is not class has {const.EXPOSE_TO_DESCENDANT} configed')
params['ancestor_context'] = self._build_ancestor_context()
# manage the creation of loader instances
for k, v in signature.parameters.items():
# >>> 2
if isinstance(v.default, Depends):
# Base: DataLoader or batch_load_fn
Loader = v.default.dependency
# check loader_instance first, if already defined in Resolver param, just take it.
if self.loader_instances and self.loader_instances.get(Loader):
loader = self.loader_instances.get(Loader)
params[k] = loader
continue
# module.kls to avoid same kls name from different module
cache_key = util.get_kls_full_path(v.default.dependency)
hit = self.loader_instance_cache.get(cache_key)
if hit:
loader = hit
else:
# >>> 2.1
# create loader instance
if isclass(Loader):
# if extra transform provides
loader = Loader()
filter_config = self.loader_filters.get(Loader, {})
for field in util.get_class_field_annotations(Loader):
# >>> 2.1.1
# class ExampleLoader(DataLoader):
# filtar_x: bool <--------------- set this field
try:
value = filter_config[field]
setattr(loader, field, value)
except KeyError: | raise LoaderFieldNotProvidedError(f'{cache_key}.{field} not found in Resolver()') | 1 | 2023-11-01 02:37:26+00:00 | 2k |
StoneMoe/ASub | app/ui/windows/subtitle_window.py | [
{
"identifier": "SRTFile",
"path": "app/core/models/srt.py",
"snippet": "class SRTFile:\r\n filepath: str\r\n entries: List[SRTEntry]\r\n\r\n def __init__(self, source: str | list):\r\n self.filepath = ''\r\n self.entries = []\r\n\r\n match source:\r\n case str()... | from PyQt5.QtWidgets import QVBoxLayout, QPushButton, QTableWidgetItem, QDialog
from qfluentwidgets import TableWidget, isDarkTheme
from qframelesswindow import FramelessWindow
from app.core.models.srt import SRTFile
from app.ui.const import CONTAINER_MARGINS
from app.core.utils.env import res_dir | 1,290 |
class SubtitleWindow(QDialog, FramelessWindow):
def __init__(self, filepath: str, parent=None):
super().__init__(parent)
self.srt_file = SRTFile(filepath)
self.hBoxLayout = QVBoxLayout(self)
self.tableView = TableWidget(self)
self.saveButton = QPushButton("Save", self)
self.saveButton.clicked.connect(self._save_subtitle_file)
self.hBoxLayout.setContentsMargins(*CONTAINER_MARGINS)
self.hBoxLayout.addWidget(self.tableView)
self.hBoxLayout.addWidget(self.saveButton)
self.init_window()
self._load_subtitle_file()
def _load_subtitle_file(self):
self.tableView.setWordWrap(False)
self.tableView.setRowCount(len(self.srt_file.entries))
self.tableView.setColumnCount(3)
for i, entry in enumerate(self.srt_file.entries):
self.tableView.setItem(i, 0, QTableWidgetItem(entry.index))
self.tableView.setItem(i, 1, QTableWidgetItem(entry.time))
self.tableView.setItem(i, 2, QTableWidgetItem(entry.text))
self.tableView.verticalHeader().hide()
self.tableView.setHorizontalHeaderLabels(['Index', 'Time', 'Text'])
self.tableView.resizeColumnsToContents()
def _save_subtitle_file(self):
for i in range(self.tableView.rowCount()):
self.srt_file.entries[i].index = self.tableView.item(i, 0).text()
self.srt_file.entries[i].time = self.tableView.item(i, 1).text()
self.srt_file.entries[i].text = self.tableView.item(i, 2).text()
self.srt_file.dump()
def init_window(self):
self.setWindowTitle(f'编辑 {self.srt_file.filepath}')
self.resize(625, 700)
self._set_qss()
def _set_qss(self):
color = 'dark' if isDarkTheme() else 'light'
|
class SubtitleWindow(QDialog, FramelessWindow):
def __init__(self, filepath: str, parent=None):
super().__init__(parent)
self.srt_file = SRTFile(filepath)
self.hBoxLayout = QVBoxLayout(self)
self.tableView = TableWidget(self)
self.saveButton = QPushButton("Save", self)
self.saveButton.clicked.connect(self._save_subtitle_file)
self.hBoxLayout.setContentsMargins(*CONTAINER_MARGINS)
self.hBoxLayout.addWidget(self.tableView)
self.hBoxLayout.addWidget(self.saveButton)
self.init_window()
self._load_subtitle_file()
def _load_subtitle_file(self):
self.tableView.setWordWrap(False)
self.tableView.setRowCount(len(self.srt_file.entries))
self.tableView.setColumnCount(3)
for i, entry in enumerate(self.srt_file.entries):
self.tableView.setItem(i, 0, QTableWidgetItem(entry.index))
self.tableView.setItem(i, 1, QTableWidgetItem(entry.time))
self.tableView.setItem(i, 2, QTableWidgetItem(entry.text))
self.tableView.verticalHeader().hide()
self.tableView.setHorizontalHeaderLabels(['Index', 'Time', 'Text'])
self.tableView.resizeColumnsToContents()
def _save_subtitle_file(self):
for i in range(self.tableView.rowCount()):
self.srt_file.entries[i].index = self.tableView.item(i, 0).text()
self.srt_file.entries[i].time = self.tableView.item(i, 1).text()
self.srt_file.entries[i].text = self.tableView.item(i, 2).text()
self.srt_file.dump()
def init_window(self):
self.setWindowTitle(f'编辑 {self.srt_file.filepath}')
self.resize(625, 700)
self._set_qss()
def _set_qss(self):
color = 'dark' if isDarkTheme() else 'light' | with open(res_dir(f'app/ui/resource/qss/{color}/style.qss'), encoding='utf-8') as f: | 2 | 2023-11-07 16:45:43+00:00 | 2k |
openshift/lightspeed-service | tests/unit/docs/test_doc_summarizer.py | [
{
"identifier": "DocsSummarizer",
"path": "ols/src/docs/docs_summarizer.py",
"snippet": "class DocsSummarizer:\n \"\"\"A class for summarizing documentation context.\"\"\"\n\n def __init__(self):\n \"\"\"Initialize the DocsSummarizer.\"\"\"\n self.logger = Logger(\"docs_summarizer\")... | import os
from unittest.mock import patch
from ols.src.docs.docs_summarizer import DocsSummarizer
from ols.utils import config
from tests.mock_classes.llm_loader import mock_llm_loader
from tests.mock_classes.mock_llama_index import MockLlamaIndex | 1,175 | """Unit tests for DocsSummarizer class."""
@patch("ols.src.docs.docs_summarizer.LLMLoader", new=mock_llm_loader(None))
@patch("ols.src.docs.docs_summarizer.ServiceContext.from_defaults")
@patch("ols.src.docs.docs_summarizer.StorageContext.from_defaults")
| """Unit tests for DocsSummarizer class."""
@patch("ols.src.docs.docs_summarizer.LLMLoader", new=mock_llm_loader(None))
@patch("ols.src.docs.docs_summarizer.ServiceContext.from_defaults")
@patch("ols.src.docs.docs_summarizer.StorageContext.from_defaults") | @patch("ols.src.docs.docs_summarizer.load_index_from_storage", new=MockLlamaIndex) | 3 | 2023-11-08 06:29:41+00:00 | 2k |
xlcaptain/LLM-Workbench | component/knowledge_chat.py | [
{
"identifier": "ElasticsearchServer",
"path": "component/pipelines/es.py",
"snippet": "class ElasticsearchServer:\n def __init__(self):\n self.client = Elasticsearch(\n ES_URL,\n verify_certs=False,\n )\n self.embedding = Embeddings()\n self.es = Ela... | import time
import os
import streamlit as st
import pandas as pd
from .pipelines.es import ElasticsearchServer
from .pipelines.utils import handle_response, create_message
from .pipelines.prompt import KNOWLEDGE_PROMPT, CHAT_EXAMPLES | 1,286 |
BAICHUAN_URL = os.getenv("BAICHUAN_URL")
def handle_kb_qa(prompt, top_k, threshold):
index_name = 'audit_index'
|
BAICHUAN_URL = os.getenv("BAICHUAN_URL")
def handle_kb_qa(prompt, top_k, threshold):
index_name = 'audit_index' | es_server = ElasticsearchServer() | 0 | 2023-11-01 07:54:03+00:00 | 2k |
NicolasZucchet/Online-learning-LR-dependencies | online_lru/rec.py | [
{
"identifier": "matrix_init",
"path": "online_lru/rec_init.py",
"snippet": "def matrix_init(key, shape, dtype=jnp.float32, normalization=1):\n return random.normal(key=key, shape=shape, dtype=dtype) / normalization"
},
{
"identifier": "truncated_normal_matrix_init",
"path": "online_lru/r... | from functools import partial
from flax import linen as nn
from .rec_init import matrix_init, truncated_normal_matrix_init, theta_init, nu_init, gamma_log_init
from flax.core.frozen_dict import unfreeze
import jax
import jax.numpy as jnp | 1,517 | A_i, b_i = q_i
A_j, b_j = q_j
return A_j * A_i, jax.lax.stop_gradient(A_j * b_i) + b_j
class LRU(nn.Module):
"""
LRU layer that updates internal elegibility traces to allow online learning.
"""
d_hidden: int # hidden state dimension
d_model: int # input and output dimensions
seq_length: int # time sequence length
gamma_norm: bool = True # use gamma normalization
exp_param: bool = True # exponential parametrization for lambda
r_min: float = 0.0 # smallest eigenvalue norm
r_max: float = 1.0 # largest eigenvalue norm
max_phase: float = 6.28 # max phase eigenvalue
training_mode: str = "bptt" # which learning algorithm that will be used
training: bool = False # TODO remove, for debugging purposes
def get_diag_lambda(self, nu=None, theta=None):
"""
Transform parameters nu and theta into the diagonal of the recurrent
Lambda matrix.
Args:
nu, theta array[N]: when set to their default values, None, the
parameters will take the values of the Module.
NOTE: these arguments are added in order to backpropagate through this
transformation.
"""
if nu is None:
nu = self.nu
if theta is None:
theta = self.theta
if self.exp_param:
theta = jnp.exp(theta)
nu = jnp.exp(nu)
return jnp.exp(-nu + 1j * theta)
def get_diag_gamma(self):
"""
Transform parameters gamma_log into the diagonal terms of the modulation matrix gamma.
"""
if self.gamma_norm:
return jnp.exp(self.gamma_log)
else:
return jnp.ones((self.d_hidden,))
def get_B(self):
"""
Get input to hidden matrix B.
"""
return self.B_re + 1j * self.B_im
def get_B_norm(self):
"""
Get modulated input to hidden matrix gamma B.
"""
return self.get_B() * jnp.expand_dims(self.get_diag_gamma(), axis=-1)
def to_output(self, inputs, hidden_states):
"""
Compute output given inputs and hidden states.
Args:
inputs array[T, H].
hidden_states array[T, N].
"""
C = self.C_re + 1j * self.C_im
D = self.D
y = jax.vmap(lambda x, u: (C @ x).real + D * u)(hidden_states, inputs)
return y
def get_hidden_states(self, inputs):
"""
Compute the hidden states corresponding to inputs
Return:
hidden_states array[T, N]
"""
# Materializing the diagonal of Lambda and projections
diag_lambda = self.get_diag_lambda()
B_norm = self.get_B_norm()
# Running the LRU + output projection
# For details on parallel scan, check discussion in Smith et al (2022).
Lambda_elements = jnp.repeat(diag_lambda[None, ...], inputs.shape[0], axis=0)
Bu_elements = jax.vmap(lambda u: B_norm @ u)(inputs)
elements = (Lambda_elements, Bu_elements)
if self.training_mode == "bptt":
_, hidden_states = jax.lax.associative_scan(binary_operator_diag, elements)
else:
_, hidden_states = jax.lax.associative_scan(binary_operator_diag_spatial, elements)
return hidden_states
def setup(self):
# Check that desired approximation is handled
if self.training_mode == "online_snap1":
raise NotImplementedError("SnAp-1 not implemented for LRU")
assert self.training_mode in [
"bptt",
"online_full",
"online_full_rec",
"online_full_rec_simpleB",
"online_snap1", # same as online_full
"online_spatial",
"online_1truncated",
"online_reservoir",
]
self.online = "online" in self.training_mode # whether we compute the gradient online
if self.online:
self.approximation_type = self.training_mode[7:]
# NOTE if exp_param is true, self.theta and self.nu actually represent the log of nu and
# theta lambda is initialized uniformly in complex plane
self.theta = self.param(
"theta",
|
# Parallel scan operations
@jax.vmap
def binary_operator_diag(q_i, q_j):
"""Binary operator for parallel scan of linear recurrence"""
A_i, b_i = q_i
A_j, b_j = q_j
return A_j * A_i, A_j * b_i + b_j
@jax.vmap
def binary_operator_diag_spatial(q_i, q_j):
"""Same as above but stop the gradient for the recurrent connection"""
A_i, b_i = q_i
A_j, b_j = q_j
return A_j * A_i, jax.lax.stop_gradient(A_j * b_i) + b_j
class LRU(nn.Module):
"""
LRU layer that updates internal elegibility traces to allow online learning.
"""
d_hidden: int # hidden state dimension
d_model: int # input and output dimensions
seq_length: int # time sequence length
gamma_norm: bool = True # use gamma normalization
exp_param: bool = True # exponential parametrization for lambda
r_min: float = 0.0 # smallest eigenvalue norm
r_max: float = 1.0 # largest eigenvalue norm
max_phase: float = 6.28 # max phase eigenvalue
training_mode: str = "bptt" # which learning algorithm that will be used
training: bool = False # TODO remove, for debugging purposes
def get_diag_lambda(self, nu=None, theta=None):
"""
Transform parameters nu and theta into the diagonal of the recurrent
Lambda matrix.
Args:
nu, theta array[N]: when set to their default values, None, the
parameters will take the values of the Module.
NOTE: these arguments are added in order to backpropagate through this
transformation.
"""
if nu is None:
nu = self.nu
if theta is None:
theta = self.theta
if self.exp_param:
theta = jnp.exp(theta)
nu = jnp.exp(nu)
return jnp.exp(-nu + 1j * theta)
def get_diag_gamma(self):
"""
Transform parameters gamma_log into the diagonal terms of the modulation matrix gamma.
"""
if self.gamma_norm:
return jnp.exp(self.gamma_log)
else:
return jnp.ones((self.d_hidden,))
def get_B(self):
"""
Get input to hidden matrix B.
"""
return self.B_re + 1j * self.B_im
def get_B_norm(self):
"""
Get modulated input to hidden matrix gamma B.
"""
return self.get_B() * jnp.expand_dims(self.get_diag_gamma(), axis=-1)
def to_output(self, inputs, hidden_states):
"""
Compute output given inputs and hidden states.
Args:
inputs array[T, H].
hidden_states array[T, N].
"""
C = self.C_re + 1j * self.C_im
D = self.D
y = jax.vmap(lambda x, u: (C @ x).real + D * u)(hidden_states, inputs)
return y
def get_hidden_states(self, inputs):
"""
Compute the hidden states corresponding to inputs
Return:
hidden_states array[T, N]
"""
# Materializing the diagonal of Lambda and projections
diag_lambda = self.get_diag_lambda()
B_norm = self.get_B_norm()
# Running the LRU + output projection
# For details on parallel scan, check discussion in Smith et al (2022).
Lambda_elements = jnp.repeat(diag_lambda[None, ...], inputs.shape[0], axis=0)
Bu_elements = jax.vmap(lambda u: B_norm @ u)(inputs)
elements = (Lambda_elements, Bu_elements)
if self.training_mode == "bptt":
_, hidden_states = jax.lax.associative_scan(binary_operator_diag, elements)
else:
_, hidden_states = jax.lax.associative_scan(binary_operator_diag_spatial, elements)
return hidden_states
def setup(self):
# Check that desired approximation is handled
if self.training_mode == "online_snap1":
raise NotImplementedError("SnAp-1 not implemented for LRU")
assert self.training_mode in [
"bptt",
"online_full",
"online_full_rec",
"online_full_rec_simpleB",
"online_snap1", # same as online_full
"online_spatial",
"online_1truncated",
"online_reservoir",
]
self.online = "online" in self.training_mode # whether we compute the gradient online
if self.online:
self.approximation_type = self.training_mode[7:]
# NOTE if exp_param is true, self.theta and self.nu actually represent the log of nu and
# theta lambda is initialized uniformly in complex plane
self.theta = self.param(
"theta", | partial(theta_init, max_phase=self.max_phase, log=self.exp_param), | 2 | 2023-11-01 13:18:32+00:00 | 2k |
uygarkurt/video-retalking | models/ENet.py | [
{
"identifier": "ResBlock",
"path": "models/base_blocks.py",
"snippet": "class ResBlock(nn.Module):\n def __init__(self, in_channels, out_channels, mode='down'):\n super(ResBlock, self).__init__()\n self.conv1 = nn.Conv2d(in_channels, in_channels, 3, 1, 1)\n self.conv2 = nn.Conv2... | import torch
import torch.nn as nn
import torch.nn.functional as F
from models.base_blocks import ResBlock, StyleConv, ToRGB | 1,390 |
class ENet(nn.Module):
def __init__(
self,
num_style_feat=512,
lnet=None,
concat=False
):
super(ENet, self).__init__()
self.low_res = lnet
for param in self.low_res.parameters():
param.requires_grad = False
channel_multiplier, narrow = 2, 1
channels = {
'4': int(512 * narrow),
'8': int(512 * narrow),
'16': int(512 * narrow),
'32': int(512 * narrow),
'64': int(256 * channel_multiplier * narrow),
'128': int(128 * channel_multiplier * narrow),
'256': int(64 * channel_multiplier * narrow),
'512': int(32 * channel_multiplier * narrow),
'1024': int(16 * channel_multiplier * narrow)
}
self.log_size = 8
first_out_size = 128
self.conv_body_first = nn.Conv2d(3, channels[f'{first_out_size}'], 1) # 256 -> 128
# downsample
in_channels = channels[f'{first_out_size}']
self.conv_body_down = nn.ModuleList()
for i in range(8, 2, -1):
out_channels = channels[f'{2**(i - 1)}']
self.conv_body_down.append(ResBlock(in_channels, out_channels, mode='down'))
in_channels = out_channels
self.num_style_feat = num_style_feat
linear_out_channel = num_style_feat
self.final_linear = nn.Linear(channels['4'] * 4 * 4, linear_out_channel)
self.final_conv = nn.Conv2d(in_channels, channels['4'], 3, 1, 1)
self.style_convs = nn.ModuleList()
self.to_rgbs = nn.ModuleList()
self.noises = nn.Module()
self.concat = concat
if concat:
in_channels = 3 + 32 # channels['64']
else:
in_channels = 3
for i in range(7, 9): # 128, 256
out_channels = channels[f'{2**i}'] #
self.style_convs.append(
StyleConv(
in_channels,
out_channels,
kernel_size=3,
num_style_feat=num_style_feat,
demodulate=True,
sample_mode='upsample'))
self.style_convs.append(
StyleConv(
out_channels,
out_channels,
kernel_size=3,
num_style_feat=num_style_feat,
demodulate=True,
sample_mode=None))
|
class ENet(nn.Module):
def __init__(
self,
num_style_feat=512,
lnet=None,
concat=False
):
super(ENet, self).__init__()
self.low_res = lnet
for param in self.low_res.parameters():
param.requires_grad = False
channel_multiplier, narrow = 2, 1
channels = {
'4': int(512 * narrow),
'8': int(512 * narrow),
'16': int(512 * narrow),
'32': int(512 * narrow),
'64': int(256 * channel_multiplier * narrow),
'128': int(128 * channel_multiplier * narrow),
'256': int(64 * channel_multiplier * narrow),
'512': int(32 * channel_multiplier * narrow),
'1024': int(16 * channel_multiplier * narrow)
}
self.log_size = 8
first_out_size = 128
self.conv_body_first = nn.Conv2d(3, channels[f'{first_out_size}'], 1) # 256 -> 128
# downsample
in_channels = channels[f'{first_out_size}']
self.conv_body_down = nn.ModuleList()
for i in range(8, 2, -1):
out_channels = channels[f'{2**(i - 1)}']
self.conv_body_down.append(ResBlock(in_channels, out_channels, mode='down'))
in_channels = out_channels
self.num_style_feat = num_style_feat
linear_out_channel = num_style_feat
self.final_linear = nn.Linear(channels['4'] * 4 * 4, linear_out_channel)
self.final_conv = nn.Conv2d(in_channels, channels['4'], 3, 1, 1)
self.style_convs = nn.ModuleList()
self.to_rgbs = nn.ModuleList()
self.noises = nn.Module()
self.concat = concat
if concat:
in_channels = 3 + 32 # channels['64']
else:
in_channels = 3
for i in range(7, 9): # 128, 256
out_channels = channels[f'{2**i}'] #
self.style_convs.append(
StyleConv(
in_channels,
out_channels,
kernel_size=3,
num_style_feat=num_style_feat,
demodulate=True,
sample_mode='upsample'))
self.style_convs.append(
StyleConv(
out_channels,
out_channels,
kernel_size=3,
num_style_feat=num_style_feat,
demodulate=True,
sample_mode=None)) | self.to_rgbs.append(ToRGB(out_channels, num_style_feat, upsample=True)) | 2 | 2023-11-02 18:25:51+00:00 | 2k |
fortelex/hiveline | hiveline/jobs/mongo.py | [
{
"identifier": "get_database",
"path": "hiveline/mongo/db.py",
"snippet": "def get_database():\n dotenv.load_dotenv()\n\n user = os.getenv(\"UP_MONGO_USER\")\n password = os.getenv(\"UP_MONGO_PASSWORD\")\n domain = os.getenv(\"UP_MONGO_DOMAIN\")\n database = os.getenv(\"UP_MONGO_DATABASE... | import datetime
import pymongo.errors
from hiveline import get_database
from hiveline.jobs.jobs import JobsDataSource, JobStatus | 1,517 |
class MongoJob:
"""
A calculation job of some sort. Used to track the status of a job. A job is uniquely identified by the key (
service_name, sim_id, job_id)
:param service_name: the name of the service
:param sim_id: the simulation ID
:param job_id: the job ID
:param status: the job status
"""
service_name: str
sim_id: str | None = None
job_id: str | None = None
status: str | None = None
created: datetime.datetime | None = None
started: datetime.datetime | None = None
finished: datetime.datetime | None = None
error: str | None = None
def __init__(self, service_name: str, sim_id: str | None = None, job_id: str | None = None,
status: str | None = None, created: datetime.datetime | None = None,
started: datetime.datetime | None = None, finished: datetime.datetime | None = None,
error: str | None = None):
self.service_name = service_name
self.sim_id = sim_id
self.job_id = job_id
self.status = status
self.created = created
self.started = started
self.finished = finished
self.error = error
def to_dict(self):
return {
"service-name": self.service_name,
"sim-id": self.sim_id,
"job-id": self.job_id,
"status": self.status,
"created": self.created,
"started": self.started,
"finished": self.finished,
"error": self.error
}
@staticmethod
def from_dict(d: dict):
return MongoJob(
service_name=d["service-name"],
sim_id=d["sim-id"],
job_id=d["job-id"],
status=d["status"],
created=d["created"],
started=d["started"],
finished=d["finished"],
error=d["error"]
)
|
class MongoJob:
"""
A calculation job of some sort. Used to track the status of a job. A job is uniquely identified by the key (
service_name, sim_id, job_id)
:param service_name: the name of the service
:param sim_id: the simulation ID
:param job_id: the job ID
:param status: the job status
"""
service_name: str
sim_id: str | None = None
job_id: str | None = None
status: str | None = None
created: datetime.datetime | None = None
started: datetime.datetime | None = None
finished: datetime.datetime | None = None
error: str | None = None
def __init__(self, service_name: str, sim_id: str | None = None, job_id: str | None = None,
status: str | None = None, created: datetime.datetime | None = None,
started: datetime.datetime | None = None, finished: datetime.datetime | None = None,
error: str | None = None):
self.service_name = service_name
self.sim_id = sim_id
self.job_id = job_id
self.status = status
self.created = created
self.started = started
self.finished = finished
self.error = error
def to_dict(self):
return {
"service-name": self.service_name,
"sim-id": self.sim_id,
"job-id": self.job_id,
"status": self.status,
"created": self.created,
"started": self.started,
"finished": self.finished,
"error": self.error
}
@staticmethod
def from_dict(d: dict):
return MongoJob(
service_name=d["service-name"],
sim_id=d["sim-id"],
job_id=d["job-id"],
status=d["status"],
created=d["created"],
started=d["started"],
finished=d["finished"],
error=d["error"]
)
| class MongoJobsDataSource(JobsDataSource): | 1 | 2023-11-07 15:34:04+00:00 | 2k |
uhppoted/uhppoted-app-home-assistant | custom_components/uhppoted/config.py | [
{
"identifier": "CONF_BIND_ADDR",
"path": "custom_components/uhppoted/const.py",
"snippet": "CONF_BIND_ADDR = 'bind_address'"
},
{
"identifier": "CONF_BROADCAST_ADDR",
"path": "custom_components/uhppoted/const.py",
"snippet": "CONF_BROADCAST_ADDR = 'broadcast_address'"
},
{
"iden... | import re
import logging
import datetime
import calendar
import uuid
from typing import Any
from uhppoted import uhppote
from .const import CONF_BIND_ADDR
from .const import CONF_BROADCAST_ADDR
from .const import CONF_LISTEN_ADDR
from .const import CONF_DEBUG
from .const import CONF_CONTROLLERS
from .const import CONF_CONTROLLER_ID
from .const import CONF_CONTROLLER_SERIAL_NUMBER
from .const import CONF_CONTROLLER_ADDR
from .const import CONF_DOORS
from .const import CONF_DOOR_ID
from .const import CONF_DOOR_CONTROLLER
from .const import CONF_DOOR_NUMBER
from .const import CONF_CARDS
from .const import CONF_CARD_UNIQUE_ID
from .const import CONF_CARD_NUMBER
from .const import CONF_CARD_NAME
from .const import CONF_CARD_STARTDATE
from .const import CONF_CARD_ENDDATE
from .const import CONF_CARD_DOORS
from .const import ERR_INVALID_CONTROLLER_ID
from .const import ERR_DUPLICATE_CONTROLLER_ID
from .const import ERR_DUPLICATE_CONTROLLER_IDS
from .const import ERR_INVALID_DOOR_ID
from .const import ERR_DUPLICATE_DOOR_ID
from .const import ERR_DUPLICATE_DOOR_IDS
from .const import ERR_INVALID_CARD_ID | 1,116 |
_LOGGER = logging.getLogger(__name__)
MAX_CARDS = 25
MAX_CARD_INDEX = 20000
MAX_ERRORS = 5
def normalise(v):
return re.sub(r'\s+', '', f'{v}', flags=re.UNICODE).lower()
def validate_controller_id(serial_no, name, options):
if not name or name.strip() == '':
raise ValueError(ERR_INVALID_CONTROLLER_ID)
if options and CONF_CONTROLLERS in options:
for v in options[CONF_CONTROLLERS]:
if normalise(v[CONF_CONTROLLER_ID]) == normalise(name):
if int(f'{v[CONF_CONTROLLER_SERIAL_NUMBER]}') != int(f'{serial_no}'):
raise ValueError(ERR_DUPLICATE_CONTROLLER_ID)
def validate_all_controllers(options):
if options and CONF_CONTROLLERS in options:
controllers = [normalise(v[CONF_CONTROLLER_ID]) for v in options[CONF_CONTROLLERS]]
if len(controllers) != len(set(controllers)):
raise ValueError(ERR_DUPLICATE_CONTROLLER_IDS)
def validate_door_id(name, options):
if not name or name.strip() == '':
raise ValueError(ERR_INVALID_DOOR_ID)
|
_LOGGER = logging.getLogger(__name__)
MAX_CARDS = 25
MAX_CARD_INDEX = 20000
MAX_ERRORS = 5
def normalise(v):
return re.sub(r'\s+', '', f'{v}', flags=re.UNICODE).lower()
def validate_controller_id(serial_no, name, options):
if not name or name.strip() == '':
raise ValueError(ERR_INVALID_CONTROLLER_ID)
if options and CONF_CONTROLLERS in options:
for v in options[CONF_CONTROLLERS]:
if normalise(v[CONF_CONTROLLER_ID]) == normalise(name):
if int(f'{v[CONF_CONTROLLER_SERIAL_NUMBER]}') != int(f'{serial_no}'):
raise ValueError(ERR_DUPLICATE_CONTROLLER_ID)
def validate_all_controllers(options):
if options and CONF_CONTROLLERS in options:
controllers = [normalise(v[CONF_CONTROLLER_ID]) for v in options[CONF_CONTROLLERS]]
if len(controllers) != len(set(controllers)):
raise ValueError(ERR_DUPLICATE_CONTROLLER_IDS)
def validate_door_id(name, options):
if not name or name.strip() == '':
raise ValueError(ERR_INVALID_DOOR_ID)
| if name.strip() != '-' and options and CONF_DOORS in options: | 8 | 2023-11-06 18:46:49+00:00 | 2k |
kyegomez/HeptapodLM | train.py | [
{
"identifier": "Autoregressive2DWrapper",
"path": "heptapod/at.py",
"snippet": "class Autoregressive2DWrapper(nn.Module):\n def __init__(self, net, matrix_size=32, pad_value=0):\n super().__init__()\n self.matrix_size = matrix_size\n self.pad_value = pad_value\n self.net ... | import gzip
import random
import numpy as np
import torch
import torch.optim as optim
import tqdm
from torch.utils.data import DataLoader, Dataset
from heptapod.at import Autoregressive2DWrapper
from heptapod.model import NonLinearTransformer | 756 |
# Constants
NUM_BATCHES = int(1e5)
BATCH_SIZE = 4
GRADIENT_ACCUMULATE_EVERY = 4
LEARNING_RATE = 2e-4
VALIDATE_EVERY = 100
GENERATE_EVERY = 500
GENERATE_LENGTH = 512
SEQ_LEN = 1024
# Helpers
def cycle(loader):
while True:
for data in loader:
yield data
def decode_token(token):
return str(chr(max(32, token)))
def decode_tokens(tokens):
return "".join(list(map(decode_token, tokens)))
# Instantiate GPT-like decoder model
|
# Constants
NUM_BATCHES = int(1e5)
BATCH_SIZE = 4
GRADIENT_ACCUMULATE_EVERY = 4
LEARNING_RATE = 2e-4
VALIDATE_EVERY = 100
GENERATE_EVERY = 500
GENERATE_LENGTH = 512
SEQ_LEN = 1024
# Helpers
def cycle(loader):
while True:
for data in loader:
yield data
def decode_token(token):
return str(chr(max(32, token)))
def decode_tokens(tokens):
return "".join(list(map(decode_token, tokens)))
# Instantiate GPT-like decoder model | model = NonLinearTransformer(vocab_size=10000, dim=512, depth=6, matrix_dim=5) | 1 | 2023-11-01 06:07:50+00:00 | 2k |
shixiaoyu0216/SAC4IR | sacd/memory/per.py | [
{
"identifier": "LazyMultiStepMemory",
"path": "sacd/memory/base.py",
"snippet": "class LazyMultiStepMemory(LazyMemory):\n\n def __init__(self, capacity, state_shape, device, gamma=0.99,\n multi_step=3):\n super(LazyMultiStepMemory, self).__init__(\n capacity, state_... | import numpy as np
import torch
from .base import LazyMultiStepMemory
from .segment_tree import SumTree, MinTree | 740 |
class LazyPrioritizedMultiStepMemory(LazyMultiStepMemory):
def __init__(self, capacity, state_shape, device, gamma=0.99,
multi_step=3, alpha=0.6, beta=0.4, beta_steps=2e5,
min_pa=0.0, max_pa=1.0, eps=0.01):
super().__init__(capacity, state_shape, device, gamma, multi_step)
self.alpha = alpha
self.beta = beta
self.beta_diff = (1.0 - beta) / beta_steps
self.min_pa = min_pa
self.max_pa = max_pa
self.eps = eps
self._cached = None
it_capacity = 1
while it_capacity < capacity:
it_capacity *= 2
self.it_sum = SumTree(it_capacity)
|
class LazyPrioritizedMultiStepMemory(LazyMultiStepMemory):
def __init__(self, capacity, state_shape, device, gamma=0.99,
multi_step=3, alpha=0.6, beta=0.4, beta_steps=2e5,
min_pa=0.0, max_pa=1.0, eps=0.01):
super().__init__(capacity, state_shape, device, gamma, multi_step)
self.alpha = alpha
self.beta = beta
self.beta_diff = (1.0 - beta) / beta_steps
self.min_pa = min_pa
self.max_pa = max_pa
self.eps = eps
self._cached = None
it_capacity = 1
while it_capacity < capacity:
it_capacity *= 2
self.it_sum = SumTree(it_capacity) | self.it_min = MinTree(it_capacity) | 2 | 2023-11-02 07:35:57+00:00 | 2k |
In-Network-Machine-Learning/QCMP | initiate_rules.py | [
{
"identifier": "init_path_weights",
"path": "q_table.py",
"snippet": "def init_path_weights(p4info_helper, ingress_sw, nhop_dmacs, nhop_ipv4s, ports):\n for i in range(50):\n write_path_weights(p4info_helper, ingress_sw=ingress_sw, value=i,\n nhop_dmac=nhop_dmacs[0], nhop_ipv4=nhop... | import sys
import argparse
import os
import pandas as pd
import grpc
import p4runtime_lib.bmv2
import p4runtime_lib.helper
from scapy.all import *
from scapy.layers.inet import _IPOption_HDR
from p4runtime_lib.error_utils import printGrpcError
from p4runtime_lib.switch import ShutdownAllSwitchConnections
from q_table import (init_path_weights, readTableRules) | 951 | # This file is part of the Planter extend project: QCMP.
# This program is a free software tool, which does ensemble in-network reinforcement learning for load balancing.
# licensed under Apache-2.0
#
# Utility: This file is used to initiate rules in the q-table
#
# Copyright (c) 2022-2023 Benjamin Rienecker Modified by Changgang Zheng
# Copyright (c) Computing Infrastructure Group, Department of Engineering Science, University of Oxford
#!/usr/bin/env python3
# Import P4Runtime lib from parent utils dir
# Probably there's a better way of doing this.
sys.path.append(
os.path.join(os.path.dirname(os.path.abspath(__file__)),
'../utils/'))
def main(p4info_file_path, bmv2_file_path):
# Instantiate a P4Runtime helper from the p4info file
p4info_helper = p4runtime_lib.helper.P4InfoHelper(p4info_file_path)
# Create a switch connection object for all switches;
# this is backed by a P4Runtime gRPC connection.
# Also, dump all P4Runtime messages sent to switch to given txt files.
s1 = p4runtime_lib.bmv2.Bmv2SwitchConnection(
name='s1',
address='127.0.0.1:50051',
device_id=0,
proto_dump_file='logs/s1-p4runtime-requests.txt')
s2 = p4runtime_lib.bmv2.Bmv2SwitchConnection(
name='s2',
address='127.0.0.1:50052',
device_id=1,
proto_dump_file='logs/s2-p4runtime-requests.txt')
s3 = p4runtime_lib.bmv2.Bmv2SwitchConnection(
name='s3',
address='127.0.0.1:50053',
device_id=2,
proto_dump_file='logs/s3-p4runtime-requests.txt')
# Send master arbitration update message to establish this controller as
# master (required by P4Runtime before performing any other write operation)
s1.MasterArbitrationUpdate()
s2.MasterArbitrationUpdate()
s3.MasterArbitrationUpdate()
nhop_dmacs = ["00:00:00:00:01:04", "00:00:00:00:01:05"]
nhop_ipv4s = ["10.0.2.0", "10.0.3.0"]
ports = [4, 5]
| # This file is part of the Planter extend project: QCMP.
# This program is a free software tool, which does ensemble in-network reinforcement learning for load balancing.
# licensed under Apache-2.0
#
# Utility: This file is used to initiate rules in the q-table
#
# Copyright (c) 2022-2023 Benjamin Rienecker Modified by Changgang Zheng
# Copyright (c) Computing Infrastructure Group, Department of Engineering Science, University of Oxford
#!/usr/bin/env python3
# Import P4Runtime lib from parent utils dir
# Probably there's a better way of doing this.
sys.path.append(
os.path.join(os.path.dirname(os.path.abspath(__file__)),
'../utils/'))
def main(p4info_file_path, bmv2_file_path):
# Instantiate a P4Runtime helper from the p4info file
p4info_helper = p4runtime_lib.helper.P4InfoHelper(p4info_file_path)
# Create a switch connection object for all switches;
# this is backed by a P4Runtime gRPC connection.
# Also, dump all P4Runtime messages sent to switch to given txt files.
s1 = p4runtime_lib.bmv2.Bmv2SwitchConnection(
name='s1',
address='127.0.0.1:50051',
device_id=0,
proto_dump_file='logs/s1-p4runtime-requests.txt')
s2 = p4runtime_lib.bmv2.Bmv2SwitchConnection(
name='s2',
address='127.0.0.1:50052',
device_id=1,
proto_dump_file='logs/s2-p4runtime-requests.txt')
s3 = p4runtime_lib.bmv2.Bmv2SwitchConnection(
name='s3',
address='127.0.0.1:50053',
device_id=2,
proto_dump_file='logs/s3-p4runtime-requests.txt')
# Send master arbitration update message to establish this controller as
# master (required by P4Runtime before performing any other write operation)
s1.MasterArbitrationUpdate()
s2.MasterArbitrationUpdate()
s3.MasterArbitrationUpdate()
nhop_dmacs = ["00:00:00:00:01:04", "00:00:00:00:01:05"]
nhop_ipv4s = ["10.0.2.0", "10.0.3.0"]
ports = [4, 5] | init_path_weights(p4info_helper, s1, nhop_dmacs, nhop_ipv4s, ports) | 0 | 2023-11-01 09:37:28+00:00 | 2k |
Fsoft-AIC/LSDM | vis_fitting_results.py | [
{
"identifier": "gen_human_meshes",
"path": "gen_human_meshes.py",
"snippet": "def gen_human_meshes(vertices_path, output_path):\n vertices = np.load(open(vertices_path, \"rb\"))\n # If your input human vertices are full resolution SMPL-X bodies, use mesh_0.obj\n # faces = trimesh.load(os.path.... | import os
import numpy as np
import argparse
import open3d as o3d
import json
from pathlib import Path
from gen_human_meshes import gen_human_meshes, gen_human_meshes_humanise
from tqdm import tqdm | 723 |
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="")
parser.add_argument("--fitting_results_path", type=str, help="Path to the fitting results of some motion sequence")
parser.add_argument("--vertices_path", type=str, help="Path to human vertices of some motion sequence")
parser.add_argument("--datatype", default="PROXD", type=str, help="Data type")
args = parser.parse_args()
input_dir = Path(args.fitting_results_path)
vertices_path = Path(args.vertices_path)
seq_name = input_dir.stem
# Check if human meshes are there
human_mesh_dir = input_dir / 'human' / 'mesh'
if not human_mesh_dir.exists():
human_mesh_dir.mkdir()
if args.datatype == "PROXD":
gen_human_meshes(vertices_path=vertices_path, output_path=human_mesh_dir)
else:
body_faces = np.load(open(args.vertices_path[:-4] + "_faces.npy", "rb"))
|
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="")
parser.add_argument("--fitting_results_path", type=str, help="Path to the fitting results of some motion sequence")
parser.add_argument("--vertices_path", type=str, help="Path to human vertices of some motion sequence")
parser.add_argument("--datatype", default="PROXD", type=str, help="Data type")
args = parser.parse_args()
input_dir = Path(args.fitting_results_path)
vertices_path = Path(args.vertices_path)
seq_name = input_dir.stem
# Check if human meshes are there
human_mesh_dir = input_dir / 'human' / 'mesh'
if not human_mesh_dir.exists():
human_mesh_dir.mkdir()
if args.datatype == "PROXD":
gen_human_meshes(vertices_path=vertices_path, output_path=human_mesh_dir)
else:
body_faces = np.load(open(args.vertices_path[:-4] + "_faces.npy", "rb")) | gen_human_meshes_humanise(vertices_path, body_faces, output_path=human_mesh_dir) | 1 | 2023-11-06 07:55:51+00:00 | 2k |
molML/traversing_chem_space | active_learning/data_prep.py | [
{
"identifier": "molecular_graph_featurizer",
"path": "active_learning/utils.py",
"snippet": "def molecular_graph_featurizer(smiles: str, y=None, structural_feats: bool = True, functional_feats: bool = True):\n\n y = torch.tensor([y]).to(torch.long)\n\n mol = Chem.MolFromSmiles(smiles, sanitize=Tr... | from active_learning.utils import molecular_graph_featurizer as smiles_to_graph
from active_learning.utils import smiles_to_ecfp, get_tanimoto_matrix, check_featurizability
from collections import OrderedDict
from rdkit.Chem.Scaffolds import MurckoScaffold
from rdkit import Chem
from tqdm import tqdm
from typing import Any
from config import ROOT_DIR
from sklearn.model_selection import train_test_split
import pandas as pd
import numpy as np
import torch
import os
import sys
import h5py
import h5py
import h5py | 1,588 |
def canonicalize(smiles: str, sanitize: bool = True):
return Chem.MolToSmiles(Chem.MolFromSmiles(smiles, sanitize=sanitize))
def get_data(random_state: int = 42, dataset: str = 'ALDH1'):
# read smiles from file and canonicalize them
with open(os.path.join(ROOT_DIR, f'data/{dataset}/original/inactives.smi')) as f:
inactives = [canonicalize(smi.strip().split()[0]) for smi in f.readlines()]
with open(os.path.join(ROOT_DIR, f'data/{dataset}/original/actives.smi')) as f:
actives = [canonicalize(smi.strip().split()[0]) for smi in f.readlines()]
# remove duplicates:
inactives = list(set(inactives))
actives = list(set(actives))
# remove intersecting molecules:
intersecting_mols = np.intersect1d(inactives, actives)
inactives = [smi for smi in inactives if smi not in intersecting_mols]
actives = [smi for smi in actives if smi not in intersecting_mols]
# remove molecules that have scaffolds that cannot be kekulized or featurized
inactives_, actives_ = [], []
for smi in tqdm(actives):
try:
if Chem.MolFromSmiles(smi_to_scaff(smi, includeChirality=False)) is not None:
|
def canonicalize(smiles: str, sanitize: bool = True):
return Chem.MolToSmiles(Chem.MolFromSmiles(smiles, sanitize=sanitize))
def get_data(random_state: int = 42, dataset: str = 'ALDH1'):
# read smiles from file and canonicalize them
with open(os.path.join(ROOT_DIR, f'data/{dataset}/original/inactives.smi')) as f:
inactives = [canonicalize(smi.strip().split()[0]) for smi in f.readlines()]
with open(os.path.join(ROOT_DIR, f'data/{dataset}/original/actives.smi')) as f:
actives = [canonicalize(smi.strip().split()[0]) for smi in f.readlines()]
# remove duplicates:
inactives = list(set(inactives))
actives = list(set(actives))
# remove intersecting molecules:
intersecting_mols = np.intersect1d(inactives, actives)
inactives = [smi for smi in inactives if smi not in intersecting_mols]
actives = [smi for smi in actives if smi not in intersecting_mols]
# remove molecules that have scaffolds that cannot be kekulized or featurized
inactives_, actives_ = [], []
for smi in tqdm(actives):
try:
if Chem.MolFromSmiles(smi_to_scaff(smi, includeChirality=False)) is not None: | if check_featurizability(smi): | 3 | 2023-11-10 08:53:40+00:00 | 2k |
yunik1004/SAiD | script/render.py | [
{
"identifier": "load_mesh",
"path": "said/util/mesh.py",
"snippet": "def load_mesh(mesh_path: str) -> trimesh.Trimesh:\n \"\"\"Load the mesh\n\n Parameters\n ----------\n filepath : str\n Path of the mesh file\n\n Returns\n -------\n trimesh.Trimesh\n Mesh object\n ... | import argparse
import os
import pathlib
import cv2
import numpy as np
from moviepy import editor as mpy
from said.util.mesh import load_mesh
from said.util.parser import parse_list
from said.util.blendshape import load_blendshape_coeffs
from rendering.render_visual import RendererObject, render_blendshape_coefficients | 1,256 | """Render the animation
"""
os.environ["PYOPENGL_PLATFORM"] = "egl"
def main() -> None:
"""Main function"""
default_data_dir = pathlib.Path(__file__).resolve().parent.parent / "data"
# Arguments
parser = argparse.ArgumentParser(description="Render the animation")
parser.add_argument(
"--neutral_path",
type=str,
default="../BlendVOCA/templates_head/FaceTalk_170731_00024_TA.obj",
help="Path of the neutral mesh",
)
parser.add_argument(
"--blendshapes_dir",
type=str,
default="../BlendVOCA/blendshapes_head/FaceTalk_170731_00024_TA",
help="Directory of the blendshape meshes",
)
parser.add_argument(
"--audio_path",
type=str,
default="../BlendVOCA/audio/FaceTalk_170731_00024_TA/sentence01.wav",
help="Path of the audio file",
)
parser.add_argument(
"--blendshape_coeffs_path",
type=str,
default="../BlendVOCA/blendshape_coeffs/FaceTalk_170731_00024_TA/sentence01.csv",
help="Path of the blendshape coefficient sequence",
)
parser.add_argument(
"--blendshape_list_path",
type=str,
default=default_data_dir / "ARKit_blendshapes.txt",
help="List of the blendshapes",
)
parser.add_argument(
"--show_difference",
type=bool,
default=False,
help="Show the vertex differences from the target blendshape coefficients as a heatmap",
)
parser.add_argument(
"--target_diff_blendshape_coeffs_path",
type=str,
default="../BlendVOCA/blendshape_coeffs/FaceTalk_170731_00024_TA/sentence01.csv",
help="Path of the target blendshape coefficient sequence to compute the vertex differences. Its length should be same as the source's.",
)
parser.add_argument(
"--max_diff",
type=float,
default=0.001,
help="Maximum threshold to visualize the vertex differences",
)
parser.add_argument(
"--fps",
type=int,
default=60,
help="FPS of the blendshape coefficients sequence",
)
parser.add_argument(
"--output_path",
type=str,
default="../out.mp4",
help="Path of the output video file",
)
parser.add_argument(
"--save_images",
type=bool,
default=False,
help="Save the image for each frame",
)
parser.add_argument(
"--output_images_dir",
type=str,
default="../out_imgs",
help="Saving directory of the output image for each frame",
)
args = parser.parse_args()
neutral_path = args.neutral_path
blendshapes_dir = args.blendshapes_dir
audio_path = args.audio_path
blendshape_coeffs_path = args.blendshape_coeffs_path
blendshape_list_path = args.blendshape_list_path
show_difference = args.show_difference
target_diff_blendshape_coeffs_path = args.target_diff_blendshape_coeffs_path
max_diff = args.max_diff
fps = args.fps
output_path = args.output_path
save_images = args.save_images
output_images_dir = args.output_images_dir
blendshape_name_list = parse_list(blendshape_list_path, str)
# Create renderer
renderer = RendererObject()
neutral_mesh = load_mesh(neutral_path)
blendshape_vectors = []
for bl_name in blendshape_name_list:
bl_path = os.path.join(blendshapes_dir, f"{bl_name}.obj")
bl_mesh = load_mesh(bl_path)
blendshape_vectors.append(bl_mesh.vertices.reshape((-1, 1)))
blendshapes_matrix = np.concatenate(blendshape_vectors, axis=1)
| """Render the animation
"""
os.environ["PYOPENGL_PLATFORM"] = "egl"
def main() -> None:
"""Main function"""
default_data_dir = pathlib.Path(__file__).resolve().parent.parent / "data"
# Arguments
parser = argparse.ArgumentParser(description="Render the animation")
parser.add_argument(
"--neutral_path",
type=str,
default="../BlendVOCA/templates_head/FaceTalk_170731_00024_TA.obj",
help="Path of the neutral mesh",
)
parser.add_argument(
"--blendshapes_dir",
type=str,
default="../BlendVOCA/blendshapes_head/FaceTalk_170731_00024_TA",
help="Directory of the blendshape meshes",
)
parser.add_argument(
"--audio_path",
type=str,
default="../BlendVOCA/audio/FaceTalk_170731_00024_TA/sentence01.wav",
help="Path of the audio file",
)
parser.add_argument(
"--blendshape_coeffs_path",
type=str,
default="../BlendVOCA/blendshape_coeffs/FaceTalk_170731_00024_TA/sentence01.csv",
help="Path of the blendshape coefficient sequence",
)
parser.add_argument(
"--blendshape_list_path",
type=str,
default=default_data_dir / "ARKit_blendshapes.txt",
help="List of the blendshapes",
)
parser.add_argument(
"--show_difference",
type=bool,
default=False,
help="Show the vertex differences from the target blendshape coefficients as a heatmap",
)
parser.add_argument(
"--target_diff_blendshape_coeffs_path",
type=str,
default="../BlendVOCA/blendshape_coeffs/FaceTalk_170731_00024_TA/sentence01.csv",
help="Path of the target blendshape coefficient sequence to compute the vertex differences. Its length should be same as the source's.",
)
parser.add_argument(
"--max_diff",
type=float,
default=0.001,
help="Maximum threshold to visualize the vertex differences",
)
parser.add_argument(
"--fps",
type=int,
default=60,
help="FPS of the blendshape coefficients sequence",
)
parser.add_argument(
"--output_path",
type=str,
default="../out.mp4",
help="Path of the output video file",
)
parser.add_argument(
"--save_images",
type=bool,
default=False,
help="Save the image for each frame",
)
parser.add_argument(
"--output_images_dir",
type=str,
default="../out_imgs",
help="Saving directory of the output image for each frame",
)
args = parser.parse_args()
neutral_path = args.neutral_path
blendshapes_dir = args.blendshapes_dir
audio_path = args.audio_path
blendshape_coeffs_path = args.blendshape_coeffs_path
blendshape_list_path = args.blendshape_list_path
show_difference = args.show_difference
target_diff_blendshape_coeffs_path = args.target_diff_blendshape_coeffs_path
max_diff = args.max_diff
fps = args.fps
output_path = args.output_path
save_images = args.save_images
output_images_dir = args.output_images_dir
blendshape_name_list = parse_list(blendshape_list_path, str)
# Create renderer
renderer = RendererObject()
neutral_mesh = load_mesh(neutral_path)
blendshape_vectors = []
for bl_name in blendshape_name_list:
bl_path = os.path.join(blendshapes_dir, f"{bl_name}.obj")
bl_mesh = load_mesh(bl_path)
blendshape_vectors.append(bl_mesh.vertices.reshape((-1, 1)))
blendshapes_matrix = np.concatenate(blendshape_vectors, axis=1) | blendshape_coeffs = load_blendshape_coeffs(blendshape_coeffs_path).numpy() | 2 | 2023-11-03 06:38:51+00:00 | 2k |
Subsets and Splits
SQL Console for tianyang/repobench_python_v1.1
Identifies repositories that have consistent code formatting levels across multiple scales (2k, 4k, 8k, 12k) and reveals the structured formatting patterns within these repositories.
SQL Console for tianyang/repobench_python_v1.1
Compares cross-file and in-file code structure patterns across different complexity levels, revealing how file organization strategies vary with code size and potentially informing better code architecture decisions.
SQL Console for tianyang/repobench_python_v1.1
Identifies repositories that have complete performance data across all seven code complexity levels, revealing consistent benchmarking patterns across different code sizes.
SQL Console for tianyang/repobench_python_v1.1
Identifies repositories that contain all 7 distinct quality levels (2k through 32k), revealing complete datasets that might be useful for comprehensive analysis.