id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
143,577
def only_until_first_boxed_from_tokens(string, tokens): idx = string.find("\\boxed") if idx < 0: idx = string.find("\\fbox") if idx < 0: return None cum_length = 0 for i, t in enumerate(tokens): cum_length += len(t) if cum_length >= idx: bre...
null
143,578
def _clean_numbers(string): """ Clean Numbers in the given string >>> _clean_numbers(None, "Hello 123") 'Hello 123' >>> _clean_numbers(None, "Hello 1234") 'Hello 1,234' >>> _clean_numbers(None, "Hello 1234324asdasd") 'Hello 1,234,324asdasd' """ num_prev_digits = 0 new_string ...
null
143,579
import sys import os from model.meta import MetaModel import argparse import torch from PIL import Image from util import misc from fairscale.nn.model_parallel import initialize as fs_init from data.alpaca import transform_val from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quant...
null
143,580
import sys import os from model.meta import MetaModel import argparse import torch from PIL import Image from util import misc from fairscale.nn.model_parallel import initialize as fs_init from data.alpaca import transform_val from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quant...
null
143,581
import sys import os from model.meta import MetaModel import argparse import torch from PIL import Image from util import misc from fairscale.nn.model_parallel import initialize as fs_init from data.alpaca import transform_val from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quant...
null
143,582
import sys import os from model.meta import MetaModel import argparse import torch from PIL import Image from util import misc from fairscale.nn.model_parallel import initialize as fs_init from data.alpaca import transform_val from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quant...
null
143,583
import argparse import numpy as np import torch import pandas as pd from tqdm import tqdm import json import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant...
null
143,584
import argparse import numpy as np import torch import pandas as pd from tqdm import tqdm import json import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant...
null
143,585
import argparse import numpy as np import torch import pandas as pd from tqdm import tqdm import json import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant...
null
143,586
import argparse import numpy as np import torch import pandas as pd from tqdm import tqdm import json import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant...
null
143,587
import argparse import json import jsonlines import re from tqdm import tqdm import torch import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quan...
null
143,588
import argparse import json import jsonlines import re from tqdm import tqdm import torch import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quan...
null
143,589
import argparse import json import jsonlines import re from tqdm import tqdm import torch import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quan...
null
143,590
import argparse import json import jsonlines import re from tqdm import tqdm import torch import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quan...
null
143,591
import torch import numpy as np import argparse from collections import defaultdict from tqdm import tqdm import pandas as pd import json import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor...
null
143,592
import torch import numpy as np import argparse from collections import defaultdict from tqdm import tqdm import pandas as pd import json import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor...
null
143,593
import torch import numpy as np import argparse from collections import defaultdict from tqdm import tqdm import pandas as pd import json import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor...
null
143,594
import torch import numpy as np import argparse from collections import defaultdict from tqdm import tqdm import pandas as pd import json import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor...
null
143,595
import argparse import torch from tqdm import tqdm import json import jsonlines import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quantize from...
null
143,596
import argparse import torch from tqdm import tqdm import json import jsonlines import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quantize from...
null
143,597
import argparse import torch from tqdm import tqdm import json import jsonlines import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quantize from...
Evaluates the functional correctness of generated samples, and writes results to f"{sample_file}_results.jsonl.gz"
143,598
import argparse import torch from tqdm import tqdm import json import jsonlines import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quantize from...
null
143,599
import argparse import torch from tqdm import tqdm import json import jsonlines import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from util.quant import quantize from...
null
143,600
import sys import os from model.meta import MetaModel import argparse import torch import openai import pandas as pd import numpy as np from collections import Counter import time from PIL import Image from util import misc from fairscale.nn.model_parallel import initialize as fs_init from data.alpaca import transform_...
null
143,601
import sys import os from model.meta import MetaModel import argparse import torch import openai import pandas as pd import numpy as np from collections import Counter import time from PIL import Image from util import misc from fairscale.nn.model_parallel import initialize as fs_init from data.alpaca import transform_...
null
143,602
import sys import os from model.meta import MetaModel import argparse import torch import openai import pandas as pd import numpy as np from collections import Counter import time from PIL import Image from util import misc from fairscale.nn.model_parallel import initialize as fs_init from data.alpaca import transform_...
null
143,603
import sys import os from model.meta import MetaModel import argparse import torch import openai import pandas as pd import numpy as np from collections import Counter import time from PIL import Image from util import misc from fairscale.nn.model_parallel import initialize as fs_init from data.alpaca import transform_...
null
143,604
import argparse import jsonlines import json import time import pandas as pd from tqdm import tqdm import torch import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from ...
null
143,605
import argparse import jsonlines import json import time import pandas as pd from tqdm import tqdm import torch import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from ...
null
143,606
import argparse import jsonlines import json import time import pandas as pd from tqdm import tqdm import torch import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from ...
null
143,607
import argparse import jsonlines import json import time import pandas as pd from tqdm import tqdm import torch import os import sys from model.meta import MetaModel from util import misc from fairscale.nn.model_parallel import initialize as fs_init from util.tensor_parallel import load_tensor_parallel_model_list from ...
null
143,608
import argparse from collections import OrderedDict import contextlib from copy import deepcopy from datetime import datetime import functools import json import logging import os import random import socket from time import time from PIL import Image from diffusers.models import AutoencoderKL import fairscale.nn.model...
null
143,609
import argparse from collections import OrderedDict import contextlib from copy import deepcopy from datetime import datetime import functools import json import logging import os import random import socket from time import time from PIL import Image from diffusers.models import AutoencoderKL import fairscale.nn.model...
null
143,610
import argparse from collections import OrderedDict import contextlib from copy import deepcopy from datetime import datetime import functools import json import logging import os import random import socket from time import time from PIL import Image from diffusers.models import AutoencoderKL import fairscale.nn.model...
Step the EMA model towards the current model.
143,611
import argparse from collections import OrderedDict import contextlib from copy import deepcopy from datetime import datetime import functools import json import logging import os import random import socket from time import time from PIL import Image from diffusers.models import AutoencoderKL import fairscale.nn.model...
End DDP training.
143,612
import argparse from collections import OrderedDict import contextlib from copy import deepcopy from datetime import datetime import functools import json import logging import os import random import socket from time import time from PIL import Image from diffusers.models import AutoencoderKL import fairscale.nn.model...
Create a logger that writes to a log file and stdout.
143,613
import argparse from collections import OrderedDict import contextlib from copy import deepcopy from datetime import datetime import functools import json import logging import os import random import socket from time import time from PIL import Image from diffusers.models import AutoencoderKL import fairscale.nn.model...
null
143,614
import argparse from collections import OrderedDict import contextlib from copy import deepcopy from datetime import datetime import functools import json import logging import os import random import socket from time import time from PIL import Image from diffusers.models import AutoencoderKL import fairscale.nn.model...
null
143,615
import argparse from collections import OrderedDict import contextlib from copy import deepcopy from datetime import datetime import functools import json import logging import os import random import socket from time import time from PIL import Image from diffusers.models import AutoencoderKL import fairscale.nn.model...
null
143,616
from PIL import Image import numpy as np import random The provided code snippet includes necessary dependencies for implementing the `center_crop_arr` function. Write a Python function `def center_crop_arr(pil_image, image_size)` to solve the following problem: Center cropping implementation from ADM. https://github....
Center cropping implementation from ADM. https://github.com/openai/guided-diffusion/blob/8fb3ad9197f16bbc40620447b2742e13458d2831/guided_diffusion/image_datasets.py#L126
143,617
from PIL import Image import numpy as np import random def center_crop(pil_image, crop_size): while pil_image.size[0] >= 2 * crop_size[0] and pil_image.size[1] >= 2 * crop_size[1]: pil_image = pil_image.resize( tuple(x // 2 for x in pil_image.size), resample=Image.BOX ) scale = max(c...
null
143,618
from PIL import Image import numpy as np import random def generate_crop_size_list(num_patches, patch_size, max_ratio=4.0): assert max_ratio >= 1. crop_size_list = [] wp, hp = num_patches, 1 while wp > 0: if max(wp, hp) / min(wp, hp) <= max_ratio: crop_size_list.append((wp * patch_s...
null
143,619
from io import BytesIO import logging import time from typing import Union from PIL import Image def init_ceph_client_if_needed(): global client if client is None: logger.info(f"initializing ceph client ...") st = time.time() from petrel_client.client import Client # noqa client...
null
143,620
import os import subprocess from time import sleep import torch import torch.distributed as dist import fairscale.nn.model_parallel.initialize as fs_init def _setup_dist_env_from_slurm(args): while not os.environ.get("MASTER_ADDR", ""): os.environ["MASTER_ADDR"] = subprocess.check_output( "sinfo...
null
143,629
import argparse import json import multiprocessing as mp import os import socket from diffusers.models import AutoencoderKL import fairscale.nn.model_parallel.initialize as fs_init import gradio as gr import torch import torch.distributed as dist from torchvision.transforms.functional import to_pil_image from transform...
null
143,630
import argparse import json import multiprocessing as mp import os import socket from diffusers.models import AutoencoderKL import fairscale.nn.model_parallel.initialize as fs_init import gradio as gr import torch import torch.distributed as dist from torchvision.transforms.functional import to_pil_image from transform...
null
143,635
import functools import math from typing import Optional, Tuple, List from apex.normalization import FusedRMSNorm as RMSNorm import fairscale.nn.model_parallel.initialize as fs_init from fairscale.nn.model_parallel.layers import ( ColumnParallelLinear, RowParallelLinear, ParallelEmbedding, ) from flash_attn import ...
null
143,636
import functools import math from typing import Optional, Tuple, List from apex.normalization import FusedRMSNorm as RMSNorm import fairscale.nn.model_parallel.initialize as fs_init from fairscale.nn.model_parallel.layers import ( ColumnParallelLinear, RowParallelLinear, ParallelEmbedding, ) from flash_attn import ...
null
143,637
import functools import math from typing import Optional, Tuple, List from apex.normalization import FusedRMSNorm as RMSNorm import fairscale.nn.model_parallel.initialize as fs_init from fairscale.nn.model_parallel.layers import ( ColumnParallelLinear, RowParallelLinear, ParallelEmbedding, ) from flash_attn import ...
null
143,638
import sys import os import torch import torch.distributed as dist import argparse import pandas as pd from typing import Tuple, List from PIL import Image import fairscale.nn.model_parallel.initialize as fs_init from accessory.model.meta import MetaModel from accessory.data.conversation import default_conversation fr...
null
143,639
import os import subprocess import sys def open_file(filename): if sys.platform == "win32": os.startfile(filename) # noqa else: opener = "open" if sys.platform == "darwin" else "xdg-open" subprocess.call([opener, filename])
null
143,640
import logging import sys from datetime import datetime from enum import Enum from logging.config import dictConfig from freemocap.system.paths_and_filenames.path_getters import get_log_file_path def ensure_min_brightness(value, threshold=50): """Ensure the RGB value is above a certain threshold.""" return max(...
Generate a consistent random color for the given value.
143,641
import logging import sys from datetime import datetime from enum import Enum from logging.config import dictConfig from freemocap.system.paths_and_filenames.path_getters import get_log_file_path logging.getLogger("tzlocal").setLevel(logging.WARNING) logging.getLogger("matplotlib").setLevel(logging.WARNING) class LogLe...
null
143,642
import logging import sys from datetime import datetime from enum import Enum from logging.config import dictConfig from freemocap.system.paths_and_filenames.path_getters import get_log_file_path def log_test_messages(logger): logger.trace("This is a TRACE message.") logger.debug("This is a DEBUG message.") ...
null
143,643
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,644
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,645
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,646
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,647
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,648
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,649
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,650
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,651
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,652
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,653
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,654
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,655
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,656
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,657
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,658
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,659
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,660
import time from datetime import datetime from pathlib import Path from typing import Union import toml from freemocap.system.paths_and_filenames.file_and_folder_names import ( LOGS_INFO_AND_SETTINGS_FOLDER_NAME, LOG_FILE_FOLDER_NAME, BASE_FREEMOCAP_DATA_FOLDER_NAME, CALIBRATIONS_FOLDER_NAME, logger...
null
143,661
import logging from pathlib import Path from typing import Optional, Union from freemocap.core_processes.export_data.blender_stuff.export_to_blender.export_to_blender import export_to_blender from freemocap.core_processes.export_data.blender_stuff.get_best_guess_of_blender_path import ( get_best_guess_of_blender_pa...
null
143,662
import logging from pathlib import Path from typing import Optional, Union from freemocap.core_processes.export_data.blender_stuff.export_to_blender.export_to_blender import export_to_blender from freemocap.core_processes.export_data.blender_stuff.get_best_guess_of_blender_path import ( get_best_guess_of_blender_pa...
null
143,663
import sys from pathlib import Path import bpy import numpy as np def get_unique_list(list: list) -> list: """Return a list of the unique elements from input list""" unique_list = [] [unique_list.append(clip) for clip in list if clip not in unique_list] return unique_list The provided code snippet incl...
Search the folder for 'mp4' files (case insensitive) and return them as a list
143,664
import json import logging import sys from copy import copy from pathlib import Path from typing import List, Union, Dict, Tuple import addon_utils import bpy import numpy as np print("Running script to create Blender file from freemocap session data: " + __file__) def get_video_paths(path_to_video_folder: Path) -> lis...
null
143,665
import json import logging import sys from copy import copy from pathlib import Path from typing import List, Union, Dict, Tuple import addon_utils import bpy import numpy as np print("Running script to create Blender file from freemocap session data: " + __file__) print(f"__name__: {__name__}") The provided code snip...
Create a key framed empty from 3d trajectory data
143,666
import json import logging import sys from copy import copy from pathlib import Path from typing import List, Union, Dict, Tuple import addon_utils import bpy import numpy as np def create_material( name: str = "Generic", color: Union[str, Tuple, List] = "#00FFFF", emission_strength: float = 1.0, ): """...
Put uv sphere meshes on the empties in `empty_names_list` with a scale of `sphere_scale`. :param empty_names_list: List of empty object names. :param parent_object: The object to be set as parent for the created spheres. :param sphere_scale: The scale of the sphere. :param material: The material to be applied to the sp...
143,667
import json import logging import sys from copy import copy from pathlib import Path from typing import List, Union, Dict, Tuple import addon_utils import bpy import numpy as np def create_stick_mesh( vertices: int = 12, radius: float = 0.02, depth: float = 1.0, location=(0, 0, 0), rotation=(0, 0, 0...
make cylinder meshes between the empties as defined in `body_connections`
143,668
import json import logging import sys from copy import copy from pathlib import Path from typing import List, Union, Dict, Tuple import addon_utils import bpy import numpy as np print("Running script to create Blender file from freemocap session data: " + __file__) print(f"__name__: {__name__}") The provided code snip...
Apply constraints to a bone (that exists in `armature_rig` based on a dictionary of constraints
143,669
import json import logging import sys from copy import copy from pathlib import Path from typing import List, Union, Dict, Tuple import addon_utils import bpy import numpy as np The provided code snippet includes necessary dependencies for implementing the `test_virtual_marker_definitions` function. Write a Python fun...
Validate the virtual marker definitions dictionary to ensure that there are the same number of marker names and weights, and that the weights sum to 1
143,670
import json import logging import sys from copy import copy from pathlib import Path from typing import List, Union, Dict, Tuple import addon_utils import bpy import numpy as np The provided code snippet includes necessary dependencies for implementing the `calculate_virtual_marker_trajectory` function. Write a Python...
Create a virtual marker from a set of component markers. A 'Virtual Marker' is a 'fake' marker created by combining the data from 'real' (measured) marker/trajectory data `trajectory_3d_frame_marker_xyz`: all trajectory data in a numpy array with shape [frame, marker, xyz] `all_trajectory_names`: list of all trajectory...
143,671
import json import logging import sys from copy import copy from pathlib import Path from typing import List, Union, Dict, Tuple import addon_utils import bpy import numpy as np print("Running script to create Blender file from freemocap session data: " + __file__) print(f"__name__: {__name__}") def creating_stick_fig...
null
143,672
import inspect import logging import subprocess from pathlib import Path from typing import Union, List from ajc27_freemocap_blender_addon.run_as_main import ajc27_run_as_main_function from freemocap.core_processes.export_data.blender_stuff.export_to_blender.methods.ajc_addon.get_numpy_path import ( get_numpy_path,...
null
143,673
import inspect import subprocess from importlib.metadata import distribution from pathlib import Path from ajc27_freemocap_blender_addon.run_as_main import ajc27_run_as_main_function from freemocap.core_processes.export_data.blender_stuff.export_to_blender.methods.ajc_addon.install.bpy_install_addon import ( INSTAL...
null
143,674
import os import shutil import tempfile from pathlib import Path from typing import Union def create_zip_from_folder(addon_folder_path: Union[Path, str]) -> str: try: addon_folder_path = Path(addon_folder_path) if not addon_folder_path.is_dir(): raise FileNotFoundError(f"Addon folder pat...
This gets called as a subprocess to install the addon to blender
143,675
import json import logging from pathlib import Path from typing import Union from freemocap.utilities.download_sample_data import get_sample_data_path def generate_jupyter_notebook(path_to_recording: Union[str, Path]): path_to_recording = Path(path_to_recording) recording_name = Path(path_to_recording).name ...
null
143,676
import itertools import logging import multiprocessing import queue import time from collections import defaultdict, Counter from copy import copy import cv2 import numpy as np import toml from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard from aniposelib.utils impo...
null
143,677
import itertools import logging import multiprocessing import queue import time from collections import defaultdict, Counter from copy import copy import cv2 import numpy as np import toml from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard from aniposelib.utils impo...
null
143,678
import itertools import logging import multiprocessing import queue import time from collections import defaultdict, Counter from copy import copy import cv2 import numpy as np import toml from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard from aniposelib.utils impo...
null
143,679
import itertools import logging import multiprocessing import queue import time from collections import defaultdict, Counter from copy import copy import cv2 import numpy as np import toml from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard from aniposelib.utils impo...
null
143,680
import itertools import logging import multiprocessing import queue import time from collections import defaultdict, Counter from copy import copy import cv2 import numpy as np import toml from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard from aniposelib.utils impo...
null
143,681
import itertools import logging import multiprocessing import queue import time from collections import defaultdict, Counter from copy import copy import cv2 import numpy as np import toml from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard from aniposelib.utils impo...
null
143,682
import itertools import logging import multiprocessing import queue import time from collections import defaultdict, Counter from copy import copy import cv2 import numpy as np import toml from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard from aniposelib.utils impo...
Rotate points by given rotation vectors and translate. Rodrigues' rotation formula is used.
143,683
import itertools import logging import multiprocessing import queue import time from collections import defaultdict, Counter from copy import copy import cv2 import numpy as np import toml from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard from aniposelib.utils impo...
null
143,684
import filecmp import logging import shutil from pathlib import Path from typing import Union from freemocap.core_processes.capture_volume_calibration.anipose_camera_calibration import ( freemocap_anipose, ) from freemocap.system.paths_and_filenames.path_getters import get_last_successful_calibration_toml_path logg...
null
143,685
import filecmp import logging import shutil from pathlib import Path from typing import Union from freemocap.core_processes.capture_volume_calibration.anipose_camera_calibration import ( freemocap_anipose, ) from freemocap.system.paths_and_filenames.path_getters import get_last_successful_calibration_toml_path logg...
null
143,686
from pathlib import Path from typing import Callable, Union from freemocap.core_processes.capture_volume_calibration.anipose_camera_calibration.anipose_camera_calibrator import ( AniposeCameraCalibrator, ) from freemocap.core_processes.capture_volume_calibration.charuco_stuff.charuco_board_definition import ( C...
null
143,687
import io import logging import zipfile from pathlib import Path import requests from freemocap.system.paths_and_filenames.file_and_folder_names import ( FREEMOCAP_TEST_DATA_RECORDING_NAME, FIGSHARE_TEST_ZIP_FILE_URL, ) from freemocap.system.paths_and_filenames.path_getters import get_recording_session_folder_p...
null
143,688
import logging import subprocess logger = logging.getLogger(__name__) def fix_opencv_conflict(): logger.warning("Conflicting versions of opencv found, uninstalling `opencv-python`") try: subprocess.run(["pip", "uninstall", "-y", "opencv-python", "opencv-contrib-python"], check=True) except subproce...
null