id int64 0 190k | prompt stringlengths 21 13.4M | docstring stringlengths 1 12k ⌀ |
|---|---|---|
143,577 |
def only_until_first_boxed_from_tokens(string, tokens):
idx = string.find("\\boxed")
if idx < 0:
idx = string.find("\\fbox")
if idx < 0:
return None
cum_length = 0
for i, t in enumerate(tokens):
cum_length += len(t)
if cum_length >= idx:
bre... | null |
143,578 | def _clean_numbers(string):
"""
Clean Numbers in the given string
>>> _clean_numbers(None, "Hello 123")
'Hello 123'
>>> _clean_numbers(None, "Hello 1234")
'Hello 1,234'
>>> _clean_numbers(None, "Hello 1234324asdasd")
'Hello 1,234,324asdasd'
"""
num_prev_digits = 0
new_string ... | null |
143,579 | import sys
import os
from model.meta import MetaModel
import argparse
import torch
from PIL import Image
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from data.alpaca import transform_val
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quant... | null |
143,580 | import sys
import os
from model.meta import MetaModel
import argparse
import torch
from PIL import Image
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from data.alpaca import transform_val
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quant... | null |
143,581 | import sys
import os
from model.meta import MetaModel
import argparse
import torch
from PIL import Image
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from data.alpaca import transform_val
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quant... | null |
143,582 | import sys
import os
from model.meta import MetaModel
import argparse
import torch
from PIL import Image
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from data.alpaca import transform_val
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quant... | null |
143,583 | import argparse
import numpy as np
import torch
import pandas as pd
from tqdm import tqdm
import json
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant... | null |
143,584 | import argparse
import numpy as np
import torch
import pandas as pd
from tqdm import tqdm
import json
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant... | null |
143,585 | import argparse
import numpy as np
import torch
import pandas as pd
from tqdm import tqdm
import json
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant... | null |
143,586 | import argparse
import numpy as np
import torch
import pandas as pd
from tqdm import tqdm
import json
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant... | null |
143,587 | import argparse
import json
import jsonlines
import re
from tqdm import tqdm
import torch
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quan... | null |
143,588 | import argparse
import json
import jsonlines
import re
from tqdm import tqdm
import torch
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quan... | null |
143,589 | import argparse
import json
import jsonlines
import re
from tqdm import tqdm
import torch
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quan... | null |
143,590 | import argparse
import json
import jsonlines
import re
from tqdm import tqdm
import torch
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quan... | null |
143,591 | import torch
import numpy as np
import argparse
from collections import defaultdict
from tqdm import tqdm
import pandas as pd
import json
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor... | null |
143,592 | import torch
import numpy as np
import argparse
from collections import defaultdict
from tqdm import tqdm
import pandas as pd
import json
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor... | null |
143,593 | import torch
import numpy as np
import argparse
from collections import defaultdict
from tqdm import tqdm
import pandas as pd
import json
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor... | null |
143,594 | import torch
import numpy as np
import argparse
from collections import defaultdict
from tqdm import tqdm
import pandas as pd
import json
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor... | null |
143,595 | import argparse
import torch
from tqdm import tqdm
import json
import jsonlines
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quantize
from... | null |
143,596 | import argparse
import torch
from tqdm import tqdm
import json
import jsonlines
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quantize
from... | null |
143,597 | import argparse
import torch
from tqdm import tqdm
import json
import jsonlines
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quantize
from... | Evaluates the functional correctness of generated samples, and writes results to f"{sample_file}_results.jsonl.gz" |
143,598 | import argparse
import torch
from tqdm import tqdm
import json
import jsonlines
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quantize
from... | null |
143,599 | import argparse
import torch
from tqdm import tqdm
import json
import jsonlines
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from util.quant import quantize
from... | null |
143,600 | import sys
import os
from model.meta import MetaModel
import argparse
import torch
import openai
import pandas as pd
import numpy as np
from collections import Counter
import time
from PIL import Image
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from data.alpaca import transform_... | null |
143,601 | import sys
import os
from model.meta import MetaModel
import argparse
import torch
import openai
import pandas as pd
import numpy as np
from collections import Counter
import time
from PIL import Image
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from data.alpaca import transform_... | null |
143,602 | import sys
import os
from model.meta import MetaModel
import argparse
import torch
import openai
import pandas as pd
import numpy as np
from collections import Counter
import time
from PIL import Image
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from data.alpaca import transform_... | null |
143,603 | import sys
import os
from model.meta import MetaModel
import argparse
import torch
import openai
import pandas as pd
import numpy as np
from collections import Counter
import time
from PIL import Image
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from data.alpaca import transform_... | null |
143,604 | import argparse
import jsonlines
import json
import time
import pandas as pd
from tqdm import tqdm
import torch
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from ... | null |
143,605 | import argparse
import jsonlines
import json
import time
import pandas as pd
from tqdm import tqdm
import torch
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from ... | null |
143,606 | import argparse
import jsonlines
import json
import time
import pandas as pd
from tqdm import tqdm
import torch
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from ... | null |
143,607 | import argparse
import jsonlines
import json
import time
import pandas as pd
from tqdm import tqdm
import torch
import os
import sys
from model.meta import MetaModel
from util import misc
from fairscale.nn.model_parallel import initialize as fs_init
from util.tensor_parallel import load_tensor_parallel_model_list
from ... | null |
143,608 | import argparse
from collections import OrderedDict
import contextlib
from copy import deepcopy
from datetime import datetime
import functools
import json
import logging
import os
import random
import socket
from time import time
from PIL import Image
from diffusers.models import AutoencoderKL
import fairscale.nn.model... | null |
143,609 | import argparse
from collections import OrderedDict
import contextlib
from copy import deepcopy
from datetime import datetime
import functools
import json
import logging
import os
import random
import socket
from time import time
from PIL import Image
from diffusers.models import AutoencoderKL
import fairscale.nn.model... | null |
143,610 | import argparse
from collections import OrderedDict
import contextlib
from copy import deepcopy
from datetime import datetime
import functools
import json
import logging
import os
import random
import socket
from time import time
from PIL import Image
from diffusers.models import AutoencoderKL
import fairscale.nn.model... | Step the EMA model towards the current model. |
143,611 | import argparse
from collections import OrderedDict
import contextlib
from copy import deepcopy
from datetime import datetime
import functools
import json
import logging
import os
import random
import socket
from time import time
from PIL import Image
from diffusers.models import AutoencoderKL
import fairscale.nn.model... | End DDP training. |
143,612 | import argparse
from collections import OrderedDict
import contextlib
from copy import deepcopy
from datetime import datetime
import functools
import json
import logging
import os
import random
import socket
from time import time
from PIL import Image
from diffusers.models import AutoencoderKL
import fairscale.nn.model... | Create a logger that writes to a log file and stdout. |
143,613 | import argparse
from collections import OrderedDict
import contextlib
from copy import deepcopy
from datetime import datetime
import functools
import json
import logging
import os
import random
import socket
from time import time
from PIL import Image
from diffusers.models import AutoencoderKL
import fairscale.nn.model... | null |
143,614 | import argparse
from collections import OrderedDict
import contextlib
from copy import deepcopy
from datetime import datetime
import functools
import json
import logging
import os
import random
import socket
from time import time
from PIL import Image
from diffusers.models import AutoencoderKL
import fairscale.nn.model... | null |
143,615 | import argparse
from collections import OrderedDict
import contextlib
from copy import deepcopy
from datetime import datetime
import functools
import json
import logging
import os
import random
import socket
from time import time
from PIL import Image
from diffusers.models import AutoencoderKL
import fairscale.nn.model... | null |
143,616 | from PIL import Image
import numpy as np
import random
The provided code snippet includes necessary dependencies for implementing the `center_crop_arr` function. Write a Python function `def center_crop_arr(pil_image, image_size)` to solve the following problem:
Center cropping implementation from ADM. https://github.... | Center cropping implementation from ADM. https://github.com/openai/guided-diffusion/blob/8fb3ad9197f16bbc40620447b2742e13458d2831/guided_diffusion/image_datasets.py#L126 |
143,617 | from PIL import Image
import numpy as np
import random
def center_crop(pil_image, crop_size):
while pil_image.size[0] >= 2 * crop_size[0] and pil_image.size[1] >= 2 * crop_size[1]:
pil_image = pil_image.resize(
tuple(x // 2 for x in pil_image.size), resample=Image.BOX
)
scale = max(c... | null |
143,618 | from PIL import Image
import numpy as np
import random
def generate_crop_size_list(num_patches, patch_size, max_ratio=4.0):
assert max_ratio >= 1.
crop_size_list = []
wp, hp = num_patches, 1
while wp > 0:
if max(wp, hp) / min(wp, hp) <= max_ratio:
crop_size_list.append((wp * patch_s... | null |
143,619 | from io import BytesIO
import logging
import time
from typing import Union
from PIL import Image
def init_ceph_client_if_needed():
global client
if client is None:
logger.info(f"initializing ceph client ...")
st = time.time()
from petrel_client.client import Client # noqa
client... | null |
143,620 | import os
import subprocess
from time import sleep
import torch
import torch.distributed as dist
import fairscale.nn.model_parallel.initialize as fs_init
def _setup_dist_env_from_slurm(args):
while not os.environ.get("MASTER_ADDR", ""):
os.environ["MASTER_ADDR"] = subprocess.check_output(
"sinfo... | null |
143,629 | import argparse
import json
import multiprocessing as mp
import os
import socket
from diffusers.models import AutoencoderKL
import fairscale.nn.model_parallel.initialize as fs_init
import gradio as gr
import torch
import torch.distributed as dist
from torchvision.transforms.functional import to_pil_image
from transform... | null |
143,630 | import argparse
import json
import multiprocessing as mp
import os
import socket
from diffusers.models import AutoencoderKL
import fairscale.nn.model_parallel.initialize as fs_init
import gradio as gr
import torch
import torch.distributed as dist
from torchvision.transforms.functional import to_pil_image
from transform... | null |
143,635 | import functools
import math
from typing import Optional, Tuple, List
from apex.normalization import FusedRMSNorm as RMSNorm
import fairscale.nn.model_parallel.initialize as fs_init
from fairscale.nn.model_parallel.layers import (
ColumnParallelLinear, RowParallelLinear, ParallelEmbedding,
)
from flash_attn import ... | null |
143,636 | import functools
import math
from typing import Optional, Tuple, List
from apex.normalization import FusedRMSNorm as RMSNorm
import fairscale.nn.model_parallel.initialize as fs_init
from fairscale.nn.model_parallel.layers import (
ColumnParallelLinear, RowParallelLinear, ParallelEmbedding,
)
from flash_attn import ... | null |
143,637 | import functools
import math
from typing import Optional, Tuple, List
from apex.normalization import FusedRMSNorm as RMSNorm
import fairscale.nn.model_parallel.initialize as fs_init
from fairscale.nn.model_parallel.layers import (
ColumnParallelLinear, RowParallelLinear, ParallelEmbedding,
)
from flash_attn import ... | null |
143,638 | import sys
import os
import torch
import torch.distributed as dist
import argparse
import pandas as pd
from typing import Tuple, List
from PIL import Image
import fairscale.nn.model_parallel.initialize as fs_init
from accessory.model.meta import MetaModel
from accessory.data.conversation import default_conversation
fr... | null |
143,639 | import os
import subprocess
import sys
def open_file(filename):
if sys.platform == "win32":
os.startfile(filename) # noqa
else:
opener = "open" if sys.platform == "darwin" else "xdg-open"
subprocess.call([opener, filename]) | null |
143,640 | import logging
import sys
from datetime import datetime
from enum import Enum
from logging.config import dictConfig
from freemocap.system.paths_and_filenames.path_getters import get_log_file_path
def ensure_min_brightness(value, threshold=50):
"""Ensure the RGB value is above a certain threshold."""
return max(... | Generate a consistent random color for the given value. |
143,641 | import logging
import sys
from datetime import datetime
from enum import Enum
from logging.config import dictConfig
from freemocap.system.paths_and_filenames.path_getters import get_log_file_path
logging.getLogger("tzlocal").setLevel(logging.WARNING)
logging.getLogger("matplotlib").setLevel(logging.WARNING)
class LogLe... | null |
143,642 | import logging
import sys
from datetime import datetime
from enum import Enum
from logging.config import dictConfig
from freemocap.system.paths_and_filenames.path_getters import get_log_file_path
def log_test_messages(logger):
logger.trace("This is a TRACE message.")
logger.debug("This is a DEBUG message.")
... | null |
143,643 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,644 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,645 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,646 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,647 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,648 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,649 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,650 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,651 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,652 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,653 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,654 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,655 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,656 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,657 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,658 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,659 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,660 | import time
from datetime import datetime
from pathlib import Path
from typing import Union
import toml
from freemocap.system.paths_and_filenames.file_and_folder_names import (
LOGS_INFO_AND_SETTINGS_FOLDER_NAME,
LOG_FILE_FOLDER_NAME,
BASE_FREEMOCAP_DATA_FOLDER_NAME,
CALIBRATIONS_FOLDER_NAME,
logger... | null |
143,661 | import logging
from pathlib import Path
from typing import Optional, Union
from freemocap.core_processes.export_data.blender_stuff.export_to_blender.export_to_blender import export_to_blender
from freemocap.core_processes.export_data.blender_stuff.get_best_guess_of_blender_path import (
get_best_guess_of_blender_pa... | null |
143,662 | import logging
from pathlib import Path
from typing import Optional, Union
from freemocap.core_processes.export_data.blender_stuff.export_to_blender.export_to_blender import export_to_blender
from freemocap.core_processes.export_data.blender_stuff.get_best_guess_of_blender_path import (
get_best_guess_of_blender_pa... | null |
143,663 | import sys
from pathlib import Path
import bpy
import numpy as np
def get_unique_list(list: list) -> list:
"""Return a list of the unique elements from input list"""
unique_list = []
[unique_list.append(clip) for clip in list if clip not in unique_list]
return unique_list
The provided code snippet incl... | Search the folder for 'mp4' files (case insensitive) and return them as a list |
143,664 | import json
import logging
import sys
from copy import copy
from pathlib import Path
from typing import List, Union, Dict, Tuple
import addon_utils
import bpy
import numpy as np
print("Running script to create Blender file from freemocap session data: " + __file__)
def get_video_paths(path_to_video_folder: Path) -> lis... | null |
143,665 | import json
import logging
import sys
from copy import copy
from pathlib import Path
from typing import List, Union, Dict, Tuple
import addon_utils
import bpy
import numpy as np
print("Running script to create Blender file from freemocap session data: " + __file__)
print(f"__name__: {__name__}")
The provided code snip... | Create a key framed empty from 3d trajectory data |
143,666 | import json
import logging
import sys
from copy import copy
from pathlib import Path
from typing import List, Union, Dict, Tuple
import addon_utils
import bpy
import numpy as np
def create_material(
name: str = "Generic",
color: Union[str, Tuple, List] = "#00FFFF",
emission_strength: float = 1.0,
):
"""... | Put uv sphere meshes on the empties in `empty_names_list` with a scale of `sphere_scale`. :param empty_names_list: List of empty object names. :param parent_object: The object to be set as parent for the created spheres. :param sphere_scale: The scale of the sphere. :param material: The material to be applied to the sp... |
143,667 | import json
import logging
import sys
from copy import copy
from pathlib import Path
from typing import List, Union, Dict, Tuple
import addon_utils
import bpy
import numpy as np
def create_stick_mesh(
vertices: int = 12,
radius: float = 0.02,
depth: float = 1.0,
location=(0, 0, 0),
rotation=(0, 0, 0... | make cylinder meshes between the empties as defined in `body_connections` |
143,668 | import json
import logging
import sys
from copy import copy
from pathlib import Path
from typing import List, Union, Dict, Tuple
import addon_utils
import bpy
import numpy as np
print("Running script to create Blender file from freemocap session data: " + __file__)
print(f"__name__: {__name__}")
The provided code snip... | Apply constraints to a bone (that exists in `armature_rig` based on a dictionary of constraints |
143,669 | import json
import logging
import sys
from copy import copy
from pathlib import Path
from typing import List, Union, Dict, Tuple
import addon_utils
import bpy
import numpy as np
The provided code snippet includes necessary dependencies for implementing the `test_virtual_marker_definitions` function. Write a Python fun... | Validate the virtual marker definitions dictionary to ensure that there are the same number of marker names and weights, and that the weights sum to 1 |
143,670 | import json
import logging
import sys
from copy import copy
from pathlib import Path
from typing import List, Union, Dict, Tuple
import addon_utils
import bpy
import numpy as np
The provided code snippet includes necessary dependencies for implementing the `calculate_virtual_marker_trajectory` function. Write a Python... | Create a virtual marker from a set of component markers. A 'Virtual Marker' is a 'fake' marker created by combining the data from 'real' (measured) marker/trajectory data `trajectory_3d_frame_marker_xyz`: all trajectory data in a numpy array with shape [frame, marker, xyz] `all_trajectory_names`: list of all trajectory... |
143,671 | import json
import logging
import sys
from copy import copy
from pathlib import Path
from typing import List, Union, Dict, Tuple
import addon_utils
import bpy
import numpy as np
print("Running script to create Blender file from freemocap session data: " + __file__)
print(f"__name__: {__name__}")
def creating_stick_fig... | null |
143,672 | import inspect
import logging
import subprocess
from pathlib import Path
from typing import Union, List
from ajc27_freemocap_blender_addon.run_as_main import ajc27_run_as_main_function
from freemocap.core_processes.export_data.blender_stuff.export_to_blender.methods.ajc_addon.get_numpy_path import (
get_numpy_path,... | null |
143,673 | import inspect
import subprocess
from importlib.metadata import distribution
from pathlib import Path
from ajc27_freemocap_blender_addon.run_as_main import ajc27_run_as_main_function
from freemocap.core_processes.export_data.blender_stuff.export_to_blender.methods.ajc_addon.install.bpy_install_addon import (
INSTAL... | null |
143,674 | import os
import shutil
import tempfile
from pathlib import Path
from typing import Union
def create_zip_from_folder(addon_folder_path: Union[Path, str]) -> str:
try:
addon_folder_path = Path(addon_folder_path)
if not addon_folder_path.is_dir():
raise FileNotFoundError(f"Addon folder pat... | This gets called as a subprocess to install the addon to blender |
143,675 | import json
import logging
from pathlib import Path
from typing import Union
from freemocap.utilities.download_sample_data import get_sample_data_path
def generate_jupyter_notebook(path_to_recording: Union[str, Path]):
path_to_recording = Path(path_to_recording)
recording_name = Path(path_to_recording).name
... | null |
143,676 | import itertools
import logging
import multiprocessing
import queue
import time
from collections import defaultdict, Counter
from copy import copy
import cv2
import numpy as np
import toml
from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard
from aniposelib.utils impo... | null |
143,677 | import itertools
import logging
import multiprocessing
import queue
import time
from collections import defaultdict, Counter
from copy import copy
import cv2
import numpy as np
import toml
from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard
from aniposelib.utils impo... | null |
143,678 | import itertools
import logging
import multiprocessing
import queue
import time
from collections import defaultdict, Counter
from copy import copy
import cv2
import numpy as np
import toml
from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard
from aniposelib.utils impo... | null |
143,679 | import itertools
import logging
import multiprocessing
import queue
import time
from collections import defaultdict, Counter
from copy import copy
import cv2
import numpy as np
import toml
from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard
from aniposelib.utils impo... | null |
143,680 | import itertools
import logging
import multiprocessing
import queue
import time
from collections import defaultdict, Counter
from copy import copy
import cv2
import numpy as np
import toml
from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard
from aniposelib.utils impo... | null |
143,681 | import itertools
import logging
import multiprocessing
import queue
import time
from collections import defaultdict, Counter
from copy import copy
import cv2
import numpy as np
import toml
from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard
from aniposelib.utils impo... | null |
143,682 | import itertools
import logging
import multiprocessing
import queue
import time
from collections import defaultdict, Counter
from copy import copy
import cv2
import numpy as np
import toml
from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard
from aniposelib.utils impo... | Rotate points by given rotation vectors and translate. Rodrigues' rotation formula is used. |
143,683 | import itertools
import logging
import multiprocessing
import queue
import time
from collections import defaultdict, Counter
from copy import copy
import cv2
import numpy as np
import toml
from aniposelib.boards import extract_points, extract_rtvecs, get_video_params, merge_rows, CharucoBoard
from aniposelib.utils impo... | null |
143,684 | import filecmp
import logging
import shutil
from pathlib import Path
from typing import Union
from freemocap.core_processes.capture_volume_calibration.anipose_camera_calibration import (
freemocap_anipose,
)
from freemocap.system.paths_and_filenames.path_getters import get_last_successful_calibration_toml_path
logg... | null |
143,685 | import filecmp
import logging
import shutil
from pathlib import Path
from typing import Union
from freemocap.core_processes.capture_volume_calibration.anipose_camera_calibration import (
freemocap_anipose,
)
from freemocap.system.paths_and_filenames.path_getters import get_last_successful_calibration_toml_path
logg... | null |
143,686 | from pathlib import Path
from typing import Callable, Union
from freemocap.core_processes.capture_volume_calibration.anipose_camera_calibration.anipose_camera_calibrator import (
AniposeCameraCalibrator,
)
from freemocap.core_processes.capture_volume_calibration.charuco_stuff.charuco_board_definition import (
C... | null |
143,687 | import io
import logging
import zipfile
from pathlib import Path
import requests
from freemocap.system.paths_and_filenames.file_and_folder_names import (
FREEMOCAP_TEST_DATA_RECORDING_NAME,
FIGSHARE_TEST_ZIP_FILE_URL,
)
from freemocap.system.paths_and_filenames.path_getters import get_recording_session_folder_p... | null |
143,688 | import logging
import subprocess
logger = logging.getLogger(__name__)
def fix_opencv_conflict():
logger.warning("Conflicting versions of opencv found, uninstalling `opencv-python`")
try:
subprocess.run(["pip", "uninstall", "-y", "opencv-python", "opencv-contrib-python"], check=True)
except subproce... | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.