id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
141,626
import os from natsort import natsorted from glob import glob def get_last_path(path, session): x = natsorted(glob(os.path.join(path,'*%s'%session)))[-1] return x
null
141,627
import torch import numpy as np import cv2 def torchPSNR(tar_img, prd_img): imdff = torch.clamp(prd_img,0,1) - torch.clamp(tar_img,0,1) rmse = (imdff**2).mean().sqrt() ps = 20*torch.log10(1/rmse) return ps
null
141,628
import torch import numpy as np import cv2 def save_img(filepath, img): cv2.imwrite(filepath,cv2.cvtColor(img, cv2.COLOR_RGB2BGR))
null
141,629
import torch import numpy as np import cv2 def numpyPSNR(tar_img, prd_img): imdff = np.float32(prd_img) - np.float32(tar_img) rmse = np.sqrt(np.mean(imdff**2)) ps = 20*np.log10(255/rmse) return ps
null
141,630
import torch import os from collections import OrderedDict def freeze(model): for p in model.parameters(): p.requires_grad=False
null
141,631
import torch import os from collections import OrderedDict def unfreeze(model): for p in model.parameters(): p.requires_grad=True
null
141,632
import torch import os from collections import OrderedDict def is_frozen(model): x = [p.requires_grad for p in model.parameters()] return not all(x)
null
141,633
import torch import os from collections import OrderedDict def save_checkpoint(model_dir, state, session): epoch = state['epoch'] model_out_path = os.path.join(model_dir,"model_epoch_{}_{}.pth".format(epoch,session)) torch.save(state, model_out_path)
null
141,634
import torch import os from collections import OrderedDict def load_checkpoint(model, weights): checkpoint = torch.load(weights) try: model.load_state_dict(checkpoint["state_dict"]) except: state_dict = checkpoint["state_dict"] new_state_dict = OrderedDict() for k, v in stat...
null
141,635
import torch import os from collections import OrderedDict def load_checkpoint_multigpu(model, weights): checkpoint = torch.load(weights) state_dict = checkpoint["state_dict"] new_state_dict = OrderedDict() for k, v in state_dict.items(): name = k[7:] # remove `module.` new_state_dict[n...
null
141,636
import torch import os from collections import OrderedDict def load_start_epoch(weights): checkpoint = torch.load(weights) epoch = checkpoint["epoch"] return epoch
null
141,637
import torch import os from collections import OrderedDict def load_optim(optimizer, weights): checkpoint = torch.load(weights) optimizer.load_state_dict(checkpoint['optimizer']) # for p in optimizer.param_groups: lr = p['lr'] # return lr
null
141,639
import os import numpy as np from glob import glob from natsort import natsorted from skimage import io import cv2 from skimage.metrics import structural_similarity from tqdm import tqdm import concurrent.futures def image_align(deblurred, gt): def compute_psnr(image_true, image_test, image_mask, data_range=None): def ...
null
141,640
import os import numpy as np from torch.utils.data import Dataset import torch from PIL import Image import torchvision.transforms.functional as TF from pdb import set_trace as stx import random def is_image_file(filename): return any(filename.endswith(extension) for extension in ['jpeg', 'JPEG', 'jpg', 'png', 'JP...
null
141,641
import os from dataset_RGB import DataLoaderTrain, DataLoaderVal, DataLoaderTest class DataLoaderTrain(Dataset): def __init__(self, rgb_dir, img_options=None): super(DataLoaderTrain, self).__init__() inp_files = sorted(os.listdir(os.path.join(rgb_dir, 'input'))) tar_files = sorted(os.listd...
null
141,661
import os from dataset_RGB import DataLoaderTrain, DataLoaderVal, DataLoaderTest class DataLoaderVal(Dataset): def __init__(self, rgb_dir, img_options=None, rgb_dir2=None): def __len__(self): def __getitem__(self, index): def get_validation_data(rgb_dir, img_options): assert os.path.exists(rgb_dir)...
null
141,676
import torch import torch.nn as nn import torch.nn.functional as F import torchvision.transforms.functional as TF from PIL import Image import os from runpy import run_path from skimage import img_as_ubyte from collections import OrderedDict from natsort import natsorted from glob import glob import cv2 import argparse...
null
141,677
import torch import torch.nn as nn import torch.nn.functional as F import torchvision.transforms.functional as TF from PIL import Image import os from runpy import run_path from skimage import img_as_ubyte from collections import OrderedDict from natsort import natsorted from glob import glob import cv2 import argparse...
null
141,678
import os import re import openai import streamlit as st from dotenv import load_dotenv from tempfile import NamedTemporaryFile import streamlit.components.v1 as components from streamlit_chat import message from langchain import LLMChain, PromptTemplate from langchain.memory import ConversationBufferWindowMemory from ...
null
141,679
import os import re import openai import streamlit as st from dotenv import load_dotenv from tempfile import NamedTemporaryFile import streamlit.components.v1 as components from streamlit_chat import message from langchain import LLMChain, PromptTemplate from langchain.memory import ConversationBufferWindowMemory from ...
null
141,680
import os import re import openai import streamlit as st from dotenv import load_dotenv from tempfile import NamedTemporaryFile import streamlit.components.v1 as components from streamlit_chat import message from langchain import LLMChain, PromptTemplate from langchain.memory import ConversationBufferWindowMemory from ...
null
141,681
import os import re import openai import streamlit as st from dotenv import load_dotenv from tempfile import NamedTemporaryFile import streamlit.components.v1 as components from streamlit_chat import message from langchain import LLMChain, PromptTemplate from langchain.memory import ConversationBufferWindowMemory from ...
null
141,682
import os import re import openai import streamlit as st from dotenv import load_dotenv from tempfile import NamedTemporaryFile import streamlit.components.v1 as components from streamlit_chat import message from langchain import LLMChain, PromptTemplate from langchain.memory import ConversationBufferWindowMemory from ...
null
141,683
import os import re import openai import streamlit as st from dotenv import load_dotenv from tempfile import NamedTemporaryFile import streamlit.components.v1 as components from streamlit_chat import message from langchain import LLMChain, PromptTemplate from langchain.memory import ConversationBufferWindowMemory from ...
null
141,684
import os import webbrowser import argparse from langchain.document_loaders.figma import FigmaFileLoader from langchain.chat_models import ChatOpenAI from langchain.indexes import VectorstoreIndexCreator from langchain.prompts.chat import ( ChatPromptTemplate, SystemMessagePromptTemplate, HumanMessagePrompt...
null
141,685
from dataclasses import dataclass, field from typing import Optional, Dict, Sequence import numpy as np import torch import transformers from torch.utils.data import Dataset from transformers import Trainer import os The provided code snippet includes necessary dependencies for implementing the `safe_save_model_for_hf...
Collects the state dict and dump to disk.
141,686
from dataclasses import dataclass, field from typing import Optional, Dict, Sequence import numpy as np import torch import transformers from torch.utils.data import Dataset from transformers import Trainer import os class ModelArguments: """filed的作用:如果有新值就会被替换掉,""" model_name_or_path: Optional[str] = field(def...
transformers.HfArgumentParser,传入的类必须有@dataclass进行装饰,这样这些参数会被自动添加到命令中
141,687
import torch import math import numpy as np from typing import List, Dict from torch.utils.data import DataLoader, Dataset import os import time class MyDataset(Dataset): def __init__(self, data_prefix, seq_length, pad_id): super(MyDataset, self).__init__() """这边要求data_prefix为完整的路径,但不包括后缀""" ...
args: - per_device_train_batch_size: 自带的 - seed: 自带的 - global_batch_distributed: "[1,2,3]"
141,688
import copy import logging from dataclasses import dataclass, field from typing import Optional, Dict, Sequence import numpy as np import torch import transformers from torch.utils.data import Dataset from transformers import Trainer import os from torch.utils.data import DataLoader from typing import List import math ...
transformers.HfArgumentParser,传入的类必须有@dataclass进行装饰,这样这些参数会被自动添加到命令中
141,689
import json import shutil import numpy as np import os from sentencepiece import SentencePieceProcessor from typing import List import argparse import multiprocessing from tqdm import tqdm import time import torch from torch.utils.data import Dataset import re import nltk def read(args): ds = MyDataset(args.rea...
null
141,690
import json import shutil import numpy as np import os from sentencepiece import SentencePieceProcessor from typing import List import argparse import multiprocessing from tqdm import tqdm import time import torch from torch.utils.data import Dataset import re import nltk The provided code snippet includes necessa...
下面这个暂时先取消掉自动合并的功能,必须要指定,也就是写入到新的文件
141,691
import json import shutil import numpy as np import os from sentencepiece import SentencePieceProcessor from typing import List import argparse import multiprocessing from tqdm import tqdm import time import torch from torch.utils.data import Dataset import re import nltk def collate_fn_from_text(text: str): r...
null
141,692
import json import shutil import numpy as np import os from sentencepiece import SentencePieceProcessor from typing import List import argparse import multiprocessing from tqdm import tqdm import time import torch from torch.utils.data import Dataset import re import nltk class Tokenizer: def __init__(self, mod...
打开文本文件
141,693
import json import shutil import numpy as np import os from sentencepiece import SentencePieceProcessor from typing import List import argparse import multiprocessing from tqdm import tqdm import time import torch from torch.utils.data import Dataset import re import nltk def write(args): """这个函数用于处理我们爬取的文件""" ...
不支持一个数据集中有不同类型的数据进行混合
141,694
import copy import logging from dataclasses import dataclass, field from typing import Optional, Dict, Sequence import numpy as np import torch import transformers from torch.utils.data import Dataset from transformers import Trainer import os from torch.utils.data.distributed import DistributedSampler class args: ...
null
141,695
from torch.utils.data import Dataset import numpy as np import os import torch import argparse import time from tqdm import tqdm from sentencepiece import SentencePieceProcessor def get_args(): parser = argparse.ArgumentParser() parser.add_argument("--read_data_path_prefix", default=None, type=str, help="会自动在结...
null
141,696
from torch.utils.data import Dataset import numpy as np import os import torch import argparse import time from tqdm import tqdm from sentencepiece import SentencePieceProcessor class MyDataset(Dataset): def __init__(self, data_prefix, seq_length, pad_id): super(MyDataset, self).__init__() """这边要求da...
只保留中文,而在之前的版本中,中文是在前面的
141,697
from torch.utils.data import Dataset import numpy as np import os import torch import argparse import time from tqdm import tqdm from sentencepiece import SentencePieceProcessor class MyDataset(Dataset): def __init__(self, data_prefix, seq_length, pad_id): super(MyDataset, self).__init__() """这边要求da...
加载待删除的索引
141,698
from torch.utils.data import Dataset import numpy as np import os import torch import argparse import time from tqdm import tqdm from sentencepiece import SentencePieceProcessor class MyDataset(Dataset): def __init__(self, data_prefix, seq_length, pad_id): def _check(self): def _load_index(self): de...
null
141,699
import os import logging import bitsandbytes as bnb import torch import transformers import argparse from transformers import ( AutoTokenizer, AutoModelForCausalLM, set_seed, Trainer, BitsAndBytesConfig, ) from peft import ( prepare_model_for_kbit_training, LoraConfig, get_peft_model, ...
null
141,700
import os import sys from typing import List import fire import torch import transformers from datasets import load_dataset from peft import ( LoraConfig, get_peft_model, get_peft_model_state_dict, prepare_model_for_kbit_training, set_peft_model_state_dict, TaskType, ) from transformers import A...
(1) ChatGLM3 uses <unk> as padding token (2) Qwen uses <|endoftext|> as padding token
141,701
import os import sys from typing import List import fire import torch import transformers from datasets import load_dataset from peft import ( LoraConfig, get_peft_model, get_peft_model_state_dict, prepare_model_for_int8_training, set_peft_model_state_dict, ) from transformers import AutoModelForCau...
null
141,702
import os import sys from typing import List import fire import torch import transformers from datasets import load_dataset from peft import ( LoraConfig, get_peft_model, get_peft_model_state_dict, prepare_model_for_int8_training, set_peft_model_state_dict, ) from transformers import LlamaForCausalL...
null
141,703
import json from fastapi import FastAPI, Request import uvicorn import datetime import torch from transformers import LlamaTokenizer from transformers import GenerationConfig from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig from utils import Prompter async def complement(request: Request): global mode...
null
141,704
import time from typing import Any, List, Optional, Union from transformers import PreTrainedTokenizer, PreTrainedTokenizerFast from transformers import LlamaTokenizer from vllm.config import (CacheConfig, ModelConfig, ParallelConfig, SchedulerConfig) from vllm.core.scheduler import Scheduler f...
Gets a tokenizer for the given model name via Huggingface.
141,705
import os import torch import uvicorn from fastapi.responses import StreamingResponse from fastapi import FastAPI, Request from transformers import AutoTokenizer, AutoConfig, AutoModelForCausalLM, pipeline from queue import Queue import json from typing import List, Dict, Optional import time import asyncio import argp...
null
141,706
import os import torch import uvicorn from fastapi.responses import StreamingResponse from fastapi import FastAPI, Request from transformers import AutoTokenizer, AutoConfig, AutoModelForCausalLM, pipeline from queue import Queue import json from typing import List, Dict, Optional import time import asyncio import argp...
null
141,707
import os import torch import uvicorn from fastapi.responses import StreamingResponse from fastapi import FastAPI, Request from transformers import AutoTokenizer, AutoConfig, AutoModelForCausalLM, pipeline from queue import Queue import json from typing import List, Dict, Optional import time import asyncio import argp...
null
141,708
import os import torch import uvicorn from fastapi.responses import StreamingResponse from fastapi import FastAPI, Request from transformers import AutoTokenizer, AutoConfig, AutoModelForCausalLM, pipeline from queue import Queue import json from typing import List, Dict, Optional import time import asyncio import argp...
null
141,709
import os import torch import uvicorn from fastapi.responses import StreamingResponse from fastapi import FastAPI, Request from transformers import AutoTokenizer, AutoConfig, AutoModelForCausalLM, pipeline from queue import Queue import json from typing import List, Dict, Optional import time import asyncio import argp...
null
141,710
import os import torch import uvicorn from fastapi.responses import StreamingResponse from fastapi import FastAPI, Request from transformers import AutoTokenizer, AutoConfig, AutoModelForCausalLM, pipeline from queue import Queue import json from typing import List, Dict, Optional import time import asyncio import argp...
null
141,711
import os import torch import uvicorn from fastapi.responses import StreamingResponse from fastapi import FastAPI, Request from transformers import AutoTokenizer, AutoConfig, AutoModelForCausalLM, pipeline from queue import Queue import json from typing import List, Dict, Optional import time import asyncio import argp...
null
141,712
import os import torch import uvicorn from fastapi.responses import StreamingResponse from fastapi import FastAPI, Request from transformers import AutoTokenizer, AutoConfig, AutoModelForCausalLM, pipeline from queue import Queue import json from typing import List, Dict, Optional import time import asyncio import argp...
null
141,713
import os import torch import uvicorn from fastapi.responses import StreamingResponse from fastapi import FastAPI, Request from transformers import AutoTokenizer, AutoConfig, AutoModelForCausalLM, pipeline from queue import Queue import json from typing import List, Dict, Optional import time import asyncio import argp...
null
141,714
import os import torch import uvicorn from fastapi.responses import StreamingResponse from fastapi import FastAPI, Request from transformers import AutoTokenizer, AutoConfig, AutoModelForCausalLM, pipeline from queue import Queue import json from typing import List, Dict, Optional import time import asyncio import argp...
null
141,715
import torch from accelerate import init_empty_weights, infer_auto_device_map, load_checkpoint_and_dispatch from transformers import GenerationConfig, LlamaForCausalLM, LlamaTokenizer, LlamaConfig, AutoModelForCausalLM, AutoConfig, AutoTokenizer from typing import List def set_limit(allocate:List[int]=None): if al...
null
141,716
import torch from accelerate import init_empty_weights, infer_auto_device_map, load_checkpoint_and_dispatch from transformers import GenerationConfig, LlamaForCausalLM, LlamaTokenizer, LlamaConfig, AutoModelForCausalLM, AutoConfig, AutoTokenizer from typing import List def get_tokenizer_and_model(base_model:str, dtype...
null
141,717
from typing import Optional import fire import torch import tqdm import transformers from peft import PeftModel The provided code snippet includes necessary dependencies for implementing the `make_diff` function. Write a Python function `def make_diff( path_raw: str, path_tuned: str, path_diff: str, device="cpu", ...
Make the weight diff. This function is given to present full transparency of how the weight diff was created. Run: python weight_diff.py make_diff --path_raw <your_path_raw> --path_tuned <your_path_tuned> --path_diff <your_path_diff>
141,718
from typing import Optional import fire import torch import tqdm import transformers from peft import PeftModel The provided code snippet includes necessary dependencies for implementing the `recover` function. Write a Python function `def recover( path_raw, path_diff, path_tuned: Optional[str] = None, ...
Recover the original weights from the released weight diff. This function is given for you to run. Things to do before running this: 1. Convert Meta's released weights into huggingface format. Follow this guide: https://huggingface.co/docs/transformers/main/model_doc/llama 2. Make sure you cloned the released weight di...
141,719
from typing import Optional import fire import torch import tqdm import transformers from peft import PeftModel The provided code snippet includes necessary dependencies for implementing the `merge` function. Write a Python function `def merge( path_zhixi, path_lora, path_sfted: Optional[str] = None, i...
Merge the released lora weight with the pretrained zhixi to be a single hf format model for further lora training or full-finetuning This function is given for you to run. Things to do before running this: 1. Getting the pretrained zhixi model. Follow this guide: https://github.com/zjunlp/KnowLM 2. Make sure you cloned...
141,720
import argparse from huggingface_hub import snapshot_download import time def _print(message): print(f"[{time.ctime()}] {message}")
null
141,721
import argparse from huggingface_hub import snapshot_download import time def add_argument(): parser = argparse.ArgumentParser(description="download") parser.add_argument('--download_path', type=str, default='./CaMA', help="storage directory") parser.add_argument('--only_lora', action='store_true', default...
null
141,722
import argparse from huggingface_hub import snapshot_download import time def check_args(args): assert args.only_lora or args.only_base or args.both, \ "Please select the file to download." assert (args.only_lora and not args.only_base and not args.both) \ or (not args.only_lora and args.onl...
null
141,723
import logging import sys from pathlib import Path import atheris import structlog def set_unblob_log_level(level=logging.CRITICAL): logger = logging.getLogger("unblob") def logger_factory(): return logger structlog.configure(logger_factory=logger_factory) logger.setLevel(level)
null
141,724
import logging import sys from pathlib import Path import atheris import structlog def extract(inpath: Path, outpath: Path): # noqa: ARG001 return
null
141,725
import logging import sys from pathlib import Path import atheris import structlog class File(mmap.mmap): access: int def from_bytes(cls, content: bytes): if not content: raise ValueError("Can't create File from empty bytes.") m = cls(-1, len(content)) m.write(content) ...
null
141,726
import os from threading import Lock __last_id = 0 __last_id_lock = Lock() def new_id(): # NOTE, that uuid4 can not be used, as there are multiple processes at run time, # and as subprocesses inherit the random number state, so uuid4 would generate colliding ids # another option that would not work is to u...
null
141,727
from typing import Optional from pyfatfs._exceptions import PyFATException from pyfatfs.PyFat import PyFat from structlog import get_logger from unblob.extractors.command import Command from unblob.file_utils import InvalidInputFormat from ...models import File, Handler, HexString, ValidChunk def get_max_offset(fs: Py...
null
141,728
import binascii import struct from typing import Optional from dissect.cstruct import Instance from unblob.extractors import Command from ...file_utils import Endian, convert_int32, get_endian from ...models import File, HexString, StructHandler, ValidChunk def swap_int32(i): return struct.unpack("<I", struct.pack...
null
141,729
import io import itertools from collections import defaultdict from enum import IntEnum from pathlib import Path from typing import Iterable, List, Optional, Tuple import attr from dissect.cstruct import Instance from structlog import get_logger from treelib import Tree from treelib.exceptions import NodeIDAbsentError ...
null
141,730
import io import itertools from collections import defaultdict from enum import IntEnum from pathlib import Path from typing import Iterable, List, Optional, Tuple import attr from dissect.cstruct import Instance from structlog import get_logger from treelib import Tree from treelib.exceptions import NodeIDAbsentError ...
File size can be encoded as 64 bits or 32 bits values. If upper 32 bits are set, it's a 64 bits integer value. Otherwise it's a 32 bits value. 0xFFFFFFFF means zero.
141,731
import io import itertools from collections import defaultdict from enum import IntEnum from pathlib import Path from typing import Iterable, List, Optional, Tuple import attr from dissect.cstruct import Instance from structlog import get_logger from treelib import Tree from treelib.exceptions import NodeIDAbsentError ...
null
141,732
import io import itertools from collections import defaultdict from enum import IntEnum from pathlib import Path from typing import Iterable, List, Optional, Tuple import attr from dissect.cstruct import Instance from structlog import get_logger from treelib import Tree from treelib.exceptions import NodeIDAbsentError ...
null
141,733
from typing import Optional from structlog import get_logger from unblob.extractors import Command from unblob.file_utils import Endian from ...models import File, HexString, StructHandler, ValidChunk The provided code snippet includes necessary dependencies for implementing the `from_733` function. Write a Python fun...
Convert from ISO 9660 7.3.3 format to uint32_t. Return the little-endian part always, to handle non-specs-compliant images.
141,734
from typing import Optional from structlog import get_logger from unblob.extractors import Command from unblob.file_utils import Endian from ...models import File, HexString, StructHandler, ValidChunk The provided code snippet includes necessary dependencies for implementing the `from_723` function. Write a Python fun...
Convert from ISO 9660 7.2.3 format to uint16_t. Return the little-endian part always, to handle non-specs-compliant images.
141,735
import io import os import stat import struct from enum import IntEnum, unique from pathlib import Path from typing import Dict, Optional from structlog import get_logger from ...file_utils import ( Endian, FileSystem, InvalidInputFormat, read_until_past, round_up, ) from ...models import ( Extr...
Apply a RomFS checksum and returns whether it's valid or not.
141,736
import io import os import stat import struct from enum import IntEnum, unique from pathlib import Path from typing import Dict, Optional from structlog import get_logger from ...file_utils import ( Endian, FileSystem, InvalidInputFormat, read_until_past, round_up, ) from ...models import ( Extr...
Read a 16 bytes aligned, null terminated string.
141,737
import io import shutil from pathlib import Path from typing import Optional import attr import lief from dissect.cstruct import Instance from structlog import get_logger from unblob.extractor import carve_chunk_to_file from unblob.file_utils import ( Endian, File, convert_int8, convert_int32, conve...
Extract the initramfs part, with a potentially 4 extra bytes. Due to alignment definition of initramfs the start-end offsets can not be exactly calculated, so the output could have a 4 extra bytes before or after the initramfs.
141,738
import io from pathlib import Path from typing import Optional import attr from dissect.cstruct import Instance from pyperscan import Flag, Pattern, Scan, StreamDatabase from structlog import get_logger from unblob.file_utils import File, iterate_file, stream_scan from unblob.models import ( Endian, Extractor, ...
null
141,739
import io from pathlib import Path from typing import Optional import attr from dissect.cstruct import Instance from pyperscan import Flag, Pattern, Scan, StreamDatabase from structlog import get_logger from unblob.file_utils import File, iterate_file, stream_scan from unblob.models import ( Endian, Extractor, ...
null
141,740
import contextlib import os import tarfile from pathlib import Path from typing import Optional from structlog import get_logger from ...file_utils import OffsetFile, SeekError, decode_int, round_up, snull from ...models import ( Extractor, ExtractResult, File, HexString, Regex, StructHandler, ...
null
141,741
import contextlib import os import tarfile from pathlib import Path from typing import Optional from structlog import get_logger from ...file_utils import OffsetFile, SeekError, decode_int, round_up, snull from ...models import ( Extractor, ExtractResult, File, HexString, Regex, StructHandler, ...
null
141,742
import binascii from pathlib import Path from typing import Optional from structlog import get_logger from unblob.extractors import Command from ...extractors.command import MultiFileCommand from ...file_utils import Endian, InvalidInputFormat, StructParser from ...models import ( DirectoryHandler, File, Gl...
null
141,743
import binascii from pathlib import Path from typing import Optional from structlog import get_logger from unblob.extractors import Command from ...extractors.command import MultiFileCommand from ...file_utils import Endian, InvalidInputFormat, StructParser from ...models import ( DirectoryHandler, File, Gl...
null
141,744
import binascii import io from pathlib import Path from typing import Iterable, Optional, Tuple, cast from dissect.cstruct import Instance from structlog import get_logger from unblob.file_utils import ( File, FileSystem, InvalidInputFormat, iterate_file, snull, ) from unblob.models import ( End...
null
141,745
import binascii import io from pathlib import Path from typing import Iterable, Optional, Tuple, cast from dissect.cstruct import Instance from structlog import get_logger from unblob.file_utils import ( File, FileSystem, InvalidInputFormat, iterate_file, snull, ) from unblob.models import ( End...
null
141,746
import binascii import io from pathlib import Path from typing import Iterable, Optional, Tuple, cast from dissect.cstruct import Instance from structlog import get_logger from unblob.file_utils import ( File, FileSystem, InvalidInputFormat, iterate_file, snull, ) from unblob.models import ( End...
null
141,747
from pathlib import Path from typing import Optional from dissect.cstruct import Instance from structlog import get_logger from unblob.file_utils import Endian, File, InvalidInputFormat, StructParser from unblob.models import Extractor, HexString, StructHandler, ValidChunk XOR_KEY = b"\xac\x78\x3c\x9e\xcf\x67\xb3\x59" ...
null
141,748
import io from pathlib import Path from typing import Optional from dissect.cstruct import Instance from structlog import get_logger from unblob.file_utils import ( Endian, File, FileSystem, InvalidInputFormat, StructParser, snull, ) from unblob.models import ( Extractor, ExtractResult, ...
null
141,749
import io from pathlib import Path from typing import Optional from dissect.cstruct import Instance from structlog import get_logger from unblob.extractor import carve_chunk_to_file from unblob.file_utils import Endian, File, InvalidInputFormat, StructParser, snull from unblob.models import Chunk, Extractor, HexString,...
null
141,750
import gzip import io import struct import zlib from pathlib import Path from typing import List, Optional from structlog import get_logger from unblob.extractors import Command from unblob.extractors.command import MultiFileCommand from unblob.models import Extractor from ...file_utils import InvalidInputFormat from ....
null
141,751
import io from typing import Optional, Tuple import attr from pyperscan import Flag, Pattern, Scan, StreamDatabase from structlog import get_logger from unblob.extractors import Command from ...file_utils import ( Endian, convert_int8, convert_int16, convert_int32, decode_multibyte_integer, read...
null
141,752
import io from typing import Optional, Tuple import attr from pyperscan import Flag, Pattern, Scan, StreamDatabase from structlog import get_logger from unblob.extractors import Command from ...file_utils import ( Endian, convert_int8, convert_int16, convert_int32, decode_multibyte_integer, read...
null
141,753
from typing import Optional import attr from pyperscan import Flag, Pattern, Scan, StreamDatabase from structlog import get_logger from unblob.extractors import Command from ...file_utils import InvalidInputFormat, SeekError, StructParser, stream_scan from ...models import File, Handler, HexString, Regex, ValidChunk d...
null
141,754
from typing import Optional import attr from pyperscan import Flag, Pattern, Scan, StreamDatabase from structlog import get_logger from unblob.extractors import Command from ...file_utils import InvalidInputFormat, SeekError, StructParser, stream_scan from ...models import File, Handler, HexString, Regex, ValidChunk ST...
null
141,755
import atexit import sys from importlib.metadata import version from pathlib import Path from typing import Dict, Iterable, List, Optional, Tuple import click from rich.console import Console from rich.panel import Panel from rich.style import Style from rich.table import Table from structlog import get_logger from unb...
null
141,756
import atexit import sys from importlib.metadata import version from pathlib import Path from typing import Dict, Iterable, List, Optional, Tuple import click from rich.console import Console from rich.panel import Panel from rich.style import Style from rich.table import Table from structlog import get_logger from unb...
null
141,757
import atexit import sys from importlib.metadata import version from pathlib import Path from typing import Dict, Iterable, List, Optional, Tuple import click from rich.console import Console from rich.panel import Panel from rich.style import Style from rich.table import Table from structlog import get_logger from unb...
null
141,758
import atexit import sys from importlib.metadata import version from pathlib import Path from typing import Dict, Iterable, List, Optional, Tuple import click from rich.console import Console from rich.panel import Panel from rich.style import Style from rich.table import Table from structlog import get_logger from unb...
null
141,759
import atexit import sys from importlib.metadata import version from pathlib import Path from typing import Dict, Iterable, List, Optional, Tuple import click from rich.console import Console from rich.panel import Panel from rich.style import Style from rich.table import Table from structlog import get_logger from unb...
null