index int64 | repo_name string | branch_name string | path string | content string | import_graph string |
|---|---|---|---|---|---|
39,078,728 | adrianogil/miru | refs/heads/main | /src/python/miru/engine/camera.py | from .transform import Transform
class Camera:
ORTOGRAPHIC = 1
PERSPECTIVE = 2
def __init__(self):
self.mode = Camera.PERSPECTIVE
self.fov = 60
self.near = 0.01
self.far = 1000
self.transform = Transform()
@staticmethod
def parse(data):
c = Came... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,729 | adrianogil/miru | refs/heads/main | /src/python/miru/raytracing/ray.py | class Ray:
def __init__(self, p1, p2):
self.p1 = p1
self.p2 = p2
self.origin = p1
self.direction = p2.minus(p1).normalized()
def __str__(self):
return "( " + str(self.origin) + " , " + str(self.direction) + " )" | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,730 | adrianogil/miru | refs/heads/main | /src/python/miru/raymarching/sdfcube.py | from miru.engine.transform import Transform
from miru.engine.color import Color
from miru.engine.vector import Vector3
class SDFCube:
def __init__(self, size):
self.size = size
self.transform = Transform()
self.color = Color.random()
def pre_render(self):
pass
def distan... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,731 | adrianogil/miru | refs/heads/main | /src/python/miru/raytracing/quadmesh.py |
from vector import Vector2
def create(mesh, vector1, vector2, initial_position):
initial_index = len(mesh.vertices)
# print("quadmesh.create - " + str(initial_index))
mesh.vertices.append(initial_position)
mesh.vertices.append(initial_position.add(vector1))
mesh.vertices.append(initial_position.... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,732 | adrianogil/miru | refs/heads/main | /src/python/miru/raymarching/scene.py | from PIL import Image
import numpy as np
import os
from miru.engine.vector import Vector3
from miru.engine.color import Color
from miru.engine.camera import Camera
from miru.engine.sceneparser import SceneParser
from miru.engine.material import Material
from miru.raymarching.sdfobjects import SDFCube, SDFSphere
t... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,733 | adrianogil/miru | refs/heads/main | /src/python/miru/raytracing/test.py | from scene import Scene
from camera import Camera
from sphere import Sphere
from plane import Plane
from cube import Cube
from vector import Vector3
from mesh import Mesh
import quadmesh
from light import Light
import random
import os
from color import Color
from meanfilter import MeanFilter
from kernelfilter im... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,734 | adrianogil/miru | refs/heads/main | /src/python/miru/raytracing/test_video.py | import imageio
from scene import Scene
from camera import Camera
from sphere import Sphere
from plane import Plane
from cube import Cube
from vector import Vector3
from mesh import Mesh
import quadmesh
from light import Light
import random
import os
from color import Color
from meanfilter import MeanFilter
from... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,735 | adrianogil/miru | refs/heads/main | /src/python/miru/raytracing/meanfilter.py |
class MeanFilter:
"""
MeanFilter 3x3 for each color channel
"""
def __init__(self):
self.filter_size_x = 5
self.filter_size_y = 5
def apply_effect(self, render_data):
for x in range(0, render_data.pixel_width):
for y in range(0, render_data.pixel_height):
... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,736 | adrianogil/miru | refs/heads/main | /src/python/miru/engine/material.py | from .color import Color
from miru.engine.shaders import UnlitShader, LambertianTintShader
class Material:
def __init__(self):
self.texture = None
self.albedo = Color(1.0, 1.0, 1.0, 1.0)
self.shader = UnlitShader()
self.debug_mode = False
self.debug_render_type = "None"
... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,737 | adrianogil/miru | refs/heads/main | /src/python/miru/raytracing/camera.py | from miru.engine.transform import Transform
class Camera:
ORTOGRAPHIC=1
PERSPECTIVE=2
def __init__(self):
self.mode = Camera.PERSPECTIVE
self.fov = 60
self.near = 0.01
self.far = 1000
self.transform = Transform()
| {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,738 | adrianogil/miru | refs/heads/main | /src/python/miru/raytracing/scenetests.py | from miru.raytracing.scene import render_scene
import os
def run_scenes_tests():
test_folder = "miru_test_results"
miru_folder = os.environ["MIRU_PROJ_PATH"]
if not os.path.exists(test_folder):
os.makedirs(test_folder)
# 1o Test: scenes/test/four_spheres_in_corners.scene
scene_file = os.path.join(miru_fol... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,739 | adrianogil/miru | refs/heads/main | /src/python/miru/raytracing/kernelfilter.py |
class KernelFilter:
"""
KernelFilter for each color channels
"""
def __init__(self, kernel, sizex, sizey):
self.kernel = kernel
self.filter_size_x = sizex
self.filter_size_y = sizey
def apply_effect(self, render_data):
for x in range(0, render_data.pixel_width):... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,740 | adrianogil/miru | refs/heads/main | /src/python/miru/raytracing/mesh.py | import sys
from vector import Vector3, Vector2
import numpy as np
import meshtools
class Mesh:
def __init__(self, vertices=[], triangles=[]):
self.vertices = vertices
self.triangles = triangles
self.uvs = None
self.normals = None
self.material = None
def add_triangle... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,078,741 | adrianogil/miru | refs/heads/main | /src/python/miru/raytracing/test_texture.py | from scene import Scene
from camera import Camera
from sphere import Sphere
from plane import Plane
from cube import Cube
from miru.engine.vector import Vector3
from mesh import Mesh
import quadmesh
from light import Light
import random
import os
from color import Color
from texture import Texture
from miru.eng... | {"/src/python/miru/engine/transform.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/raytracing/scene.py": ["/src/python/miru/raytracing/ray.py"], "/src/python/miru/engine/color.py": ["/src/python/miru/engine/vector.py"], "/src/python/miru/engine/light.py": ["/src/python/miru/engine/transform.py", "/src/py... |
39,088,705 | andirigible/wwplando | refs/heads/main | /randomizers/items.py |
import os
import re
from collections import OrderedDict
from fs_helpers import *
import tweaks
def randomize_items(self):
print("Randomizing items...")
if self.options.get("race_mode"):
randomize_boss_rewards(self)
if not self.options.get("keylunacy"):
randomize_dungeon_items(self)
randomize... | {"/wwlib/rarc.py": ["/wwlib/bmg.py", "/wwlib/bdl.py"], "/randomizer.py": ["/wwlib/rarc.py", "/wwlib/gcm.py", "/customizer.py"]} |
39,088,706 | andirigible/wwplando | refs/heads/main | /randomizers/charts.py |
import copy
def randomize_charts(self):
# Shuffles around which chart points to each sector.
randomizable_charts = [chart for chart in self.chart_list.charts if chart.type in [0, 1, 2, 6]]
original_charts = copy.deepcopy(randomizable_charts)
# Sort the charts by their texture ID so we get the same resul... | {"/wwlib/rarc.py": ["/wwlib/bmg.py", "/wwlib/bdl.py"], "/randomizer.py": ["/wwlib/rarc.py", "/wwlib/gcm.py", "/customizer.py"]} |
39,088,707 | andirigible/wwplando | refs/heads/main | /wwlib/j3d.py |
import os
from enum import Enum
from io import BytesIO
from collections import OrderedDict
from wwlib.bti import BTI
from fs_helpers import *
IMPLEMENTED_CHUNK_TYPES = [
"TEX1",
"MAT3",
"MDL3",
"TRK1",
]
class J3DFile:
def __init__(self, data):
if try_read_str(data, 0, 4) == "Yaz0":
data = Yaz0... | {"/wwlib/rarc.py": ["/wwlib/bmg.py", "/wwlib/bdl.py"], "/randomizer.py": ["/wwlib/rarc.py", "/wwlib/gcm.py", "/customizer.py"]} |
39,088,708 | andirigible/wwplando | refs/heads/main | /randomizer.py |
import os
from io import BytesIO
import shutil
from pathlib import Path
import re
from random import Random
from collections import OrderedDict
import hashlib
import yaml
from fs_helpers import *
from wwlib.yaz0 import Yaz0
from wwlib.rarc import RARC
from wwlib.dol import DOL
from wwlib.rel import REL, RELRelocation... | {"/wwlib/rarc.py": ["/wwlib/bmg.py", "/wwlib/bdl.py"], "/randomizer.py": ["/wwlib/rarc.py", "/wwlib/gcm.py", "/customizer.py"]} |
39,105,054 | sahilsid/EWSNet | refs/heads/main | /convert_js.py | import os
import sys
import pandas as pd
from sklearn.model_selection import train_test_split
import numpy as np
from tensorflow.keras import backend as K
from tensorflow.keras.layers import Conv1D, BatchNormalization, GlobalAveragePooling1D, Permute, Dropout, Flatten
from tensorflow.keras.layers import Input, Dense, L... | {"/src/model_training/exp_utils.py": ["/src/utils/generic_utils.py"]} |
39,105,055 | sahilsid/EWSNet | refs/heads/main | /ewsnet.py |
import os
import tensorflow as tf
from tensorflow.keras import backend as K
from tensorflow.keras.layers import Conv1D, BatchNormalization, GlobalAveragePooling1D, Permute, Dropout, Flatten
from tensorflow.keras.layers import Input, Dense, LSTM, concatenate, Activation, GRU, SimpleRNN
from tensorflow.keras.models imp... | {"/src/model_training/exp_utils.py": ["/src/utils/generic_utils.py"]} |
39,110,314 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/transforms/_transform.py | import collections.abc
import inspect
import re
from typing import Any, Callable, Dict, Optional, Type, Union, cast, Set, Collection
import torch
from torch import nn
from torchvision.prototype import features
from torchvision.prototype.utils._internal import add_suggestion
class Transform(nn.Module):
"""Base cl... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,315 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/datasets/_builtin/imagenet.py | import functools
import io
import pathlib
import re
from typing import Any, Callable, Dict, List, Optional, Tuple, cast
import torch
from torchdata.datapipes.iter import IterDataPipe, LineReader, IterKeyZipper, Mapper, TarArchiveReader, Filter
from torchvision.prototype.datasets.utils import (
Dataset,
Dataset... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,316 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/transforms/_container.py | from typing import Any, List
import torch
from torch import nn
from torchvision.prototype.transforms import Transform
class ContainerTransform(nn.Module):
def supports(self, obj: Any) -> bool:
raise NotImplementedError()
def forward(self, *inputs: Any) -> Any:
raise NotImplementedError()
... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,317 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/datasets/utils/__init__.py | from . import _internal
from ._dataset import DatasetType, DatasetConfig, DatasetInfo, Dataset
from ._query import SampleQuery
from ._resource import OnlineResource, HttpResource, GDriveResource, ManualDownloadResource
| {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,318 | Hsuxu/vision | refs/heads/master | /test/test_prototype_features.py | import functools
import itertools
import pytest
import torch
from torch.testing import make_tensor as _make_tensor, assert_close
from torchvision.prototype import features
from torchvision.prototype.utils._internal import sequence_to_str
make_tensor = functools.partial(_make_tensor, device="cpu", dtype=torch.float32... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,319 | Hsuxu/vision | refs/heads/master | /test/builtin_dataset_mocks.py | import functools
import gzip
import json
import lzma
import pathlib
import pickle
import tempfile
from collections import defaultdict
from typing import Any, Dict, Tuple
import numpy as np
import PIL.Image
import pytest
import torch
from datasets_utils import create_image_folder, make_tar, make_zip
from torch.testing ... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,320 | Hsuxu/vision | refs/heads/master | /test/test_prototype_transforms.py | import pytest
from torchvision.prototype import transforms, features
from torchvision.prototype.utils._internal import sequence_to_str
FEATURE_TYPES = {
feature_type
for name, feature_type in features.__dict__.items()
if not name.startswith("_")
and isinstance(feature_type, type)
and issubclass(fe... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,321 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/utils/_internal.py | import collections.abc
import difflib
import enum
import functools
import inspect
import os
import os.path
import textwrap
import warnings
from typing import Collection, Sequence, Callable, Any, Iterator, NoReturn, Mapping, TypeVar, Iterable, Tuple, cast
__all__ = [
"StrEnum",
"sequence_to_str",
"add_sugge... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,322 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/features/__init__.py | from ._bounding_box import BoundingBoxFormat, BoundingBox
from ._feature import Feature, DEFAULT
from ._image import Image, ColorSpace
from ._label import Label
| {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,323 | Hsuxu/vision | refs/heads/master | /torchvision/models/segmentation/segmentation.py | import warnings
# Import all methods/classes for BC:
from . import * # noqa: F401, F403
warnings.warn(
"The 'torchvision.models.segmentation.segmentation' module is deprecated. Please use directly the parent module "
"instead."
)
| {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,324 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/transforms/__init__.py | from ._transform import Transform
from ._container import Compose, RandomApply, RandomChoice, RandomOrder # usort: skip
from ._geometry import Resize, RandomResize, HorizontalFlip, Crop, CenterCrop, RandomCrop
from ._misc import Identity, Normalize
from ._presets import CocoEval, ImageNetEval, VocEval, Kinect400Eval,... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,325 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/datasets/_builtin/semeion.py | import functools
import io
from typing import Any, Callable, Dict, List, Optional, Tuple
import torch
from torchdata.datapipes.iter import (
IterDataPipe,
Mapper,
CSVParser,
)
from torchvision.prototype.datasets.decoder import raw
from torchvision.prototype.datasets.utils import (
Dataset,
DatasetC... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,326 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/features/_bounding_box.py | import enum
import functools
from typing import Callable, Union, Tuple, Dict, Any, Optional, cast
import torch
from torchvision.prototype.utils._internal import StrEnum
from ._feature import Feature, DEFAULT
class BoundingBoxFormat(StrEnum):
# this is just for test purposes
_SENTINEL = -1
XYXY = enum.au... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,327 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/features/_image.py | from typing import Dict, Any, Union, Tuple
import torch
from torchvision.prototype.utils._internal import StrEnum
from ._feature import Feature, DEFAULT
class ColorSpace(StrEnum):
# this is just for test purposes
_SENTINEL = -1
OTHER = 0
GRAYSCALE = 1
RGB = 3
class Image(Feature):
color_sp... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,328 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/datasets/decoder.py | import io
import PIL.Image
import torch
from torchvision.prototype import features
from torchvision.transforms.functional import pil_to_tensor
__all__ = ["raw", "pil"]
def raw(buffer: io.IOBase) -> torch.Tensor:
raise RuntimeError("This is just a sentinel and should never be called.")
def pil(buffer: io.IOBas... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,329 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/transforms/_geometry.py | from typing import Any, Dict, Tuple, Union
import torch
from torch.nn.functional import interpolate
from torchvision.prototype.datasets.utils import SampleQuery
from torchvision.prototype.features import BoundingBox, Image, Label
from torchvision.prototype.transforms import Transform
class HorizontalFlip(Transform):... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,330 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/features/_label.py | from typing import Dict, Any, Optional, Tuple
from ._feature import Feature, DEFAULT
class Label(Feature):
category: Optional[str]
@classmethod
def _parse_meta_data(
cls,
category: Optional[str] = DEFAULT, # type: ignore[assignment]
) -> Dict[str, Tuple[Any, Any]]:
return di... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,331 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/features/_feature.py | from typing import Tuple, cast, TypeVar, Set, Dict, Any, Callable, Optional, Mapping, Type, Sequence
import torch
from torch._C import _TensorBase, DisableTorchFunction
from torchvision.prototype.utils._internal import add_suggestion
F = TypeVar("F", bound="Feature")
DEFAULT = object()
class Feature(torch.Tensor... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,332 | Hsuxu/vision | refs/heads/master | /test/test_prototype_builtin_datasets.py | import io
import builtin_dataset_mocks
import pytest
import torch
from torch.utils.data.datapipes.iter.grouping import ShardingFilterIterDataPipe as ShardingFilter
from torch.utils.data.graph import traverse
from torchdata.datapipes.iter import IterDataPipe, Shuffler
from torchvision.prototype import datasets, transfo... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,333 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/transforms/_misc.py | from typing import Any, Dict, Sequence
import torch
from torchvision.prototype.features import Image, BoundingBox, Label
from torchvision.prototype.transforms import Transform
class Identity(Transform):
"""Identity transform that supports all built-in :class:`~torchvision.prototype.features.Feature`'s."""
d... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,110,334 | Hsuxu/vision | refs/heads/master | /torchvision/prototype/datasets/_builtin/celeba.py | import csv
import functools
import io
from typing import Any, Callable, Dict, List, Optional, Tuple, Iterator, Sequence
import torch
from torchdata.datapipes.iter import (
IterDataPipe,
Mapper,
Filter,
Zipper,
IterKeyZipper,
)
from torchvision.prototype.datasets.utils import (
Dataset,
Data... | {"/torchvision/prototype/transforms/_transform.py": ["/torchvision/prototype/transforms/_utils.py", "/torchvision/prototype/__init__.py", "/torchvision/prototype/utils/_internal.py"], "/torchvision/prototype/transforms/_meta.py": ["/torchvision/prototype/__init__.py", "/torchvision/prototype/transforms/__init__.py"], "... |
39,118,944 | crandersn/temp_probe_pi_code | refs/heads/master | /ConnectedTemperatureSensor.py | from threading import Timer
import pyrebase
from TempReader import TempReader
from datetime import datetime
import RPi.GPIO as GPIO
import board
import digitalio
import adafruit_character_lcd.character_lcd as characterlcd
import time
import datetime
import boto3
import os
class ConnectedTempSensor:
def __init... | {"/ConnectedTemperatureSensor.py": ["/TempReader.py"], "/RunTempSensor.py": ["/ConnectedTemperatureSensor.py"]} |
39,118,945 | crandersn/temp_probe_pi_code | refs/heads/master | /AwsTimeStream.py | import boto3
import time
class TimeStream:
def __init__ (self, dbName, tableName, region):
self.dbName = dbName
self.tableName = tableName
self.region = region
def write(self, temp):
CURRENT_TIME = str(int(time.time() *1000))
client = boto3.client('timestream-... | {"/ConnectedTemperatureSensor.py": ["/TempReader.py"], "/RunTempSensor.py": ["/ConnectedTemperatureSensor.py"]} |
39,118,946 | crandersn/temp_probe_pi_code | refs/heads/master | /TempReader.py | import glob
class TempReader:
base_dir = '/sys/bus/w1/devices/'
device_folder = glob.glob(base_dir + '28*')[0]
device_file = device_folder + '/w1_slave'
@staticmethod
def readTempFile(): #Fundtion to read the value of Temperature
try:
file = open(TempReader.d... | {"/ConnectedTemperatureSensor.py": ["/TempReader.py"], "/RunTempSensor.py": ["/ConnectedTemperatureSensor.py"]} |
39,118,947 | crandersn/temp_probe_pi_code | refs/heads/master | /RunTempSensor.py | from ConnectedTemperatureSensor import ConnectedTempSensor
tempSensor = ConnectedTempSensor()
tempSensor.run() | {"/ConnectedTemperatureSensor.py": ["/TempReader.py"], "/RunTempSensor.py": ["/ConnectedTemperatureSensor.py"]} |
39,220,242 | Venko15/MLT | refs/heads/master | /bot/cogs/autorole.py | from re import L
from types import coroutine
from discord import client
import pymongo
import discord
import asyncio
from discord.ext import commands
from discord import Client
from discord import channel, guild
from discord.ext.commands.core import command
import os, sys
import typing as t
class AutoRbot(commands.Cog... | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,243 | Venko15/MLT | refs/heads/master | /env/Lib/site-packages/httpcore/_sync/connection.py | from ssl import SSLContext
from typing import List, Optional, Tuple, cast
from .._backends.sync import SyncBackend, SyncLock, SyncSocketStream, SyncBackend
from .._exceptions import ConnectError, ConnectTimeout
from .._types import URL, Headers, Origin, TimeoutDict
from .._utils import exponential_backoff, get_logger,... | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,244 | Venko15/MLT | refs/heads/master | /env/Lib/site-packages/youtubesearchpython/handlers/componenthandler.py | from typing import List, Union
from youtubesearchpython.internal.constants import *
class ComponentHandler:
def _getVideoComponent(self, element: dict, shelfTitle: str = None) -> dict:
video = element[videoElementKey]
component = {
'type': 'video',
... | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,245 | Venko15/MLT | refs/heads/master | /bot/glbs.py | queue = {} | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,246 | Venko15/MLT | refs/heads/master | /env/Lib/site-packages/prawcore/auth.py | """Provides Authentication and Authorization classes."""
import time
from requests import Request
from requests.status_codes import codes
from . import const
from .exceptions import InvalidInvocation, OAuthException, ResponseException
class BaseAuthenticator(object):
"""Provide the base authenticator object tha... | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,247 | Venko15/MLT | refs/heads/master | /bot/cogs/memes.py | import random
from discord import embeds
import datetime as dt
from discord.ext import commands
import discord
import requests
class Meme(commands.Cog):
def __init__(self, bot):
self.bot = bot
@commands.command(name='memepls', aliases = ["meme"])
async def memepls(self,ctx):
res = requests.... | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,248 | Venko15/MLT | refs/heads/master | /bot/cogs/weatherbot.py | from re import I
import discord
from discord.ext import commands
from discord.ext.commands.core import command
import requests
import json
import datetime
import typing as t
import unittest
class NoCityName(commands.CommandError):
pass
class NoCityFound(commands.CommandError):
pass
class WeatherBot(commands.C... | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,249 | Venko15/MLT | refs/heads/master | /launcher.py | from bot import MLT
def main():
bot=MLT()
bot.run()
if __name__ == "__main__":
main() | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,250 | Venko15/MLT | refs/heads/master | /bot/__init__.py | from .dsBot import MLT
from . import glbs | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,251 | Venko15/MLT | refs/heads/master | /env/Lib/site-packages/youtubesearchpython/__future__/internal/streamurlfetcher.py | isPyTubeInstalled = False
import asyncio
import httpx
try:
from youtubesearchpython.__future__.internal.json import loads
from pytube.extract import apply_descrambler, apply_signature
from pytube import YouTube, extract
from urllib.parse import parse_qs
isPyTubeInstalled = True
except:
class Yo... | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,252 | Venko15/MLT | refs/heads/master | /bot/cogs/musicv3.py | from cache_decorator.cache import cache
from discord.activity import Spotify
from discord.player import FFmpegAudio
from youtube_dl import YoutubeDL
import discord
import asyncio
from discord import FFmpegPCMAudio
from discord.ext.commands.cog import Cog
from discord.ext.commands.core import command
from discord.ext im... | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,253 | Venko15/MLT | refs/heads/master | /bot/dsBot.py | from pathlib import Path
from discord import channel, guild
from discord.ext.commands.core import command
import pymongo
import discord
from discord.ext import commands
import asyncio
import datetime
import random
from discord import Client
from dotenv import load_dotenv
import os
import dns
import unittest
load_dotenv... | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,254 | Venko15/MLT | refs/heads/master | /env/Lib/site-packages/youtubesearchpython/__init__.py | from youtubesearchpython.search import Search, VideosSearch, ChannelsSearch, PlaylistsSearch, CustomSearch
from youtubesearchpython.extras import Video, Playlist, Suggestions
from youtubesearchpython.streamurlfetcher import StreamURLFetcher
from youtubesearchpython.internal.constants import *
__title__ = 'yout... | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,220,255 | Venko15/MLT | refs/heads/master | /env/Lib/site-packages/httpcore/_async/http11.py | from ssl import SSLContext
from typing import AsyncIterator, List, Tuple, Union, cast
import h11
from .._backends.auto import AsyncSocketStream
from .._bytestreams import AsyncIteratorByteStream
from .._exceptions import LocalProtocolError, RemoteProtocolError, map_exceptions
from .._types import URL, Headers, Timeou... | {"/dsBot.py": ["/music.py", "/memes.py"], "/launcher.py": ["/bot/__init__.py"], "/bot/__init__.py": ["/bot/dsBot.py"], "/bot/cogs/musicv3.py": ["/bot/glbs.py"], "/bot/dsBot.py": ["/bot/glbs.py", "/bot/cogs/musicv3.py"]} |
39,254,963 | Muhammad-Elgendi/Athena | refs/heads/production | /assistant/views.py | from .models import Question,Choice
from django.http import JsonResponse
from django.template import loader
from django.http import HttpResponse, HttpResponseRedirect , Http404
from django.shortcuts import get_object_or_404, render
from django.urls import reverse
from django.conf import settings
from django.core.files.... | {"/assistant/views.py": ["/assistant/modules/FaceRecognizer.py", "/assistant/modules/Chatbot.py"], "/assistant/modules/Chatbot.py": ["/assistant/modules/EmotionAnalyser.py"]} |
39,254,964 | Muhammad-Elgendi/Athena | refs/heads/production | /assistant/modules/EmotionAnalyser.py | # -*- coding: utf-8 -*-
import nltk.classify.util
from nltk.classify import NaiveBayesClassifier
import nltk
from nltk.corpus import stopwords
from nltk import ngrams
from nltk.tokenize import TweetTokenizer
import csv
import _pickle as cPickle
class EmotionAnalyser:
__instance = None
@staticmethod
def ge... | {"/assistant/views.py": ["/assistant/modules/FaceRecognizer.py", "/assistant/modules/Chatbot.py"], "/assistant/modules/Chatbot.py": ["/assistant/modules/EmotionAnalyser.py"]} |
39,254,965 | Muhammad-Elgendi/Athena | refs/heads/production | /assistant/urls.py | from django.urls import path
from . import views
from django.conf import settings
from django.conf.urls.static import static
app_name = 'assistant'
urlpatterns = [
# ex: /assistant/
path('', views.index, name='index'),
# ex: /assistant/5/
# path('<int:question_id>/', views.detail, name='detail'),
#... | {"/assistant/views.py": ["/assistant/modules/FaceRecognizer.py", "/assistant/modules/Chatbot.py"], "/assistant/modules/Chatbot.py": ["/assistant/modules/EmotionAnalyser.py"]} |
39,254,966 | Muhammad-Elgendi/Athena | refs/heads/production | /assistant/modules/Chatbot.py | import nltk
from nltk.corpus import stopwords
import numpy as np
import random
import string
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.metrics.pairwise import cosine_similarity
from .EmotionAnalyser import EmotionAnalyser
# first-time use only
# nltk.download('punkt')
# nltk.download('w... | {"/assistant/views.py": ["/assistant/modules/FaceRecognizer.py", "/assistant/modules/Chatbot.py"], "/assistant/modules/Chatbot.py": ["/assistant/modules/EmotionAnalyser.py"]} |
39,254,967 | Muhammad-Elgendi/Athena | refs/heads/production | /assistant/modules/EmotionAnalyser_scikit_learn.py | from sklearn.feature_extraction.text import CountVectorizer
from sklearn.model_selection import train_test_split
# class EmotionAnalyser:
# pass
count_vect = CountVectorizer()
| {"/assistant/views.py": ["/assistant/modules/FaceRecognizer.py", "/assistant/modules/Chatbot.py"], "/assistant/modules/Chatbot.py": ["/assistant/modules/EmotionAnalyser.py"]} |
39,254,968 | Muhammad-Elgendi/Athena | refs/heads/production | /assistant/modules/FaceRecognizer.py | import cv2 as cv2
import numpy as np
import glob
from os import listdir
from os.path import isfile, join,split
class FaceRecognizer():
def __init__(self,base_path):
#create empth face list
self.faceSamples = []
#create empty ID list
self.ids = []
#create empty Persons list
... | {"/assistant/views.py": ["/assistant/modules/FaceRecognizer.py", "/assistant/modules/Chatbot.py"], "/assistant/modules/Chatbot.py": ["/assistant/modules/EmotionAnalyser.py"]} |
39,265,170 | LeResKP/pyramid_sqladmin | refs/heads/master | /setup.py | from setuptools import setup, find_packages
import sys, os
# Hack to prevent TypeError: 'NoneType' object is not callable error
# on exit of python setup.py test
try:
import multiprocessing
except ImportError:
pass
version = '0.1'
setup(name='pyramid_sqladmin/',
version=version,
description="Simp... | {"/tests/test_init.py": ["/pyramid_sqladmin/__init__.py"]} |
39,292,865 | juliarobles/graffitisWeb | refs/heads/master | /graffitiApp/views.py | from django.shortcuts import render
from rest_framework import serializers
from rest_framework_mongoengine import viewsets
from .serializers import PublicacionSerializer, UsuarioSerializer, GraffitiSerializer
from .models import Publicacion, Usuario, Graffiti
from django.http import HttpResponse
from bson import Object... | {"/graffitiApp/admin.py": ["/graffitiApp/models.py"], "/graffitiApp/Apiviews/ComentarioAPIView.py": ["/graffitiApp/models.py", "/graffitiApp/Apiviews/PublicacionAPIView.py", "/graffitiApp/serializers.py"], "/graffitiApp/urls.py": ["/graffitiApp/views.py", "/graffitiApp/Apiviews/UserAPIView.py", "/graffitiApp/Apiviews/P... |
39,297,862 | roblee357/Clarios_mold_OCR | refs/heads/master | /pil_scratch.py | from PIL import Image
import PIL.ImageTk
capture_image = PIL.Image.open('Android_Web\\flask\\examples\\tutorial\\flaskr\\images\\2021_03_26___13_06_45.png') | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,863 | roblee357/Clarios_mold_OCR | refs/heads/master | /OCR.py | import numpy as np
import cv2
import string
import sys
import os
import pytesseract
from pytesseract import Output
from threading import Thread, Event
from time import sleep
import struct
# import redis
import time
import preprocess
class OCR():
def __init__(self, threshold=None):
self.DD = self.Del()
... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,864 | roblee357/Clarios_mold_OCR | refs/heads/master | /mold_No_button.py | #! python3
import pyautogui, sys, keyboard, time, json, os, win32clipboard
from tkinter import *
import win_loc
global wormhole_active, CBdata, prev_x, prev_y, config
keyboard.is_pressed('F8')
configPath = 'config.json'
configExists = os.path.exists(configPath)
if keyboard.is_pressed('F8') or not configExists:
i... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,865 | roblee357/Clarios_mold_OCR | refs/heads/master | /key_poller.py | from pynput.keyboard import Key, Listener
import time
from threading import Thread
class Key_listener():
def __init__(self):
self.queue = [2,1,1,1,1,1,1,1]
self.itime = time.time()
self.scanned = False
start_thread = Thread(target=self.start)
start_thread.start()
def ... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,866 | roblee357/Clarios_mold_OCR | refs/heads/master | /win_loc.py | import win32gui
def callback(hwnd, pos):
rect = win32gui.GetWindowRect(hwnd)
x = rect[0]
y = rect[1]
w = rect[2] - x
h = rect[3] - y
if 'PSMII Bluesheet' in win32gui.GetWindowText(hwnd):
# print("Window %s:" % win32gui.GetWindowText(hwnd))
# print("\tLocation: (%d, %d)" % (x, y)... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,867 | roblee357/Clarios_mold_OCR | refs/heads/master | /gui.py | from tkinter import *
from PIL import Image
from PIL import ImageTk
import cv2, time
import numpy as np
from datetime import datetime
import subprocess, os, sys
from multiprocessing import Process
import capture, preprocess, OCR
root = Tk()
panelA = None
panelB = None
image = Image.open('redis_not_started.jpg')
ROI... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,868 | roblee357/Clarios_mold_OCR | refs/heads/master | /zmqimage_server_test.py | import zmqimage, cv2
zmq = zmqimage.zmqImageShowServer(open_port="tcp://*:5555")
print("Starting zmqImageShow Server...")
print(" press Ctrl-C to stop")
i = 0
while True: # Until Ctrl-C is pressed, will repeatedly
arrayname, image = zmq.getArray()
cv2.imshow(arrayname, image)
cv2.waitKey(1)
pri... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,869 | roblee357/Clarios_mold_OCR | refs/heads/master | /capture.py | import cv2, time, json
import struct
##import redis
import numpy as np
import ctypes
import argparse
from random import randrange
import pickle as pkl
# from multiprocessing import Process, Value, Array
from threading import Thread, Event
def parseCmdLineArgs ():
# parse the command line
parser = argparse.Argu... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,870 | roblee357/Clarios_mold_OCR | refs/heads/master | /sub_img.py | from ctypes import create_string_buffer
import sys, json, pickle
import zmq
import numpy as np
import time, cv2
# Socket to talk to server
context = zmq.Context()
socket = context.socket(zmq.SUB)
print("Collecting updates from weather server...")
socket.connect("tcp://localhost:5555")
zip_filter = 'raw'
socket.sets... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,871 | roblee357/Clarios_mold_OCR | refs/heads/master | /preprocess.py | import numpy as np
import cv2, string, sys, os, json, time
import pytesseract
from pytesseract import Output
from multiprocessing import Process
from time import sleep
import struct
import argparse
import pickle as pkl
from threading import Thread
import capture
configPath = 'config.json'
class Processor():
def ... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,872 | roblee357/Clarios_mold_OCR | refs/heads/master | /zmqimage_client_test.py | import numpy as np
import cv2
import zmqimage
print("Connecting to zmqShowImage Server ... ")
zmq = zmqimage.zmqConnect(connect_to="tcp://127.0.0.1:5555")
for i in range(255):
image = np.ones((500, 500), dtype="uint8")*i
zmq.imshow("Zero Image 500 x 500", image)
# build a rectangular mask & display it
m... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,873 | roblee357/Clarios_mold_OCR | refs/heads/master | /thread_play.py | from threading import Thread
from queue import Queue
import time
def thread1(threadname, q):
#read variable "a" modify by thread 2
while True:
a = q.get()
if a is None: return # Poison pill
print(a)
def thread2(threadname, q):
a = 0
for _ in range(10):
a += 1
q.... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,874 | roblee357/Clarios_mold_OCR | refs/heads/master | /Live_phone.py | import os
import pathlib
import cv2
from threading import Thread, Event
class Live_phone:
def __init__(self):
self.img = None
def get(self, img):
# self.img = img
i = 0
while True:
try:
path = 'Android_Web\\flask\\examples\\tutorial\\flaskr\\images\\l... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,875 | roblee357/Clarios_mold_OCR | refs/heads/master | /zmqimage.py | # zmqimage.py -- classes to send, receive and display cv2 images via zmq
# based on serialization in pyzmq docs and pyzmq/examples/serialization
'''
PURPOSE:
These classes allow a headless (no display) computer running OpenCV code
to display OpenCV images on another computer with a display.
For example... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,876 | roblee357/Clarios_mold_OCR | refs/heads/master | /gui2.py | from io import RawIOBase
import capture, preprocess, OCR
import time, cv2
cap = capture.Capture(1)
print('before run')
cap.start()
t = time.time()
while cap.img is None:
print('waiting for camera to connect',time.time()-t)
time.sleep(.25)
proc = preprocess.Processor()
proc.start(cap)
while proc.img is None:
... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,877 | roblee357/Clarios_mold_OCR | refs/heads/master | /badges.py | import pandas as pd
import tkinter, datetime, pyautogui, time
class Badge:
def __init__(self, master):
self.master = master
self.path = 'badges.csv'
self.df = pd.read_csv(self.path,index_col='index') # ,header=None , comment = '#',names=["code", "firstname", "lastname", "date_added"]
... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,297,878 | roblee357/Clarios_mold_OCR | refs/heads/master | /multiprocess_test.py | from multiprocessing import Process, Lock
from multiprocessing.sharedctypes import Value, Array
from ctypes import Structure, c_double
import numpy as np
import ctypes
class Point(Structure):
_fields_ = [('x', c_double), ('y', c_double)]
# _fields_ = np.array([[1.875,-6.25], [-5.75,2.0], [2.375,9.5]])
def mod... | {"/gui_web.py": ["/capture.py", "/preprocess.py", "/OCR.py", "/win_loc.py", "/badges.py", "/Live_phone.py"], "/OCR.py": ["/preprocess.py"], "/mold_No_button.py": ["/win_loc.py"], "/gui.py": ["/capture.py", "/preprocess.py", "/OCR.py"], "/zmqimage_server_test.py": ["/zmqimage.py"], "/preprocess.py": ["/capture.py"], "/z... |
39,356,916 | kalizhaankyzy/STUDGUIDE-WebDevProject | refs/heads/master | /api/urls.py | from django.urls import path
from rest_framework_jwt.views import obtain_jwt_token
from rest_framework.urlpatterns import format_suffix_patterns
from .views import author_detail, author_list, category_list, category_detail, category_news, course_detail, course_list, level_detail, level_list, news_detail, news_list
url... | {"/sg_back/api/serializers.py": ["/sg_back/api/models.py"], "/sg_back/api/urls.py": ["/sg_back/api/views.py"], "/sg_back/api/views.py": ["/sg_back/api/models.py", "/sg_back/api/serializers.py"]} |
39,484,224 | amanmyrats/taksi | refs/heads/master | /api/urls.py | from django.urls import path
from .views import saher_ara_view, saher_ici_view, etrap_obalary_view
urlpatterns = [
path('api/saherara/', saher_ara_view),
path('api/saherici/', saher_ici_view),
path('api/etrapobalary/', etrap_obalary_view)
]
| {"/taksist/urls.py": ["/taksist/views.py"], "/yolagcy/views.py": ["/yolagcy/filters.py", "/api/views.py", "/taksist/models.py"], "/api/urls.py": ["/api/views.py"], "/api/views.py": ["/api/serializers.py", "/taksist/models.py"], "/yolagcy/filters.py": ["/taksist/models.py"], "/taksist/migrations of sqlite3/0001_initial.... |
39,484,225 | amanmyrats/taksi | refs/heads/master | /api/views.py | from django.http import JsonResponse
from .serializers import TaxiSerializer
from taksist.models import TaxiProfile
def saher_ara_view(request):
if request.method=='GET':
taksistler = TaxiProfile.objects.all()
serializer = TaxiSerializer(taksistler, many=True)
return JsonResponse(serializ... | {"/taksist/urls.py": ["/taksist/views.py"], "/yolagcy/views.py": ["/yolagcy/filters.py", "/api/views.py", "/taksist/models.py"], "/api/urls.py": ["/api/views.py"], "/api/views.py": ["/api/serializers.py", "/taksist/models.py"], "/yolagcy/filters.py": ["/taksist/models.py"], "/taksist/migrations of sqlite3/0001_initial.... |
39,484,226 | amanmyrats/taksi | refs/heads/master | /api/serializers.py | from rest_framework import serializers
from taksist.models import TaxiProfile, User, TaxiStatus, Category, Status
class TaxiSerializer(serializers.ModelSerializer):
username = serializers.SerializerMethodField('get_taksist_name')
# xstatus = serializers.SerializerMethodField('get_taksist_status')
# xcateg... | {"/taksist/urls.py": ["/taksist/views.py"], "/yolagcy/views.py": ["/yolagcy/filters.py", "/api/views.py", "/taksist/models.py"], "/api/urls.py": ["/api/views.py"], "/api/views.py": ["/api/serializers.py", "/taksist/models.py"], "/yolagcy/filters.py": ["/taksist/models.py"], "/taksist/migrations of sqlite3/0001_initial.... |
39,484,227 | amanmyrats/taksi | refs/heads/master | /test.py | # import os
# from twilio.rest import Client
# account_sid=os.environ['ACa0b0646156f095b60fc6a159cad21464']
# auth_token=os.environ['1ba680297b4b5ec2f695a961e88e8985']
# client=Client(account_sid, auth_token)
# message=client.messages.create(body="Join Earth's mightiest heroes. Like Kevin Bacon.",from_='+12057360756... | {"/taksist/urls.py": ["/taksist/views.py"], "/yolagcy/views.py": ["/yolagcy/filters.py", "/api/views.py", "/taksist/models.py"], "/api/urls.py": ["/api/views.py"], "/api/views.py": ["/api/serializers.py", "/taksist/models.py"], "/yolagcy/filters.py": ["/taksist/models.py"], "/taksist/migrations of sqlite3/0001_initial.... |
39,484,228 | amanmyrats/taksi | refs/heads/master | /taksist/views.py | from django.shortcuts import render, redirect
from django.contrib.auth.forms import UserCreationForm
from django.contrib.auth import login, authenticate, logout
from django.contrib.auth.models import User
from django.http import JsonResponse
from .models import TaxiProfile , TaxiStatus, TaxiCategory
from .forms import... | {"/taksist/urls.py": ["/taksist/views.py"], "/yolagcy/views.py": ["/yolagcy/filters.py", "/api/views.py", "/taksist/models.py"], "/api/urls.py": ["/api/views.py"], "/api/views.py": ["/api/serializers.py", "/taksist/models.py"], "/yolagcy/filters.py": ["/taksist/models.py"], "/taksist/migrations of sqlite3/0001_initial.... |
39,484,229 | amanmyrats/taksi | refs/heads/master | /taksist/admin.py | from django.contrib import admin
from .models import TaxiProfile , TaxiCategory, TaxiStatus
admin.site.register( TaxiProfile)
admin.site.register( TaxiCategory)
admin.site.register( TaxiStatus) | {"/taksist/urls.py": ["/taksist/views.py"], "/yolagcy/views.py": ["/yolagcy/filters.py", "/api/views.py", "/taksist/models.py"], "/api/urls.py": ["/api/views.py"], "/api/views.py": ["/api/serializers.py", "/taksist/models.py"], "/yolagcy/filters.py": ["/taksist/models.py"], "/taksist/migrations of sqlite3/0001_initial.... |
39,484,230 | amanmyrats/taksi | refs/heads/master | /taksist/forms.py | from django import forms
from .models import TaxiProfile , TaxiStatus, TaxiCategory
from django.contrib.auth.models import User
class TaxiProfileModelForm(forms.ModelForm):
class Meta:
model=TaxiProfile
fields=['car_photo', 'user_photo', 'mobile']
class TaxiCategoryModelForm(forms.ModelForm):
... | {"/taksist/urls.py": ["/taksist/views.py"], "/yolagcy/views.py": ["/yolagcy/filters.py", "/api/views.py", "/taksist/models.py"], "/api/urls.py": ["/api/views.py"], "/api/views.py": ["/api/serializers.py", "/taksist/models.py"], "/yolagcy/filters.py": ["/taksist/models.py"], "/taksist/migrations of sqlite3/0001_initial.... |
39,484,231 | amanmyrats/taksi | refs/heads/master | /yolagcy/views.py | from django.shortcuts import render
from django.contrib.auth.models import User
def search_view(request):
taxi_drivers=User.objects.all()
return render(request, 'search.html', {'taxi_drivers':taxi_drivers})
| {"/taksist/urls.py": ["/taksist/views.py"], "/yolagcy/views.py": ["/yolagcy/filters.py", "/api/views.py", "/taksist/models.py"], "/api/urls.py": ["/api/views.py"], "/api/views.py": ["/api/serializers.py", "/taksist/models.py"], "/yolagcy/filters.py": ["/taksist/models.py"], "/taksist/migrations of sqlite3/0001_initial.... |
39,486,895 | conskourris/fyp | refs/heads/master | /patterns_final.py | import numpy as np
from definitions.doji import doji_60
from definitions.close_near_high import c_high60
from definitions.close_near_low import c_low122
def no_definition(df, index) :
return True
downtrend = no_definition
uptrend = no_definition
is_tall = no_definition
is_doji = doji_60
close_near_high = c_high6... | {"/main.py": ["/patterns_final.py", "/tools.py", "/evaluate_findings.py", "/evaluation_tools.py", "/plotting_tools.py", "/trading_tools.py", "/definitions/positive_fit.py", "/definitions/positive_mav4.py", "/definitions/positive_mav5.py", "/definitions/negative_fit.py", "/definitions/negative_mav3.py", "/definitions/ne... |
39,486,896 | conskourris/fyp | refs/heads/master | /tools.py | import os
import datetime as dt
import matplotlib.pyplot as plt
from matplotlib import style
import matplotlib
import numpy as np
from math import floor
from scipy.stats import skew
import mplfinance as mpf
import pandas as pd
import pandas_datareader as web
import random
import pickle
import json
import statistics
... | {"/main.py": ["/patterns_final.py", "/tools.py", "/evaluate_findings.py", "/evaluation_tools.py", "/plotting_tools.py", "/trading_tools.py", "/definitions/positive_fit.py", "/definitions/positive_mav4.py", "/definitions/positive_mav5.py", "/definitions/negative_fit.py", "/definitions/negative_mav3.py", "/definitions/ne... |
39,486,897 | conskourris/fyp | refs/heads/master | /trading_strategies.py | from trading_strategies import *
from patterns_final import *
from tools import *
from evaluation_tools import *
from plotting_tools import *
def sell_after_1(pattern) :
indexes = get_pattern_indexes(pattern)
for ticker in indexes :
for index in indexes[ticker] :
| {"/main.py": ["/patterns_final.py", "/tools.py", "/evaluate_findings.py", "/evaluation_tools.py", "/plotting_tools.py", "/trading_tools.py", "/definitions/positive_fit.py", "/definitions/positive_mav4.py", "/definitions/positive_mav5.py", "/definitions/negative_fit.py", "/definitions/negative_mav3.py", "/definitions/ne... |
39,486,898 | conskourris/fyp | refs/heads/master | /main.py | import os
import datetime as dt
import matplotlib.pyplot as plt
from matplotlib import style
import matplotlib
import numpy as np
from math import floor
from scipy.stats import skew
import mplfinance as mpf
import pandas as pd
import pandas_datareader as web
import random
import pickle
import json
from trading_strat... | {"/main.py": ["/patterns_final.py", "/tools.py", "/evaluate_findings.py", "/evaluation_tools.py", "/plotting_tools.py", "/trading_tools.py", "/definitions/positive_fit.py", "/definitions/positive_mav4.py", "/definitions/positive_mav5.py", "/definitions/negative_fit.py", "/definitions/negative_mav3.py", "/definitions/ne... |
39,495,702 | xdiabetes/xproject_api | refs/heads/master | /xapp/api_v1/consumers/DispatchConsumer.py | import asyncio
from django.http.response import Http404
from rest_framework.exceptions import APIException
from xapp.api_v1.consumers.BasicConsumer import BasicConsumer
from xapp.api_v1.consumers.helpers import SocketException, ConsumerException
from .ConsumerView import Path, ConsumerRequest, WatcherList, ConsumerRe... | {"/user_profile/models.py": ["/location/models.py", "/user_profile/helpers.py"], "/diabo/tests.py": ["/diabo/models.py", "/job/models.py", "/user_profile/serializers.py", "/user_profile/tests.py"], "/job/tests.py": ["/job/models.py", "/user_profile/tests.py"], "/diabo/views.py": ["/diabo/models.py", "/job/models.py", "... |
39,495,703 | xdiabetes/xproject_api | refs/heads/master | /sms/migrations/0007_auto_20191010_1300.py | # Generated by Django 2.2.4 on 2019-10-10 13:00
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('sms', '0006_auto_20190623_1410'),
]
operations = [
migrations.AlterModelOptions(
name='message',
options={'verbose_n... | {"/user_profile/models.py": ["/location/models.py", "/user_profile/helpers.py"], "/diabo/tests.py": ["/diabo/models.py", "/job/models.py", "/user_profile/serializers.py", "/user_profile/tests.py"], "/job/tests.py": ["/job/models.py", "/user_profile/tests.py"], "/diabo/views.py": ["/diabo/models.py", "/job/models.py", "... |
39,495,704 | xdiabetes/xproject_api | refs/heads/master | /xapp/api_v1/consumers/ConsumerView.py | from abc import ABC
from django.utils.translation import gettext as _
from rest_framework import serializers, status
from rest_framework.permissions import BasePermission
from xapp.api_v1.consumers.helpers import ConsumerException, MODIFY_METHODS, WATCH, get_object_or_404
class NotImplementedConsumer(ConsumerExcept... | {"/user_profile/models.py": ["/location/models.py", "/user_profile/helpers.py"], "/diabo/tests.py": ["/diabo/models.py", "/job/models.py", "/user_profile/serializers.py", "/user_profile/tests.py"], "/job/tests.py": ["/job/models.py", "/user_profile/tests.py"], "/diabo/views.py": ["/diabo/models.py", "/job/models.py", "... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.