code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
import pathlib
import os.path
cwd = pathlib.Path(__file__).parent.absolute()
cwd = os.path.abspath(cwd)
fp = os.path.join(cwd, 'ProjectSettings', 'InputManager.asset')
def output_unity_axis(f, joy_idx, name, button='', axis_type=0, joy_axis=0):
f.write(' - serializedVersion: 3\n')
f.write(' m_Name: joyst... | [
"pathlib.Path"
] | [((37, 59), 'pathlib.Path', 'pathlib.Path', (['__file__'], {}), '(__file__)\n', (49, 59), False, 'import pathlib\n')] |
# Copyright 2019 The Johns Hopkins University Applied Physics Laboratory
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required ... | [
"numpy.load",
"numpy.save",
"argparse.ArgumentParser",
"numpy.zeros",
"numpy.array",
"itertools.product"
] | [((862, 881), 'numpy.load', 'np.load', (['args.input'], {}), '(args.input)\n', (869, 881), True, 'import numpy as np\n'), ((1066, 1090), 'numpy.zeros', 'np.zeros', (['cell_map.shape'], {}), '(cell_map.shape)\n', (1074, 1090), True, 'import numpy as np\n'), ((1804, 1872), 'argparse.ArgumentParser', 'argparse.ArgumentPar... |
# Copyright (C) 2010 Google Inc. All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are
# met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the ... | [
"webkitpy.tool.steps.abstractstep.AbstractStep.options",
"urlparse.urlparse",
"re.match"
] | [((1864, 1886), 'webkitpy.tool.steps.abstractstep.AbstractStep.options', 'AbstractStep.options', ([], {}), '()\n', (1884, 1886), False, 'from webkitpy.tool.steps.abstractstep import AbstractStep\n'), ((2516, 2539), 'urlparse.urlparse', 'urlparse', (['user_response'], {}), '(user_response)\n', (2524, 2539), False, 'from... |
from PIL import Image
from pylab import *
img = Image.open('images/profile.jpg').convert('L')
print(array(img)[500])
imgArray = array(img)
figure()
hist(imgArray.flatten(), 300)
show()
# img.show()
| [
"PIL.Image.open"
] | [((52, 84), 'PIL.Image.open', 'Image.open', (['"""images/profile.jpg"""'], {}), "('images/profile.jpg')\n", (62, 84), False, 'from PIL import Image\n')] |
# -*- coding: utf-8 -*-
##
# @file data_context.py
# @brief
# @author wondereamer
# @version 0.1
# @date 2016-11-27
import datetime
from quantity.digger.engine.series import SeriesBase, NumberSeries, DateTimeSeries
from quantity.digger.technicals.base import TechnicalBase
from quantity.digger.util import elogger as ... | [
"quantity.digger.datastruct.Bar",
"datetime.datetime",
"quantity.digger.engine.series.DateTimeSeries",
"quantity.digger.engine.series.NumberSeries",
"quantity.digger.util.elogger.error"
] | [((659, 697), 'quantity.digger.engine.series.NumberSeries', 'NumberSeries', (['data.open.values', '"""open"""'], {}), "(data.open.values, 'open')\n", (671, 697), False, 'from quantity.digger.engine.series import SeriesBase, NumberSeries, DateTimeSeries\n'), ((719, 759), 'quantity.digger.engine.series.NumberSeries', 'Nu... |
#! /usr/bin/env python
'''
Solvers and example test cases for Day 5 of the Advent of Code 2021.
Problem description: <https://adventofcode.com/2021/day/5>
'''
from collections import Counter
from dataclasses import dataclass
from typing import Iterable, List, Tuple
import unittest
def part1(lines: Iterable[str]) ->... | [
"collections.Counter",
"dataclasses.dataclass"
] | [((974, 996), 'dataclasses.dataclass', 'dataclass', ([], {'frozen': '(True)'}), '(frozen=True)\n', (983, 996), False, 'from dataclasses import dataclass\n'), ((429, 438), 'collections.Counter', 'Counter', ([], {}), '()\n', (436, 438), False, 'from collections import Counter\n'), ((799, 808), 'collections.Counter', 'Cou... |
__copyright__ = """This code is licensed under the 3-clause BSD license.
Copyright ETH Zurich, Laboratory of Physical Chemistry, Reiher Group.
See LICENSE.txt for details.
"""
import pytest
import scine_utilities as scine
import numpy as np
import os
class SigmaVectorEvaluatorPython(scine.SigmaVectorEvaluator):
d... | [
"scine_utilities.NonOrthogonalDavidson",
"scine_utilities.OrthogonalDavidson",
"numpy.diag_indices_from",
"scine_utilities.core.Log.silent",
"numpy.transpose",
"numpy.ones",
"numpy.linalg.eig",
"scine_utilities.SigmaVectorEvaluator.__init__",
"numpy.identity",
"scine_utilities.IndirectSigmaVectorE... | [((687, 711), 'numpy.random.rand', 'np.random.rand', (['(100)', '(100)'], {}), '(100, 100)\n', (701, 711), True, 'import numpy as np\n'), ((927, 969), 'scine_utilities.IndirectSigmaVectorEvaluator', 'scine.IndirectSigmaVectorEvaluator', (['matrix'], {}), '(matrix)\n', (961, 969), True, 'import scine_utilities as scine\... |
import os
from keras import layers
from keras.layers import Input, merge
from keras.layers.convolutional import (AveragePooling2D, Conv2D, MaxPooling2D,
ZeroPadding2D)
from keras.layers.core import Activation, Dense, Flatten
from keras.layers.normalization import BatchNormalizat... | [
"keras.layers.core.Dense",
"keras.layers.convolutional.AveragePooling2D",
"keras.layers.add",
"keras.layers.core.Activation",
"keras.layers.convolutional.MaxPooling2D",
"keras.models.Model",
"keras.utils.plot_model",
"keras.layers.convolutional.Conv2D",
"keras.layers.core.Flatten",
"keras.layers.I... | [((987, 1014), 'keras.layers.add', 'layers.add', (['[out, shortcut]'], {}), '([out, shortcut])\n', (997, 1014), False, 'from keras import layers\n'), ((1772, 1799), 'keras.layers.add', 'layers.add', (['[out, shortcut]'], {}), '([out, shortcut])\n', (1782, 1799), False, 'from keras import layers\n'), ((1877, 1903), 'ker... |
"""
<NAME> 2014-2016
Python Progress Indicator Utility
Author: <NAME> <<EMAIL>>
License: BSD 3 clause
Contributors: https://github.com/rasbt/pyprind/graphs/contributors
Code Repository: https://github.com/rasbt/pyprind
PyPI: https://pypi.python.org/pypi/PyPrind
"""
import sys
import time
import pyprind
n = 100
slee... | [
"pyprind.ProgPercent",
"time.sleep"
] | [((372, 394), 'pyprind.ProgPercent', 'pyprind.ProgPercent', (['n'], {}), '(n)\n', (391, 394), False, 'import pyprind\n'), ((502, 543), 'pyprind.ProgPercent', 'pyprind.ProgPercent', (['n'], {'stream': 'sys.stdout'}), '(n, stream=sys.stdout)\n', (521, 543), False, 'import pyprind\n'), ((773, 809), 'pyprind.ProgPercent', ... |
# coding: utf-8
# # 使用预训练的VGG模型Fine-tune CNN
# In[1]:
# Import packs
import numpy as np
import os
import scipy.io
from scipy.misc import imread, imresize
import matplotlib.pyplot as plt
import skimage.io
import skimage.transform
import tensorflow as tf
get_ipython().magic(u'matplotlib inline')
cwd = os.getcwd()
pri... | [
"tensorflow.matmul",
"os.path.isfile",
"numpy.random.randint",
"numpy.mean",
"os.path.join",
"numpy.ndarray",
"tensorflow.nn.relu",
"tensorflow.nn.softmax_cross_entropy_with_logits",
"numpy.transpose",
"tensorflow.placeholder",
"tensorflow.cast",
"numpy.reshape",
"tensorflow.initialize_all_v... | [((305, 316), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (314, 316), False, 'import os\n'), ((1812, 1840), 'numpy.ndarray', 'np.ndarray', (['(imgcnt, nclass)'], {}), '((imgcnt, nclass))\n', (1822, 1840), True, 'import numpy as np\n'), ((2630, 2668), 'numpy.random.randint', 'np.random.randint', (['imgcnt'], {'size': 'i... |
from setuptools import setup
requires = [
'click==6.7',
'bucketstore==0.1.1'
]
setup(
name="s3env",
version="0.0.4",
author="<NAME>",
description="Manipulate a key/value JSON object file in an S3 bucket through the CLI",
author_email="<EMAIL>",
url='https://github.com/cameronmaske/s3e... | [
"setuptools.setup"
] | [((90, 456), 'setuptools.setup', 'setup', ([], {'name': '"""s3env"""', 'version': '"""0.0.4"""', 'author': '"""<NAME>"""', 'description': '"""Manipulate a key/value JSON object file in an S3 bucket through the CLI"""', 'author_email': '"""<EMAIL>"""', 'url': '"""https://github.com/cameronmaske/s3env"""', 'py_modules': ... |
"""Unit tests for the testplan.testing.multitest.result module."""
import collections
import mock
import pytest
from testplan.testing.multitest import result as result_mod
from testplan.testing.multitest.suite import testcase, testsuite
from testplan.testing.multitest import MultiTest
from testplan.common.utils impor... | [
"testplan.testing.multitest.result.DictNamespace",
"testplan.testing.multitest.result.FixNamespace",
"testplan.common.utils.testing.FixMessage",
"mock.MagicMock",
"collections.deque"
] | [((1615, 1631), 'mock.MagicMock', 'mock.MagicMock', ([], {}), '()\n', (1629, 1631), False, 'import mock\n'), ((1658, 1677), 'collections.deque', 'collections.deque', ([], {}), '()\n', (1675, 1677), False, 'import collections\n'), ((1689, 1726), 'testplan.testing.multitest.result.DictNamespace', 'result_mod.DictNamespac... |
#!/usr/bin/python3
# -*- coding: UTF-8 -*-
# This module gets that daya from SoaringSpot and prepaeres the infor for the REAL TIME SCORING
#
# Author: <NAME> - May 2021
#
#import sys
import json
import urllib.request
import urllib.error
import urllib.parse
import hmac
import hashlib
import base64
import os
import socke... | [
"ognddbfuncs.getognreg",
"json.load",
"ognddbfuncs.getognflarmid",
"os.getcwd",
"hmac.new",
"pycountry.countries.get",
"json.dumps",
"socket.gethostname",
"datetime.datetime.utcnow",
"base64.b64encode",
"simplehal.HalDocument.from_python",
"datetime.datetime.now",
"os.urandom"
] | [((1218, 1230), 'json.load', 'json.load', (['r'], {}), '(r)\n', (1227, 1230), False, 'import json\n'), ((3076, 3090), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (3088, 3090), False, 'from datetime import datetime\n'), ((3136, 3156), 'socket.gethostname', 'socket.gethostname', ([], {}), '()\n', (3154, 31... |
#!/usr/bin/env python3
import json
from pathlib import PurePath, Path
import subprocess
import tempfile
import zipfile
wasm_pack = Path("~/.cargo/bin/wasm-pack").expanduser()
root_files = ["module.json", "README.md", "CHANGELOG.md", "LICENSE"]
wasm_files = ["gridless_pathfinding_bg.wasm", "gridless_pathfinding.js"]
... | [
"subprocess.run",
"json.load",
"tempfile.TemporaryDirectory",
"zipfile.ZipFile",
"pathlib.Path",
"pathlib.PurePath"
] | [((333, 349), 'pathlib.Path', 'Path', (['"""artifact"""'], {}), "('artifact')\n", (337, 349), False, 'from pathlib import PurePath, Path\n'), ((419, 431), 'pathlib.Path', 'Path', (['"""wasm"""'], {}), "('wasm')\n", (423, 431), False, 'from pathlib import PurePath, Path\n'), ((443, 452), 'pathlib.Path', 'Path', (['"""."... |
import functools
import inspect
import equinox
from typing import Callable, Iterable
def compose(workflow: Iterable[Callable], do_jit: bool=False) -> Callable:
def pipeline(*args, **kwargs): # *args are for grad, **kwargs are the rest
res = dict([])
for f in workflow:
sig = inspect.... | [
"inspect.Signature",
"inspect.signature",
"equinox.filter_jit"
] | [((1842, 1873), 'inspect.signature', 'inspect.signature', (['workflow[-1]'], {}), '(workflow[-1])\n', (1859, 1873), False, 'import inspect\n'), ((1939, 1993), 'inspect.Signature', 'inspect.Signature', (['workflow_pars'], {'return_annotation': 'an'}), '(workflow_pars, return_annotation=an)\n', (1956, 1993), False, 'impo... |
import requests
base_url = "https://stepic.org/media/attachments/course67/3.6.3/"
with open("solutions/week-3/dataset_3378_3.txt") as f:
first_url = f.readline().strip()
r = requests.get(first_url)
answer = r.text.strip()
count = 1
while not answer.startswith("We"):
r = requests.get(f"{base_url}{answer}")
... | [
"requests.get"
] | [((181, 204), 'requests.get', 'requests.get', (['first_url'], {}), '(first_url)\n', (193, 204), False, 'import requests\n'), ((283, 318), 'requests.get', 'requests.get', (['f"""{base_url}{answer}"""'], {}), "(f'{base_url}{answer}')\n", (295, 318), False, 'import requests\n')] |
# Copyright (c) Facebook, Inc. and its affiliates.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
import subprocess
import sys
from contextlib import contextmanager
from signal import Signals
from subprocess import Popen as _Popen
from typi... | [
"subprocess.Popen"
] | [((1873, 1896), 'subprocess.Popen', '_Popen', (['*args'], {}), '(*args, **kwargs)\n', (1879, 1896), True, 'from subprocess import Popen as _Popen\n')] |
# Twitter Parser
# Parses Data and uploads to MongoDB
import twitter_access
import mongo_access
from tweepy.streaming import StreamListener
from tweepy import OAuthHandler
from tweepy import Stream
import json
import re
import tweepy
import time
from datetime import datetime
from dateutil.parser import parse
from py... | [
"pymongo.MongoClient",
"dateutil.parser.parse",
"json.loads",
"time.time",
"pandas.read_excel",
"vaderSentiment.vaderSentiment.SentimentIntensityAnalyzer",
"tweepy.Stream",
"tweepy.OAuthHandler",
"re.sub"
] | [((497, 526), 'pymongo.MongoClient', 'MongoClient', (['mongo_access.URL'], {}), '(mongo_access.URL)\n', (508, 526), False, 'from pymongo import MongoClient\n'), ((2883, 2950), 'tweepy.OAuthHandler', 'OAuthHandler', (['twitter_access.API_KEY', 'twitter_access.API_SECRET_KEY'], {}), '(twitter_access.API_KEY, twitter_acce... |
# Imports the Google Cloud client library
from google.cloud import language
from google.cloud.language import enums
from google.cloud.language import types
# Instantiates a client
client = language.LanguageServiceClient()
ENTITY_TYPE_TO_IGNORE = [8, 9, 10, 11, 12]
# The text to analyze
text = "Aside from course work,... | [
"google.cloud.language.LanguageServiceClient",
"google.cloud.language.types.Document"
] | [((190, 222), 'google.cloud.language.LanguageServiceClient', 'language.LanguageServiceClient', ([], {}), '()\n', (220, 222), False, 'from google.cloud import language\n'), ((1282, 1347), 'google.cloud.language.types.Document', 'types.Document', ([], {'content': 'text', 'type': 'enums.Document.Type.PLAIN_TEXT'}), '(cont... |
from selenium import webdriver
from selenium.webdriver.common.by import By
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.support import expected_conditions
from selenium.common.exceptions import NoSuchElementException
import time, math
class PageGetSet:
def __init__(self):
... | [
"selenium.webdriver.support.ui.WebDriverWait",
"time.sleep",
"selenium.webdriver.support.expected_conditions.visibility_of_element_located",
"selenium.webdriver.Firefox"
] | [((1176, 1195), 'selenium.webdriver.Firefox', 'webdriver.Firefox', ([], {}), '()\n', (1193, 1195), False, 'from selenium import webdriver\n'), ((2213, 2231), 'time.sleep', 'time.sleep', (['length'], {}), '(length)\n', (2223, 2231), False, 'import time, math\n'), ((1765, 1827), 'selenium.webdriver.support.expected_condi... |
from unicodedata import name
from accounts import views as account_views
from django.urls import path, reverse
from .views import (AboutPageView, AuctionCreateView, AuctionDetailView, AuctionListView, BidCreateView, BidDetailView,
DashboardPageView, DatasourceView, HomePageView, StripeConnectionVi... | [
"accounts.views.MyCompanyDetailView.as_view",
"django.urls.path",
"accounts.views.AccountDetailView.as_view",
"accounts.views.CompanyUpdateView.as_view",
"accounts.views.CompanyCreateView.as_view"
] | [((1608, 1673), 'django.urls.path', 'path', (['"""dashboard/connect/"""', 'account_views.connect'], {'name': '"""connect"""'}), "('dashboard/connect/', account_views.connect, name='connect')\n", (1612, 1673), False, 'from django.urls import path, reverse\n'), ((1327, 1370), 'accounts.views.MyCompanyDetailView.as_view',... |
"""A population model that creates samples with more and more variants. Suitable for the aligner paper experiments
^ = intersection
E = subset
vx ^ v0 = v0
vx ^ v1 = v0
...
vx ^ vn = v0
v0 E v1
v1 E v2
v2 E v3
...
v(n-1) E vn
This plugin does not honor the site frequency spectrum model and ignores the original 'p' v... | [
"numpy.random.RandomState"
] | [((2112, 2143), 'numpy.random.RandomState', 'np.random.RandomState', (['rng_seed'], {}), '(rng_seed)\n', (2133, 2143), True, 'import numpy as np\n')] |
#!/usr/bin/env python
"""Test that GoSubDag contains ancestors from only the user-specified relationships"""
# tests/test_gosubdag_relationships_i126.py
# goatools/gosubdag/gosubdag.py
# goatools/gosubdag/godag_rcnt.py
# goatools/gosubdag/godag_rcnt_init.py
# goatools/godag/go_tasks.py
# goatools/obo_parser.py
from __... | [
"goatools.test_data.wr_subobo.WrSubObo.prt_goterms",
"goatools.gosubdag.gosubdag.GoSubDag",
"os.path.join",
"goatools.base.get_godag"
] | [((1116, 1168), 'os.path.join', 'join', (['REPO', '"""tests/data/i126/viral_gene_silence.obo"""'], {}), "(REPO, 'tests/data/i126/viral_gene_silence.obo')\n", (1120, 1168), False, 'from os.path import join\n'), ((1203, 1239), 'goatools.base.get_godag', 'get_godag', (['fin_obo'], {'loading_bar': 'None'}), '(fin_obo, load... |
from .wizards import Wizard
from .data import Dataset, DatasetInfo, AutomaticAnnotator
from .simulation import Renderer, DatasetRenderer, RobotLookupCreator
from .prediction import Predictor, LiveCamera
from .paths import Paths
from .prediction.analysis import Grapher
from .prediction.synthetic import SyntheticPredicto... | [
"logging.basicConfig"
] | [((420, 459), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (439, 459), False, 'import logging\n')] |
"""
Functions to rotate a point by a known euler pole.
"""
import numpy as np
from . import fault_vector_functions
def point_rotation_by_Euler_Pole(Point, Euler_Pole):
"""
Compute the velocity of rotation of a point about an Euler pole on a spherical earth.
This function is useful for computing the veloc... | [
"numpy.deg2rad",
"numpy.cross",
"numpy.sin",
"numpy.cos",
"numpy.dot",
"numpy.sqrt"
] | [((1109, 1133), 'numpy.cross', 'np.cross', (['omega', 'R_point'], {}), '(omega, R_point)\n', (1117, 1133), True, 'import numpy as np\n'), ((2334, 2349), 'numpy.deg2rad', 'np.deg2rad', (['lon'], {}), '(lon)\n', (2344, 2349), True, 'import numpy as np\n'), ((2469, 2511), 'numpy.sqrt', 'np.sqrt', (['(R_fixed * R_fixed - X... |
# Variables
base_list = 'List_1' # this is the base list, each item in this list is checked for a match in the other
list_2 = 'List_2' # List_2 is the name of the list in the excel file
xlfile = 'DATA_IN.xlsx'
# Importing Libs
import pandas as pd
import numpy as np
# Smart Stuff
df_0 = pd.read_excel(xlfile, dtype=str... | [
"pandas.read_excel",
"pandas.DataFrame"
] | [((289, 321), 'pandas.read_excel', 'pd.read_excel', (['xlfile'], {'dtype': 'str'}), '(xlfile, dtype=str)\n', (302, 321), True, 'import pandas as pd\n'), ((426, 440), 'pandas.DataFrame', 'pd.DataFrame', ([], {}), '()\n', (438, 440), True, 'import pandas as pd\n')] |
""" Common functions and parameters amongst the experiments. """
from edo.distributions import Uniform
from sklearn.preprocessing import MinMaxScaler
def scale_dataframe(individual):
""" Scale the individual's dataframe to the unit square for calculating
fitness. """
original = individual.dataframe.copy... | [
"sklearn.preprocessing.MinMaxScaler"
] | [((339, 353), 'sklearn.preprocessing.MinMaxScaler', 'MinMaxScaler', ([], {}), '()\n', (351, 353), False, 'from sklearn.preprocessing import MinMaxScaler\n')] |
import numpy as np
import random
import os, sys
from scipy import ndimage
import healpy as hp
from astropy.io import fits
import matplotlib.pyplot as plt
from scipy.signal import savgol_filter
from astropy.io import fits
from importlib import reload
from pycs.misc.cosmostat_init import *
from pycs.misc.mr_prog import... | [
"healpy.write_map",
"healpy.alm2map",
"healpy.mollview",
"numpy.copy",
"healpy.map2alm",
"numpy.deg2rad",
"healpy.ud_grade",
"healpy.nside2npix",
"numpy.where",
"healpy.npix2nside",
"healpy.map2alm_spin",
"healpy.alm2map_spin",
"healpy.smoothing",
"healpy.read_map",
"numpy.sqrt"
] | [((634, 699), 'numpy.where', 'np.where', (['(bincount > 0.5)', '(bincount_weighted / bincount)', 'hp.UNSEEN'], {}), '(bincount > 0.5, bincount_weighted / bincount, hp.UNSEEN)\n', (642, 699), True, 'import numpy as np\n'), ((938, 953), 'healpy.read_map', 'hp.read_map', (['FN'], {}), '(FN)\n', (949, 953), True, 'import h... |
"""Contract test cases for cases for all organization catalogs."""
import json
import pytest
import requests
from tests import responses
@pytest.mark.contract
@pytest.mark.docker
def test_all_catalogs(docker_service: str) -> None:
"""Should return the all_catalogs response."""
url = f"{docker_service}/organ... | [
"json.loads",
"requests.get"
] | [((352, 369), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (364, 369), False, 'import requests\n'), ((701, 718), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (713, 718), False, 'import requests\n'), ((1119, 1136), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (1131, 1136), False, '... |
"""
Builds a NEXUS file form a long-table format CSV.
"""
import csv
from pathlib import Path
from collections import defaultdict
BASE = Path(__file__).parents[1] / "data"
with open(BASE / "ielex.csv", encoding="utf-8") as h:
data = list(csv.DictReader(h))
taxa = sorted(set([row["LANGUAGE"] for row in data]))
... | [
"collections.defaultdict",
"csv.DictReader",
"pathlib.Path"
] | [((328, 345), 'collections.defaultdict', 'defaultdict', (['list'], {}), '(list)\n', (339, 345), False, 'from collections import defaultdict\n'), ((357, 373), 'collections.defaultdict', 'defaultdict', (['set'], {}), '(set)\n', (368, 373), False, 'from collections import defaultdict\n'), ((246, 263), 'csv.DictReader', 'c... |
from graia.application.message.chain import MessageChain
from graia.application.message.elements.internal import Plain
from graia.application.message.elements.internal import At
from SAGIRIBOT.basics.aio_mysql_excute import execute_sql
async def get_rank(group_id: int, memberList: list) -> list:
sql = "select * ... | [
"SAGIRIBOT.basics.aio_mysql_excute.execute_sql",
"graia.application.message.elements.internal.Plain"
] | [((402, 418), 'SAGIRIBOT.basics.aio_mysql_excute.execute_sql', 'execute_sql', (['sql'], {}), '(sql)\n', (413, 418), False, 'from SAGIRIBOT.basics.aio_mysql_excute import execute_sql\n'), ((953, 969), 'graia.application.message.elements.internal.Plain', 'Plain', ([], {'text': 'text'}), '(text=text)\n', (958, 969), False... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup
import re
import os
import sys
name = "django-db-queue-exports"
package = "django_dbq_exports"
description = "An extension to django-db-queue for monitoring long running jobs"
url = "https://www.dabapps.com/"
project_urls = {"Source": "https:... | [
"os.walk",
"re.search",
"os.path.join"
] | [((736, 806), 're.search', 're.search', (['"""^__version__ = [\'"]([^\'"]+)[\'"]"""', 'init_py', 're.MULTILINE'], {}), '(\'^__version__ = [\\\'"]([^\\\'"]+)[\\\'"]\', init_py, re.MULTILINE)\n', (745, 806), False, 'import re\n'), ((994, 1010), 'os.walk', 'os.walk', (['package'], {}), '(package)\n', (1001, 1010), False, ... |
import mobula.layers as L
import numpy as np
def test_sigmoid():
X = ((np.arange(10000) - 5000) / 1000.0).reshape((-1, 1, 1, 1))
data = L.Data(X, "data")
data.reshape()
l = L.Sigmoid(data)
l.reshape()
assert l.Y.shape == X.shape
l.forward()
l.dY = np.random.random(l.Y.shape) * 10
l.... | [
"mobula.layers.PReLU",
"mobula.layers.Tanh",
"numpy.allclose",
"mobula.layers.ReLU",
"numpy.zeros",
"numpy.square",
"numpy.random.random",
"numpy.arange",
"numpy.exp",
"mobula.layers.Data",
"mobula.layers.SELU",
"mobula.layers.Sigmoid"
] | [((145, 162), 'mobula.layers.Data', 'L.Data', (['X', '"""data"""'], {}), "(X, 'data')\n", (151, 162), True, 'import mobula.layers as L\n'), ((190, 205), 'mobula.layers.Sigmoid', 'L.Sigmoid', (['data'], {}), '(data)\n', (199, 205), True, 'import mobula.layers as L\n'), ((342, 352), 'numpy.exp', 'np.exp', (['(-X)'], {}),... |
# Copyright 2019 <NAME>
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to i... | [
"elasticsearch.Elasticsearch",
"rhoci.database.Database.find_one",
"rhoci.database.Database.find"
] | [((1728, 1783), 'elasticsearch.Elasticsearch', 'Elasticsearch', (["app.config['custom']['elk']['es']['url']"], {}), "(app.config['custom']['elk']['es']['url'])\n", (1741, 1783), False, 'from elasticsearch import Elasticsearch\n'), ((3374, 3419), 'rhoci.database.Database.find', 'Database.find', ([], {'collection': '"""D... |
from prometheus_client.core import GaugeMetricFamily
def make_metrics(queue):
list_metrics = []
# Total items in Queue
metric = GaugeMetricFamily(
'jenkins_queue_total',
'Total items in Queue',
labels=None
)
metric.add_metric(
labels=[],
value=queue.get_to... | [
"prometheus_client.core.GaugeMetricFamily"
] | [((144, 221), 'prometheus_client.core.GaugeMetricFamily', 'GaugeMetricFamily', (['"""jenkins_queue_total"""', '"""Total items in Queue"""'], {'labels': 'None'}), "('jenkins_queue_total', 'Total items in Queue', labels=None)\n", (161, 221), False, 'from prometheus_client.core import GaugeMetricFamily\n'), ((420, 548), '... |
from datetime import datetime
from typing import List
from sqlalchemy import insert, func, select, distinct
from src import db
from src.database.models import Artifact, LockedArtifact, ArtifactLabelRelation, FlaggedArtifact, LabelingData
from src.helper.tools_common import string_none_or_empty, who_is_signed_in
def... | [
"sqlalchemy.insert",
"src.database.models.LockedArtifact.query.all",
"src.database.models.LockedArtifact",
"src.database.models.LockedArtifact.query.filter_by",
"src.db.session.execute",
"sqlalchemy.distinct",
"sqlalchemy.select",
"datetime.datetime.utcnow",
"src.helper.tools_common.who_is_signed_in... | [((999, 1018), 'src.db.session.commit', 'db.session.commit', ([], {}), '()\n', (1016, 1018), False, 'from src import db\n'), ((1769, 1788), 'src.db.session.commit', 'db.session.commit', ([], {}), '()\n', (1786, 1788), False, 'from src import db\n'), ((2105, 2131), 'src.database.models.LockedArtifact.query.all', 'Locked... |
from demo.exceptions import AuthException
def token_auth(fn):
def wrapper(*args, **kwargs):
from flask import request
if request.headers.get("token", None) != "123":
raise AuthException("Fail to get access")
return fn(*args, **kwargs)
return wrapper
| [
"demo.exceptions.AuthException",
"flask.request.headers.get"
] | [((144, 178), 'flask.request.headers.get', 'request.headers.get', (['"""token"""', 'None'], {}), "('token', None)\n", (163, 178), False, 'from flask import request\n'), ((207, 242), 'demo.exceptions.AuthException', 'AuthException', (['"""Fail to get access"""'], {}), "('Fail to get access')\n", (220, 242), False, 'from... |
import matplotlib as mpl
import matplotlib.pyplot as plt
import numpy as np
import scipy as sci
import pandas as pd
import numbers
import csv
import sys
filepath = sys.argv[1]
filename = filepath[filepath.rfind('/')+1:]
print(filepath)
data = pd.read_csv(filepath)
print("Adding year column")
data['year'] = data.apply... | [
"pandas.read_csv"
] | [((244, 265), 'pandas.read_csv', 'pd.read_csv', (['filepath'], {}), '(filepath)\n', (255, 265), True, 'import pandas as pd\n'), ((1192, 1237), 'pandas.read_csv', 'pd.read_csv', (['"""categorized_ucc_dictionary.csv"""'], {}), "('categorized_ucc_dictionary.csv')\n", (1203, 1237), True, 'import pandas as pd\n')] |
import arcpy
from arcpy import env
from arcpy.sa import *
import file_functions
from file_functions import *
import create_centerline
import create_station_lines
from create_station_lines import create_station_lines_function
import os
from os import listdir
from os.path import isfile, join
import xlrd
import shutil
fro... | [
"os.remove",
"create_centerline.make_centerline",
"arcpy.MultipartToSinglepart_management",
"arcpy.SaveToLayerFile_management",
"shutil.rmtree",
"arcpy.MakeRasterLayer_management",
"arcpy.JoinField_management",
"os.path.join",
"arcpy.Intersect_analysis",
"arcpy.ExtractValuesToTable_ga",
"arcpy.C... | [((10192, 10212), 'os.path.dirname', 'os.path.dirname', (['dem'], {}), '(dem)\n', (10207, 10212), False, 'import os\n'), ((1074, 1097), 'os.path.exists', 'os.path.exists', (['laspath'], {}), '(laspath)\n', (1088, 1097), False, 'import os\n'), ((1107, 1127), 'os.makedirs', 'os.makedirs', (['laspath'], {}), '(laspath)\n'... |
from model_new.config import Config
from model_new.keras_model import KerasModel
from model_new.utils import load_json
config = Config()
model = KerasModel(config)
train_set = load_json('dataset/train.json')
# dev_set = load_json('dataset/dev.json')
# sub_set = dev_set[:config.batch_size * 50]
model.train(train_set,... | [
"model_new.utils.load_json",
"model_new.keras_model.KerasModel",
"model_new.config.Config"
] | [((129, 137), 'model_new.config.Config', 'Config', ([], {}), '()\n', (135, 137), False, 'from model_new.config import Config\n'), ((146, 164), 'model_new.keras_model.KerasModel', 'KerasModel', (['config'], {}), '(config)\n', (156, 164), False, 'from model_new.keras_model import KerasModel\n'), ((178, 209), 'model_new.u... |
from django.conf.urls import patterns, url
from Standings import views
urlpatterns = patterns(
'',
# ex: /standings/
url(r'^$', views.standings_index, name='index'),
) | [
"django.conf.urls.url"
] | [((131, 177), 'django.conf.urls.url', 'url', (['"""^$"""', 'views.standings_index'], {'name': '"""index"""'}), "('^$', views.standings_index, name='index')\n", (134, 177), False, 'from django.conf.urls import patterns, url\n')] |
#!/usr/bin/python3
# This script runs whenever a user tries to commit something in this repo.
# It checks the commit for any text that resembled an encoded JSON web token,
# and asks the user to verify that they want to commit a JWT if it finds any.
import sys
import subprocess
import re
import base64
import binascii
... | [
"subprocess.check_output",
"re.findall",
"base64.urlsafe_b64decode",
"re.compile"
] | [((790, 830), 're.compile', 're.compile', (['"""JWT|iat|name|sub|alg|exp|k"""'], {}), "('JWT|iat|name|sub|alg|exp|k')\n", (800, 830), False, 'import re\n'), ((987, 1026), 're.findall', 're.findall', (['"""[A-Za-z0-9_=-]{20,}"""', 'line'], {}), "('[A-Za-z0-9_=-]{20,}', line)\n", (997, 1026), False, 'import re\n'), ((194... |
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not... | [
"custos.clients.super_tenant_management_client.SuperTenantManagementClient",
"custos.clients.tenant_management_client.TenantManagementClient",
"logging.StreamHandler",
"custos.clients.utils.utilities.get_token",
"custos.clients.identity_management_client.IdentityManagementClient",
"custos.transport.settin... | [((1170, 1197), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1187, 1197), False, 'import logging\n'), ((1289, 1312), 'logging.StreamHandler', 'logging.StreamHandler', ([], {}), '()\n', (1310, 1312), False, 'import logging\n'), ((1364, 1392), 'custos.transport.settings.CustosServerClien... |
# uncompyle6 version 2.13.2
# Python bytecode 3.5 (3351)
# Decompiled from: Python 3.5.3 (default, Jan 19 2017, 14:11:04)
# [GCC 6.3.0 20170118]
# Embedded file name: db\tables\licenses.py
from sqlalchemy import *
from sqlalchemy.orm import mapper
from db.tables import metadata
LicensesTable = Table('licenses', metada... | [
"sqlalchemy.orm.mapper"
] | [((632, 665), 'sqlalchemy.orm.mapper', 'mapper', (['LicenseRow', 'LicensesTable'], {}), '(LicenseRow, LicensesTable)\n', (638, 665), False, 'from sqlalchemy.orm import mapper\n')] |
#!/usr/bin/env python
# -*- coding: UTF-8 -*-
"""
Author: enen92
License: I don't care version 3.0
"""
import xbmc,xbmcgui,xbmcaddon,os
addon_id = 'script.retrogames'
selfAddon = xbmcaddon.Addon(id=addon_id)
datapath = xbmc.translatePath(selfAddon.getAddonInfo('profile')).decode('utf-8')
addonfolder = xbmc.trans... | [
"xbmcgui.Dialog",
"os.path.join",
"xbmcaddon.Addon"
] | [((186, 214), 'xbmcaddon.Addon', 'xbmcaddon.Addon', ([], {'id': 'addon_id'}), '(id=addon_id)\n', (201, 214), False, 'import xbmc, xbmcgui, xbmcaddon, os\n'), ((389, 434), 'os.path.join', 'os.path.join', (['addonfolder', '"""resources"""', '"""img"""'], {}), "(addonfolder, 'resources', 'img')\n", (401, 434), False, 'imp... |
"""Test Palmetto."""
import pytest
from palmettopy.palmetto import Palmetto
from palmettopy.exceptions import CoherenceTypeNotAvailable, EndpointDown, WrongContentType
@pytest.fixture
def words():
"""Load test data fixture."""
words = ["cake", "apple", "banana", "cherry", "chocolate"]
return words
@py... | [
"pytest.raises",
"palmettopy.palmetto.Palmetto"
] | [((759, 769), 'palmettopy.palmetto.Palmetto', 'Palmetto', ([], {}), '()\n', (767, 769), False, 'from palmettopy.palmetto import Palmetto\n'), ((921, 931), 'palmettopy.palmetto.Palmetto', 'Palmetto', ([], {}), '()\n', (929, 931), False, 'from palmettopy.palmetto import Palmetto\n'), ((1076, 1119), 'palmettopy.palmetto.P... |
import model3 as M
import numpy as np
import tensorflow as tf
params = np.load('lstmpm_d1.npy').item()
params2 = np.load('lstmpm_d2.npy').item()
def get_conv(name):
res = []
# print(params[name])
res.append(params[name]['weights'])
res.append(params[name]['bias'])
# print(res[0].shape)
return res
def get_c... | [
"model3.Saver",
"numpy.load",
"model3.AvgPool",
"model3.MaxPool",
"tensorflow.pad",
"numpy.transpose",
"tensorflow.concat",
"numpy.ones",
"tensorflow.Variable",
"tensorflow.tanh",
"tensorflow.sigmoid"
] | [((7984, 8013), 'numpy.transpose', 'np.transpose', (['x', '[0, 3, 1, 2]'], {}), '(x, [0, 3, 1, 2])\n', (7996, 8013), True, 'import numpy as np\n'), ((8365, 8394), 'numpy.transpose', 'np.transpose', (['x', '[0, 3, 1, 2]'], {}), '(x, [0, 3, 1, 2])\n', (8377, 8394), True, 'import numpy as np\n'), ((8500, 8513), 'model3.Sa... |
from django.urls import path, re_path
from . import views
# 这是应用Book中的url具体配置,请求到这里才能调用该应用的视图,上一级首先是调用BookManager的url,如果没有匹配admin
# 那么进入应用的匹配Book.urls,之后才会进入这里应用的url
urlpatterns = [
# http://127.0.0.1:8000/admin/ 匹配
# 正则匹配,对请求地址进行正则匹配,如果路径中包含admin,就把后台站点中的url信息包含到这个项目中,指明下一集路径如何匹配
# 如果匹配成功,那么直接调用... | [
"django.urls.re_path",
"django.urls.path"
] | [((403, 444), 'django.urls.path', 'path', (['"""stockplot/"""', 'views.showlinediagram'], {}), "('stockplot/', views.showlinediagram)\n", (407, 444), False, 'from django.urls import path, re_path\n'), ((451, 494), 'django.urls.path', 'path', (['"""index3"""', 'views.index3'], {'name': '"""index3"""'}), "('index3', view... |
from __future__ import unicode_literals
from __future__ import print_function
from __future__ import division
from __future__ import absolute_import
from future import standard_library
standard_library.install_aliases()
from builtins import *
class OJAIList(list):
def __init__(self):
super(OJAIList, self)._... | [
"mapr.ojai.ojai.OJAIDocument.OJAIDocument",
"future.standard_library.install_aliases",
"mapr.ojai.ojai.OJAITagsBuilder.OJAITagsBuilder"
] | [((185, 219), 'future.standard_library.install_aliases', 'standard_library.install_aliases', ([], {}), '()\n', (217, 219), False, 'from future import standard_library\n'), ((521, 538), 'mapr.ojai.ojai.OJAITagsBuilder.OJAITagsBuilder', 'OJAITagsBuilder', ([], {}), '()\n', (536, 538), False, 'from mapr.ojai.ojai.OJAITags... |
import os
from copy import copy
from typing import Any, List, Tuple
from nubia import context, eventbus
from nubia.internal import cmdloader
from nubia.internal.cmdbase import AutoCommand
from pygments.token import Name, Token
from rich import box, inspect
from rich.align import Align
from rich.console import Console
... | [
"copy.copy",
"nubia.internal.cmdloader.load_commands",
"sortedcontainers.SortedSet",
"nubia.internal.cmdbase.AutoCommand",
"rich.console.Console",
"rich.align.Align.center"
] | [((718, 729), 'sortedcontainers.SortedSet', 'SortedSet', ([], {}), '()\n', (727, 729), False, 'from sortedcontainers import SortedSet\n'), ((1816, 1857), 'nubia.internal.cmdloader.load_commands', 'cmdloader.load_commands', (['COMMAND_PACKAGES'], {}), '(COMMAND_PACKAGES)\n', (1839, 1857), False, 'from nubia.internal imp... |
# -*- coding: utf-8 -*-
from pims import ImageReader
from load_data.ILoadSupervised import ILoadSupervised
from os.path import join, exists
import csv
class LoadPokemon(ILoadSupervised):
def __init__(self, path="train_data/Folder_Videojuegos/pokemon-images-and-types"):
self.path = path
self.classe... | [
"pims.ImageReader",
"csv.DictReader",
"os.path.join",
"os.path.exists"
] | [((526, 549), 'csv.DictReader', 'csv.DictReader', (['csv_obj'], {}), '(csv_obj)\n', (540, 549), False, 'import csv\n'), ((452, 482), 'os.path.join', 'join', (['self.path', '"""pokemon.csv"""'], {}), "(self.path, 'pokemon.csv')\n", (456, 482), False, 'from os.path import join, exists\n'), ((613, 660), 'os.path.join', 'j... |
import discord
from discord.ext import commands
import random
import asyncio
class GameManager():
def __init__(self):
self.setup()
def setup(self):
print("GameManager: Loaded")
class GameManagerCog(commands.Cog):
def __init__(self, client):
self.client = clie... | [
"discord.utils.get",
"discord.ext.commands.has_role",
"discord.ext.commands.command",
"discord.PermissionOverwrite"
] | [((895, 1316), 'discord.ext.commands.command', 'commands.command', ([], {'aliases': "['Creategame', 'CreateGame', 'cg', 'Cg', 'cG', 'CG', 'gamecreate',\n 'Gamecreate', 'GameCreate', 'gc', 'Gc', 'gC', 'GC']", 'brief': '"""Makes the necessary channels and roles for a game."""', 'description': '"""/creategame [arg1] [a... |
# Copyright 2016 Isotoma Limited
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing... | [
"touchdown.core.utils.force_bytes",
"datetime.datetime.now"
] | [((1300, 1323), 'datetime.datetime.now', 'datetime.datetime.now', ([], {}), '()\n', (1321, 1323), False, 'import datetime\n'), ((1441, 1463), 'touchdown.core.utils.force_bytes', 'force_bytes', (['user_data'], {}), '(user_data)\n', (1452, 1463), False, 'from touchdown.core.utils import force_bytes, force_str\n')] |
import sys
import click
from .function import LambdaPoolFunction
from . import utils
from tabulate import tabulate
from lambdapool import exceptions
@click.group()
def cli():
pass
@cli.command()
@click.option('--requirements', '-r', type=click.Path(exists=True), help="Specifies the dependencies to be installed ... | [
"click.argument",
"click.option",
"click.echo",
"tabulate.tabulate",
"click.Path",
"click.group",
"sys.exit"
] | [((153, 166), 'click.group', 'click.group', ([], {}), '()\n', (164, 166), False, 'import click\n'), ((347, 449), 'click.option', 'click.option', (['"""--memory"""'], {'type': 'click.INT', 'help': '"""Sets the memory size of the function environment"""'}), "('--memory', type=click.INT, help=\n 'Sets the memory size o... |
import time
import random
def print_pause(message_to_print):
print(message_to_print)
time.sleep(2)
def intro(item, option):
print_pause("You find yourself standing in an open field, filled "
"with grass and yellow wildflowers.\n")
print_pause("Rumor has it that a " + option + " is some... | [
"random.choice",
"time.sleep"
] | [((94, 107), 'time.sleep', 'time.sleep', (['(2)'], {}), '(2)\n', (104, 107), False, 'import time\n'), ((3986, 4049), 'random.choice', 'random.choice', (["['pirate', 'fairy', 'dragon', 'gorgon', 'troll']"], {}), "(['pirate', 'fairy', 'dragon', 'gorgon', 'troll'])\n", (3999, 4049), False, 'import random\n')] |
from builtins import str
from django.test import TestCase
from measure_mate.tests.factories import TemplateFactory, AttributeFactory, RatingFactory
class RatingTestCases(TestCase):
def test_creation_of_rating(self):
template = TemplateFactory()
attribute = AttributeFactory(template=template)
... | [
"builtins.str",
"measure_mate.tests.factories.TemplateFactory",
"measure_mate.tests.factories.AttributeFactory",
"measure_mate.tests.factories.RatingFactory"
] | [((242, 259), 'measure_mate.tests.factories.TemplateFactory', 'TemplateFactory', ([], {}), '()\n', (257, 259), False, 'from measure_mate.tests.factories import TemplateFactory, AttributeFactory, RatingFactory\n'), ((280, 315), 'measure_mate.tests.factories.AttributeFactory', 'AttributeFactory', ([], {'template': 'templ... |
"""
File: mirror_lake.py
----------------------------------
This file reads in mt-rainier.jpg and
makes a new image that creates a mirror
lake vibe by placing an inverse image of
mt-rainier.jpg below the original one.
"""
from simpleimage import SimpleImage
def reflect(filename):
"""
:param filename: str, the... | [
"simpleimage.SimpleImage",
"simpleimage.SimpleImage.blank"
] | [((409, 430), 'simpleimage.SimpleImage', 'SimpleImage', (['filename'], {}), '(filename)\n', (420, 430), False, 'from simpleimage import SimpleImage\n'), ((447, 491), 'simpleimage.SimpleImage.blank', 'SimpleImage.blank', (['img.width', '(img.height * 2)'], {}), '(img.width, img.height * 2)\n', (464, 491), False, 'from s... |
"""
Tests the subsets_exp module.
"""
import mock
import unittest
from .context import subsets_exp
from .context import config
from .context import runner
from .context import test_utils as tu
class Subsets_ExperimentTestCase(unittest.TestCase):
def setUp(self):
config_obj = tu.sample_config()
moc... | [
"unittest.main",
"mock.call",
"unittest.TestLoader",
"mock.Mock"
] | [((3303, 3318), 'unittest.main', 'unittest.main', ([], {}), '()\n', (3316, 3318), False, 'import unittest\n'), ((335, 359), 'mock.Mock', 'mock.Mock', (['runner.Runner'], {}), '(runner.Runner)\n', (344, 359), False, 'import mock\n'), ((1426, 1437), 'mock.Mock', 'mock.Mock', ([], {}), '()\n', (1435, 1437), False, 'import... |
from timberscale import Timber
import csv
import math
##### REPORT DATA MODULE
class Report(object):
LOG_RANGE_LIST = [["40+ ft", range(41, 121)], ["31-40 ft", range(31, 41)], ["21-30 ft", range(21, 31)],
["11-20 ft", range(11, 21)], ["1-10 ft", range(1, 11)]]
def __init__(self, CSV, St... | [
"csv.reader",
"timberscale.Timber",
"math.sqrt"
] | [((937, 958), 'csv.reader', 'csv.reader', (['tree_data'], {}), '(tree_data)\n', (947, 958), False, 'import csv\n'), ((3322, 3343), 'csv.reader', 'csv.reader', (['tree_data'], {}), '(tree_data)\n', (3332, 3343), False, 'import csv\n'), ((5691, 5744), 'math.sqrt', 'math.sqrt', (['(master[key][1] / master[key][0] / 0.0054... |
# -*- coding: UTF-8 -*-
import ply.yacc
from collections import OrderedDict
from .lexer import tokens, lex
# a:4:{s:4:"date";s:10:"2019-12-29";s:10:"type_fonds";s:11:"arko_seriel";s:4:"ref1";i:12;s:4:"ref2";i:4669;}
from .models import Object
start = 'expression'
def p_expression(p):
"""expression : atom
... | [
"collections.OrderedDict"
] | [((1191, 1204), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (1202, 1204), False, 'from collections import OrderedDict\n')] |
import base64
import json
import typing
import marshmallow
from boto3.dynamodb.conditions import Key
from drf_yasg2.utils import swagger_auto_schema
from flag_engine.api.schemas import APITraitSchema
from flag_engine.identities.builders import (
build_identity_dict,
build_identity_model,
)
from rest_framework ... | [
"projects.exceptions.DynamoNotEnabledError",
"base64.b64decode",
"environments.identities.models.Identity.dynamo_wrapper.get_all_items",
"rest_framework.response.Response",
"environments.models.Environment.objects.get",
"rest_framework.exceptions.NotFound",
"edge_api.identities.serializers.EdgeIdentityF... | [((1215, 1231), 'flag_engine.api.schemas.APITraitSchema', 'APITraitSchema', ([], {}), '()\n', (1229, 1231), False, 'from flag_engine.api.schemas import APITraitSchema\n'), ((4174, 4234), 'rest_framework.decorators.action', 'action', ([], {'detail': '(True)', 'methods': "['get']", 'url_path': '"""list-traits"""'}), "(de... |
import torch
from collections import defaultdict
from uninas.data.abstract import AbstractDataSet
from uninas.models.networks.abstract import AbstractNetwork
from uninas.training.result import ResultValue
from uninas.utils.args import ArgsInterface, Namespace, Argument
class AbstractMetric(ArgsInterface):
"""
... | [
"collections.defaultdict",
"torch.argmax",
"uninas.utils.args.Argument",
"torch.no_grad"
] | [((7330, 7347), 'collections.defaultdict', 'defaultdict', (['dict'], {}), '(dict)\n', (7341, 7347), False, 'from collections import defaultdict\n'), ((2332, 2361), 'torch.argmax', 'torch.argmax', (['targets'], {'dim': '(-1)'}), '(targets, dim=-1)\n', (2344, 2361), False, 'import torch\n'), ((6432, 6447), 'torch.no_grad... |
import subprocess
path_oommf = 'C:/Users/jmank/Desktop/oommf12b4_20200930_86_x64/oommf/oommf.tcl'
mif_file = 'C:/Users/jmank/Desktop/oommf12b4_20200930_86_x64/oommf/Skyrmion/skyrmionDome2.mif'
length = 2
param_string = ' boxsi '#-parameters "integer_length % s" ' % length
threads_string = ' -threads 28 '
oommf_comma... | [
"subprocess.call"
] | [((391, 433), 'subprocess.call', 'subprocess.call', (['oommf_command'], {'shell': '(True)'}), '(oommf_command, shell=True)\n', (406, 433), False, 'import subprocess\n')] |
from __future__ import annotations
from abc import abstractclassmethod, abstractmethod, abstractproperty
from typing import (
TYPE_CHECKING,
Any,
Generator,
Generic,
Optional,
Type,
TypeVar,
Union,
)
import numpy as np
from pandas import DataFrame
from tanuki.data_store.data_type impo... | [
"typing.TypeVar"
] | [((567, 600), 'typing.TypeVar', 'TypeVar', (['"""B"""'], {'bound': '"""DataBackend"""'}), "('B', bound='DataBackend')\n", (574, 600), False, 'from typing import TYPE_CHECKING, Any, Generator, Generic, Optional, Type, TypeVar, Union\n')] |
SERVER_HOST = '127.0.0.1'
SERVER_PORT = 1335
MAX_GET_REQUESTS = 10
import re, socket
import random
import string
MSG_AUTH_RE = re.compile('''^AUTH +(.+) +(.+)''')
MSG_GET_RE = re.compile('''^GET +(.+) +(.+)''')
MSG_RVK_RE = re.compile('''^RVK +(.+)''')
def serve(srv):
while 1:
print('[S] Waiting for new... | [
"socket.socket",
"random.choice",
"re.compile"
] | [((129, 160), 're.compile', 're.compile', (['"""^AUTH +(.+) +(.+)"""'], {}), "('^AUTH +(.+) +(.+)')\n", (139, 160), False, 'import re, socket\n'), ((178, 208), 're.compile', 're.compile', (['"""^GET +(.+) +(.+)"""'], {}), "('^GET +(.+) +(.+)')\n", (188, 208), False, 'import re, socket\n'), ((226, 250), 're.compile', 'r... |
#!/usr/bin/python3
import json
state_file_path = "resources/state.json"
version = 0
class StateVersionException(Exception):
pass
def _get_state():
with open(state_file_path, 'r') as state_file:
data = json.load(state_file)
if data['version'] == 0:
data.pop('version', None)
... | [
"json.dump",
"json.load"
] | [((222, 243), 'json.load', 'json.load', (['state_file'], {}), '(state_file)\n', (231, 243), False, 'import json\n'), ((582, 633), 'json.dump', 'json.dump', (['dc', 'state_file'], {'sort_keys': '(True)', 'indent': '(4)'}), '(dc, state_file, sort_keys=True, indent=4)\n', (591, 633), False, 'import json\n')] |
from rain_alert.utils import RECEIVERS_FILE_PATH
import unittest
from .context import utils
import os
class TestGetReceivers(unittest.TestCase):
def test_no_file(self):
# set the path of the receivers to this folder because in tests folder
# we dont have the receivers file
utils.RECEIVE... | [
"unittest.main",
"os.path.split"
] | [((504, 519), 'unittest.main', 'unittest.main', ([], {}), '()\n', (517, 519), False, 'import unittest\n'), ((335, 375), 'os.path.split', 'os.path.split', (['utils.RECEIVERS_FILE_PATH'], {}), '(utils.RECEIVERS_FILE_PATH)\n', (348, 375), False, 'import os\n')] |
# Generated by Django 3.2.7 on 2021-10-18 12:21
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
('perfil', '0001_initial')... | [
"django.db.models.CharField",
"django.db.models.OneToOneField",
"django.db.migrations.swappable_dependency"
] | [((227, 284), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (258, 284), False, 'from django.db import migrations, models\n'), ((459, 515), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(50)', '... |
import gpflow
import matplotlib.pyplot as plt
import numpy as np
from robustgp import ConditionalVariance
X = np.random.rand(150, 1)
Y = 0.8 * np.cos(10 * X) + 1.2 * np.sin(8 * X + 0.3) + np.cos(17 * X) * 1.2 + np.random.randn(*X.shape) * 0.1
gpr = gpflow.models.GPR((X, Y), gpflow.kernels.SquaredExponential())
opt = ... | [
"robustgp.ConditionalVariance",
"gpflow.kernels.SquaredExponential",
"matplotlib.pyplot.show",
"numpy.random.randn",
"numpy.argmax",
"gpflow.optimizers.Scipy",
"gpflow.utilities.read_values",
"numpy.sin",
"numpy.linspace",
"numpy.cos",
"numpy.random.rand",
"matplotlib.pyplot.subplots"
] | [((111, 133), 'numpy.random.rand', 'np.random.rand', (['(150)', '(1)'], {}), '(150, 1)\n', (125, 133), True, 'import numpy as np\n'), ((320, 345), 'gpflow.optimizers.Scipy', 'gpflow.optimizers.Scipy', ([], {}), '()\n', (343, 345), False, 'import gpflow\n'), ((446, 481), 'gpflow.kernels.SquaredExponential', 'gpflow.kern... |
import sys
import time
from src.github import Github
from src.gitlab import Gitlab
def exit(message):
print(message)
sys.exit()
if __name__ == "__main__":
# Get all gitlab repositories
gitlab = Gitlab()
gitlab_repos = gitlab.repositories()
if gitlab_repos == None:
exit('Not abl... | [
"time.sleep",
"src.github.Github",
"sys.exit",
"src.gitlab.Gitlab"
] | [((126, 136), 'sys.exit', 'sys.exit', ([], {}), '()\n', (134, 136), False, 'import sys\n'), ((216, 224), 'src.gitlab.Gitlab', 'Gitlab', ([], {}), '()\n', (222, 224), False, 'from src.gitlab import Gitlab\n'), ((575, 583), 'src.github.Github', 'Github', ([], {}), '()\n', (581, 583), False, 'from src.github import Github... |
# -*- coding: utf-8 -*-
from flask import make_response, request, jsonify
from flask_login import login_required
import json
from werkzeug.utils import secure_filename
import os
from app.models.fragment import Fragment
from app.models.branch import Branch
from app.models.tag import Tag
from app.api import api
from app... | [
"app.api.api.route",
"app.models.tag.Tag.add",
"json.dumps",
"werkzeug.utils.secure_filename",
"app.models.branch.Branch.add",
"flask.jsonify",
"app.whoosh.search_helper.search",
"os.path.join"
] | [((498, 538), 'app.api.api.route', 'api.route', (['"""/add_tag/"""'], {'methods': "['POST']"}), "('/add_tag/', methods=['POST'])\n", (507, 538), False, 'from app.api import api\n'), ((1013, 1056), 'app.api.api.route', 'api.route', (['"""/add_branch/"""'], {'methods': "['POST']"}), "('/add_branch/', methods=['POST'])\n"... |
from env.evrp import EVRPEnv
from utils.config import read_config
import numpy as np
from env.worker import VectorizedEVRP
from utils.plot import convert_plt_to_numpy, convert_plt_to_tf, plot_tours
from utils import try_import_tensorflow
tf = try_import_tensorflow()
if __name__ == "__main__":
config = rea... | [
"utils.config.read_config",
"env.evrp.EVRPEnv",
"utils.plot.plot_tours",
"utils.try_import_tensorflow"
] | [((249, 272), 'utils.try_import_tensorflow', 'try_import_tensorflow', ([], {}), '()\n', (270, 272), False, 'from utils import try_import_tensorflow\n'), ((317, 343), 'utils.config.read_config', 'read_config', (['"""config.json"""'], {}), "('config.json')\n", (328, 343), False, 'from utils.config import read_config\n'),... |
# -*- coding: utf-8 -*- #
# Copyright 2019 Google LLC. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | [
"googlecloudsdk.calliope.base.ASYNC_FLAG.AddToParser",
"googlecloudsdk.calliope.arg_parsers.ArgDict"
] | [((3367, 3402), 'googlecloudsdk.calliope.base.ASYNC_FLAG.AddToParser', 'base.ASYNC_FLAG.AddToParser', (['parser'], {}), '(parser)\n', (3394, 3402), False, 'from googlecloudsdk.calliope import base\n'), ((2955, 3011), 'googlecloudsdk.calliope.arg_parsers.ArgDict', 'arg_parsers.ArgDict', ([], {'spec': "{'network': str}",... |
from flask import jsonify
from pyawsstarter import Logger
from wordservice import create_app
# Call the Application Factory function to construct a Flask application instance
# using the standard configuration defined in /instance/flask.cfg
application = create_app('flask.cfg')
@application.errorhandler(Exception)
... | [
"pyawsstarter.Logger.get_logger",
"wordservice.create_app"
] | [((257, 280), 'wordservice.create_app', 'create_app', (['"""flask.cfg"""'], {}), "('flask.cfg')\n", (267, 280), False, 'from wordservice import create_app\n'), ((565, 597), 'pyawsstarter.Logger.get_logger', 'Logger.get_logger', (['"""wordservice"""'], {}), "('wordservice')\n", (582, 597), False, 'from pyawsstarter impo... |
from django.conf.urls import url
from django.contrib import admin
import views
admin.autodiscover()
urlpatterns = [
url(r'^pass_forget/$', views.pass_forget, name='pass_forget'),
# url(r'^pass_rec/$', views.pass_rec, name='pass_rec'),
]
| [
"django.contrib.admin.autodiscover",
"django.conf.urls.url"
] | [((81, 101), 'django.contrib.admin.autodiscover', 'admin.autodiscover', ([], {}), '()\n', (99, 101), False, 'from django.contrib import admin\n'), ((123, 183), 'django.conf.urls.url', 'url', (['"""^pass_forget/$"""', 'views.pass_forget'], {'name': '"""pass_forget"""'}), "('^pass_forget/$', views.pass_forget, name='pass... |
from django.db import models
from .utils import upload_to_file, generate_random_string
from .validators import validate_file
class CompanyManager(models.Manager):
def get_or_none(self, **kwargs):
try:
return self.get(**kwargs)
except Company.DoesNotExist:
return None
cl... | [
"django.db.models.FileField",
"django.db.models.TextField",
"django.db.models.CharField",
"django.db.models.ForeignKey",
"django.db.models.BooleanField",
"django.db.models.EmailField",
"django.db.models.IntegerField",
"django.db.models.DateTimeField"
] | [((392, 455), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(60)', 'default': '"""default"""', 'unique': '(True)'}), "(max_length=60, default='default', unique=True)\n", (408, 455), False, 'from django.db import models\n'), ((474, 513), 'django.db.models.TextField', 'models.TextField', ([], {'b... |
# Copyright 2020-present Kensho Technologies, LLC.
import logging
import sys
import click
import gpg
from ..signing import (
DAYS_WARNING_FOR_KEY_EXPIRATION,
add_trusted_keys_to_gpg_home_dir,
get_days_until_expiry,
)
from ..utils import get_temporary_directory
logger = logging.getLogger(__name__)
logger... | [
"click.argument",
"logging.getLogger",
"click.command",
"gpg.Context",
"sys.exit"
] | [((286, 313), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (303, 313), False, 'import logging\n'), ((1637, 1652), 'click.command', 'click.command', ([], {}), '()\n', (1650, 1652), False, 'import click\n'), ((1654, 1707), 'click.argument', 'click.argument', (['"""days_before_warning"""']... |
from os import remove
from os.path import isdir, join
from pathlib import Path
from gdown import download
from zipfile import ZipFile
def download_ffhq(path):
path = join(path, 'ffhq')
if not isdir(path):
Path(path).mkdir(parents=True, exist_ok=True)
path_zip = join(path, 'ffhq.zip')
do... | [
"os.remove",
"zipfile.ZipFile",
"os.path.isdir",
"gdown.download",
"pathlib.Path",
"os.path.join"
] | [((171, 189), 'os.path.join', 'join', (['path', '"""ffhq"""'], {}), "(path, 'ffhq')\n", (175, 189), False, 'from os.path import isdir, join\n'), ((201, 212), 'os.path.isdir', 'isdir', (['path'], {}), '(path)\n', (206, 212), False, 'from os.path import isdir, join\n'), ((287, 309), 'os.path.join', 'join', (['path', '"""... |
import pdb
import time
import lib.tf_silent
import numpy as np
import tensorflow as tf
import matplotlib.pyplot as plt
import matplotlib.cm as cm
from matplotlib.colors import Normalize
from matplotlib.gridspec import GridSpec
import os
import pickle
import argparse
from lib.pinn import PINN
from lib.network import Ne... | [
"matplotlib.pyplot.title",
"pickle.dump",
"numpy.abs",
"argparse.ArgumentParser",
"numpy.floor",
"os.path.isfile",
"matplotlib.pyplot.figure",
"matplotlib.pyplot.contourf",
"matplotlib.pyplot.contour",
"pickle.load",
"numpy.round",
"matplotlib.pyplot.tight_layout",
"numpy.meshgrid",
"matpl... | [((395, 420), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (418, 420), False, 'import argparse\n'), ((1189, 1204), 'tensorflow.constant', 'tf.constant', (['xy'], {}), '(xy)\n', (1200, 1204), True, 'import tensorflow as tf\n'), ((2285, 2302), 'matplotlib.pyplot.subplot', 'plt.subplot', (['grid... |
"""The command line application for running the advent of code solutions."""
import argparse
import importlib
import pathlib
from typing import cast
from aoc2021.lib import ModSolution
def main() -> int:
"""Run the main CLI entry point."""
parser = argparse.ArgumentParser(
description="Run the advent... | [
"pathlib.Path",
"argparse.ArgumentParser",
"importlib.import_module"
] | [((260, 339), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Run the advent of code puzzle solutions."""'}), "(description='Run the advent of code puzzle solutions.')\n", (283, 339), False, 'import argparse\n'), ((646, 687), 'pathlib.Path', 'pathlib.Path', (['f"""input/{args.day:02d}.txt... |
import csv
import logging
import random
from argparse import ArgumentParser
from irc import IRC
from irc.messages import IRCMessage
from vaderSentiment.vaderSentiment import SentimentIntensityAnalyzer
positives = [
"(˶‾᷄ ⁻̫ ‾᷅˵)",
"(っˆڡˆς)",
"♥‿♥",
"(づ。◕‿‿◕。)づ",
"٩( ๑╹ ꇴ╹)۶",
"ᕕ( ᐛ )ᕗ",
"... | [
"argparse.ArgumentParser",
"logging.basicConfig",
"random.choice",
"irc.IRC",
"vaderSentiment.vaderSentiment.SentimentIntensityAnalyzer"
] | [((604, 730), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""[%(asctime)s] [%(levelname)-5s] %(message)s"""', 'level': 'logging.INFO', 'datefmt': '"""%Y-%m-%d %H:%M:%S"""'}), "(format='[%(asctime)s] [%(levelname)-5s] %(message)s',\n level=logging.INFO, datefmt='%Y-%m-%d %H:%M:%S')\n", (623, 730), ... |
import gzip
import pandas as pd
import numpy as np
import io
import os
import re
import torch
import torch.utils.data as data_utils
import subprocess
import zipfile
import zlib
from Bio import AlignIO
from Bio.SeqIO.FastaIO import FastaIterator, as_fasta
from Bio.Align.Applications import MuscleCommandline
class Ind... | [
"torch.from_numpy",
"io.StringIO",
"zipfile.ZipFile",
"numpy.random.seed",
"torch.utils.data.DataLoader",
"io.BytesIO",
"pandas.read_csv",
"Bio.SeqIO.FastaIO.FastaIterator",
"torch.load",
"numpy.floor",
"gzip.open",
"pandas.isnull",
"Bio.Align.Applications.MuscleCommandline",
"torch.save",... | [((2335, 2379), 'pandas.read_csv', 'pd.read_csv', (['input[1]'], {'index_col': '(0)', 'header': '(0)'}), '(input[1], index_col=0, header=0)\n', (2346, 2379), True, 'import pandas as pd\n'), ((2393, 2414), 'pandas.read_csv', 'pd.read_csv', (['input[0]'], {}), '(input[0])\n', (2404, 2414), True, 'import pandas as pd\n'),... |
"""construct landmark models """
import json
def read_json(fname):
with open(fname) as fid:
data = json.load(fid)
return data
def write_json(model, fname):
with open(fname, "w") as fid:
json.dump(model, fid)
index = dict(
jaw=[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 1... | [
"json.dump",
"json.load"
] | [((113, 127), 'json.load', 'json.load', (['fid'], {}), '(fid)\n', (122, 127), False, 'import json\n'), ((218, 239), 'json.dump', 'json.dump', (['model', 'fid'], {}), '(model, fid)\n', (227, 239), False, 'import json\n')] |
from django.db import models
import datetime
from easy_thumbnails.files import get_thumbnailer
from filer.fields.image import FilerImageField
import shortuuid
class Calendar(models.Model):
name = models.CharField(max_length=250)
uuid = models.CharField(max_length=22)
YEAR_CHOICES = [(r, r) for r in r... | [
"django.db.models.URLField",
"filer.fields.image.FilerImageField",
"django.db.models.CharField",
"shortuuid.uuid",
"django.db.models.ForeignKey",
"easy_thumbnails.files.get_thumbnailer",
"datetime.date.today",
"datetime.datetime.now"
] | [((207, 239), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(250)'}), '(max_length=250)\n', (223, 239), False, 'from django.db import models\n'), ((251, 282), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(22)'}), '(max_length=22)\n', (267, 282), False, 'from django.db ... |
from flask import Blueprint
ctr = Blueprint('ctr', __name__)
from . import views
| [
"flask.Blueprint"
] | [((35, 61), 'flask.Blueprint', 'Blueprint', (['"""ctr"""', '__name__'], {}), "('ctr', __name__)\n", (44, 61), False, 'from flask import Blueprint\n')] |
import tkinter as tk
from megawidget.tree import Hook
class TreeHook(Hook):
def __init__(self, parent_view, nodebar_builder, host):
self._parent_view = parent_view
self._nodebar_builder = nodebar_builder
self._host = host
self._stringvar_expander = tk.StringVar()
self._stri... | [
"tkinter.StringVar",
"tkinter.Entry",
"tkinter.Button",
"tkinter.Frame"
] | [((287, 301), 'tkinter.StringVar', 'tk.StringVar', ([], {}), '()\n', (299, 301), True, 'import tkinter as tk\n'), ((334, 348), 'tkinter.StringVar', 'tk.StringVar', ([], {}), '()\n', (346, 348), True, 'import tkinter as tk\n'), ((2019, 2123), 'tkinter.Button', 'tk.Button', (['frame'], {'name': '"""treeExpanderButton"""'... |
#!/usr/bin/python
# This file is licensed under MIT license.
# See the LICENSE file in the project root for more information.
import unittest
import rostest
import rosunit
import numpy as np
from numpy.testing import assert_almost_equal
from std_msgs.msg import Header
from geometry_msgs.msg import PoseStamped, Pose,... | [
"geometry_msgs.msg.PoseStamped",
"car_core.common.msgs_helpers.path_poses_to_array",
"car_core.common.msgs_helpers.array_to_point",
"rosunit.unitrun",
"car_core.common.msgs_helpers.array_to_path_poses",
"car_core.common.geom_helpers.get_closest_path_point",
"car_core.common.msgs_helpers.point_to_array",... | [((3932, 3997), 'rosunit.unitrun', 'rosunit.unitrun', (['"""car_core"""', '"""test_msgs_helpers"""', 'TestMsgsHelpers'], {}), "('car_core', 'test_msgs_helpers', TestMsgsHelpers)\n", (3947, 3997), False, 'import rosunit\n'), ((4002, 4067), 'rosunit.unitrun', 'rosunit.unitrun', (['"""car_core"""', '"""test_geom_helpers""... |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
__author__ = 'ipetrash'
from PyQt5.QtGui import QPixmap, QPainter, QFont
from PyQt5.QtWidgets import QApplication, QLabel
from PyQt5.QtCore import Qt, QRect
app = QApplication([])
text = "Hello World!"
pixmap = QPixmap(180, 130)
pixmap.fill(Qt.white)
painter = QPai... | [
"PyQt5.QtGui.QPainter",
"PyQt5.QtWidgets.QLabel",
"PyQt5.QtCore.QRect",
"PyQt5.QtGui.QFont",
"PyQt5.QtGui.QPixmap",
"PyQt5.QtWidgets.QApplication"
] | [((215, 231), 'PyQt5.QtWidgets.QApplication', 'QApplication', (['[]'], {}), '([])\n', (227, 231), False, 'from PyQt5.QtWidgets import QApplication, QLabel\n'), ((265, 282), 'PyQt5.QtGui.QPixmap', 'QPixmap', (['(180)', '(130)'], {}), '(180, 130)\n', (272, 282), False, 'from PyQt5.QtGui import QPixmap, QPainter, QFont\n'... |
from django.dispatch import Signal
# Arguments: "msg"
email_sent = Signal()
| [
"django.dispatch.Signal"
] | [((68, 76), 'django.dispatch.Signal', 'Signal', ([], {}), '()\n', (74, 76), False, 'from django.dispatch import Signal\n')] |
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.
import numpy as np
from typing import Union
from sklearn.manifold import TSNE
def tsne(
embedding: np.ndarray,
num_components: int = 2,
perplexity: float = 30.0,
early_exaggeration: float = 12.0,
learning... | [
"sklearn.manifold.TSNE"
] | [((6728, 7078), 'sklearn.manifold.TSNE', 'TSNE', ([], {'n_components': 'num_components', 'perplexity': 'perplexity', 'early_exaggeration': 'early_exaggeration', 'learning_rate': 'learning_rate', 'n_iter': 'num_iterations', 'n_iter_without_progress': 'num_iterations_without_progress', 'min_grad_norm': 'min_grad_norm', '... |
from django import forms
import voxel_globe.meta.models as models
class TiePointForm(forms.Form):
image_set = forms.ModelChoiceField(label="Image Set",
queryset=models.ImageSet.objects.all().order_by('name'))
class PointCloudForm(forms.Form):
point_cloud = forms.ModelChoiceField(label="Point Cloud",
... | [
"voxel_globe.meta.models.CameraSet.objects.all",
"voxel_globe.meta.models.Image.objects.all",
"voxel_globe.meta.models.ImageSet.objects.all",
"voxel_globe.meta.models.PointCloud.objects.all"
] | [((171, 200), 'voxel_globe.meta.models.ImageSet.objects.all', 'models.ImageSet.objects.all', ([], {}), '()\n', (198, 200), True, 'import voxel_globe.meta.models as models\n'), ((330, 361), 'voxel_globe.meta.models.PointCloud.objects.all', 'models.PointCloud.objects.all', ([], {}), '()\n', (359, 361), True, 'import voxe... |
import cv2
import numpy as np
imagen = cv2.imread('imagen.jpg')
imagen = cv2.cvtColor(imagen,cv2.COLOR_BGR2RGB)
print(imagen.shape)
print(imagen[0][0][0])
imagen = cv2.resize(imagen,(256, 256))
imagen = cv2.imread('imagen.jpg')
imagen = cv2.cvtColor(imagen,cv2.COLOR_BGR2GRAY)
print(imagen.shape)
print(imagen[0][0])
... | [
"cv2.cvtColor",
"cv2.imwrite",
"numpy.zeros",
"cv2.imread",
"cv2.resize"
] | [((40, 64), 'cv2.imread', 'cv2.imread', (['"""imagen.jpg"""'], {}), "('imagen.jpg')\n", (50, 64), False, 'import cv2\n'), ((74, 113), 'cv2.cvtColor', 'cv2.cvtColor', (['imagen', 'cv2.COLOR_BGR2RGB'], {}), '(imagen, cv2.COLOR_BGR2RGB)\n', (86, 113), False, 'import cv2\n'), ((165, 195), 'cv2.resize', 'cv2.resize', (['ima... |
# --------------------------------------------------------
# Faster R-CNN
# Copyright (c) 2015 Microsoft
# Licensed under The MIT License [see LICENSE for details]
# Written by <NAME> and <NAME>
# --------------------------------------------------------
# --------------------------------------------------------
# Reorg... | [
"torch.FloatTensor",
"torch.cat",
"torch.nonzero",
"torch.max",
"torch.arange",
"numpy.random.permutation",
"numpy.random.rand",
"numpy.round",
"torch.from_numpy"
] | [((1065, 1114), 'torch.FloatTensor', 'torch.FloatTensor', (['cfg.TRAIN.BBOX_NORMALIZE_MEANS'], {}), '(cfg.TRAIN.BBOX_NORMALIZE_MEANS)\n', (1082, 1114), False, 'import torch\n'), ((1150, 1198), 'torch.FloatTensor', 'torch.FloatTensor', (['cfg.TRAIN.BBOX_NORMALIZE_STDS'], {}), '(cfg.TRAIN.BBOX_NORMALIZE_STDS)\n', (1167, ... |
# -*- coding: utf-8 -*-
"""Model.ipynb
Automatically generated by Colaboratory.
Original file is located at
https://colab.research.google.com/drive/1QPnK5YOh8kRYPOOue6txwrgUqwKOMS0I
"""
# # Use seaborn for pairplot
# !pip install -q seaborn
# !pip install tensorflow==2.0.0
# # Use some functions from tensorflow_... | [
"pandas.DataFrame",
"matplotlib.rc",
"warnings.filterwarnings",
"pandas.read_csv",
"datetime.strptime",
"tensorflow.keras.layers.Dense",
"pandas.isnull",
"sklearn.preprocessing.LabelEncoder",
"matplotlib.pyplot.style.use",
"numpy.int64",
"tensorflow.keras.optimizers.RMSprop"
] | [((1373, 1405), 'matplotlib.pyplot.style.use', 'plt.style.use', (['"""fivethirtyeight"""'], {}), "('fivethirtyeight')\n", (1386, 1405), True, 'import matplotlib.pyplot as plt\n'), ((1406, 1455), 'matplotlib.rc', 'mpl.rc', (['"""patch"""'], {'edgecolor': '"""dimgray"""', 'linewidth': '(1)'}), "('patch', edgecolor='dimgr... |
from datetime import datetime
from utils.api import API
from time import sleep
from config import *
import random
def load_file(file):
try:
l = []
with open(file, 'r') as f:
for line in f:
l.append(line.rstrip())
return l
except FileNotFoundErr... | [
"utils.api.API",
"datetime.datetime.now",
"random.randint",
"time.sleep"
] | [((440, 517), 'utils.api.API', 'API', (['REDDIT_CLIENT_ID', 'REDDIT_CLIENT_SECRET', 'REDDIT_USERNAME', 'REDDIT_PASSWORD'], {}), '(REDDIT_CLIENT_ID, REDDIT_CLIENT_SECRET, REDDIT_USERNAME, REDDIT_PASSWORD)\n', (443, 517), False, 'from utils.api import API\n'), ((782, 790), 'time.sleep', 'sleep', (['(1)'], {}), '(1)\n', (... |
# -*- coding: utf-8 -*-
# Define here the models for your scraped items
#
# See documentation in:
# https://doc.scrapy.org/en/latest/topics/items.html
import scrapy
from scrapy.loader.processors import TakeFirst, Join, MapCompose
from datetime import datetime, timedelta
def comments_strip(string,loader_context):... | [
"scrapy.loader.processors.Join",
"scrapy.Field",
"json.loads",
"datetime.datetime.fromtimestamp"
] | [((3018, 3037), 'json.loads', 'json.loads', (['date[0]'], {}), '(date[0])\n', (3028, 3037), False, 'import json\n'), ((3681, 3709), 'json.loads', 'json.loads', (['post_id[::-1][0]'], {}), '(post_id[::-1][0])\n', (3691, 3709), False, 'import json\n'), ((3825, 3839), 'scrapy.Field', 'scrapy.Field', ([], {}), '()\n', (383... |
#!/usr/bin/env python3
import os, sys, subprocess, socket
#import cgroups
def start_mjpg_streamer():
print("Starting up mjpg_streamer.")
# TODO: Add notification if either mjpg-streamer or
# cvfilter_py.so aren't installed
# TODO: Detect any error if process exits,
# such as the uvcvid... | [
"line_follower.mjs_filter",
"os.path.realpath",
"socket.socket",
"car_control.car_control"
] | [((1140, 1168), 'line_follower.mjs_filter', 'line_follower.mjs_filter', (['dc'], {}), '(dc)\n', (1164, 1168), False, 'import line_follower\n'), ((1402, 1427), 'car_control.car_control', 'car_control.car_control', ([], {}), '()\n', (1425, 1427), False, 'import car_control\n'), ((1549, 1597), 'socket.socket', 'socket.soc... |
import os
import json
from flask import Flask
from tqdm import tqdm
from ..validator.schema_validator import Validator
from ..utils.paths import get_path
app = Flask(__name__)
# Function to accept user-generated feedstock
# Args:
# path: Path to the feedstock
# remove_old: Should fully accepted feedstock be rem... | [
"os.remove",
"json.loads",
"os.path.basename",
"flask.Flask",
"os.path.join",
"os.listdir"
] | [((163, 178), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (168, 178), False, 'from flask import Flask\n'), ((1036, 1051), 'os.remove', 'os.remove', (['path'], {}), '(path)\n', (1045, 1051), False, 'import os\n'), ((1664, 1680), 'os.listdir', 'os.listdir', (['path'], {}), '(path)\n', (1674, 1680), False,... |
import os
os.environ['MANTLE'] = 'lattice'
from magma import *
from mantle import And, XOr
from simulator import testvectors
main = DefineCircuit('main', "a", In(Bit), "b", In(Bit), "c", In(Bit), "d", Out(Bit), 'CLK', In(Bit))
t = And(2)(main.a,main.b)
d = XOr(2)(t,main.c)
wire(d,main.d)
EndCircuit()
print(testvecto... | [
"simulator.testvectors",
"mantle.XOr",
"mantle.And"
] | [((233, 239), 'mantle.And', 'And', (['(2)'], {}), '(2)\n', (236, 239), False, 'from mantle import And, XOr\n'), ((259, 265), 'mantle.XOr', 'XOr', (['(2)'], {}), '(2)\n', (262, 265), False, 'from mantle import And, XOr\n'), ((311, 328), 'simulator.testvectors', 'testvectors', (['main'], {}), '(main)\n', (322, 328), Fals... |
from __future__ import absolute_import
# --------------------------------------------------------
# Spatial Attention Network withFeature Mimicking
# Copyright (c) 2018 University of Illinois
# Licensed under The MIT License [see LICENSE for details]
# --------------------------------------------------------
# --------... | [
"numpy.amax",
"numpy.argsort",
"numpy.zeros"
] | [((1534, 1559), 'numpy.amax', 'np.amax', (['cls_prob'], {'axis': '(1)'}), '(cls_prob, axis=1)\n', (1541, 1559), True, 'import numpy as np\n'), ((1775, 1826), 'numpy.zeros', 'np.zeros', (['(proposals.shape[0], 1)'], {'dtype': 'np.float32'}), '((proposals.shape[0], 1), dtype=np.float32)\n', (1783, 1826), True, 'import nu... |