content
stringlengths
27
928k
path
stringlengths
4
230
size
int64
27
928k
nl_text
stringlengths
21
396k
nl_size
int64
21
396k
nl_language
stringlengths
2
3
nl_language_score
float64
0.04
1
# Copyright 2012, VMware, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock import testtools from neutron.agent.linux import utils from neutron.tests import base _marker = object() class AgentUtilsExecuteTest(base.BaseTestCase): def setUp(self): super(AgentUtilsExecuteTest, self).setUp() self.test_file = self.get_temp_file_path('test_execute.tmp') open(self.test_file, 'w').close() self.process = mock.patch('eventlet.green.subprocess.Popen').start() self.process.return_value.returncode = 0 self.mock_popen = self.process.return_value.communicate def test_without_helper(self): expected = "%s\n" % self.test_file self.mock_popen.return_value = [expected, ""] result = utils.execute(["ls", self.test_file]) self.assertEqual(result, expected) def test_with_helper(self): expected = "ls %s\n" % self.test_file self.mock_popen.return_value = [expected, ""] self.config(group='AGENT', root_helper='echo') result = utils.execute(["ls", self.test_file], run_as_root=True) self.assertEqual(result, expected) def test_stderr_true(self): expected = "%s\n" % self.test_file self.mock_popen.return_value = [expected, ""] out = utils.execute(["ls", self.test_file], return_stderr=True) self.assertIsInstance(out, tuple) self.assertEqual(out, (expected, "")) def test_check_exit_code(self): self.mock_popen.return_value = ["", ""] stdout = utils.execute(["ls", self.test_file[:-1]], check_exit_code=False) self.assertEqual(stdout, "") def test_execute_raises(self): self.mock_popen.side_effect = RuntimeError self.assertRaises(RuntimeError, utils.execute, ["ls", self.test_file[:-1]]) def test_process_input(self): expected = "%s\n" % self.test_file[:-1] self.mock_popen.return_value = [expected, ""] result = utils.execute(["cat"], process_input="%s\n" % self.test_file[:-1]) self.assertEqual(result, expected) def test_with_addl_env(self): expected = "%s\n" % self.test_file self.mock_popen.return_value = [expected, ""] result = utils.execute(["ls", self.test_file], addl_env={'foo': 'bar'}) self.assertEqual(result, expected) def test_return_code_log_error_raise_runtime(self): self.mock_popen.return_value = ('', '') self.process.return_value.returncode = 1 with mock.patch.object(utils, 'LOG') as log: self.assertRaises(RuntimeError, utils.execute, ['ls']) self.assertTrue(log.error.called) def test_return_code_log_error_no_raise_runtime(self): self.mock_popen.return_value = ('', '') self.process.return_value.returncode = 1 with mock.patch.object(utils, 'LOG') as log: utils.execute(['ls'], check_exit_code=False) self.assertTrue(log.error.called) def test_return_code_log_debug(self): self.mock_popen.return_value = ('', '') with mock.patch.object(utils, 'LOG') as log: utils.execute(['ls']) self.assertTrue(log.debug.called) def test_return_code_raise_runtime_do_not_log_fail_as_error(self): self.mock_popen.return_value = ('', '') self.process.return_value.returncode = 1 with mock.patch.object(utils, 'LOG') as log: self.assertRaises(RuntimeError, utils.execute, ['ls'], log_fail_as_error=False) self.assertFalse(log.error.called) class AgentUtilsGetInterfaceMAC(base.BaseTestCase): def test_get_interface_mac(self): expect_val = '01:02:03:04:05:06' with mock.patch('fcntl.ioctl') as ioctl: ioctl.return_value = ''.join(['\x00' * 18, '\x01\x02\x03\x04\x05\x06', '\x00' * 232]) actual_val = utils.get_interface_mac('eth0') self.assertEqual(actual_val, expect_val) class AgentUtilsReplaceFile(base.BaseTestCase): def test_replace_file(self): # make file to replace with mock.patch('tempfile.NamedTemporaryFile') as ntf: ntf.return_value.name = '/baz' with mock.patch('os.chmod') as chmod: with mock.patch('os.rename') as rename: utils.replace_file('/foo', 'bar') expected = [mock.call('w+', dir='/', delete=False), mock.call().write('bar'), mock.call().close()] ntf.assert_has_calls(expected) chmod.assert_called_once_with('/baz', 0o644) rename.assert_called_once_with('/baz', '/foo') class TestFindChildPids(base.BaseTestCase): def test_returns_empty_list_for_exit_code_1(self): with mock.patch.object(utils, 'execute', side_effect=RuntimeError('Exit code: 1')): self.assertEqual(utils.find_child_pids(-1), []) def test_returns_empty_list_for_no_output(self): with mock.patch.object(utils, 'execute', return_value=''): self.assertEqual(utils.find_child_pids(-1), []) def test_returns_list_of_child_process_ids_for_good_ouput(self): with mock.patch.object(utils, 'execute', return_value=' 123 \n 185\n'): self.assertEqual(utils.find_child_pids(-1), ['123', '185']) def test_raises_unknown_exception(self): with testtools.ExpectedException(RuntimeError): with mock.patch.object(utils, 'execute', side_effect=RuntimeError()): utils.find_child_pids(-1) class TestGetRoothelperChildPid(base.BaseTestCase): def _test_get_root_helper_child_pid(self, expected=_marker, run_as_root=False, pids=None): def _find_child_pids(x): if not pids: return [] pids.pop(0) return pids mock_pid = object() with mock.patch.object(utils, 'find_child_pids', side_effect=_find_child_pids): actual = utils.get_root_helper_child_pid(mock_pid, run_as_root) if expected is _marker: expected = str(mock_pid) self.assertEqual(expected, actual) def test_returns_process_pid_not_root(self): self._test_get_root_helper_child_pid() def test_returns_child_pid_as_root(self): self._test_get_root_helper_child_pid(expected='2', pids=['1', '2'], run_as_root=True) def test_returns_last_child_pid_as_root(self): self._test_get_root_helper_child_pid(expected='3', pids=['1', '2', '3'], run_as_root=True) def test_returns_none_as_root(self): self._test_get_root_helper_child_pid(expected=None, run_as_root=True) class TestPathUtilities(base.BaseTestCase): def test_remove_abs_path(self): self.assertEqual(['ping', '8.8.8.8'], utils.remove_abs_path(['/usr/bin/ping', '8.8.8.8'])) def test_cmdlines_are_equal(self): self.assertTrue(utils.cmdlines_are_equal( ['ping', '8.8.8.8'], ['/usr/bin/ping', '8.8.8.8'])) def test_cmdlines_are_equal_different_commands(self): self.assertFalse(utils.cmdlines_are_equal( ['ping', '8.8.8.8'], ['/usr/bin/ping6', '8.8.8.8']))
neutron/tests/unit/agent/linux/test_utils.py
8,292
Copyright 2012, VMware, Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. make file to replace
599
en
0.859696
import logging from urllib.parse import urljoin import scrapy from scrapy import Request from scrapy_selenium import SeleniumRequest from selenium.webdriver.common.by import By from selenium.webdriver.support import expected_conditions as EC from sainsburys.items import SainsburysItem logger = logging.getLogger('spam_application') logger.setLevel(logging.DEBUG) class BasicSpider(scrapy.Spider): name = 'basic' allowed_domains = ['www.sainsburys.co.uk'] start_urls = ['https://www.sainsburys.co.uk/shop/gb/groceries/meat-fish'] def parse(self, response): urls = response.css("ul.categories.departments li a::attr(href)").extract() for url in urls: yield response.follow(url, callback=self.parse_department) def parse_department(self, response): products = response.css("ul.productLister.gridView").extract() if products: for product in self.handle_product_listing(response): yield product pages = response.css("ul.categories.shelf li a::attr(href)").extract() if not pages: pages = response.css("ul.categories.aisles li a::attr(href)").extract() if not pages: return for url in pages: yield response.follow(url, callback=self.parse_department) def handle_product_listing(self, response): selector = 'button.ln-c-button' urls = response.css("ul.productLister.gridView li.gridItem h3 a::attr(href)").extract() for url in urls: # yield response.follow(url, callback=self.parse_product) yield SeleniumRequest(url=url, callback=self.parse_product, wait_time=10, wait_until=EC.element_to_be_clickable((By.CSS_SELECTOR, selector))) next_page = response.css("#productLister > div.pagination.paginationBottom > ul > li.next > a::attr(href)").extract() if next_page: yield response.follow(next_page, callback=self.handle_product_listing) def parse_product(self, response): product_name = response.css("h1.pd__header::text").extract()[0] product_image = response.css("img.pd__image::attr(src)").extract()[0] item = SainsburysItem() item["url"] = response.url item["name"] = product_name item["image"] = product_image yield item
sainsburys/sainsburys/spiders/basic.py
2,385
yield response.follow(url, callback=self.parse_product)
55
en
0.157362
import pickle import torch from torch.utils.data import Dataset, DataLoader import numpy as np import torch.nn as nn def load_pickle(pickle_file): try: with open(pickle_file, 'rb') as f: pickle_data = pickle.load(f) except UnicodeDecodeError as e: with open(pickle_file, 'rb') as f: pickle_data = pickle.load(f, encoding='latin1') except Exception as e: print('Unable to load data ', pickle_file, ':', e) raise return pickle_data class HumorDataset(Dataset): def __init__(self, id_list, max_context_len=5, max_sen_len=20): self.id_list = id_list openface_file = "openface_features_sdk.pkl" covarep_file = "covarep_features_sdk.pkl" language_file = "language_sdk.pkl" word_embedding_list_file = "word_embedding_list.pkl" humor_label_file = "humor_label_sdk.pkl" self.word_aligned_openface_sdk = load_pickle(openface_file) self.word_aligned_covarep_sdk = load_pickle(covarep_file) self.language_sdk = load_pickle(language_file) self.word_embedding_list_sdk = load_pickle(word_embedding_list_file) self.humor_label_sdk = load_pickle(humor_label_file) self.of_d = 371 self.cvp_d = 81 self.glove_d = 300 self.total_dim = self.glove_d + self.of_d + self.cvp_d self.max_context_len = max_context_len self.max_sen_len = max_sen_len # left padding with zero vector upto maximum number of words in a sentence * glove embedding dimension def paded_word_idx(self, seq, max_sen_len=20, left_pad=1): seq = seq[0:max_sen_len] pad_w = np.concatenate((np.zeros(max_sen_len - len(seq)), seq), axis=0) pad_w = np.array([self.word_embedding_list_sdk[int(w_id)] for w_id in pad_w]) return pad_w # left padding with zero vector upto maximum number of words in a sentence * covarep dimension def padded_covarep_features(self, seq, max_sen_len=20, left_pad=1): seq = seq[0:max_sen_len] return np.concatenate((np.zeros((max_sen_len - len(seq), self.cvp_d)), seq), axis=0) # left padding with zero vector upto maximum number of words in a sentence * openface dimension def padded_openface_features(self, seq, max_sen_len=20, left_pad=1): seq = seq[0:max_sen_len] return np.concatenate((np.zeros(((max_sen_len - len(seq)), self.of_d)), seq), axis=0) # left padding with zero vectors upto maximum number of sentences in context * maximum num of words in a sentence * 456 def padded_context_features(self, context_w, context_of, context_cvp, max_context_len=5, max_sen_len=20): context_w = context_w[-max_context_len:] context_of = context_of[-max_context_len:] context_cvp = context_cvp[-max_context_len:] padded_context = [] for i in range(len(context_w)): p_seq_w = self.paded_word_idx(context_w[i], max_sen_len) p_seq_cvp = self.padded_covarep_features( context_cvp[i], max_sen_len) p_seq_of = self.padded_openface_features( context_of[i], max_sen_len) padded_context.append(np.concatenate( (p_seq_w, p_seq_cvp, p_seq_of), axis=1)) pad_c_len = max_context_len - len(padded_context) padded_context = np.array(padded_context) # if there is no context if not padded_context.any(): return np.zeros((max_context_len, max_sen_len, self.total_dim)) return np.concatenate((np.zeros((pad_c_len, max_sen_len, self.total_dim)), padded_context), axis=0) def padded_punchline_features(self, punchline_w, punchline_of, punchline_cvp, max_sen_len=20, left_pad=1): p_seq_w = self.paded_word_idx(punchline_w, max_sen_len) p_seq_cvp = self.padded_covarep_features(punchline_cvp, max_sen_len) p_seq_of = self.padded_openface_features(punchline_of, max_sen_len) return np.concatenate((p_seq_w, p_seq_cvp, p_seq_of), axis=1) def __len__(self): return len(self.id_list) def __getitem__(self, index): hid = self.id_list[index] punchline_w = np.array( self.language_sdk[hid]['punchline_embedding_indexes']) punchline_of = np.array( self.word_aligned_openface_sdk[hid]['punchline_features']) punchline_cvp = np.array( self.word_aligned_covarep_sdk[hid]['punchline_features']) context_w = np.array( self.language_sdk[hid]['context_embedding_indexes']) context_of = np.array( self.word_aligned_openface_sdk[hid]['context_features']) context_cvp = np.array( self.word_aligned_covarep_sdk[hid]['context_features']) # punchline feature x_p = torch.LongTensor( self.padded_punchline_features(punchline_w, punchline_of, punchline_cvp, self.max_sen_len)) # context feature x_c = torch.LongTensor( self.padded_context_features(context_w, context_of, context_cvp, self.max_context_len, self.max_sen_len)) y = torch.FloatTensor([self.humor_label_sdk[hid]]) return x_p, x_c, y
deprecated/dataloaders/affect/humor_dataset.py
5,222
left padding with zero vector upto maximum number of words in a sentence * glove embedding dimension left padding with zero vector upto maximum number of words in a sentence * covarep dimension left padding with zero vector upto maximum number of words in a sentence * openface dimension left padding with zero vectors upto maximum number of sentences in context * maximum num of words in a sentence * 456 if there is no context punchline feature context feature
465
en
0.689097
#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "travel.settings") try: from django.core.management import execute_from_command_line except ImportError: # The above import may fail for some other reason. Ensure that the # issue is really that Django is missing to avoid masking other # exceptions on Python 2. try: import django except ImportError: raise ImportError( "Couldn't import Django. Are you sure it's installed and " "available on your PYTHONPATH environment variable? Did you " "forget to activate a virtual environment?" ) raise execute_from_command_line(sys.argv)
travel/manage.py
804
!/usr/bin/env python The above import may fail for some other reason. Ensure that the issue is really that Django is missing to avoid masking other exceptions on Python 2.
171
en
0.863733
class Solution: def maxChunksToSorted(self, arr): """ :type arr: List[int] :rtype: int """ stacks = [] for num in arr: if not stacks: stacks.append([num]) elif num >= stacks[-1][0]: stacks.append([num]) else: # num < stacks[-1][0] stacks[-1].append(num) while len(stacks) >= 2: if num < stacks[-2][0]: stacks[-2][0] = max(stacks[-2][0], stacks[-1][0]) stacks[-2].extend(stacks.pop()) else: break # print(stacks) return len(stacks) sol = Solution().maxChunksToSorted print(sol([5, 4, 3, 2, 1])) print(sol([2, 1, 3, 4, 4])) print(sol([5, 3, 1, 2, 4])) print(sol([1, 0, 1, 3, 2])) print(sol([5, 1, 1, 8, 1, 6, 5, 9, 7, 8]))
p768_max_chunks_to_make_sorted_ii.py
919
:type arr: List[int] :rtype: int num < stacks[-1][0] print(stacks)
68
en
0.244743
from operator import attrgetter import pyangbind.lib.xpathhelper as xpathhelper from pyangbind.lib.yangtypes import RestrictedPrecisionDecimalType, RestrictedClassType, TypedListType from pyangbind.lib.yangtypes import YANGBool, YANGListType, YANGDynClass, ReferenceType from pyangbind.lib.base import PybindBase from decimal import Decimal from bitarray import bitarray import __builtin__ class lsp_admin_group_include_any(PybindBase): """ This class was auto-generated by the PythonClass plugin for PYANG from YANG module brocade-mpls - based on the path /brocade_mpls_rpc/show-mpls-lsp-extensive/output/lsp/show-mpls-lsp-extensive-info/show-mpls-lsp-sec-path-info/sec-path/lsp-sec-path-config-admin-groups/lsp-admin-group/lsp-admin-group-include-any. Each member element of the container is represented as a class variable - with a specific YANG type. """ __slots__ = ('_pybind_generated_by', '_path_helper', '_yang_name', '_rest_name', '_extmethods', '__lsp_admin_group_include_any_group_id',) _yang_name = 'lsp-admin-group-include-any' _rest_name = 'lsp-admin-group-include-any' _pybind_generated_by = 'container' def __init__(self, *args, **kwargs): path_helper_ = kwargs.pop("path_helper", None) if path_helper_ is False: self._path_helper = False elif path_helper_ is not None and isinstance(path_helper_, xpathhelper.YANGPathHelper): self._path_helper = path_helper_ elif hasattr(self, "_parent"): path_helper_ = getattr(self._parent, "_path_helper", False) self._path_helper = path_helper_ else: self._path_helper = False extmethods = kwargs.pop("extmethods", None) if extmethods is False: self._extmethods = False elif extmethods is not None and isinstance(extmethods, dict): self._extmethods = extmethods elif hasattr(self, "_parent"): extmethods = getattr(self._parent, "_extmethods", None) self._extmethods = extmethods else: self._extmethods = False self.__lsp_admin_group_include_any_group_id = YANGDynClass(base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name="lsp-admin-group-include-any-group-id", rest_name="lsp-admin-group-include-any-group-id", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, is_keyval=True, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True) load = kwargs.pop("load", None) if args: if len(args) > 1: raise TypeError("cannot create a YANG container with >1 argument") all_attr = True for e in self._pyangbind_elements: if not hasattr(args[0], e): all_attr = False break if not all_attr: raise ValueError("Supplied object did not have the correct attributes") for e in self._pyangbind_elements: nobj = getattr(args[0], e) if nobj._changed() is False: continue setmethod = getattr(self, "_set_%s" % e) if load is None: setmethod(getattr(args[0], e)) else: setmethod(getattr(args[0], e), load=load) def _path(self): if hasattr(self, "_parent"): return self._parent._path()+[self._yang_name] else: return [u'brocade_mpls_rpc', u'show-mpls-lsp-extensive', u'output', u'lsp', u'show-mpls-lsp-extensive-info', u'show-mpls-lsp-sec-path-info', u'sec-path', u'lsp-sec-path-config-admin-groups', u'lsp-admin-group', u'lsp-admin-group-include-any'] def _rest_path(self): if hasattr(self, "_parent"): if self._rest_name: return self._parent._rest_path()+[self._rest_name] else: return self._parent._rest_path() else: return [u'show-mpls-lsp-extensive', u'output', u'lsp', u'sec-path', u'lsp-sec-path-config-admin-groups', u'lsp-admin-group-include-any'] def _get_lsp_admin_group_include_any_group_id(self): """ Getter method for lsp_admin_group_include_any_group_id, mapped from YANG variable /brocade_mpls_rpc/show_mpls_lsp_extensive/output/lsp/show_mpls_lsp_extensive_info/show_mpls_lsp_sec_path_info/sec_path/lsp_sec_path_config_admin_groups/lsp_admin_group/lsp_admin_group_include_any/lsp_admin_group_include_any_group_id (uint32) YANG Description: Include any admin group id """ return self.__lsp_admin_group_include_any_group_id def _set_lsp_admin_group_include_any_group_id(self, v, load=False): """ Setter method for lsp_admin_group_include_any_group_id, mapped from YANG variable /brocade_mpls_rpc/show_mpls_lsp_extensive/output/lsp/show_mpls_lsp_extensive_info/show_mpls_lsp_sec_path_info/sec_path/lsp_sec_path_config_admin_groups/lsp_admin_group/lsp_admin_group_include_any/lsp_admin_group_include_any_group_id (uint32) If this variable is read-only (config: false) in the source YANG file, then _set_lsp_admin_group_include_any_group_id is considered as a private method. Backends looking to populate this variable should do so via calling thisObj._set_lsp_admin_group_include_any_group_id() directly. YANG Description: Include any admin group id """ parent = getattr(self, "_parent", None) if parent is not None and load is False: raise AttributeError("Cannot set keys directly when" + " within an instantiated list") if hasattr(v, "_utype"): v = v._utype(v) try: t = YANGDynClass(v,base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name="lsp-admin-group-include-any-group-id", rest_name="lsp-admin-group-include-any-group-id", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, is_keyval=True, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True) except (TypeError, ValueError): raise ValueError({ 'error-string': """lsp_admin_group_include_any_group_id must be of a type compatible with uint32""", 'defined-type': "uint32", 'generated-type': """YANGDynClass(base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name="lsp-admin-group-include-any-group-id", rest_name="lsp-admin-group-include-any-group-id", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, is_keyval=True, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True)""", }) self.__lsp_admin_group_include_any_group_id = t if hasattr(self, '_set'): self._set() def _unset_lsp_admin_group_include_any_group_id(self): self.__lsp_admin_group_include_any_group_id = YANGDynClass(base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name="lsp-admin-group-include-any-group-id", rest_name="lsp-admin-group-include-any-group-id", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, is_keyval=True, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True) lsp_admin_group_include_any_group_id = __builtin__.property(_get_lsp_admin_group_include_any_group_id, _set_lsp_admin_group_include_any_group_id) _pyangbind_elements = {'lsp_admin_group_include_any_group_id': lsp_admin_group_include_any_group_id, }
pybind/slxos/v17r_1_01a/brocade_mpls_rpc/show_mpls_lsp_extensive/output/lsp/show_mpls_lsp_extensive_info/show_mpls_lsp_sec_path_info/sec_path/lsp_sec_path_config_admin_groups/lsp_admin_group/lsp_admin_group_include_any/__init__.py
7,563
This class was auto-generated by the PythonClass plugin for PYANG from YANG module brocade-mpls - based on the path /brocade_mpls_rpc/show-mpls-lsp-extensive/output/lsp/show-mpls-lsp-extensive-info/show-mpls-lsp-sec-path-info/sec-path/lsp-sec-path-config-admin-groups/lsp-admin-group/lsp-admin-group-include-any. Each member element of the container is represented as a class variable - with a specific YANG type. Getter method for lsp_admin_group_include_any_group_id, mapped from YANG variable /brocade_mpls_rpc/show_mpls_lsp_extensive/output/lsp/show_mpls_lsp_extensive_info/show_mpls_lsp_sec_path_info/sec_path/lsp_sec_path_config_admin_groups/lsp_admin_group/lsp_admin_group_include_any/lsp_admin_group_include_any_group_id (uint32) YANG Description: Include any admin group id Setter method for lsp_admin_group_include_any_group_id, mapped from YANG variable /brocade_mpls_rpc/show_mpls_lsp_extensive/output/lsp/show_mpls_lsp_extensive_info/show_mpls_lsp_sec_path_info/sec_path/lsp_sec_path_config_admin_groups/lsp_admin_group/lsp_admin_group_include_any/lsp_admin_group_include_any_group_id (uint32) If this variable is read-only (config: false) in the source YANG file, then _set_lsp_admin_group_include_any_group_id is considered as a private method. Backends looking to populate this variable should do so via calling thisObj._set_lsp_admin_group_include_any_group_id() directly. YANG Description: Include any admin group id
1,436
en
0.470591
# -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! # source: sprint.proto import sys _b=sys.version_info[0]<3 and (lambda x:x) or (lambda x:x.encode('latin1')) from google.protobuf import descriptor as _descriptor from google.protobuf import message as _message from google.protobuf import reflection as _reflection from google.protobuf import symbol_database as _symbol_database # @@protoc_insertion_point(imports) _sym_db = _symbol_database.Default() from easy_work_service_sdk.model.topboard import product_basic_pb2 as easy__work__service__sdk_dot_model_dot_topboard_dot_product__basic__pb2 from easy_work_service_sdk.model.topboard import issue_basic_pb2 as easy__work__service__sdk_dot_model_dot_topboard_dot_issue__basic__pb2 DESCRIPTOR = _descriptor.FileDescriptor( name='sprint.proto', package='topboard', syntax='proto3', serialized_options=_b('ZBgo.easyops.local/contracts/protorepo-models/easyops/model/topboard'), serialized_pb=_b('\n\x0csprint.proto\x12\x08topboard\x1a\x38\x65\x61sy_work_service_sdk/model/topboard/product_basic.proto\x1a\x36\x65\x61sy_work_service_sdk/model/topboard/issue_basic.proto\"\xca\x01\n\x06Sprint\x12\'\n\x07product\x18\x01 \x03(\x0b\x32\x16.topboard.ProductBasic\x12$\n\x06issues\x18\x02 \x03(\x0b\x32\x14.topboard.IssueBasic\x12\x0c\n\x04name\x18\x03 \x01(\t\x12\x12\n\ninstanceId\x18\x04 \x01(\t\x12\r\n\x05title\x18\x05 \x01(\t\x12\x0e\n\x06status\x18\x06 \x01(\t\x12\x0c\n\x04goal\x18\x07 \x01(\t\x12\x11\n\tstartTime\x18\x08 \x01(\t\x12\x0f\n\x07\x65ndTime\x18\t \x01(\tBDZBgo.easyops.local/contracts/protorepo-models/easyops/model/topboardb\x06proto3') , dependencies=[easy__work__service__sdk_dot_model_dot_topboard_dot_product__basic__pb2.DESCRIPTOR,easy__work__service__sdk_dot_model_dot_topboard_dot_issue__basic__pb2.DESCRIPTOR,]) _SPRINT = _descriptor.Descriptor( name='Sprint', full_name='topboard.Sprint', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='product', full_name='topboard.Sprint.product', index=0, number=1, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='issues', full_name='topboard.Sprint.issues', index=1, number=2, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='name', full_name='topboard.Sprint.name', index=2, number=3, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='instanceId', full_name='topboard.Sprint.instanceId', index=3, number=4, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='title', full_name='topboard.Sprint.title', index=4, number=5, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='status', full_name='topboard.Sprint.status', index=5, number=6, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='goal', full_name='topboard.Sprint.goal', index=6, number=7, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='startTime', full_name='topboard.Sprint.startTime', index=7, number=8, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), _descriptor.FieldDescriptor( name='endTime', full_name='topboard.Sprint.endTime', index=8, number=9, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=141, serialized_end=343, ) _SPRINT.fields_by_name['product'].message_type = easy__work__service__sdk_dot_model_dot_topboard_dot_product__basic__pb2._PRODUCTBASIC _SPRINT.fields_by_name['issues'].message_type = easy__work__service__sdk_dot_model_dot_topboard_dot_issue__basic__pb2._ISSUEBASIC DESCRIPTOR.message_types_by_name['Sprint'] = _SPRINT _sym_db.RegisterFileDescriptor(DESCRIPTOR) Sprint = _reflection.GeneratedProtocolMessageType('Sprint', (_message.Message,), { 'DESCRIPTOR' : _SPRINT, '__module__' : 'sprint_pb2' # @@protoc_insertion_point(class_scope:topboard.Sprint) }) _sym_db.RegisterMessage(Sprint) DESCRIPTOR._options = None # @@protoc_insertion_point(module_scope)
easy_work_service_sdk/model/topboard/sprint_pb2.py
6,211
-*- coding: utf-8 -*- Generated by the protocol buffer compiler. DO NOT EDIT! source: sprint.proto @@protoc_insertion_point(imports) @@protoc_insertion_point(class_scope:topboard.Sprint) @@protoc_insertion_point(module_scope)
226
en
0.539846
from __future__ import print_function from __future__ import absolute_import from __future__ import division try: basestring except NameError: basestring = str import compas_rhino from compas.utilities import iterable_like from compas_rhino.artists._primitiveartist import PrimitiveArtist __all__ = ['LineArtist'] class LineArtist(PrimitiveArtist): """Artist for drawing lines. Parameters ---------- primitive : :class:`compas.geometry.Line` A COMPAS line. Notes ----- See :class:`compas_rhino.artists.PrimitiveArtist` for all other parameters. """ def draw(self): """Draw the line. Returns ------- list The GUIDs of the created Rhino objects. """ start = list(self.primitive.start) end = list(self.primitive.end) lines = [{'start': start, 'end': end, 'color': self.color, 'name': self.name}] guids = compas_rhino.draw_lines(lines, layer=self.layer, clear=False, redraw=False) self._guids = guids return guids @staticmethod def draw_collection(collection, names=None, colors=None, layer=None, clear=False, add_to_group=False, group_name=None): """Draw a collection of lines. Parameters ---------- collection: list of compas.geometry.Line A collection of ``Line`` objects. names : list of str, optional Individual names for the lines. colors : color or list of color, optional A color specification for the lines as a single color or a list of individual colors. layer : str, optional A layer path. clear : bool, optional Clear the layer before drawing. add_to_group : bool, optional Add the frames to a group. group_name : str, optional Name of the group. Returns ------- guids: list A list of GUIDs if the collection is not grouped. groupname: str The name of the group if the collection objects are grouped. """ lines = [{'start': list(line[0]), 'end': list(line[1])} for line in collection] if colors: if isinstance(colors[0], (int, float)): colors = iterable_like(collection, [colors], colors) else: colors = iterable_like(collection, colors, colors[0]) for line, rgb in zip(lines, colors): line['color'] = rgb if names: if isinstance(names, basestring): names = iterable_like(collection, [names], names) else: names = iterable_like(collection, names, names[0]) for line, name in zip(lines, names): line['name'] = name guids = compas_rhino.draw_lines(lines, layer=layer, clear=clear) if not add_to_group: return guids group = compas_rhino.rs.AddGroup(group_name) if group: compas_rhino.rs.AddObjectsToGroup(guids, group) return group # ============================================================================== # Main # ============================================================================== if __name__ == "__main__": pass
src/compas_rhino/artists/lineartist.py
3,307
Artist for drawing lines. Parameters ---------- primitive : :class:`compas.geometry.Line` A COMPAS line. Notes ----- See :class:`compas_rhino.artists.PrimitiveArtist` for all other parameters. Draw the line. Returns ------- list The GUIDs of the created Rhino objects. Draw a collection of lines. Parameters ---------- collection: list of compas.geometry.Line A collection of ``Line`` objects. names : list of str, optional Individual names for the lines. colors : color or list of color, optional A color specification for the lines as a single color or a list of individual colors. layer : str, optional A layer path. clear : bool, optional Clear the layer before drawing. add_to_group : bool, optional Add the frames to a group. group_name : str, optional Name of the group. Returns ------- guids: list A list of GUIDs if the collection is not grouped. groupname: str The name of the group if the collection objects are grouped. ============================================================================== Main ==============================================================================
1,145
en
0.478219
# Copyright (C) 2021 Xilinx, Inc # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # This file is used to configure your project. # Read more about the various options under: # http://setuptools.readthedocs.io/en/latest/setuptools.html#configuring-setup-using-setup-cfg-files # -*- coding: utf-8 -*- """ Setup file for logicnets. Use setup.cfg to configure your project. This file was generated with PyScaffold 3.2.3. PyScaffold helps you to put up the scaffold of your new Python project. Learn more under: https://pyscaffold.org/ """ import sys from pkg_resources import VersionConflict, require from setuptools import setup try: require('setuptools>=38.3') except VersionConflict: print("Error: version of setuptools is too old (<38.3)!") sys.exit(1) if __name__ == "__main__": setup(use_pyscaffold=True)
setup.py
1,357
Setup file for logicnets. Use setup.cfg to configure your project. This file was generated with PyScaffold 3.2.3. PyScaffold helps you to put up the scaffold of your new Python project. Learn more under: https://pyscaffold.org/ Copyright (C) 2021 Xilinx, Inc Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. This file is used to configure your project. Read more about the various options under: http://setuptools.readthedocs.io/en/latest/setuptools.htmlconfiguring-setup-using-setup-cfg-files -*- coding: utf-8 -*-
1,001
en
0.873573
# Copyright (C) 2017-2019 Apple Inc. All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions # are met: # 1. Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # 2. Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # THIS SOFTWARE IS PROVIDED BY APPLE INC. AND ITS CONTRIBUTORS ``AS IS'' AND ANY # EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR ITS CONTRIBUTORS BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON # ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. import socket class Device(object): def __init__(self, platform_device): self.platform_device = platform_device self.listening_socket = None def listening_port(self): if not self.listening_socket: return None return self.listening_socket.getsockname()[1] def install_app(self, app_path, env=None): return self.platform_device.install_app(app_path, env) def install_dylibs(self, path_to_dylibs): if hasattr(self.platform_device, 'install_dylibs'): return self.platform_device.install_dylibs(path_to_dylibs=path_to_dylibs) return True def launch_app(self, bundle_id, args, env=None): return self.platform_device.launch_app(bundle_id, args, env) def prepare_for_testing(self, ports_to_forward, test_app_bundle_id, layout_test_directory): if not self.listening_socket: self.listening_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) self.listening_socket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) self.listening_socket.bind(('127.0.0.1', 0)) if hasattr(self.platform_device, 'prepare_for_testing'): self.platform_device.prepare_for_testing( ports_to_forward=ports_to_forward + [self.listening_port()], test_app_bundle_id=test_app_bundle_id, layout_test_directory=layout_test_directory, ) def finished_testing(self): if hasattr(self.platform_device, 'finished_testing'): self.platform_device.finished_testing() self.listening_socket = None def symbolicate_crash_log_if_needed(self, path): if hasattr(self.platform_device, 'symbolicate_crash_log_if_needed'): return self.platform_device.symbolicate_crash_log_if_needed(path) return self.filesystem.read_text_file(path) def release_worker_resources(self): if hasattr(self.platform_device, 'release_worker_resources'): return self.platform_device.release_worker_resources() return None @property def executive(self): return self.platform_device.executive @property def filesystem(self): return self.platform_device.filesystem @property def user(self): return self.platform_device.user @property def platform(self): return self.platform_device.platform @property def workspace(self): return self.platform_device.workspace @property def udid(self): return self.platform_device.udid @property def device_type(self): return self.platform_device.device_type @property def build_version(self): return self.platform_device.build_version def __nonzero__(self): return self.platform_device is not None def __eq__(self, other): return self.udid == other.udid def __ne__(self, other): return not self.__eq__(other) def __repr__(self): return u'{}'.format(self.platform_device) def __hash__(self): return hash(self.udid)
Tools/Scripts/webkitpy/port/device.py
4,453
Copyright (C) 2017-2019 Apple Inc. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. THIS SOFTWARE IS PROVIDED BY APPLE INC. AND ITS CONTRIBUTORS ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR ITS CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
1,291
en
0.892219
import datetime from datetime import timedelta import csv from module.stock import Stock, Market # Historic Files DAX = 'DAX.csv' DOW = 'Dow.csv' FTSE = 'FTSE.csv' # Output File OUTPUT = 'output.txt' def import_csv(file_name): data = [] with open(file_name) as file: reader = csv.reader(file) data = list(reader) return(data) # Import historic stock performance Dax_data = import_csv(DAX) Dow_data = import_csv(DOW) Ftse_data = import_csv(FTSE) Ftse = Market('FTSE', Ftse_data) Ftse.p.sort(key=lambda x: x.d, reverse=False) # Sorts into date order vanguard = Stock('Vanguard', 1, 100000) # Name, price, units vanguard.sim(Ftse, datetime.date(2007, 1, 1), datetime.date(2012, 1, 1)) print(vanguard.v) ''' Simulation ''' '''for i in range(0,1000): Ftse = Market('Random', []) Ftse.p.sort(key=lambda x: x.d, reverse=False) vanguard = Stock('Vanguard', 1, 100000) # Name, price, units vanguard.sim(Ftse, datetime.date(2019, 2, 1), datetime.date(2022, 1, 1)) print(vanguard.v) if(vanguard.v > 100000): break '''
d.py
1,082
Historic Files Output File Import historic stock performance Sorts into date order Name, price, units
101
en
0.710626
# -*- coding: utf-8 -*- from __future__ import unicode_literals import flask from flask import Blueprint, request import flask_restx as restx # Add a dummy Resource to verify that the app is properly set. class HelloWorld(restx.Resource): def get(self): return {} class GoodbyeWorld(restx.Resource): def __init__(self, err): self.err = err def get(self): flask.abort(self.err) class APIWithBlueprintTest(object): def test_api_base(self, app): blueprint = Blueprint('test', __name__) api = restx.Api(blueprint) app.register_blueprint(blueprint) assert api.urls == {} assert api.prefix == '' assert api.default_mediatype == 'application/json' def test_api_delayed_initialization(self, app): blueprint = Blueprint('test', __name__) api = restx.Api() api.init_app(blueprint) app.register_blueprint(blueprint) api.add_resource(HelloWorld, '/', endpoint="hello") def test_add_resource_endpoint(self, app, mocker): blueprint = Blueprint('test', __name__) api = restx.Api(blueprint) view = mocker.Mock(**{'as_view.return_value.__name__': str('test_view')}) api.add_resource(view, '/foo', endpoint='bar') app.register_blueprint(blueprint) view.as_view.assert_called_with('bar', api) def test_add_resource_endpoint_after_registration(self, app, mocker): blueprint = Blueprint('test', __name__) api = restx.Api(blueprint) app.register_blueprint(blueprint) view = mocker.Mock(**{'as_view.return_value.__name__': str('test_view')}) api.add_resource(view, '/foo', endpoint='bar') view.as_view.assert_called_with('bar', api) def test_url_with_api_prefix(self, app): blueprint = Blueprint('test', __name__) api = restx.Api(blueprint, prefix='/api') api.add_resource(HelloWorld, '/hi', endpoint='hello') app.register_blueprint(blueprint) with app.test_request_context('/api/hi'): assert request.endpoint == 'test.hello' def test_url_with_blueprint_prefix(self, app): blueprint = Blueprint('test', __name__, url_prefix='/bp') api = restx.Api(blueprint) api.add_resource(HelloWorld, '/hi', endpoint='hello') app.register_blueprint(blueprint) with app.test_request_context('/bp/hi'): assert request.endpoint == 'test.hello' def test_url_with_registration_prefix(self, app): blueprint = Blueprint('test', __name__) api = restx.Api(blueprint) api.add_resource(HelloWorld, '/hi', endpoint='hello') app.register_blueprint(blueprint, url_prefix='/reg') with app.test_request_context('/reg/hi'): assert request.endpoint == 'test.hello' def test_registration_prefix_overrides_blueprint_prefix(self, app): blueprint = Blueprint('test', __name__, url_prefix='/bp') api = restx.Api(blueprint) api.add_resource(HelloWorld, '/hi', endpoint='hello') app.register_blueprint(blueprint, url_prefix='/reg') with app.test_request_context('/reg/hi'): assert request.endpoint == 'test.hello' def test_url_with_api_and_blueprint_prefix(self, app): blueprint = Blueprint('test', __name__, url_prefix='/bp') api = restx.Api(blueprint, prefix='/api') api.add_resource(HelloWorld, '/hi', endpoint='hello') app.register_blueprint(blueprint) with app.test_request_context('/bp/api/hi'): assert request.endpoint == 'test.hello' def test_error_routing(self, app, mocker): blueprint = Blueprint('test', __name__) api = restx.Api(blueprint) api.add_resource(HelloWorld, '/hi', endpoint="hello") api.add_resource(GoodbyeWorld(404), '/bye', endpoint="bye") app.register_blueprint(blueprint) with app.test_request_context('/hi', method='POST'): assert api._should_use_fr_error_handler() is True assert api._has_fr_route() is True with app.test_request_context('/bye'): api._should_use_fr_error_handler = mocker.Mock(return_value=False) assert api._has_fr_route() is True def test_non_blueprint_rest_error_routing(self, app, mocker): blueprint = Blueprint('test', __name__) api = restx.Api(blueprint) api.add_resource(HelloWorld, '/hi', endpoint="hello") api.add_resource(GoodbyeWorld(404), '/bye', endpoint="bye") app.register_blueprint(blueprint, url_prefix='/blueprint') api2 = restx.Api(app) api2.add_resource(HelloWorld(api), '/hi', endpoint="hello") api2.add_resource(GoodbyeWorld(404), '/bye', endpoint="bye") with app.test_request_context('/hi', method='POST'): assert api._should_use_fr_error_handler() is False assert api2._should_use_fr_error_handler() is True assert api._has_fr_route() is False assert api2._has_fr_route() is True with app.test_request_context('/blueprint/hi', method='POST'): assert api._should_use_fr_error_handler() is True assert api2._should_use_fr_error_handler() is False assert api._has_fr_route() is True assert api2._has_fr_route() is False api._should_use_fr_error_handler = mocker.Mock(return_value=False) api2._should_use_fr_error_handler = mocker.Mock(return_value=False) with app.test_request_context('/bye'): assert api._has_fr_route() is False assert api2._has_fr_route() is True with app.test_request_context('/blueprint/bye'): assert api._has_fr_route() is True assert api2._has_fr_route() is False def test_non_blueprint_non_rest_error_routing(self, app, mocker): blueprint = Blueprint('test', __name__) api = restx.Api(blueprint) api.add_resource(HelloWorld, '/hi', endpoint="hello") api.add_resource(GoodbyeWorld(404), '/bye', endpoint="bye") app.register_blueprint(blueprint, url_prefix='/blueprint') @app.route('/hi') def hi(): return 'hi' @app.route('/bye') def bye(): flask.abort(404) with app.test_request_context('/hi', method='POST'): assert api._should_use_fr_error_handler() is False assert api._has_fr_route() is False with app.test_request_context('/blueprint/hi', method='POST'): assert api._should_use_fr_error_handler() is True assert api._has_fr_route() is True api._should_use_fr_error_handler = mocker.Mock(return_value=False) with app.test_request_context('/bye'): assert api._has_fr_route() is False with app.test_request_context('/blueprint/bye'): assert api._has_fr_route() is True
tests/legacy/test_api_with_blueprint.py
6,919
-*- coding: utf-8 -*- Add a dummy Resource to verify that the app is properly set.
82
en
0.844923
# pyOCD debugger # Copyright (c) 2019 Arm Limited # SPDX-License-Identifier: Apache-2.0 # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from collections import namedtuple from . import target_kinetis FamilyInfo = namedtuple("FamilyInfo", "vendor matches klass") ## @brief Lookup table to convert from CMSIS-Pack family names to a family class. # # The vendor name must be an exact match to the 'Dvendor' attribute of the CMSIS-Pack family # element. # # At least one of the regexes must match the entirety of either the CMSIS-Pack 'Dfamily' or # 'DsubFamily' (if present) attributes. FAMILIES = [ FamilyInfo("NXP", [re.compile(r'MK[LEVWS]?.*')], target_kinetis.Kinetis), ]
pyocd/target/family/__init__.py
1,208
pyOCD debugger Copyright (c) 2019 Arm Limited SPDX-License-Identifier: Apache-2.0 Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. @brief Lookup table to convert from CMSIS-Pack family names to a family class. The vendor name must be an exact match to the 'Dvendor' attribute of the CMSIS-Pack family element. At least one of the regexes must match the entirety of either the CMSIS-Pack 'Dfamily' or 'DsubFamily' (if present) attributes.
910
en
0.8254
import argparse import RDT import time import rdt_3_0 def makePigLatin(word): m = len(word) vowels = "a", "e", "i", "o", "u", "y" if m < 3 or word == "the": return word else: for i in vowels: if word.find(i) < m and word.find(i) != -1: m = word.find(i) if m == 0: return word + "way" else: return word[m:] + word[:m] + "ay" def piglatinize(message): essagemay = "" message = message.strip(".") for word in message.split(' '): essagemay += " " + makePigLatin(word) return essagemay.strip() + "." if __name__ == '__main__': parser = argparse.ArgumentParser(description='Pig Latin conversion server.') parser.add_argument('port', help='Port.', type=int) args = parser.parse_args() timeout = 5 # close connection if no new data within 5 seconds time_of_last_data = time.time() rdt = rdt_3_0.RDT('server', None, args.port) while (True): # try to receiver message before timeout # msg_S = rdt.rdt_2_1_receive() msg_S = rdt.rdt_3_0_receive() if msg_S is None: if time_of_last_data + timeout < time.time(): break else: continue time_of_last_data = time.time() # convert and reply rep_msg_S = piglatinize(msg_S) print('Converted %s \nto \n%s\n' % (msg_S, rep_msg_S)) # rdt.rdt_2_1_send(rep_msg_S) rdt.rdt_3_0_send(rep_msg_S) rdt.disconnect()
submission_files/server_3_0.py
1,535
close connection if no new data within 5 seconds try to receiver message before timeout msg_S = rdt.rdt_2_1_receive() convert and reply rdt.rdt_2_1_send(rep_msg_S)
163
en
0.599335
# coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. # # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is # regenerated. # -------------------------------------------------------------------------- from .catalog_item import CatalogItem class USqlExternalDataSource(CatalogItem): """A Data Lake Analytics catalog U-SQL external datasource item. :param compute_account_name: the name of the Data Lake Analytics account. :type compute_account_name: str :param version: the version of the catalog item. :type version: str :param database_name: the name of the database. :type database_name: str :param name: the name of the external data source. :type name: str :param provider: the name of the provider for the external data source. :type provider: str :param provider_string: the name of the provider string for the external data source. :type provider_string: str :param pushdown_types: the list of types to push down from the external data source. :type pushdown_types: list of str """ _attribute_map = { 'compute_account_name': {'key': 'computeAccountName', 'type': 'str'}, 'version': {'key': 'version', 'type': 'str'}, 'database_name': {'key': 'databaseName', 'type': 'str'}, 'name': {'key': 'externalDataSourceName', 'type': 'str'}, 'provider': {'key': 'provider', 'type': 'str'}, 'provider_string': {'key': 'providerString', 'type': 'str'}, 'pushdown_types': {'key': 'pushdownTypes', 'type': '[str]'}, } def __init__(self, compute_account_name=None, version=None, database_name=None, name=None, provider=None, provider_string=None, pushdown_types=None): super(USqlExternalDataSource, self).__init__(compute_account_name=compute_account_name, version=version) self.database_name = database_name self.name = name self.provider = provider self.provider_string = provider_string self.pushdown_types = pushdown_types
venv/lib/python3.8/site-packages/azure/mgmt/datalake/analytics/catalog/models/usql_external_data_source.py
2,339
A Data Lake Analytics catalog U-SQL external datasource item. :param compute_account_name: the name of the Data Lake Analytics account. :type compute_account_name: str :param version: the version of the catalog item. :type version: str :param database_name: the name of the database. :type database_name: str :param name: the name of the external data source. :type name: str :param provider: the name of the provider for the external data source. :type provider: str :param provider_string: the name of the provider string for the external data source. :type provider_string: str :param pushdown_types: the list of types to push down from the external data source. :type pushdown_types: list of str coding=utf-8 -------------------------------------------------------------------------- Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT License. See License.txt in the project root for license information. Code generated by Microsoft (R) AutoRest Code Generator. Changes may cause incorrect behavior and will be lost if the code is regenerated. --------------------------------------------------------------------------
1,157
en
0.558444
# This module just exists as a shortcut for running the main Kivy app if __name__ == '__main__': from naturtag.app.app import NaturtagApp NaturtagApp().run()
naturtag/ui.py
167
This module just exists as a shortcut for running the main Kivy app
67
en
0.809556
#!C:\Users\John\PycharmProjects\arcadePython\venv\Scripts\python.exe # EASY-INSTALL-ENTRY-SCRIPT: 'pip==10.0.1','console_scripts','pip3.6' __requires__ = 'pip==10.0.1' import re import sys from pkg_resources import load_entry_point if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit( load_entry_point('pip==10.0.1', 'console_scripts', 'pip3.6')() )
venv/Scripts/pip3.6-script.py
421
!C:\Users\John\PycharmProjects\arcadePython\venv\Scripts\python.exe EASY-INSTALL-ENTRY-SCRIPT: 'pip==10.0.1','console_scripts','pip3.6'
135
en
0.371472
from django.db import models from osc_bge.users import models as user_models # Create your models here. class TimeStampedModel(models.Model): created_at = models.DateTimeField(auto_now_add=True, null=True) updated_at = models.DateTimeField(auto_now=True, null=True) class Meta: abstract = True #Agent Head Table class AgencyHead(TimeStampedModel): PROGRAM_CHOICES = ( ('secondary', 'Secondary'), ('college', 'College'), ('camp', 'Camp'), ) name = models.CharField(max_length=80, null=True, blank=True) location = models.CharField(max_length=140, null=True, blank=True) number_branches = models.CharField(max_length=80, null=True, blank=True) capacity_students = models.CharField(max_length=255, null=True, blank=True) commission = models.CharField(max_length=140, null=True, blank=True) promotion = models.CharField(max_length=255, null=True, blank=True) others = models.CharField(max_length=255, null=True, blank=True) comment = models.TextField(null=True, blank=True) def __str__(self): return "{}".format(self.name) class AgencyProgram(TimeStampedModel): head = models.ForeignKey(AgencyHead, on_delete=models.CASCADE, null=True) program = models.CharField(max_length=80, null=True, blank=True) #Agent Branch Table class Agency(TimeStampedModel): head = models.ForeignKey(AgencyHead, on_delete=models.CASCADE, null=True, related_name='agent_branch') name = models.CharField(max_length=140, null=True, blank=True) location = models.CharField(max_length=140, null=True, blank=True) capacity_students = models.CharField(max_length=255, null=True, blank=True) commission = models.CharField(max_length=140, null=True, blank=True) promotion = models.CharField(max_length=255, null=True, blank=True) others = models.CharField(max_length=255, null=True, blank=True) comment = models.TextField(null=True, blank=True) def __str__(self): return "{}".format(self.name) class AgencyBranchProgram(TimeStampedModel): branch = models.ForeignKey(Agency, on_delete=models.CASCADE, null=True) program = models.CharField(max_length=80, null=True, blank=True) def set_filename_format(now, instance, filename): return "{schoolname}-{microsecond}".format( agentname=instance.agency, microsecond=now.microsecond, ) def agent_directory_path(instance, filename): now = datetime.datetime.now() path = "agents/{agentname}/{filename}".format( agentname=instance.agency, filename=set_filename_format(now, instance, filename), ) return path class AgencyHeadContactInfo(TimeStampedModel): LEVEL_CHOICES = ( ('s', 'S'), ('a', 'A'), ('b', 'B'), ('c', 'C'), ('d', 'D'), ) agent = models.ForeignKey(AgencyHead, on_delete=models.CASCADE, null=True) name = models.CharField(max_length=80, null=True, blank=True) contracted_date = models.DateTimeField(auto_now=True, null=True) phone = models.CharField(max_length=80, null=True, blank=True) email = models.CharField(max_length=140, null=True, blank=True) skype = models.CharField(max_length=80, null=True, blank=True) wechat = models.CharField(max_length=80, null=True, blank=True) location = models.CharField(max_length=140, null=True, blank=True) level = models.CharField(max_length=80, null=True, blank=True) image = models.ImageField(upload_to=agent_directory_path, null=True, blank=True) def __str__(self): return "{}".format(self.name) class AgentRelationshipHistory(TimeStampedModel): head = models.ForeignKey(AgencyHead, on_delete=models.CASCADE, null=True) writer = models.CharField(max_length=80, null=True, blank=True) name = models.CharField(max_length=80, null=True, blank=True) date = models.DateField(null=True, blank=True) location = models.CharField(max_length=140, null=True, blank=True) category = models.CharField(max_length=80, null=True, blank=True) priority = models.IntegerField(null=True, blank=True) comment = models.TextField(null=True, blank=True) class SecodnaryProgram(TimeStampedModel): agent = models.ForeignKey(AgencyHead, on_delete=models.CASCADE, null=True) preriod = models.CharField(max_length=80, null=True, blank=True) target = models.IntegerField(null=True, blank=True) new_students_fall = models.IntegerField(null=True, blank=True) new_students_spring = models.IntegerField(null=True, blank=True) total_new_students_bge = models.IntegerField(null=True, blank=True) total_students_bge = models.IntegerField(null=True, blank=True) terminating_students = models.IntegerField(null=True, blank=True) comments = models.TextField(null=True, blank=True) class Camp(TimeStampedModel): agent = models.ForeignKey(AgencyHead, on_delete=models.CASCADE, null=True) preriod = models.CharField(max_length=80, null=True, blank=True) target = models.IntegerField(null=True, blank=True) summer_camp = models.IntegerField(null=True, blank=True) winter_camp = models.IntegerField(null=True, blank=True) comments = models.TextField(null=True, blank=True) class CollegeApplication(TimeStampedModel): agent = models.ForeignKey(AgencyHead, on_delete=models.CASCADE, null=True) preriod = models.CharField(max_length=80, null=True, blank=True) college_application = models.IntegerField(null=True, blank=True) other_program = models.IntegerField(null=True, blank=True) comments = models.TextField(null=True, blank=True)
osc_bge/agent/models.py
5,611
Create your models here.Agent Head TableAgent Branch Table
58
en
0.56847
''' First Version: Tweets contents from subreddits ''' from time import sleep from reddit import Reddit from twitter import Twitter def main(): ''' connects the pieces to grab posts from reddit and throw them on twitter ''' reddit = Reddit() twitter = Twitter() tweets = reddit.get_tweets() print("sending {} tweets".format(len(tweets))) for tweet in reddit.get_tweets(): status = twitter.send_tweet(tweet.Primary) if tweet.Second: twitter.send_tweet(tweet.Second, status.id) sleep(90) if __name__ == '__main__': main()
tweet_bot.py
596
connects the pieces to grab posts from reddit and throw them on twitter First Version: Tweets contents from subreddits
118
en
0.892743
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright © 2017 Matthew Stone <mstone5@mgh.harvard.edu> # Distributed under terms of the MIT license. """ Calculate enrichment of clipped reads or discordant pairs at SV breakpoints. """ import argparse import sys import pysam import pandas as pd from svtk.pesr import SRTestRunner, PETestRunner, PETest, SRTest def sr_test(argv): parser = argparse.ArgumentParser( description="Calculate enrichment of clipped reads at SV breakpoints.", prog='svtk sr-test', formatter_class=argparse.RawDescriptionHelpFormatter) parser.add_argument('vcf', help='VCF of variant calls. Standardized to include ' 'CHR2, END, SVTYPE, STRANDS in INFO.') parser.add_argument('countfile', help='Tabix indexed file of split counts.' ' Columns: chrom,pos,clip,count,sample') parser.add_argument('fout', help='Output table of most significant start/end' 'positions.') parser.add_argument('-w', '--window', type=int, default=100, help='Window around variant start/end to consider for ' 'split read support. [100]') parser.add_argument('--common', default=False, action='store_true', help='Ignore background for common AF') parser.add_argument('-b', '--background', type=int, default=160, help='Number of background samples to choose for ' 'comparison in t-test. [160]') parser.add_argument('-s', '--samples', type=argparse.FileType('r'), default=None, help='Whitelist of samples to restrict testing to.') parser.add_argument('--index', default=None, help='Tabix index of discordant pair file. Required if ' 'discordant pair file is hosted remotely.') # TODO: add normalization parser.add_argument('--medianfile', default=None, help='Median coverage statistics for each library ' '(optional). If provided, each sample\'s split ' 'counts will be normalized accordingly. ' 'Same format as RdTest, one column per sample.') parser.add_argument('--log', action='store_true', default=False, help='Print progress log to stderr.') # Print help if no arguments specified if len(argv) == 0: parser.print_help() sys.exit(1) args = parser.parse_args(argv) vcf = pysam.VariantFile(args.vcf) if args.index is not None: countfile = pysam.TabixFile(args.countfile, index=args.index, parser=pysam.asTuple()) else: if args.countfile.startswith('http'): raise Exception('Must provide tabix index with remote URL') countfile = pysam.TabixFile(args.countfile, parser=pysam.asTuple()) if args.fout in '- stdout'.split(): fout = sys.stdout else: fout = open(args.fout, 'w') header = 'name coord pos log_pval called_median bg_median bg_frac'.split() fout.write('\t'.join(header) + '\n') if args.samples is not None: whitelist = [s.strip() for s in args.samples.readlines()] else: whitelist = None if args.medianfile is not None: medians = pd.read_table(args.medianfile) medians = pd.melt(medians, var_name='sample', value_name='median_cov') else: medians = None runner = SRTestRunner(vcf, countfile, fout, args.background, args.common, args.window, whitelist, medians=medians, log=args.log) runner.run() def pe_test(argv): parser = argparse.ArgumentParser( description="Calculate enrichment of discordant pairs at SV breakpoints.", prog='svtk pe-test', formatter_class=argparse.RawDescriptionHelpFormatter) parser.add_argument('vcf', help='Variants.') parser.add_argument('disc', help='Table of discordant pair coordinates.') parser.add_argument('fout', type=argparse.FileType('w'), help='Output table of PE counts.') parser.add_argument('-o', '--window-out', type=int, default=500, help='Window outside breakpoint to query for ' 'discordant pairs. [500]') parser.add_argument('-i', '--window-in', type=int, default=50, help='Window inside breakpoint to query for ' 'discordant pairs. [50]') parser.add_argument('-b', '--background', type=int, default=160, help='Number of background samples to sample for PE ' 'evidence. [160]') parser.add_argument('--common', default=False, action='store_true', help='Ignore background for common AF') parser.add_argument('-s', '--samples', type=argparse.FileType('r'), default=None, help='Whitelist of samples to restrict testing to.') parser.add_argument('--index', default=None, help='Tabix index of discordant pair file. Required if ' 'discordant pair file is hosted remotely.') parser.add_argument('--medianfile', default=None, help='Median coverage statistics for each library ' '(optional). If provided, each sample\'s split ' 'counts will be normalized accordingly. ' 'Same format as RdTest, one column per sample.') parser.add_argument('--log', action='store_true', default=False, help='Print progress log to stderr.') if len(argv) == 0: parser.print_help() sys.exit(1) args = parser.parse_args(argv) if args.vcf in '- stdin'.split(): vcf = pysam.VariantFile(sys.stdin) else: vcf = pysam.VariantFile(args.vcf) if args.fout in '- stdout'.split(): fout = sys.stdout else: fout = args.fout header = 'name log_pval called_median bg_median bg_frac'.split() args.fout.write('\t'.join(header) + '\n') if args.samples is not None: whitelist = [s.strip() for s in args.samples.readlines()] else: whitelist = None if args.index is not None: discfile = pysam.TabixFile(args.disc, index=args.index) else: if args.disc.startswith('http'): raise Exception('Must provide tabix index with remote URL') discfile = pysam.TabixFile(args.disc) if args.medianfile is not None: medians = pd.read_table(args.medianfile) medians = pd.melt(medians, var_name='sample', value_name='median_cov') else: medians = None runner = PETestRunner(vcf, discfile, fout, args.background, args.common, args.window_in, args.window_out, whitelist, medians=medians, log=args.log) runner.run() def count_pe(argv): parser = argparse.ArgumentParser( description="Count discordant pairs supporting a SV breakpoints.", prog='svtk count-pe', formatter_class=argparse.RawDescriptionHelpFormatter) parser.add_argument('vcf', help='Variants.') parser.add_argument('disc', help='Table of discordant pair coordinates.') parser.add_argument('fout', type=argparse.FileType('w'), help='Output table of PE counts.') parser.add_argument('-o', '--window-out', type=int, default=500, help='Window outside breakpoint to query for ' 'discordant pairs. [500]') parser.add_argument('-i', '--window-in', type=int, default=50, help='Window inside breakpoint to query for ' 'discordant pairs. [50]') parser.add_argument('--common', default=False, action='store_true', help='Ignore background for common AF') parser.add_argument('-s', '--samples', type=argparse.FileType('r'), default=None, help='Whitelist of samples to restrict testing to.') parser.add_argument('--index', default=None, help='Tabix index of discordant pair file. Required if ' 'discordant pair file is hosted remotely.') parser.add_argument('--medianfile', default=None, help='Median coverage statistics for each library ' '(optional). If provided, each sample\'s split ' 'counts will be normalized accordingly. ' 'Same format as RdTest, one column per sample.') if len(argv) == 0: parser.print_help() sys.exit(1) args = parser.parse_args(argv) if args.vcf in '- stdin'.split(): vcf = pysam.VariantFile(sys.stdin) else: vcf = pysam.VariantFile(args.vcf) if args.fout in '- stdout'.split(): fout = sys.stdout else: fout = args.fout header = 'name sample count'.split() args.fout.write('\t'.join(header) + '\n') if args.samples is not None: whitelist = [s.strip() for s in args.samples.readlines()] else: whitelist = [s for s in vcf.header.samples] if args.index is not None: discfile = pysam.TabixFile(args.disc, index=args.index) else: if args.disc.startswith('http'): raise Exception('Must provide tabix index with remote URL') discfile = pysam.TabixFile(args.disc) if args.medianfile is not None: medians = pd.read_table(args.medianfile) medians = pd.melt(medians, var_name='sample', value_name='median_cov') else: medians = None petest = PETest(discfile, args.common, args.window_in, args.window_out, medians=medians) for record in vcf: counts = petest.load_counts(record, args.window_in, args.window_out) counts = petest.normalize_counts(counts) counts = counts.set_index('sample') counts = counts.reindex(whitelist).fillna(0).astype(int) counts = counts.reset_index() counts['name'] = record.id cols = 'name sample count'.split() for row in counts[cols].as_matrix(): fout.write('\t'.join([str(x) for x in row]) + '\n') # counts[cols].to_csv(fout, header=False, index=False, sep='\t', na_rep='NA') def count_sr(argv): parser = argparse.ArgumentParser( description="Count clipped reads at SV breakpoints. Unwindowed.", prog='svtk count-sr', formatter_class=argparse.RawDescriptionHelpFormatter) parser.add_argument('vcf', help='VCF of variant calls. Standardized to include ' 'CHR2, END, SVTYPE, STRANDS in INFO.') parser.add_argument('countfile', help='Tabix indexed file of split counts.' ' Columns: chrom,pos,clip,count,sample') parser.add_argument('fout', help='Output table of split read counts.') parser.add_argument('--common', default=False, action='store_true', help='Ignore background for common AF') parser.add_argument('-s', '--samples', type=argparse.FileType('r'), default=None, help='Whitelist of samples to restrict testing to.') parser.add_argument('--index', default=None, help='Tabix index of discordant pair file. Required if ' 'discordant pair file is hosted remotely.') # TODO: add normalization parser.add_argument('--medianfile', default=None, help='Median coverage statistics for each library ' '(optional). If provided, each sample\'s split ' 'counts will be normalized accordingly. ' 'Same format as RdTest, one column per sample.') # Print help if no arguments specified if len(argv) == 0: parser.print_help() sys.exit(1) args = parser.parse_args(argv) vcf = pysam.VariantFile(args.vcf) if args.index is not None: countfile = pysam.TabixFile(args.countfile, index=args.index, parser=pysam.asTuple()) else: if args.countfile.startswith('http'): raise Exception('Must provide tabix index with remote URL') countfile = pysam.TabixFile(args.countfile, parser=pysam.asTuple()) if args.fout in '- stdout'.split(): fout = sys.stdout else: fout = open(args.fout, 'w') header = 'name coord sample count'.split() fout.write('\t'.join(header) + '\n') if args.samples is not None: whitelist = [s.strip() for s in args.samples.readlines()] else: whitelist = [s for s in vcf.header.samples] if args.medianfile is not None: medians = pd.read_table(args.medianfile) medians = pd.melt(medians, var_name='sample', value_name='median_cov') else: medians = None srtest = SRTest(countfile, args.common, window=0, medians=medians) for record in vcf: for coord in 'start end'.split(): if coord == 'start': pos, strand, chrom = record.pos, record.info['STRANDS'][0], record.chrom else: # TODO: With a properly formatted VCF, should be using END2 instead of END here pos, strand, chrom = record.stop, record.info['STRANDS'][1], record.info['CHR2'] counts = srtest.load_counts(chrom, pos, strand) counts = srtest.normalize_counts(counts) counts = counts['sample count'.split()] counts = counts.set_index('sample') counts = counts.reindex(whitelist).fillna(0).astype(int) counts = counts.reset_index() counts['name'] = record.id counts['coord'] = coord for row in counts[header].values: fout.write('\t'.join([str(x) for x in row]) + '\n') # counts[header].to_csv(fout, header=False, index=False, sep='\t', na_rep='NA')
src/svtk/svtk/cli/pesr_test.py
14,259
Calculate enrichment of clipped reads or discordant pairs at SV breakpoints. !/usr/bin/env python -*- coding: utf-8 -*- Copyright © 2017 Matthew Stone <mstone5@mgh.harvard.edu> Distributed under terms of the MIT license. TODO: add normalization Print help if no arguments specified counts[cols].to_csv(fout, header=False, index=False, sep='\t', na_rep='NA') TODO: add normalization Print help if no arguments specified TODO: With a properly formatted VCF, should be using END2 instead of END here counts[header].to_csv(fout, header=False, index=False, sep='\t', na_rep='NA')
577
en
0.295601
#!/usr/bin/python3 def lowestPositiveInt(A): A.sort() index = -1 try: index = A.index(1) except ValueError: return 1 # Remove all negatives and zero if 1 is found # if index >= 0: A = A[index:] length = len(A) i = 1 found = 0 while i < length and found == 0: if A[i] > (A[i-1]+1): found = A[i-1]+1 i = i+1 if found > 0: return found else: return A[length-1]+1 def main(): sample1 = [-1, -3, -47, -10, 0, 27, 27, 0, 1, 1] sample2 = [0, 1, 2, 3, 7, 25, -12, -14, -17, 4, 4] sample3 = [1, 1, 5, 27, -100, -234, 3, 2, 3, 4, 4] print("Sample 1: {}".format(lowestPositiveInt(sample1))) print("Sample 2: {}".format(lowestPositiveInt(sample2))) print("Sample 3: {}".format(lowestPositiveInt(sample3))) main()
codingTestIterative.py
891
!/usr/bin/python3 Remove all negatives and zero if 1 is found if index >= 0:
76
en
0.526975
from keras_vggface.utils import preprocess_input from keras_vggface.vggface import VGGFace import numpy as np import pickle from sklearn.metrics.pairwise import cosine_similarity import cv2 from mtcnn import MTCNN from PIL import Image feature_list = np.array(pickle.load(open('artifacts/extracted_features/embedding.pkl','rb'))) filenames = pickle.load(open('artifacts/pickle_format_data/img_PICKLE_file.pkl','rb')) model = VGGFace(model='resnet50',include_top=False,input_shape=(224,224,3),pooling='avg') #detect face detector = MTCNN() # load img -> face detection sample_img = cv2.imread('samples/saif_dup.jpg') results = detector.detect_faces(sample_img) x,y,width,height = results[0]['box'] face = sample_img[y:y+height,x:x+width] # extract its features image = Image.fromarray(face) image = image.resize((224,224)) face_array = np.asarray(image) face_array = face_array.astype('float32') expanded_img = np.expand_dims(face_array,axis=0) preprocessed_img = preprocess_input(expanded_img) result = model.predict(preprocessed_img).flatten() # print(result) # print(result.shape) # print(result.reshape(1,-1)) # find the cosine distance of current image with all the 8664 features similarity = [] for i in range(len(feature_list)): similarity.append(cosine_similarity(result.reshape(1,-1),feature_list[i].reshape(1,-1))[0][0]) # print(len(similarity)) index_pos = sorted(list(enumerate(similarity)),reverse=True,key=lambda x:x[1])[0][0] #recommend that image temp_img = cv2.imread(filenames[index_pos]) cv2.imshow('output',temp_img) cv2.waitKey(0)
src/testing.py
1,571
detect face load img -> face detection extract its features print(result) print(result.shape) print(result.reshape(1,-1)) find the cosine distance of current image with all the 8664 features print(len(similarity))recommend that image
234
en
0.80239
#!/usr/bin/env python # GenerateJSONOutput.py # # Command-line interface for turning output root files into JSON files for further processing. # # By: Larry Lee - Dec 2017 import argparse import sys import os import ROOT ROOT.gSystem.Load(f"{os.getenv('HISTFITTER')}/lib/libSusyFitter.so") ROOT.gROOT.SetBatch() parser = argparse.ArgumentParser() parser.add_argument( "--inputFiles", "-i", type=str, nargs="+", help="input ROOT files -- if you give me just the nominal, I'll try to find the theory variations and the upper limit automatically", required=True, ) parser.add_argument( "--format", "-f", type=str, help="format of object names", default="hypo_SU_%f_%f_0_10", ) parser.add_argument( "--interpretation", "-p", type=str, help="interpretation of object name", default="m0:m12", ) parser.add_argument( "--addCoordinates", "-a", type=str, help="add additional coordinates to json using a python dictionary {existing json value name and math : new value name}", default='{"m0":"x","m12":"y"}', ) parser.add_argument("--cut", "-c", type=str, help="cut string", default="1") parser.add_argument( "--noAddTabs", "-n", help="don't convert JSON to human readable file", action="store_true", default=False, ) args = parser.parse_args() # Print out the settings for arg in vars(args): user_input = getattr(args, arg) print(f">>> ... Setting: {arg: >20} {str(user_input): >40}") print("") def main(): for filename in args.inputFiles: processFile(filename) if args.addCoordinates != "": addCoordinates(filename, args.addCoordinates) if "Nominal" in filename: print(">>> Attempting to find theory variation files") try: newfilename = filename.replace("Nominal", "Up") if newfilename == filename: raise processFile(newfilename) if args.addCoordinates != "": addCoordinates(newfilename, args.addCoordinates) except: print( ">>> WARNING: Can't find file: %s" % filename.replace("Nominal", "Up") ) try: newfilename = filename.replace("Nominal", "Down") if newfilename == filename: raise processFile(newfilename) if args.addCoordinates != "": addCoordinates(newfilename, args.addCoordinates) except: print( ">>> WARNING: Can't find file: %s" % filename.replace("Nominal", "Down") ) try: newfilename = filename.replace( "_fixSigXSecNominal_hypotest", "_upperlimit" ) if newfilename == filename: raise processFile(newfilename) if args.addCoordinates != "": addCoordinates(newfilename, args.addCoordinates) except: print( ">>> WARNING: Can't find file: %s" % filename.replace("_fixSigXSecNominal_hypotest", "_upperlimit") ) try: newfilename = filename.replace("_Nominal", "_upperlimit") if newfilename == filename: raise processFile(newfilename) if args.addCoordinates != "": addCoordinates(newfilename, args.addCoordinates) except: print( ">>> WARNING: Can't find file: %s" % filename.replace("_fixSigXSecNominal_hypotest", "_upperlimit") ) if not args.noAddTabs: cleanUpJSON() print(">>>") print(">>> Done!") print(">>>") return def processFile(file): print("") if os.path.isfile(file): ROOT.CollectAndWriteHypoTestResults( file, args.format, args.interpretation, args.cut ) else: print(">>> ERROR: File does not exist: %s" % file) sys.exit(1) print("") return def cleanUpJSON(): import json import glob for file in glob.glob("./*json"): print(">>> Making file human readable: %s" % file) data = json.load(open(file)) with open(file, "w") as f: f.write(json.dumps(data, indent=4)) return def addCoordinates(fileName, coordString): import json import re coordDict = json.loads(coordString) jsonFileName = fileName.split("/")[-1] # grab just the filename jsonFileName = jsonFileName.replace(".root", "__1_harvest_list.json") data = json.load(open(jsonFileName)) for i, hypo_test in enumerate(data): # an entry is one hypo test result for key in coordDict: # each item of the result # parse input arguments, thanks to Larry for regex suggestions total = eval(re.sub(r"\b([a-zA-Z]+[0-9]*)\b", r'hypo_test["\g<1>"]', key)) # assign new key to value hypo_test[coordDict[key]] = total with open(jsonFileName, "w") as f: f.write(json.dumps(data)) if __name__ == "__main__": main()
scripts/GenerateJSONOutput.py
5,344
!/usr/bin/env python GenerateJSONOutput.py Command-line interface for turning output root files into JSON files for further processing. By: Larry Lee - Dec 2017 Print out the settings grab just the filename an entry is one hypo test result each item of the result parse input arguments, thanks to Larry for regex suggestions assign new key to value
348
en
0.646568
import csv import logging import random import re from collections import OrderedDict from enum import IntEnum from pathlib import Path from typing import Optional, Tuple, List import attr from torch.utils.data import Dataset from syntok import segmenter from sklearn.model_selection import train_test_split logger = logging.getLogger(__name__) DATA_DIR = Path(__file__).parent / 'data' PRONOUNS = {'she', 'her', 'hers', 'he', 'him', 'his'} PRONOUNS_GENDER = {'she': 'F', 'her': 'F', 'hers': 'F', 'he': 'M', 'him': 'M', 'his': 'M'} class GAPLabel(IntEnum): A, B, NEITHER = 0, 1, 2 @attr.s(auto_attribs=True) class GAPExample(object): id: str url: str tokens: List[str] pronoun_index: int a_start: int a_end: int # exclusive b_start: int b_end: int # exclusive label: Optional[GAPLabel] def load_train_val_examples(random_seed, train_size=0.9) -> Tuple[List[GAPExample], List[GAPExample]]: examples = [] for tsv_file in ('gap-development.tsv', 'gap-validation.tsv', 'gap-test.tsv'): examples.extend(_load_gap(DATA_DIR / tsv_file)) examples_gender = [PRONOUNS_GENDER[e.tokens[e.pronoun_index].lower()] for e in examples] train_examples, val_examples = train_test_split( examples, random_state=random_seed, train_size=train_size, shuffle=True, stratify=examples_gender) return train_examples, val_examples def load_test_examples(tsv_path: Path = DATA_DIR / 'test_stage_2.tsv') -> List[GAPExample]: examples = _load_gap(tsv_path) return examples def _load_gap(tsv_path: Path) -> List[GAPExample]: examples: List[GAPExample] = [] with tsv_path.open() as f: reader = csv.DictReader(f, delimiter='\t') for row in reader: examples.append(_create_example(row)) logger.info('Loaded %d examples from %s', len(examples), tsv_path) return examples def _create_example(row: OrderedDict): label = None a_coref, b_coref = map(lambda x: row.get(f'{x}-coref', '').upper(), 'AB') if a_coref == 'TRUE' and b_coref == 'FALSE': label = GAPLabel.A elif b_coref == 'TRUE' and a_coref == 'FALSE': label = GAPLabel.B elif a_coref == 'FALSE' and b_coref == 'FALSE': label = GAPLabel.NEITHER tokens = _word_tokenizer(row['Text']) pronoun_index = _char_to_token_offset( row['Text'], row['Pronoun'], int(row['Pronoun-offset']), tokens) assert tokens[pronoun_index].lower() in PRONOUNS a_start = _char_to_token_offset( row['Text'], row['A'], int(row['A-offset']), tokens) a_end = a_start + len(_word_tokenizer(row['A'])) # exclusive b_start = _char_to_token_offset( row['Text'], row['B'], int(row['B-offset']), tokens) b_end = b_start + len(_word_tokenizer(row['B'])) # exclusive example = GAPExample( id=row['ID'], url=row['URL'], tokens=tokens, pronoun_index=pronoun_index, a_start=a_start, a_end=a_end, b_start=b_start, b_end=b_end, label=label) return example def _word_tokenizer(text: str) -> List[str]: tokens: List[str] = [] for paragraph in segmenter.analyze(text): for sentence in paragraph: for token in sentence: # Split tokens on additional characters not handled by syntok token_value = token.value for c in ('/', r'\*', "'", r'\.', '--', ':'): token_value = re.sub(rf'({c})', r' \1 ', token_value) tokens.extend(token_value.split()) return tokens def _char_to_token_offset( text: str, mention: str, char_offset: int, text_tokens: List[str]) -> int: char_index = token_index = 0 while char_index < char_offset: if text[char_index:].startswith(text_tokens[token_index]): char_index += len(text_tokens[token_index]) token_index += 1 else: char_index += 1 # whitespace return token_index class GAPDataset(Dataset): def __init__(self, examples: List[GAPExample], flip_prob: float = 0.0) -> None: super().__init__() self._examples = examples assert 0.0 <= flip_prob <= 1.0 self._flip_prob = flip_prob def __getitem__(self, index: int) -> GAPExample: example = self._examples[index] if (self._flip_prob == 1.0 or (self._flip_prob > 0.0 and random.random() <= self._flip_prob)): example = self._flip_example(example) return example def __len__(self) -> int: return len(self._examples) def _flip_example(self, example: GAPExample) -> GAPExample: new_label = example.label if example.label == GAPLabel.A: new_label = GAPLabel.B elif example.label == GAPLabel.B: new_label = GAPLabel.A new_example = GAPExample( id=example.id, url=example.url, tokens=example.tokens, pronoun_index=example.pronoun_index, a_start=example.b_start, a_end=example.b_end, b_start=example.a_start, b_end=example.a_end, label=new_label) return new_example
gendered-pronoun-resolution/data.py
5,257
exclusive exclusive exclusive exclusive Split tokens on additional characters not handled by syntok whitespace
110
en
0.751839
# Generated by Django 2.2.24 on 2021-12-13 07:54 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ("taskflow3", "0017_auto_20211116_1526"), ] operations = [ migrations.AlterModelOptions( name="autoretrynodestrategy", options={ "verbose_name": "节点自动重试策略 AutoRetryNodeStrategy", "verbose_name_plural": "节点自动重试策略 AutoRetryNodeStrategy", }, ), migrations.CreateModel( name="TimeoutNodeConfig", fields=[ ("task_id", models.BigIntegerField(verbose_name="taskflow id")), ("root_pipeline_id", models.CharField(max_length=64, verbose_name="root pipeline id")), ( "action", models.CharField( choices=[("forced_fail", "强制失败"), ("forced_fail_and_skip", "强制失败并跳过")], max_length=32, verbose_name="action", ), ), ( "node_id", models.CharField(max_length=64, primary_key=True, serialize=False, verbose_name="task node id"), ), ("timeout", models.IntegerField(verbose_name="node timeout time")), ], options={ "verbose_name": "节点超时配置 TimeoutNodeConfig", "verbose_name_plural": "节点超时配置 TimeoutNodeConfig", "index_together": {("root_pipeline_id", "node_id")}, }, ), ]
gcloud/taskflow3/migrations/0018_auto_20211213_1554.py
1,695
Generated by Django 2.2.24 on 2021-12-13 07:54
46
en
0.735742
""" Django settings for app project. Generated by 'django-admin startproject' using Django 2.1.15. For more information on this file, see https://docs.djangoproject.com/en/2.1/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/2.1/ref/settings/ """ import os # Build paths inside the project like this: os.path.join(BASE_DIR, ...) BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/2.1/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = 'j*imh*c$v&r#enk)ja+^g#r%+=8)7z^!&&3e!e9=f+3-4zh&87' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True ALLOWED_HOSTS = [] # Application definition INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'rest_framework.authtoken', 'core', 'users', 'series', ] MIDDLEWARE = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ] ROOT_URLCONF = 'app.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ], }, }, ] WSGI_APPLICATION = 'app.wsgi.application' # Database # https://docs.djangoproject.com/en/2.1/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'HOST': os.environ.get('DB_HOST'), 'NAME': os.environ.get('DB_NAME'), 'USER': os.environ.get('DB_USER'), 'PASSWORD': os.environ.get('DB_PASS'), } } # Password validation # https://docs.djangoproject.com/en/2.1/ref/settings/#auth-password-validators AUTH_PASSWORD_VALIDATORS = [ { 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator', }, { 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator', }, { 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator', }, { 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator', }, ] # Internationalization # https://docs.djangoproject.com/en/2.1/topics/i18n/ LANGUAGE_CODE = 'en-us' TIME_ZONE = 'UTC' USE_I18N = True USE_L10N = True USE_TZ = True # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/2.1/howto/static-files/ STATIC_URL = '/static/' MEDIA_URL = '/media/' # 127.0.0.1:8000/static will map to static directories, llly media will map. # Tells django where to look and store media. MEDIA_ROOT = '/vol/web/media' # collect static in django collects static files from STATIC_ROOT = '/vol/web/static' # custom User model authorization AUTH_USER_MODEL = 'core.User'
app/app/settings.py
3,625
Django settings for app project. Generated by 'django-admin startproject' using Django 2.1.15. For more information on this file, see https://docs.djangoproject.com/en/2.1/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/2.1/ref/settings/ Build paths inside the project like this: os.path.join(BASE_DIR, ...) Quick-start development settings - unsuitable for production See https://docs.djangoproject.com/en/2.1/howto/deployment/checklist/ SECURITY WARNING: keep the secret key used in production secret! SECURITY WARNING: don't run with debug turned on in production! Application definition Database https://docs.djangoproject.com/en/2.1/ref/settings/databases Password validation https://docs.djangoproject.com/en/2.1/ref/settings/auth-password-validators Internationalization https://docs.djangoproject.com/en/2.1/topics/i18n/ Static files (CSS, JavaScript, Images) https://docs.djangoproject.com/en/2.1/howto/static-files/ 127.0.0.1:8000/static will map to static directories, llly media will map. Tells django where to look and store media. collect static in django collects static files from custom User model authorization
1,188
en
0.662332
""" Provide classes to perform the groupby aggregate operations. These are not exposed to the user and provide implementations of the grouping operations, primarily in cython. These classes (BaseGrouper and BinGrouper) are contained *in* the SeriesGroupBy and DataFrameGroupBy objects. """ from __future__ import annotations import collections import functools from typing import ( Callable, Generic, Hashable, Iterator, Sequence, final, overload, ) import numpy as np from pandas._libs import ( NaT, lib, ) import pandas._libs.groupby as libgroupby import pandas._libs.reduction as libreduction from pandas._typing import ( ArrayLike, DtypeObj, NDFrameT, Shape, npt, ) from pandas.errors import AbstractMethodError from pandas.util._decorators import cache_readonly from pandas.core.dtypes.cast import ( maybe_cast_pointwise_result, maybe_downcast_to_dtype, ) from pandas.core.dtypes.common import ( ensure_float64, ensure_int64, ensure_platform_int, is_1d_only_ea_obj, is_bool_dtype, is_categorical_dtype, is_complex_dtype, is_datetime64_any_dtype, is_float_dtype, is_integer_dtype, is_numeric_dtype, is_sparse, is_timedelta64_dtype, needs_i8_conversion, ) from pandas.core.dtypes.dtypes import ExtensionDtype from pandas.core.dtypes.missing import ( isna, maybe_fill, ) from pandas.core.arrays import ( DatetimeArray, ExtensionArray, PeriodArray, TimedeltaArray, ) from pandas.core.arrays.boolean import BooleanDtype from pandas.core.arrays.floating import ( Float64Dtype, FloatingDtype, ) from pandas.core.arrays.integer import ( Int64Dtype, _IntegerDtype, ) from pandas.core.arrays.masked import ( BaseMaskedArray, BaseMaskedDtype, ) from pandas.core.arrays.string_ import StringDtype from pandas.core.frame import DataFrame from pandas.core.generic import NDFrame from pandas.core.groupby import grouper from pandas.core.indexes.api import ( CategoricalIndex, Index, MultiIndex, ensure_index, ) from pandas.core.series import Series from pandas.core.sorting import ( compress_group_index, decons_obs_group_ids, get_flattened_list, get_group_index, get_group_index_sorter, get_indexer_dict, ) class WrappedCythonOp: """ Dispatch logic for functions defined in _libs.groupby """ # Functions for which we do _not_ attempt to cast the cython result # back to the original dtype. cast_blocklist = frozenset(["rank", "count", "size", "idxmin", "idxmax"]) def __init__(self, kind: str, how: str): self.kind = kind self.how = how _CYTHON_FUNCTIONS = { "aggregate": { "add": "group_add", "prod": "group_prod", "min": "group_min", "max": "group_max", "mean": "group_mean", "median": "group_median", "var": "group_var", "first": "group_nth", "last": "group_last", "ohlc": "group_ohlc", }, "transform": { "cumprod": "group_cumprod", "cumsum": "group_cumsum", "cummin": "group_cummin", "cummax": "group_cummax", "rank": "group_rank", }, } _MASKED_CYTHON_FUNCTIONS = {"cummin", "cummax", "min", "max"} _cython_arity = {"ohlc": 4} # OHLC # Note: we make this a classmethod and pass kind+how so that caching # works at the class level and not the instance level @classmethod @functools.lru_cache(maxsize=None) def _get_cython_function( cls, kind: str, how: str, dtype: np.dtype, is_numeric: bool ): dtype_str = dtype.name ftype = cls._CYTHON_FUNCTIONS[kind][how] # see if there is a fused-type version of function # only valid for numeric f = getattr(libgroupby, ftype) if is_numeric: return f # error: Non-overlapping equality check (left operand type: "dtype[Any]", right # operand type: "Literal['object']") elif dtype == object: # type: ignore[comparison-overlap] if "object" not in f.__signatures__: # raise NotImplementedError here rather than TypeError later raise NotImplementedError( f"function is not implemented for this dtype: " f"[how->{how},dtype->{dtype_str}]" ) return f def get_cython_func_and_vals(self, values: np.ndarray, is_numeric: bool): """ Find the appropriate cython function, casting if necessary. Parameters ---------- values : np.ndarray is_numeric : bool Returns ------- func : callable values : np.ndarray """ how = self.how kind = self.kind if how in ["median", "cumprod"]: # these two only have float64 implementations if is_numeric: values = ensure_float64(values) else: raise NotImplementedError( f"function is not implemented for this dtype: " f"[how->{how},dtype->{values.dtype.name}]" ) func = getattr(libgroupby, f"group_{how}_float64") return func, values func = self._get_cython_function(kind, how, values.dtype, is_numeric) if values.dtype.kind in ["i", "u"]: if how in ["add", "var", "prod", "mean", "ohlc"]: # result may still include NaN, so we have to cast values = ensure_float64(values) return func, values def _disallow_invalid_ops(self, dtype: DtypeObj, is_numeric: bool = False): """ Check if we can do this operation with our cython functions. Raises ------ NotImplementedError This is either not a valid function for this dtype, or valid but not implemented in cython. """ how = self.how if is_numeric: # never an invalid op for those dtypes, so return early as fastpath return if is_categorical_dtype(dtype): # NotImplementedError for methods that can fall back to a # non-cython implementation. if how in ["add", "prod", "cumsum", "cumprod"]: raise TypeError(f"{dtype} type does not support {how} operations") raise NotImplementedError(f"{dtype} dtype not supported") elif is_sparse(dtype): # categoricals are only 1d, so we # are not setup for dim transforming raise NotImplementedError(f"{dtype} dtype not supported") elif is_datetime64_any_dtype(dtype): # we raise NotImplemented if this is an invalid operation # entirely, e.g. adding datetimes if how in ["add", "prod", "cumsum", "cumprod"]: raise TypeError(f"datetime64 type does not support {how} operations") elif is_timedelta64_dtype(dtype): if how in ["prod", "cumprod"]: raise TypeError(f"timedelta64 type does not support {how} operations") def _get_output_shape(self, ngroups: int, values: np.ndarray) -> Shape: how = self.how kind = self.kind arity = self._cython_arity.get(how, 1) out_shape: Shape if how == "ohlc": out_shape = (ngroups, 4) elif arity > 1: raise NotImplementedError( "arity of more than 1 is not supported for the 'how' argument" ) elif kind == "transform": out_shape = values.shape else: out_shape = (ngroups,) + values.shape[1:] return out_shape def get_out_dtype(self, dtype: np.dtype) -> np.dtype: how = self.how if how == "rank": out_dtype = "float64" else: if is_numeric_dtype(dtype): out_dtype = f"{dtype.kind}{dtype.itemsize}" else: out_dtype = "object" return np.dtype(out_dtype) @overload def _get_result_dtype(self, dtype: np.dtype) -> np.dtype: ... # pragma: no cover @overload def _get_result_dtype(self, dtype: ExtensionDtype) -> ExtensionDtype: ... # pragma: no cover def _get_result_dtype(self, dtype: DtypeObj) -> DtypeObj: """ Get the desired dtype of a result based on the input dtype and how it was computed. Parameters ---------- dtype : np.dtype or ExtensionDtype Input dtype. Returns ------- np.dtype or ExtensionDtype The desired dtype of the result. """ how = self.how if how in ["add", "cumsum", "sum", "prod"]: if dtype == np.dtype(bool): return np.dtype(np.int64) elif isinstance(dtype, (BooleanDtype, _IntegerDtype)): return Int64Dtype() elif how in ["mean", "median", "var"]: if isinstance(dtype, (BooleanDtype, _IntegerDtype)): return Float64Dtype() elif is_float_dtype(dtype) or is_complex_dtype(dtype): return dtype elif is_numeric_dtype(dtype): return np.dtype(np.float64) return dtype def uses_mask(self) -> bool: return self.how in self._MASKED_CYTHON_FUNCTIONS @final def _ea_wrap_cython_operation( self, values: ExtensionArray, min_count: int, ngroups: int, comp_ids: np.ndarray, **kwargs, ) -> ArrayLike: """ If we have an ExtensionArray, unwrap, call _cython_operation, and re-wrap if appropriate. """ # TODO: general case implementation overridable by EAs. if isinstance(values, BaseMaskedArray) and self.uses_mask(): return self._masked_ea_wrap_cython_operation( values, min_count=min_count, ngroups=ngroups, comp_ids=comp_ids, **kwargs, ) if isinstance(values, (DatetimeArray, PeriodArray, TimedeltaArray)): # All of the functions implemented here are ordinal, so we can # operate on the tz-naive equivalents npvalues = values._ndarray.view("M8[ns]") elif isinstance(values.dtype, (BooleanDtype, _IntegerDtype)): # IntegerArray or BooleanArray npvalues = values.to_numpy("float64", na_value=np.nan) elif isinstance(values.dtype, FloatingDtype): # FloatingArray npvalues = values.to_numpy(values.dtype.numpy_dtype, na_value=np.nan) elif isinstance(values.dtype, StringDtype): # StringArray npvalues = values.to_numpy(object, na_value=np.nan) else: raise NotImplementedError( f"function is not implemented for this dtype: {values.dtype}" ) res_values = self._cython_op_ndim_compat( npvalues, min_count=min_count, ngroups=ngroups, comp_ids=comp_ids, mask=None, **kwargs, ) if self.how in ["rank"]: # i.e. how in WrappedCythonOp.cast_blocklist, since # other cast_blocklist methods dont go through cython_operation return res_values return self._reconstruct_ea_result(values, res_values) def _reconstruct_ea_result(self, values, res_values): """ Construct an ExtensionArray result from an ndarray result. """ # TODO: allow EAs to override this logic if isinstance( values.dtype, (BooleanDtype, _IntegerDtype, FloatingDtype, StringDtype) ): dtype = self._get_result_dtype(values.dtype) cls = dtype.construct_array_type() return cls._from_sequence(res_values, dtype=dtype) elif needs_i8_conversion(values.dtype): i8values = res_values.view("i8") return type(values)(i8values, dtype=values.dtype) raise NotImplementedError @final def _masked_ea_wrap_cython_operation( self, values: BaseMaskedArray, min_count: int, ngroups: int, comp_ids: np.ndarray, **kwargs, ) -> BaseMaskedArray: """ Equivalent of `_ea_wrap_cython_operation`, but optimized for masked EA's and cython algorithms which accept a mask. """ orig_values = values # Copy to ensure input and result masks don't end up shared mask = values._mask.copy() result_mask = np.zeros(ngroups, dtype=bool) arr = values._data res_values = self._cython_op_ndim_compat( arr, min_count=min_count, ngroups=ngroups, comp_ids=comp_ids, mask=mask, result_mask=result_mask, **kwargs, ) dtype = self._get_result_dtype(orig_values.dtype) assert isinstance(dtype, BaseMaskedDtype) cls = dtype.construct_array_type() if self.kind != "aggregate": return cls(res_values.astype(dtype.type, copy=False), mask) else: return cls(res_values.astype(dtype.type, copy=False), result_mask) @final def _cython_op_ndim_compat( self, values: np.ndarray, *, min_count: int, ngroups: int, comp_ids: np.ndarray, mask: np.ndarray | None = None, result_mask: np.ndarray | None = None, **kwargs, ) -> np.ndarray: if values.ndim == 1: # expand to 2d, dispatch, then squeeze if appropriate values2d = values[None, :] if mask is not None: mask = mask[None, :] if result_mask is not None: result_mask = result_mask[None, :] res = self._call_cython_op( values2d, min_count=min_count, ngroups=ngroups, comp_ids=comp_ids, mask=mask, result_mask=result_mask, **kwargs, ) if res.shape[0] == 1: return res[0] # otherwise we have OHLC return res.T return self._call_cython_op( values, min_count=min_count, ngroups=ngroups, comp_ids=comp_ids, mask=mask, result_mask=result_mask, **kwargs, ) @final def _call_cython_op( self, values: np.ndarray, # np.ndarray[ndim=2] *, min_count: int, ngroups: int, comp_ids: np.ndarray, mask: np.ndarray | None, result_mask: np.ndarray | None, **kwargs, ) -> np.ndarray: # np.ndarray[ndim=2] orig_values = values dtype = values.dtype is_numeric = is_numeric_dtype(dtype) is_datetimelike = needs_i8_conversion(dtype) if is_datetimelike: values = values.view("int64") is_numeric = True elif is_bool_dtype(dtype): values = values.astype("int64") elif is_integer_dtype(dtype): # e.g. uint8 -> uint64, int16 -> int64 dtype_str = dtype.kind + "8" values = values.astype(dtype_str, copy=False) elif is_numeric: if not is_complex_dtype(dtype): values = ensure_float64(values) values = values.T if mask is not None: mask = mask.T if result_mask is not None: result_mask = result_mask.T out_shape = self._get_output_shape(ngroups, values) func, values = self.get_cython_func_and_vals(values, is_numeric) out_dtype = self.get_out_dtype(values.dtype) result = maybe_fill(np.empty(out_shape, dtype=out_dtype)) if self.kind == "aggregate": counts = np.zeros(ngroups, dtype=np.int64) if self.how in ["min", "max", "mean"]: func( result, counts, values, comp_ids, min_count, mask=mask, result_mask=result_mask, is_datetimelike=is_datetimelike, ) elif self.how in ["add"]: # We support datetimelike func( result, counts, values, comp_ids, min_count, datetimelike=is_datetimelike, ) else: func(result, counts, values, comp_ids, min_count) else: # TODO: min_count if self.uses_mask(): func( result, values, comp_ids, ngroups, is_datetimelike, mask=mask, **kwargs, ) else: func(result, values, comp_ids, ngroups, is_datetimelike, **kwargs) if self.kind == "aggregate": # i.e. counts is defined. Locations where count<min_count # need to have the result set to np.nan, which may require casting, # see GH#40767 if is_integer_dtype(result.dtype) and not is_datetimelike: cutoff = max(1, min_count) empty_groups = counts < cutoff if empty_groups.any(): # Note: this conversion could be lossy, see GH#40767 result = result.astype("float64") result[empty_groups] = np.nan result = result.T if self.how not in self.cast_blocklist: # e.g. if we are int64 and need to restore to datetime64/timedelta64 # "rank" is the only member of cast_blocklist we get here res_dtype = self._get_result_dtype(orig_values.dtype) op_result = maybe_downcast_to_dtype(result, res_dtype) else: op_result = result # error: Incompatible return value type (got "Union[ExtensionArray, ndarray]", # expected "ndarray") return op_result # type: ignore[return-value] @final def cython_operation( self, *, values: ArrayLike, axis: int, min_count: int = -1, comp_ids: np.ndarray, ngroups: int, **kwargs, ) -> ArrayLike: """ Call our cython function, with appropriate pre- and post- processing. """ if values.ndim > 2: raise NotImplementedError("number of dimensions is currently limited to 2") elif values.ndim == 2: assert axis == 1, axis elif not is_1d_only_ea_obj(values): # Note: it is *not* the case that axis is always 0 for 1-dim values, # as we can have 1D ExtensionArrays that we need to treat as 2D assert axis == 0 dtype = values.dtype is_numeric = is_numeric_dtype(dtype) # can we do this operation with our cython functions # if not raise NotImplementedError self._disallow_invalid_ops(dtype, is_numeric) if not isinstance(values, np.ndarray): # i.e. ExtensionArray return self._ea_wrap_cython_operation( values, min_count=min_count, ngroups=ngroups, comp_ids=comp_ids, **kwargs, ) return self._cython_op_ndim_compat( values, min_count=min_count, ngroups=ngroups, comp_ids=comp_ids, mask=None, **kwargs, ) class BaseGrouper: """ This is an internal Grouper class, which actually holds the generated groups Parameters ---------- axis : Index groupings : Sequence[Grouping] all the grouping instances to handle in this grouper for example for grouper list to groupby, need to pass the list sort : bool, default True whether this grouper will give sorted result or not group_keys : bool, default True mutated : bool, default False indexer : np.ndarray[np.intp], optional the indexer created by Grouper some groupers (TimeGrouper) will sort its axis and its group_info is also sorted, so need the indexer to reorder """ axis: Index def __init__( self, axis: Index, groupings: Sequence[grouper.Grouping], sort: bool = True, group_keys: bool = True, mutated: bool = False, indexer: npt.NDArray[np.intp] | None = None, dropna: bool = True, ): assert isinstance(axis, Index), axis self.axis = axis self._groupings: list[grouper.Grouping] = list(groupings) self._sort = sort self.group_keys = group_keys self.mutated = mutated self.indexer = indexer self.dropna = dropna @property def groupings(self) -> list[grouper.Grouping]: return self._groupings @property def shape(self) -> Shape: return tuple(ping.ngroups for ping in self.groupings) def __iter__(self): return iter(self.indices) @property def nkeys(self) -> int: return len(self.groupings) def get_iterator( self, data: NDFrameT, axis: int = 0 ) -> Iterator[tuple[Hashable, NDFrameT]]: """ Groupby iterator Returns ------- Generator yielding sequence of (name, subsetted object) for each group """ splitter = self._get_splitter(data, axis=axis) keys = self.group_keys_seq for key, group in zip(keys, splitter): yield key, group.__finalize__(data, method="groupby") @final def _get_splitter(self, data: NDFrame, axis: int = 0) -> DataSplitter: """ Returns ------- Generator yielding subsetted objects __finalize__ has not been called for the subsetted objects returned. """ ids, _, ngroups = self.group_info return get_splitter(data, ids, ngroups, axis=axis) def _get_grouper(self): """ We are a grouper as part of another's groupings. We have a specific method of grouping, so cannot convert to a Index for our grouper. """ return self.groupings[0].grouping_vector @final @cache_readonly def group_keys_seq(self): if len(self.groupings) == 1: return self.levels[0] else: ids, _, ngroups = self.group_info # provide "flattened" iterator for multi-group setting return get_flattened_list(ids, ngroups, self.levels, self.codes) @final def apply( self, f: Callable, data: DataFrame | Series, axis: int = 0 ) -> tuple[list, bool]: mutated = self.mutated splitter = self._get_splitter(data, axis=axis) group_keys = self.group_keys_seq result_values = [] # This calls DataSplitter.__iter__ zipped = zip(group_keys, splitter) for key, group in zipped: object.__setattr__(group, "name", key) # group might be modified group_axes = group.axes res = f(group) if not mutated and not _is_indexed_like(res, group_axes, axis): mutated = True result_values.append(res) # getattr pattern for __name__ is needed for functools.partial objects if len(group_keys) == 0 and getattr(f, "__name__", None) not in [ "idxmin", "idxmax", "nanargmin", "nanargmax", ]: # If group_keys is empty, then no function calls have been made, # so we will not have raised even if this is an invalid dtype. # So do one dummy call here to raise appropriate TypeError. f(data.iloc[:0]) return result_values, mutated @cache_readonly def indices(self) -> dict[Hashable, npt.NDArray[np.intp]]: """dict {group name -> group indices}""" if len(self.groupings) == 1 and isinstance(self.result_index, CategoricalIndex): # This shows unused categories in indices GH#38642 return self.groupings[0].indices codes_list = [ping.codes for ping in self.groupings] keys = [ping.group_index for ping in self.groupings] return get_indexer_dict(codes_list, keys) @final @property def codes(self) -> list[np.ndarray]: return [ping.codes for ping in self.groupings] @property def levels(self) -> list[Index]: return [ping.group_index for ping in self.groupings] @property def names(self) -> list[Hashable]: return [ping.name for ping in self.groupings] @final def size(self) -> Series: """ Compute group sizes. """ ids, _, ngroups = self.group_info if ngroups: out = np.bincount(ids[ids != -1], minlength=ngroups) else: out = [] return Series(out, index=self.result_index, dtype="int64") @cache_readonly def groups(self) -> dict[Hashable, np.ndarray]: """dict {group name -> group labels}""" if len(self.groupings) == 1: return self.groupings[0].groups else: to_groupby = zip(*(ping.grouping_vector for ping in self.groupings)) index = Index(to_groupby) return self.axis.groupby(index) @final @cache_readonly def is_monotonic(self) -> bool: # return if my group orderings are monotonic return Index(self.group_info[0]).is_monotonic @cache_readonly def group_info(self) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp], int]: comp_ids, obs_group_ids = self._get_compressed_codes() ngroups = len(obs_group_ids) comp_ids = ensure_platform_int(comp_ids) return comp_ids, obs_group_ids, ngroups @final @cache_readonly def codes_info(self) -> npt.NDArray[np.intp]: # return the codes of items in original grouped axis ids, _, _ = self.group_info if self.indexer is not None: sorter = np.lexsort((ids, self.indexer)) ids = ids[sorter] ids = ensure_platform_int(ids) # TODO: if numpy annotates np.lexsort, this ensure_platform_int # may become unnecessary return ids @final def _get_compressed_codes(self) -> tuple[np.ndarray, npt.NDArray[np.intp]]: # The first returned ndarray may have any signed integer dtype if len(self.groupings) > 1: group_index = get_group_index(self.codes, self.shape, sort=True, xnull=True) return compress_group_index(group_index, sort=self._sort) ping = self.groupings[0] return ping.codes, np.arange(len(ping.group_index), dtype=np.intp) @final @cache_readonly def ngroups(self) -> int: return len(self.result_index) @property def reconstructed_codes(self) -> list[np.ndarray]: codes = self.codes ids, obs_ids, _ = self.group_info return decons_obs_group_ids(ids, obs_ids, self.shape, codes, xnull=True) @final @cache_readonly def result_arraylike(self) -> ArrayLike: """ Analogous to result_index, but returning an ndarray/ExtensionArray allowing us to retain ExtensionDtypes not supported by Index. """ # TODO(ExtensionIndex): once Index supports arbitrary EAs, this can # be removed in favor of result_index if len(self.groupings) == 1: return self.groupings[0].group_arraylike # result_index is MultiIndex return self.result_index._values @cache_readonly def result_index(self) -> Index: if len(self.groupings) == 1: return self.groupings[0].result_index.rename(self.names[0]) codes = self.reconstructed_codes levels = [ping.result_index for ping in self.groupings] return MultiIndex( levels=levels, codes=codes, verify_integrity=False, names=self.names ) @final def get_group_levels(self) -> list[ArrayLike]: # Note: only called from _insert_inaxis_grouper_inplace, which # is only called for BaseGrouper, never for BinGrouper if len(self.groupings) == 1: return [self.groupings[0].group_arraylike] name_list = [] for ping, codes in zip(self.groupings, self.reconstructed_codes): codes = ensure_platform_int(codes) levels = ping.group_arraylike.take(codes) name_list.append(levels) return name_list # ------------------------------------------------------------ # Aggregation functions @final def _cython_operation( self, kind: str, values, how: str, axis: int, min_count: int = -1, **kwargs, ) -> ArrayLike: """ Returns the values of a cython operation. """ assert kind in ["transform", "aggregate"] cy_op = WrappedCythonOp(kind=kind, how=how) ids, _, _ = self.group_info ngroups = self.ngroups return cy_op.cython_operation( values=values, axis=axis, min_count=min_count, comp_ids=ids, ngroups=ngroups, **kwargs, ) @final def agg_series( self, obj: Series, func: Callable, preserve_dtype: bool = False ) -> ArrayLike: """ Parameters ---------- obj : Series func : function taking a Series and returning a scalar-like preserve_dtype : bool Whether the aggregation is known to be dtype-preserving. Returns ------- np.ndarray or ExtensionArray """ # test_groupby_empty_with_category gets here with self.ngroups == 0 # and len(obj) > 0 if len(obj) == 0: # SeriesGrouper would raise if we were to call _aggregate_series_fast result = self._aggregate_series_pure_python(obj, func) elif not isinstance(obj._values, np.ndarray): result = self._aggregate_series_pure_python(obj, func) # we can preserve a little bit more aggressively with EA dtype # because maybe_cast_pointwise_result will do a try/except # with _from_sequence. NB we are assuming here that _from_sequence # is sufficiently strict that it casts appropriately. preserve_dtype = True else: result = self._aggregate_series_pure_python(obj, func) npvalues = lib.maybe_convert_objects(result, try_float=False) if preserve_dtype: out = maybe_cast_pointwise_result(npvalues, obj.dtype, numeric_only=True) else: out = npvalues return out @final def _aggregate_series_pure_python( self, obj: Series, func: Callable ) -> npt.NDArray[np.object_]: ids, _, ngroups = self.group_info counts = np.zeros(ngroups, dtype=int) result = np.empty(ngroups, dtype="O") initialized = False # equiv: splitter = self._get_splitter(obj, axis=0) splitter = get_splitter(obj, ids, ngroups, axis=0) for i, group in enumerate(splitter): res = func(group) res = libreduction.extract_result(res) if not initialized: # We only do this validation on the first iteration libreduction.check_result_array(res, group.dtype) initialized = True counts[i] = group.shape[0] result[i] = res return result class BinGrouper(BaseGrouper): """ This is an internal Grouper class Parameters ---------- bins : the split index of binlabels to group the item of axis binlabels : the label list mutated : bool, default False indexer : np.ndarray[np.intp] Examples -------- bins: [2, 4, 6, 8, 10] binlabels: DatetimeIndex(['2005-01-01', '2005-01-03', '2005-01-05', '2005-01-07', '2005-01-09'], dtype='datetime64[ns]', freq='2D') the group_info, which contains the label of each item in grouped axis, the index of label in label list, group number, is (array([0, 0, 1, 1, 2, 2, 3, 3, 4, 4]), array([0, 1, 2, 3, 4]), 5) means that, the grouped axis has 10 items, can be grouped into 5 labels, the first and second items belong to the first label, the third and forth items belong to the second label, and so on """ bins: npt.NDArray[np.int64] binlabels: Index mutated: bool def __init__( self, bins, binlabels, mutated: bool = False, indexer=None, ): self.bins = ensure_int64(bins) self.binlabels = ensure_index(binlabels) self.mutated = mutated self.indexer = indexer # These lengths must match, otherwise we could call agg_series # with empty self.bins, which would raise in libreduction. assert len(self.binlabels) == len(self.bins) @cache_readonly def groups(self): """dict {group name -> group labels}""" # this is mainly for compat # GH 3881 result = { key: value for key, value in zip(self.binlabels, self.bins) if key is not NaT } return result @property def nkeys(self) -> int: # still matches len(self.groupings), but we can hard-code return 1 def _get_grouper(self): """ We are a grouper as part of another's groupings. We have a specific method of grouping, so cannot convert to a Index for our grouper. """ return self def get_iterator(self, data: NDFrame, axis: int = 0): """ Groupby iterator Returns ------- Generator yielding sequence of (name, subsetted object) for each group """ if axis == 0: slicer = lambda start, edge: data.iloc[start:edge] else: slicer = lambda start, edge: data.iloc[:, start:edge] length = len(data.axes[axis]) start = 0 for edge, label in zip(self.bins, self.binlabels): if label is not NaT: yield label, slicer(start, edge) start = edge if start < length: yield self.binlabels[-1], slicer(start, None) @cache_readonly def indices(self): indices = collections.defaultdict(list) i = 0 for label, bin in zip(self.binlabels, self.bins): if i < bin: if label is not NaT: indices[label] = list(range(i, bin)) i = bin return indices @cache_readonly def group_info(self) -> tuple[npt.NDArray[np.intp], npt.NDArray[np.intp], int]: ngroups = self.ngroups obs_group_ids = np.arange(ngroups, dtype=np.intp) rep = np.diff(np.r_[0, self.bins]) rep = ensure_platform_int(rep) if ngroups == len(self.bins): comp_ids = np.repeat(np.arange(ngroups), rep) else: comp_ids = np.repeat(np.r_[-1, np.arange(ngroups)], rep) return ( ensure_platform_int(comp_ids), obs_group_ids, ngroups, ) @cache_readonly def reconstructed_codes(self) -> list[np.ndarray]: # get unique result indices, and prepend 0 as groupby starts from the first return [np.r_[0, np.flatnonzero(self.bins[1:] != self.bins[:-1]) + 1]] @cache_readonly def result_index(self): if len(self.binlabels) != 0 and isna(self.binlabels[0]): return self.binlabels[1:] return self.binlabels @property def levels(self) -> list[Index]: return [self.binlabels] @property def names(self) -> list[Hashable]: return [self.binlabels.name] @property def groupings(self) -> list[grouper.Grouping]: lev = self.binlabels ping = grouper.Grouping(lev, lev, in_axis=False, level=None) return [ping] def _aggregate_series_fast(self, obj: Series, func: Callable) -> np.ndarray: # -> np.ndarray[object] raise NotImplementedError( "This should not be reached; use _aggregate_series_pure_python" ) def _is_indexed_like(obj, axes, axis: int) -> bool: if isinstance(obj, Series): if len(axes) > 1: return False return obj.axes[axis].equals(axes[axis]) elif isinstance(obj, DataFrame): return obj.axes[axis].equals(axes[axis]) return False # ---------------------------------------------------------------------- # Splitting / application class DataSplitter(Generic[NDFrameT]): def __init__( self, data: NDFrameT, labels: npt.NDArray[np.intp], ngroups: int, axis: int = 0, ): self.data = data self.labels = ensure_platform_int(labels) # _should_ already be np.intp self.ngroups = ngroups self.axis = axis assert isinstance(axis, int), axis @cache_readonly def slabels(self) -> npt.NDArray[np.intp]: # Sorted labels return self.labels.take(self._sort_idx) @cache_readonly def _sort_idx(self) -> npt.NDArray[np.intp]: # Counting sort indexer return get_group_index_sorter(self.labels, self.ngroups) def __iter__(self): sdata = self.sorted_data if self.ngroups == 0: # we are inside a generator, rather than raise StopIteration # we merely return signal the end return starts, ends = lib.generate_slices(self.slabels, self.ngroups) for start, end in zip(starts, ends): yield self._chop(sdata, slice(start, end)) @cache_readonly def sorted_data(self) -> NDFrameT: return self.data.take(self._sort_idx, axis=self.axis) def _chop(self, sdata, slice_obj: slice) -> NDFrame: raise AbstractMethodError(self) class SeriesSplitter(DataSplitter): def _chop(self, sdata: Series, slice_obj: slice) -> Series: # fastpath equivalent to `sdata.iloc[slice_obj]` mgr = sdata._mgr.get_slice(slice_obj) # __finalize__ not called here, must be applied by caller if applicable # fastpath equivalent to: # `return sdata._constructor(mgr, name=sdata.name, fastpath=True)` obj = type(sdata)._from_mgr(mgr) object.__setattr__(obj, "_flags", sdata._flags) object.__setattr__(obj, "_name", sdata._name) return obj class FrameSplitter(DataSplitter): def _chop(self, sdata: DataFrame, slice_obj: slice) -> DataFrame: # Fastpath equivalent to: # if self.axis == 0: # return sdata.iloc[slice_obj] # else: # return sdata.iloc[:, slice_obj] mgr = sdata._mgr.get_slice(slice_obj, axis=1 - self.axis) # __finalize__ not called here, must be applied by caller if applicable # fastpath equivalent to `return sdata._constructor(mgr)` obj = type(sdata)._from_mgr(mgr) object.__setattr__(obj, "_flags", sdata._flags) return obj def get_splitter( data: NDFrame, labels: np.ndarray, ngroups: int, axis: int = 0 ) -> DataSplitter: if isinstance(data, Series): klass: type[DataSplitter] = SeriesSplitter else: # i.e. DataFrame klass = FrameSplitter return klass(data, labels, ngroups, axis)
pandas/core/groupby/ops.py
40,148
This is an internal Grouper class, which actually holds the generated groups Parameters ---------- axis : Index groupings : Sequence[Grouping] all the grouping instances to handle in this grouper for example for grouper list to groupby, need to pass the list sort : bool, default True whether this grouper will give sorted result or not group_keys : bool, default True mutated : bool, default False indexer : np.ndarray[np.intp], optional the indexer created by Grouper some groupers (TimeGrouper) will sort its axis and its group_info is also sorted, so need the indexer to reorder This is an internal Grouper class Parameters ---------- bins : the split index of binlabels to group the item of axis binlabels : the label list mutated : bool, default False indexer : np.ndarray[np.intp] Examples -------- bins: [2, 4, 6, 8, 10] binlabels: DatetimeIndex(['2005-01-01', '2005-01-03', '2005-01-05', '2005-01-07', '2005-01-09'], dtype='datetime64[ns]', freq='2D') the group_info, which contains the label of each item in grouped axis, the index of label in label list, group number, is (array([0, 0, 1, 1, 2, 2, 3, 3, 4, 4]), array([0, 1, 2, 3, 4]), 5) means that, the grouped axis has 10 items, can be grouped into 5 labels, the first and second items belong to the first label, the third and forth items belong to the second label, and so on Dispatch logic for functions defined in _libs.groupby Returns the values of a cython operation. Check if we can do this operation with our cython functions. Raises ------ NotImplementedError This is either not a valid function for this dtype, or valid but not implemented in cython. If we have an ExtensionArray, unwrap, call _cython_operation, and re-wrap if appropriate. We are a grouper as part of another's groupings. We have a specific method of grouping, so cannot convert to a Index for our grouper. We are a grouper as part of another's groupings. We have a specific method of grouping, so cannot convert to a Index for our grouper. Get the desired dtype of a result based on the input dtype and how it was computed. Parameters ---------- dtype : np.dtype or ExtensionDtype Input dtype. Returns ------- np.dtype or ExtensionDtype The desired dtype of the result. Returns ------- Generator yielding subsetted objects __finalize__ has not been called for the subsetted objects returned. Equivalent of `_ea_wrap_cython_operation`, but optimized for masked EA's and cython algorithms which accept a mask. Construct an ExtensionArray result from an ndarray result. Parameters ---------- obj : Series func : function taking a Series and returning a scalar-like preserve_dtype : bool Whether the aggregation is known to be dtype-preserving. Returns ------- np.ndarray or ExtensionArray Call our cython function, with appropriate pre- and post- processing. Find the appropriate cython function, casting if necessary. Parameters ---------- values : np.ndarray is_numeric : bool Returns ------- func : callable values : np.ndarray Groupby iterator Returns ------- Generator yielding sequence of (name, subsetted object) for each group Groupby iterator Returns ------- Generator yielding sequence of (name, subsetted object) for each group dict {group name -> group labels} dict {group name -> group labels} dict {group name -> group indices} Analogous to result_index, but returning an ndarray/ExtensionArray allowing us to retain ExtensionDtypes not supported by Index. Compute group sizes. Provide classes to perform the groupby aggregate operations. These are not exposed to the user and provide implementations of the grouping operations, primarily in cython. These classes (BaseGrouper and BinGrouper) are contained *in* the SeriesGroupBy and DataFrameGroupBy objects. Functions for which we do _not_ attempt to cast the cython result back to the original dtype. OHLC Note: we make this a classmethod and pass kind+how so that caching works at the class level and not the instance level see if there is a fused-type version of function only valid for numeric error: Non-overlapping equality check (left operand type: "dtype[Any]", right operand type: "Literal['object']") type: ignore[comparison-overlap] raise NotImplementedError here rather than TypeError later these two only have float64 implementations result may still include NaN, so we have to cast never an invalid op for those dtypes, so return early as fastpath NotImplementedError for methods that can fall back to a non-cython implementation. categoricals are only 1d, so we are not setup for dim transforming we raise NotImplemented if this is an invalid operation entirely, e.g. adding datetimes pragma: no cover pragma: no cover TODO: general case implementation overridable by EAs. All of the functions implemented here are ordinal, so we can operate on the tz-naive equivalents IntegerArray or BooleanArray FloatingArray StringArray i.e. how in WrappedCythonOp.cast_blocklist, since other cast_blocklist methods dont go through cython_operation TODO: allow EAs to override this logic Copy to ensure input and result masks don't end up shared expand to 2d, dispatch, then squeeze if appropriate otherwise we have OHLC np.ndarray[ndim=2] np.ndarray[ndim=2] e.g. uint8 -> uint64, int16 -> int64 We support datetimelike TODO: min_count i.e. counts is defined. Locations where count<min_count need to have the result set to np.nan, which may require casting, see GH40767 Note: this conversion could be lossy, see GH40767 e.g. if we are int64 and need to restore to datetime64/timedelta64 "rank" is the only member of cast_blocklist we get here error: Incompatible return value type (got "Union[ExtensionArray, ndarray]", expected "ndarray") type: ignore[return-value] Note: it is *not* the case that axis is always 0 for 1-dim values, as we can have 1D ExtensionArrays that we need to treat as 2D can we do this operation with our cython functions if not raise NotImplementedError i.e. ExtensionArray provide "flattened" iterator for multi-group setting This calls DataSplitter.__iter__ group might be modified getattr pattern for __name__ is needed for functools.partial objects If group_keys is empty, then no function calls have been made, so we will not have raised even if this is an invalid dtype. So do one dummy call here to raise appropriate TypeError. This shows unused categories in indices GH38642 return if my group orderings are monotonic return the codes of items in original grouped axis TODO: if numpy annotates np.lexsort, this ensure_platform_int may become unnecessary The first returned ndarray may have any signed integer dtype TODO(ExtensionIndex): once Index supports arbitrary EAs, this can be removed in favor of result_index result_index is MultiIndex Note: only called from _insert_inaxis_grouper_inplace, which is only called for BaseGrouper, never for BinGrouper ------------------------------------------------------------ Aggregation functions test_groupby_empty_with_category gets here with self.ngroups == 0 and len(obj) > 0 SeriesGrouper would raise if we were to call _aggregate_series_fast we can preserve a little bit more aggressively with EA dtype because maybe_cast_pointwise_result will do a try/except with _from_sequence. NB we are assuming here that _from_sequence is sufficiently strict that it casts appropriately. equiv: splitter = self._get_splitter(obj, axis=0) We only do this validation on the first iteration These lengths must match, otherwise we could call agg_series with empty self.bins, which would raise in libreduction. this is mainly for compat GH 3881 still matches len(self.groupings), but we can hard-code get unique result indices, and prepend 0 as groupby starts from the first -> np.ndarray[object] ---------------------------------------------------------------------- Splitting / application _should_ already be np.intp Sorted labels Counting sort indexer we are inside a generator, rather than raise StopIteration we merely return signal the end fastpath equivalent to `sdata.iloc[slice_obj]` __finalize__ not called here, must be applied by caller if applicable fastpath equivalent to: `return sdata._constructor(mgr, name=sdata.name, fastpath=True)` Fastpath equivalent to: if self.axis == 0: return sdata.iloc[slice_obj] else: return sdata.iloc[:, slice_obj] __finalize__ not called here, must be applied by caller if applicable fastpath equivalent to `return sdata._constructor(mgr)` i.e. DataFrame
8,479
en
0.754825
# Copyright 2015 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Basic arithmetic operators. See the [python/math_ops](python/math_ops) guide. """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import numpy as np from six.moves import xrange # pylint: disable=redefined-builtin from tensorflow.python.eager import context from tensorflow.python.framework import common_shapes from tensorflow.python.framework import constant_op from tensorflow.python.framework import dtypes from tensorflow.python.framework import graph_util from tensorflow.python.framework import ops from tensorflow.python.framework import sparse_tensor from tensorflow.python.framework import tensor_shape from tensorflow.python.ops import array_ops from tensorflow.python.ops import gen_data_flow_ops from tensorflow.python.ops import gen_math_ops from tensorflow.python.ops import gen_nn_ops from tensorflow.python.ops import gen_sparse_ops # go/tf-wildcard-import # pylint: disable=wildcard-import from tensorflow.python.ops.gen_math_ops import * # pylint: enable=wildcard-import from tensorflow.python.platform import tf_logging as logging from tensorflow.python.util import compat from tensorflow.python.util import deprecation from tensorflow.python.util import dispatch from tensorflow.python.util import nest from tensorflow.python.util.tf_export import tf_export # Aliases for some automatically-generated names. linspace = gen_math_ops.lin_space nextafter = gen_math_ops.next_after arg_max = deprecation.deprecated(None, "Use `tf.math.argmax` instead")(arg_max) # pylint: disable=used-before-assignment arg_min = deprecation.deprecated(None, "Use `tf.math.argmin` instead")(arg_min) # pylint: disable=used-before-assignment tf_export(v1=["arg_max"])(arg_max) tf_export(v1=["arg_min"])(arg_min) # This is set by resource_variable_ops.py. It is included in this way since # there is a circular dependency between math_ops and resource_variable_ops _resource_variable_type = None def _set_doc(doc): def _decorator(func): func.__doc__ = doc return func return _decorator # pylint: disable=redefined-builtin @tf_export(v1=["math.argmax", "argmax"]) @deprecation.deprecated_args(None, "Use the `axis` argument instead", "dimension") @_set_doc( gen_math_ops.arg_max.__doc__.replace("dimensions", "axes").replace( "dimension", "axis")) def argmax(input, axis=None, name=None, dimension=None, output_type=dtypes.int64): axis = deprecation.deprecated_argument_lookup( "axis", axis, "dimension", dimension) return argmax_v2(input, axis, output_type, name) @tf_export("math.argmax", "argmax", v1=[]) def argmax_v2(input, axis=None, output_type=dtypes.int64, name=None): """Returns the index with the largest value across axes of a tensor. Note that in case of ties the identity of the return value is not guaranteed. Args: input: A `Tensor`. Must be one of the following types: `float32`, `float64`, `int32`, `uint8`, `int16`, `int8`, `complex64`, `int64`, `qint8`, `quint8`, `qint32`, `bfloat16`, `uint16`, `complex128`, `half`, `uint32`, `uint64`. axis: A `Tensor`. Must be one of the following types: `int32`, `int64`. int32 or int64, must be in the range `-rank(input), rank(input))`. Describes which axis of the input Tensor to reduce across. For vectors, use axis = 0. output_type: An optional `tf.DType` from: `tf.int32, tf.int64`. Defaults to `tf.int64`. name: A name for the operation (optional). Returns: A `Tensor` of type `output_type`. """ if axis is None: axis = 0 return gen_math_ops.arg_max(input, axis, name=name, output_type=output_type) @tf_export(v1=["math.argmin", "argmin"]) @deprecation.deprecated_args(None, "Use the `axis` argument instead", "dimension") @_set_doc( gen_math_ops.arg_min.__doc__.replace("dimensions", "axes").replace( "dimension", "axis")) def argmin(input, axis=None, name=None, dimension=None, output_type=dtypes.int64): axis = deprecation.deprecated_argument_lookup( "axis", axis, "dimension", dimension) return argmin_v2(input, axis, output_type, name) @tf_export("math.argmin", "argmin", v1=[]) def argmin_v2(input, axis=None, output_type=dtypes.int64, name=None): """Returns the index with the smallest value across axes of a tensor. Note that in case of ties the identity of the return value is not guaranteed. Args: input: A `Tensor`. Must be one of the following types: `float32`, `float64`, `int32`, `uint8`, `int16`, `int8`, `complex64`, `int64`, `qint8`, `quint8`, `qint32`, `bfloat16`, `uint16`, `complex128`, `half`, `uint32`, `uint64`. axis: A `Tensor`. Must be one of the following types: `int32`, `int64`. int32 or int64, must be in the range `-rank(input), rank(input))`. Describes which axis of the input Tensor to reduce across. For vectors, use axis = 0. output_type: An optional `tf.DType` from: `tf.int32, tf.int64`. Defaults to `tf.int64`. name: A name for the operation (optional). Returns: A `Tensor` of type `output_type`. """ if axis is None: axis = 0 return gen_math_ops.arg_min(input, axis, name=name, output_type=output_type) # pylint: enable=redefined-builtin # pylint: disable=anomalous-backslash-in-string,protected-access # pylint: disable=g-docstring-has-escape @tf_export("math.abs", "abs") @dispatch.add_dispatch_support def abs(x, name=None): # pylint: disable=redefined-builtin r"""Computes the absolute value of a tensor. Given a tensor `x` of complex numbers, this operation returns a tensor of type `float32` or `float64` that is the absolute value of each element in `x`. All elements in `x` must be complex numbers of the form \\(a + bj\\). The absolute value is computed as \\( \sqrt{a^2 + b^2}\\). For example: ```python x = tf.constant([[-2.25 + 4.75j], [-3.25 + 5.75j]]) tf.abs(x) # [5.25594902, 6.60492229] ``` Args: x: A `Tensor` or `SparseTensor` of type `float16`, `float32`, `float64`, `int32`, `int64`, `complex64` or `complex128`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` the same size and type as `x` with absolute values. Note, for `complex64` or `complex128` input, the returned `Tensor` will be of type `float32` or `float64`, respectively. """ with ops.name_scope(name, "Abs", [x]) as name: x = ops.convert_to_tensor(x, name="x") if x.dtype.is_complex: return gen_math_ops.complex_abs(x, Tout=x.dtype.real_dtype, name=name) return gen_math_ops._abs(x, name=name) # pylint: enable=g-docstring-has-escape # pylint: disable=redefined-builtin def _bucketize(input, boundaries, name=None): return gen_math_ops.bucketize(input=input, boundaries=boundaries, name=name) # pylint: enable=redefined-builtin class DivideDelegateWithName(object): """Use Python2/Python3 division delegation to implement divide for tensors.""" def __init__(self, x, name): """Construct DivideDelegateWithName. Args: x: Tensor to use as left operand in operator overloads name: The name that is preferred for the op created. """ self.x = x self.name = name def __truediv__(self, y): return _truediv_python3(self.x, y, self.name) def __floordiv__(self, y): return floordiv(self.x, y, self.name) def __div__(self, y): return _div_python2(self.x, y, self.name) @tf_export("math.divide", "divide") @dispatch.add_dispatch_support def divide(x, y, name=None): """Computes Python style division of `x` by `y`.""" if name is not None: # Cannot use tensors operator overload, because it has no way to track # override names. Use a dummy class to track the runtime division behavior return DivideDelegateWithName(x, name) / y else: return x / y @tf_export("math.multiply", "multiply") @dispatch.add_dispatch_support def multiply(x, y, name=None): return gen_math_ops.mul(x, y, name) multiply.__doc__ = gen_math_ops.mul.__doc__.replace("Multiply", "`tf.multiply`") # TODO(aselle): put deprecation in after another round of global code changes @deprecation.deprecated( "2016-12-30", "`tf.mul(x, y)` is deprecated, please use `tf.multiply(x, y)` or `x * y`") def _mul(x, y, name=None): return gen_math_ops.mul(x, y, name) _mul.__doc__ = ( gen_math_ops.mul.__doc__ + ("" if _mul.__doc__ is None else _mul.__doc__)) @tf_export("math.subtract", "subtract") @dispatch.add_dispatch_support def subtract(x, y, name=None): return gen_math_ops.sub(x, y, name) subtract.__doc__ = gen_math_ops.sub.__doc__.replace("`Sub`", "`tf.subtract`") # TODO(aselle): put deprecation in after another round of global code changes @deprecation.deprecated( "2016-12-30", "`tf.sub(x, y)` is deprecated, please use `tf.subtract(x, y)` or `x - y`") def _sub(x, y, name=None): return gen_math_ops.sub(x, y, name) _sub.__doc__ = ( gen_math_ops.sub.__doc__ + ("" if _sub.__doc__ is None else _sub.__doc__)) negative = gen_math_ops.neg # pylint: disable=g-docstring-has-escape @deprecation.deprecated( "2016-12-30", "`tf.neg(x)` is deprecated, please use `tf.negative(x)` or `-x`") def _neg(x, name=None): """Computes numerical negative value element-wise. I.e., \\(y = -x\\). Args: x: A `Tensor` or `SparseTensor`. Must be one of the following types: `half`, `float32`, `float64`, `int32`, `int64`, `complex64`, `complex128`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor`, respectively. Has the same type as `x`. """ return negative(x, name) # pylint: enable=g-docstring-has-escape @tf_export(v1=["math.scalar_mul", "scalar_mul"]) def scalar_mul(scalar, x, name=None): """Multiplies a scalar times a `Tensor` or `IndexedSlices` object. Intended for use in gradient code which might deal with `IndexedSlices` objects, which are easy to multiply by a scalar but more expensive to multiply with arbitrary tensors. Args: scalar: A 0-D scalar `Tensor`. Must have known shape. x: A `Tensor` or `IndexedSlices` to be scaled. name: A name for the operation (optional). Returns: `scalar * x` of the same type (`Tensor` or `IndexedSlices`) as `x`. Raises: ValueError: if scalar is not a 0-D `scalar`. """ scalar = ops.convert_to_tensor( scalar, dtype=x.dtype.base_dtype, name="scalar") shape = scalar.get_shape() if shape.ndims == 0: if isinstance(x, ops.IndexedSlices): return ops.IndexedSlices(gen_math_ops.mul(scalar, x.values, name), x.indices, x.dense_shape) else: return gen_math_ops.mul(scalar, x, name) else: raise ValueError("Only scalar multiply works, got shape %s" % shape) @tf_export("math.scalar_mul", "scalar_mul", v1=[]) @_set_doc(scalar_mul.__doc__) def scalar_mul_v2(scalar, x, name=None): with ops.name_scope(name, "scalar_mul", [x]) as name: return scalar_mul(scalar, x, name) @tf_export("math.pow", "pow") @dispatch.add_dispatch_support def pow(x, y, name=None): # pylint: disable=redefined-builtin r"""Computes the power of one value to another. Given a tensor `x` and a tensor `y`, this operation computes \\(x^y\\) for corresponding elements in `x` and `y`. For example: ```python x = tf.constant([[2, 2], [3, 3]]) y = tf.constant([[8, 16], [2, 3]]) tf.pow(x, y) # [[256, 65536], [9, 27]] ``` Args: x: A `Tensor` of type `float16`, `float32`, `float64`, `int32`, `int64`, `complex64`, or `complex128`. y: A `Tensor` of type `float16`, `float32`, `float64`, `int32`, `int64`, `complex64`, or `complex128`. name: A name for the operation (optional). Returns: A `Tensor`. """ with ops.name_scope(name, "Pow", [x]) as name: return gen_math_ops._pow(x, y, name=name) # pylint: disable=redefined-builtin,redefined-outer-name @tf_export("dtypes.complex", "complex") @dispatch.add_dispatch_support def complex(real, imag, name=None): r"""Converts two real numbers to a complex number. Given a tensor `real` representing the real part of a complex number, and a tensor `imag` representing the imaginary part of a complex number, this operation returns complex numbers elementwise of the form \\(a + bj\\), where *a* represents the `real` part and *b* represents the `imag` part. The input tensors `real` and `imag` must have the same shape. For example: ```python real = tf.constant([2.25, 3.25]) imag = tf.constant([4.75, 5.75]) tf.complex(real, imag) # [[2.25 + 4.75j], [3.25 + 5.75j]] ``` Args: real: A `Tensor`. Must be one of the following types: `float32`, `float64`. imag: A `Tensor`. Must have the same type as `real`. name: A name for the operation (optional). Returns: A `Tensor` of type `complex64` or `complex128`. """ real = ops.convert_to_tensor(real, name="real") imag = ops.convert_to_tensor(imag, name="imag") with ops.name_scope(name, "Complex", [real, imag]) as name: input_types = (real.dtype, imag.dtype) if input_types == (dtypes.float64, dtypes.float64): Tout = dtypes.complex128 elif input_types == (dtypes.float32, dtypes.float32): Tout = dtypes.complex64 else: raise TypeError("real and imag have incorrect types: " "{} {}".format(real.dtype.name, imag.dtype.name)) return gen_math_ops._complex(real, imag, Tout=Tout, name=name) @tf_export("math.real", v1=["math.real", "real"]) @deprecation.deprecated_endpoints("real") @dispatch.add_dispatch_support def real(input, name=None): r"""Returns the real part of a complex (or real) tensor. Given a tensor `input`, this operation returns a tensor of type `float` that is the real part of each element in `input` considered as a complex number. For example: ```python x = tf.constant([-2.25 + 4.75j, 3.25 + 5.75j]) tf.real(x) # [-2.25, 3.25] ``` If `input` is already real, it is returned unchanged. Args: input: A `Tensor`. Must have numeric type. name: A name for the operation (optional). Returns: A `Tensor` of type `float32` or `float64`. """ with ops.name_scope(name, "Real", [input]) as name: if input.dtype.is_complex: real_dtype = input.dtype.real_dtype return gen_math_ops.real(input, Tout=real_dtype, name=name) else: return input @tf_export("math.imag", v1=["math.imag", "imag"]) @deprecation.deprecated_endpoints("imag") @dispatch.add_dispatch_support def imag(input, name=None): r"""Returns the imaginary part of a complex (or real) tensor. Given a tensor `input`, this operation returns a tensor of type `float` that is the imaginary part of each element in `input` considered as a complex number. If `input` is real, a tensor of all zeros is returned. For example: ```python x = tf.constant([-2.25 + 4.75j, 3.25 + 5.75j]) tf.imag(x) # [4.75, 5.75] ``` Args: input: A `Tensor`. Must be one of the following types: `float`, `double`, `complex64`, `complex128`. name: A name for the operation (optional). Returns: A `Tensor` of type `float32` or `float64`. """ with ops.name_scope(name, "Imag", [input]) as name: if input.dtype.is_complex: return gen_math_ops.imag(input, Tout=input.dtype.real_dtype, name=name) else: return array_ops.zeros_like(input) @tf_export("math.angle", v1=["math.angle", "angle"]) @deprecation.deprecated_endpoints("angle") @dispatch.add_dispatch_support def angle(input, name=None): r"""Returns the element-wise argument of a complex (or real) tensor. Given a tensor `input`, this operation returns a tensor of type `float` that is the argument of each element in `input` considered as a complex number. The elements in `input` are considered to be complex numbers of the form \\(a + bj\\), where *a* is the real part and *b* is the imaginary part. If `input` is real then *b* is zero by definition. The argument returned by this function is of the form \\(atan2(b, a)\\). If `input` is real, a tensor of all zeros is returned. For example: ``` # tensor 'input' is [-2.25 + 4.75j, 3.25 + 5.75j] tf.angle(input) ==> [2.0132, 1.056] ``` Args: input: A `Tensor`. Must be one of the following types: `float`, `double`, `complex64`, `complex128`. name: A name for the operation (optional). Returns: A `Tensor` of type `float32` or `float64`. """ with ops.name_scope(name, "Angle", [input]) as name: if input.dtype.is_complex: return gen_math_ops.angle(input, Tout=input.dtype.real_dtype, name=name) else: return array_ops.zeros_like(input) # pylint: enable=redefined-outer-name,redefined-builtin @tf_export("math.round", "round") @dispatch.add_dispatch_support def round(x, name=None): # pylint: disable=redefined-builtin """Rounds the values of a tensor to the nearest integer, element-wise. Rounds half to even. Also known as bankers rounding. If you want to round according to the current system rounding mode use tf::cint. For example: ```python x = tf.constant([0.9, 2.5, 2.3, 1.5, -4.5]) tf.round(x) # [ 1.0, 2.0, 2.0, 2.0, -4.0 ] ``` Args: x: A `Tensor` of type `float16`, `float32`, `float64`, `int32`, or `int64`. name: A name for the operation (optional). Returns: A `Tensor` of same shape and type as `x`. """ x = ops.convert_to_tensor(x, name="x") if x.dtype.is_integer: return x else: return gen_math_ops.round(x, name=name) @tf_export("dtypes.cast", "cast") @dispatch.add_dispatch_support def cast(x, dtype, name=None): """Casts a tensor to a new type. The operation casts `x` (in case of `Tensor`) or `x.values` (in case of `SparseTensor` or `IndexedSlices`) to `dtype`. For example: ```python x = tf.constant([1.8, 2.2], dtype=tf.float32) tf.cast(x, tf.int32) # [1, 2], dtype=tf.int32 ``` The operation supports data types (for `x` and `dtype`) of `uint8`, `uint16`, `uint32`, `uint64`, `int8`, `int16`, `int32`, `int64`, `float16`, `float32`, `float64`, `complex64`, `complex128`, `bfloat16`. In case of casting from complex types (`complex64`, `complex128`) to real types, only the real part of `x` is returned. In case of casting from real types to complex types (`complex64`, `complex128`), the imaginary part of the returned value is set to `0`. The handling of complex types here matches the behavior of numpy. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices` of numeric type. It could be `uint8`, `uint16`, `uint32`, `uint64`, `int8`, `int16`, `int32`, `int64`, `float16`, `float32`, `float64`, `complex64`, `complex128`, `bfloat16`. dtype: The destination type. The list of supported dtypes is the same as `x`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` and same type as `dtype`. Raises: TypeError: If `x` cannot be cast to the `dtype`. """ base_type = dtypes.as_dtype(dtype).base_dtype if isinstance(x, (ops.Tensor, _resource_variable_type)) and base_type == x.dtype: return x with ops.name_scope(name, "Cast", [x]) as name: if isinstance(x, sparse_tensor.SparseTensor): values_cast = cast(x.values, base_type, name=name) x = sparse_tensor.SparseTensor(x.indices, values_cast, x.dense_shape) elif isinstance(x, ops.IndexedSlices): values_cast = cast(x.values, base_type, name=name) x = ops.IndexedSlices(values_cast, x.indices, x.dense_shape) else: # TODO(josh11b): If x is not already a Tensor, we could return # ops.convert_to_tensor(x, dtype=dtype, ...) here, but that # allows some conversions that cast() can't do, e.g. casting numbers to # strings. x = ops.convert_to_tensor(x, name="x") if x.dtype.base_dtype != base_type: x = gen_math_ops.cast(x, base_type, name=name) if x.dtype.is_complex and base_type.is_floating: logging.warn("Casting complex to real discards imaginary part.") return x @tf_export("dtypes.saturate_cast", "saturate_cast") @dispatch.add_dispatch_support def saturate_cast(value, dtype, name=None): """Performs a safe saturating cast of `value` to `dtype`. This function casts the input to `dtype` without applying any scaling. If there is a danger that values would over or underflow in the cast, this op applies the appropriate clamping before the cast. Args: value: A `Tensor`. dtype: The desired output `DType`. name: A name for the operation (optional). Returns: `value` safely cast to `dtype`. """ # When casting to a type with smaller representable range, clamp. # Note that this covers casting to unsigned types as well. with ops.name_scope(name, "saturate_cast", [value]) as name: value = ops.convert_to_tensor(value, name="value") dtype = dtypes.as_dtype(dtype).base_dtype if value.dtype.min < dtype.min: value = gen_math_ops.maximum(value, ops.convert_to_tensor( dtype.min, dtype=value.dtype, name="min")) if value.dtype.max > dtype.max: value = gen_math_ops.minimum(value, ops.convert_to_tensor( dtype.max, dtype=value.dtype, name="max")) return cast(value, dtype, name=name) @deprecation.deprecated(date=None, instructions="Use tf.cast instead.") @tf_export(v1=["to_float"]) def to_float(x, name="ToFloat"): """Casts a tensor to type `float32`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `float32`. Raises: TypeError: If `x` cannot be cast to the `float32`. """ return cast(x, dtypes.float32, name=name) @deprecation.deprecated(date=None, instructions="Use tf.cast instead.") @tf_export(v1=["to_double"]) def to_double(x, name="ToDouble"): """Casts a tensor to type `float64`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `float64`. Raises: TypeError: If `x` cannot be cast to the `float64`. """ return cast(x, dtypes.float64, name=name) @deprecation.deprecated(date=None, instructions="Use tf.cast instead.") @tf_export(v1=["to_int32"]) def to_int32(x, name="ToInt32"): """Casts a tensor to type `int32`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `int32`. Raises: TypeError: If `x` cannot be cast to the `int32`. """ return cast(x, dtypes.int32, name=name) @deprecation.deprecated(date=None, instructions="Use tf.cast instead.") @tf_export(v1=["to_int64"]) def to_int64(x, name="ToInt64"): """Casts a tensor to type `int64`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `int64`. Raises: TypeError: If `x` cannot be cast to the `int64`. """ return cast(x, dtypes.int64, name=name) @deprecation.deprecated(date=None, instructions="Use tf.cast instead.") @tf_export(v1=["to_bfloat16"]) def to_bfloat16(x, name="ToBFloat16"): """Casts a tensor to type `bfloat16`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `bfloat16`. Raises: TypeError: If `x` cannot be cast to the `bfloat16`. """ return cast(x, dtypes.bfloat16, name=name) @deprecation.deprecated(date=None, instructions="Use tf.cast instead.") @tf_export(v1=["to_complex64"]) def to_complex64(x, name="ToComplex64"): """Casts a tensor to type `complex64`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `complex64`. Raises: TypeError: If `x` cannot be cast to the `complex64`. """ return cast(x, dtypes.complex64, name=name) @deprecation.deprecated(date=None, instructions="Use tf.cast instead.") @tf_export(v1=["to_complex128"]) def to_complex128(x, name="ToComplex128"): """Casts a tensor to type `complex128`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `complex128`. Raises: TypeError: If `x` cannot be cast to the `complex128`. """ return cast(x, dtypes.complex128, name=name) ops.Tensor._override_operator("__neg__", gen_math_ops.neg) ops.Tensor._override_operator("__abs__", abs) # __invert__ corresponds to the ~ operator. Here we follow the numpy convention # ~ marks an elementwise bit-wise inverse. This is only implemented for boolean # tensors and will throw a TypeError if used on nonboolean arrays ops.Tensor._override_operator("__invert__", gen_math_ops.logical_not) def _OverrideBinaryOperatorHelper(func, op_name, clazz_object=ops.Tensor): """Register operators with different tensor and scalar versions. If `clazz_object` is `SparseTensor`, assumes `func` takes `(sp_indices, sp_values, sp_shape, dense)` and outputs `(new_sp_values)`. Args: func: the operator op_name: name of the operator being overridden clazz_object: class to override for. Either `Tensor` or `SparseTensor`. """ def binary_op_wrapper(x, y): with ops.name_scope(None, op_name, [x, y]) as name: if isinstance(x, ops.Tensor) and isinstance(y, ops.Tensor): return func(x, y, name=name) elif not isinstance(y, sparse_tensor.SparseTensor): try: y = ops.convert_to_tensor_v2(y, dtype_hint=x.dtype.base_dtype, name="y") except TypeError: # If the RHS is not a tensor, it might be a tensor aware object # that can implement the operator with knowledge of itself # and the tensor. if hasattr(type(y), "__r%s__" % op_name): return NotImplemented else: raise return func(x, y, name=name) def binary_op_wrapper_sparse(sp_x, y): with ops.name_scope(None, op_name, [sp_x, y]) as name: y = ops.convert_to_tensor(y, dtype=sp_x.dtype.base_dtype, name="y") return sparse_tensor.SparseTensor(sp_x.indices, func( sp_x.indices, sp_x.values, sp_x.dense_shape, y, name=name), sp_x.dense_shape) def r_binary_op_wrapper(y, x): with ops.name_scope(None, op_name, [x, y]) as name: x = ops.convert_to_tensor(x, dtype=y.dtype.base_dtype, name="x") return func(x, y, name=name) # Propagate func.__doc__ to the wrappers try: doc = func.__doc__ except AttributeError: doc = None binary_op_wrapper.__doc__ = doc r_binary_op_wrapper.__doc__ = doc binary_op_wrapper_sparse.__doc__ = doc if clazz_object is ops.Tensor: clazz_object._override_operator("__%s__" % op_name, binary_op_wrapper) del binary_op_wrapper clazz_object._override_operator("__r%s__" % op_name, r_binary_op_wrapper) del r_binary_op_wrapper else: clazz_object._override_operator("__%s__" % op_name, binary_op_wrapper_sparse) del binary_op_wrapper_sparse # Conversion table for __truediv__. None entries mean no conversion required. _TRUEDIV_TABLE = { dtypes.uint8: dtypes.float32, dtypes.int8: dtypes.float32, dtypes.uint16: dtypes.float32, dtypes.int16: dtypes.float32, dtypes.int32: dtypes.float64, dtypes.int64: dtypes.float64, dtypes.bfloat16: None, dtypes.float16: None, dtypes.float32: None, dtypes.float64: None, dtypes.complex64: None, dtypes.complex128: None, } # NOTE: the support of "sparse (true)div dense" is currently not baked in into # "tf.(true_)div()". Until such an API decision is made, the supported usage is # to explicitly use the "/" operator to invoke either truediv or div. def _sparse_dense_truediv(sp_indices, sp_values, sp_shape, y, name=None): """Internal helper function for 'sp_t / dense_t'.""" with ops.name_scope(name, "truediv", [sp_indices, sp_values, sp_shape, y]) as name: sp_values = ops.convert_to_tensor(sp_values, name="sp_values") y = ops.convert_to_tensor(y, name="y") x_dtype = sp_values.dtype.base_dtype y_dtype = y.dtype.base_dtype if x_dtype != y_dtype: raise TypeError("x and y must have the same dtype, got %r != %r" % (x_dtype, y_dtype)) try: dtype = _TRUEDIV_TABLE[x_dtype] except KeyError: raise TypeError("Invalid dtype %r in __truediv__" % x_dtype) if dtype is not None: sp_values = cast(sp_values, dtype) y = cast(y, dtype) return gen_sparse_ops.sparse_dense_cwise_div( sp_indices, sp_values, sp_shape, y, name=name) def _truediv_python3(x, y, name=None): with ops.name_scope(name, "truediv", [x, y]) as name: x = ops.convert_to_tensor(x, name="x") y = ops.convert_to_tensor(y, name="y") x_dtype = x.dtype.base_dtype y_dtype = y.dtype.base_dtype if x_dtype != y_dtype: raise TypeError("x and y must have the same dtype, got %r != %r" % (x_dtype, y_dtype)) try: dtype = _TRUEDIV_TABLE[x_dtype] except KeyError: raise TypeError("Invalid dtype %r in __truediv__" % x_dtype) if dtype is not None: x = cast(x, dtype) y = cast(y, dtype) return gen_math_ops.real_div(x, y, name=name) def _div_python2(x, y, name=None): """Divide two values using Python 2 semantics. Used for Tensor.__div__. Args: x: `Tensor` numerator of real numeric type. y: `Tensor` denominator of real numeric type. name: A name for the operation (optional). Returns: `x / y` returns the quotient of x and y. """ with ops.name_scope(name, "div", [x, y]) as name: x = ops.convert_to_tensor(x, name="x") y = ops.convert_to_tensor(y, name="y", dtype=x.dtype.base_dtype) x_dtype = x.dtype.base_dtype y_dtype = y.dtype.base_dtype if x_dtype != y_dtype: raise TypeError("x and y must have the same dtype, got %r != %r" % (x_dtype, y_dtype)) if x_dtype.is_floating or x_dtype.is_complex: return gen_math_ops.real_div(x, y, name=name) else: return gen_math_ops.floor_div(x, y, name=name) @tf_export("math.truediv", "truediv") @dispatch.add_dispatch_support def truediv(x, y, name=None): """Divides x / y elementwise (using Python 3 division operator semantics). NOTE: Prefer using the Tensor operator or tf.divide which obey Python division operator semantics. This function forces Python 3 division operator semantics where all integer arguments are cast to floating types first. This op is generated by normal `x / y` division in Python 3 and in Python 2.7 with `from __future__ import division`. If you want integer division that rounds down, use `x // y` or `tf.math.floordiv`. `x` and `y` must have the same numeric type. If the inputs are floating point, the output will have the same type. If the inputs are integral, the inputs are cast to `float32` for `int8` and `int16` and `float64` for `int32` and `int64` (matching the behavior of Numpy). Args: x: `Tensor` numerator of numeric type. y: `Tensor` denominator of numeric type. name: A name for the operation (optional). Returns: `x / y` evaluated in floating point. Raises: TypeError: If `x` and `y` have different dtypes. """ return _truediv_python3(x, y, name) @deprecation.deprecated( date=None, instructions="Deprecated in favor of operator or tf.math.divide.") @tf_export(v1=["div"]) def div(x, y, name=None): """Divides x / y elementwise (using Python 2 division operator semantics). NOTE: Prefer using the Tensor division operator or tf.divide which obey Python division operator semantics. This function divides `x` and `y`, forcing Python 2.7 semantics. That is, if one of `x` or `y` is a float, then the result will be a float. Otherwise, the output will be an integer type. Flooring semantics are used for integer division. Args: x: `Tensor` numerator of real numeric type. y: `Tensor` denominator of real numeric type. name: A name for the operation (optional). Returns: `x / y` returns the quotient of x and y. """ return _div_python2(x, y, name) @tf_export("div_no_nan") @dispatch.add_dispatch_support def div_no_nan(x, y, name=None): """Computes an unsafe divide which returns 0 if the y is zero. Args: x: A `Tensor`. Must be one of the following types: `float32`, `float64`. y: A `Tensor` whose dtype is compatible with `x`. name: A name for the operation (optional). Returns: The element-wise value of the x divided by y. """ with ops.name_scope(name, "div_no_nan", [x, y]) as name: x = ops.convert_to_tensor(x, name="x") y = ops.convert_to_tensor(y, name="y", dtype=x.dtype.base_dtype) x_dtype = x.dtype.base_dtype y_dtype = y.dtype.base_dtype if x_dtype != y_dtype: raise TypeError("x and y must have the same dtype, got %r != %r" % (x_dtype, y_dtype)) return gen_math_ops.div_no_nan(x, y, name=name) # TODO(aselle): This should be removed mod = gen_math_ops.floor_mod # TODO(aselle): Deprecate this once all internal functionality uses # tf.truncatediv @tf_export("math.floordiv", v1=["math.floordiv", "floordiv"]) @dispatch.add_dispatch_support @deprecation.deprecated_endpoints("floordiv") def floordiv(x, y, name=None): """Divides `x / y` elementwise, rounding toward the most negative integer. The same as `tf.div(x,y)` for integers, but uses `tf.floor(tf.div(x,y))` for floating point arguments so that the result is always an integer (though possibly an integer represented as floating point). This op is generated by `x // y` floor division in Python 3 and in Python 2.7 with `from __future__ import division`. `x` and `y` must have the same type, and the result will have the same type as well. Args: x: `Tensor` numerator of real numeric type. y: `Tensor` denominator of real numeric type. name: A name for the operation (optional). Returns: `x / y` rounded down. Raises: TypeError: If the inputs are complex. """ with ops.name_scope(name, "floordiv", [x, y]) as name: return gen_math_ops.floor_div(x, y, name=name) realdiv = gen_math_ops.real_div truncatediv = gen_math_ops.truncate_div # TODO(aselle): Rename this to floordiv when we can. floor_div = gen_math_ops.floor_div truncatemod = gen_math_ops.truncate_mod floormod = gen_math_ops.floor_mod def _mul_dispatch(x, y, name=None): """Dispatches cwise mul for "Dense*Dense" and "Dense*Sparse".""" is_tensor_y = isinstance(y, ops.Tensor) if is_tensor_y: return gen_math_ops.mul(x, y, name=name) else: assert isinstance(y, sparse_tensor.SparseTensor) # Case: Dense * Sparse. new_vals = gen_sparse_ops.sparse_dense_cwise_mul(y.indices, y.values, y.dense_shape, x, name) return sparse_tensor.SparseTensor(y.indices, new_vals, y.dense_shape) # NOTE(aselle): When integer division is added for sparse_dense_cwise, # div, truediv, and floordiv should be delegated appropriately for # Python sematnics, analogous to dense cwise tensor operations. _OverrideBinaryOperatorHelper(gen_sparse_ops.sparse_dense_cwise_div, "div", sparse_tensor.SparseTensor) _OverrideBinaryOperatorHelper(_sparse_dense_truediv, "truediv", sparse_tensor.SparseTensor) _OverrideBinaryOperatorHelper(gen_sparse_ops.sparse_dense_cwise_mul, "mul", sparse_tensor.SparseTensor) _OverrideBinaryOperatorHelper(gen_math_ops.add, "add") _OverrideBinaryOperatorHelper(gen_math_ops.sub, "sub") _OverrideBinaryOperatorHelper(_mul_dispatch, "mul") _OverrideBinaryOperatorHelper(_div_python2, "div") _OverrideBinaryOperatorHelper(_truediv_python3, "truediv") _OverrideBinaryOperatorHelper(floordiv, "floordiv") _OverrideBinaryOperatorHelper(gen_math_ops.floor_mod, "mod") _OverrideBinaryOperatorHelper(pow, "pow") @tf_export("math.logical_xor", v1=["math.logical_xor", "logical_xor"]) @dispatch.add_dispatch_support @deprecation.deprecated_endpoints("logical_xor") def logical_xor(x, y, name="LogicalXor"): """x ^ y = (x | y) & ~(x & y).""" # TODO(alemi) Make this a cwise op if people end up relying on it. return gen_math_ops.logical_and( gen_math_ops.logical_or(x, y), gen_math_ops.logical_not(gen_math_ops.logical_and(x, y)), name=name) _OverrideBinaryOperatorHelper(gen_math_ops.logical_and, "and") _OverrideBinaryOperatorHelper(gen_math_ops.logical_or, "or") _OverrideBinaryOperatorHelper(logical_xor, "xor") ops.Tensor._override_operator("__lt__", gen_math_ops.less) ops.Tensor._override_operator("__le__", gen_math_ops.less_equal) ops.Tensor._override_operator("__gt__", gen_math_ops.greater) ops.Tensor._override_operator("__ge__", gen_math_ops.greater_equal) @tf_export("range") def range(start, limit=None, delta=1, dtype=None, name="range"): # pylint: disable=redefined-builtin """Creates a sequence of numbers. Creates a sequence of numbers that begins at `start` and extends by increments of `delta` up to but not including `limit`. The dtype of the resulting tensor is inferred from the inputs unless it is provided explicitly. Like the Python builtin `range`, `start` defaults to 0, so that `range(n) = range(0, n)`. For example: ```python start = 3 limit = 18 delta = 3 tf.range(start, limit, delta) # [3, 6, 9, 12, 15] start = 3 limit = 1 delta = -0.5 tf.range(start, limit, delta) # [3, 2.5, 2, 1.5] limit = 5 tf.range(limit) # [0, 1, 2, 3, 4] ``` Args: start: A 0-D `Tensor` (scalar). Acts as first entry in the range if `limit` is not None; otherwise, acts as range limit and first entry defaults to 0. limit: A 0-D `Tensor` (scalar). Upper limit of sequence, exclusive. If None, defaults to the value of `start` while the first entry of the range defaults to 0. delta: A 0-D `Tensor` (scalar). Number that increments `start`. Defaults to 1. dtype: The type of the elements of the resulting tensor. name: A name for the operation. Defaults to "range". Returns: An 1-D `Tensor` of type `dtype`. @compatibility(numpy) Equivalent to np.arange @end_compatibility """ if limit is None: start, limit = 0, start with ops.name_scope(name, "Range", [start, limit, delta]) as name: start = ops.convert_to_tensor(start, dtype=dtype, name="start") limit = ops.convert_to_tensor(limit, dtype=dtype, name="limit") delta = ops.convert_to_tensor(delta, dtype=dtype, name="delta") # infer dtype if not explicitly provided if dtype is None: dtype_hierarchy = [ dtypes.int32, dtypes.int64, dtypes.float32, dtypes.float64 ] assert all(arg.dtype in dtype_hierarchy for arg in [start, limit, delta]) inferred_dtype = max( [arg.dtype for arg in [start, limit, delta]], key=dtype_hierarchy.index) start = cast(start, inferred_dtype) limit = cast(limit, inferred_dtype) delta = cast(delta, inferred_dtype) return gen_math_ops._range(start, limit, delta, name=name) # Reduction operations def _ReductionDims(x, axis, reduction_indices=None): # pylint: disable=invalid-name """Returns range(0, rank(x)) if reduction_indices is None.""" # TODO(aselle): Remove this after deprecation if reduction_indices is not None: if axis is not None: raise ValueError("Can't specify both axis' and 'reduction_indices'.") axis = reduction_indices if axis is not None: return axis else: # Fast path: avoid creating Rank and Range ops if ndims is known. rank = common_shapes.rank(x) if rank is not None: return constant_op.constant(np.arange(rank), dtype=dtypes.int32) if (isinstance(x, sparse_tensor.SparseTensor) and x.dense_shape.shape.is_fully_defined()): rank = x.dense_shape.shape.dims[0].value # sparse.dense_shape is 1-D. return constant_op.constant(np.arange(rank), dtype=dtypes.int32) # Otherwise, we rely on Range and Rank to do the right thing at run-time. return range(0, array_ops.rank(x)) def _may_reduce_to_scalar(keepdims, axis, output): """Set a reduction's output shape to be a scalar if we are certain.""" if not common_shapes.has_fully_defined_shape(output) and (not keepdims) and ( axis is None): output.set_shape(()) return output @tf_export(v1=["math.reduce_sum", "reduce_sum"]) @deprecation.deprecated_args( None, "keep_dims is deprecated, use keepdims instead", "keep_dims") def reduce_sum_v1(input_tensor, axis=None, keepdims=None, name=None, reduction_indices=None, keep_dims=None): """Computes the sum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1, 1, 1], [1, 1, 1]]) tf.reduce_sum(x) # 6 tf.reduce_sum(x, 0) # [2, 2, 2] tf.reduce_sum(x, 1) # [3, 3] tf.reduce_sum(x, 1, keepdims=True) # [[3], [3]] tf.reduce_sum(x, [0, 1]) # 6 ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor, of the same dtype as the input_tensor. @compatibility(numpy) Equivalent to np.sum apart the fact that numpy upcast uint8 and int32 to int64 while tensorflow returns the same dtype as the input. @end_compatibility """ axis = deprecation.deprecated_argument_lookup( "axis", axis, "reduction_indices", reduction_indices) keepdims = deprecation.deprecated_argument_lookup("keepdims", keepdims, "keep_dims", keep_dims) return reduce_sum(input_tensor, axis, keepdims, name) @tf_export("math.reduce_sum", "reduce_sum", v1=[]) @dispatch.add_dispatch_support def reduce_sum(input_tensor, axis=None, keepdims=False, name=None): """Computes the sum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1, 1, 1], [1, 1, 1]]) tf.reduce_sum(x) # 6 tf.reduce_sum(x, 0) # [2, 2, 2] tf.reduce_sum(x, 1) # [3, 3] tf.reduce_sum(x, 1, keepdims=True) # [[3], [3]] tf.reduce_sum(x, [0, 1]) # 6 ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor, of the same dtype as the input_tensor. @compatibility(numpy) Equivalent to np.sum apart the fact that numpy upcast uint8 and int32 to int64 while tensorflow returns the same dtype as the input. @end_compatibility """ keepdims = False if keepdims is None else keepdims return _may_reduce_to_scalar( keepdims, axis, gen_math_ops._sum( input_tensor, _ReductionDims(input_tensor, axis), keepdims, name=name)) @tf_export(v1=["math.count_nonzero", "count_nonzero"]) @deprecation.deprecated_args( None, "keep_dims is deprecated, use keepdims instead", "keep_dims") @deprecation.deprecated_args( None, "reduction_indices is deprecated, use axis instead", "axis") def count_nonzero(input_tensor, axis=None, keepdims=None, dtype=dtypes.int64, name=None, reduction_indices=None, keep_dims=None): """Computes number of nonzero elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` has no entries, all dimensions are reduced, and a tensor with a single element is returned. **NOTE** Floating point comparison to zero is done by exact floating point equality check. Small values are **not** rounded to zero for purposes of the nonzero check. For example: ```python x = tf.constant([[0, 1, 0], [1, 1, 0]]) tf.count_nonzero(x) # 3 tf.count_nonzero(x, 0) # [1, 2, 0] tf.count_nonzero(x, 1) # [1, 2] tf.count_nonzero(x, 1, keepdims=True) # [[1], [2]] tf.count_nonzero(x, [0, 1]) # 3 ``` **NOTE** Strings are compared against zero-length empty string `""`. Any string with a size greater than zero is already considered as nonzero. For example: ```python x = tf.constant(["", "a", " ", "b", ""]) tf.count_nonzero(x) # 3, with "a", " ", and "b" as nonzero strings. ``` Args: input_tensor: The tensor to reduce. Should be of numeric type, `bool`, or `string`. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. dtype: The output dtype; defaults to `tf.int64`. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor (number of nonzero values). """ keepdims = deprecation.deprecated_argument_lookup("keepdims", keepdims, "keep_dims", keep_dims) axis = deprecation.deprecated_argument_lookup( "axis", axis, "reduction_indices", reduction_indices ) return count_nonzero_v2(input_tensor, axis, keepdims, dtype, name) @tf_export("math.count_nonzero", v1=[]) def count_nonzero_v2(input, # pylint: disable=redefined-builtin axis=None, keepdims=None, dtype=dtypes.int64, name=None): """Computes number of nonzero elements across dimensions of a tensor. Reduces `input` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` has no entries, all dimensions are reduced, and a tensor with a single element is returned. **NOTE** Floating point comparison to zero is done by exact floating point equality check. Small values are **not** rounded to zero for purposes of the nonzero check. For example: ```python x = tf.constant([[0, 1, 0], [1, 1, 0]]) tf.count_nonzero(x) # 3 tf.count_nonzero(x, 0) # [1, 2, 0] tf.count_nonzero(x, 1) # [1, 2] tf.count_nonzero(x, 1, keepdims=True) # [[1], [2]] tf.count_nonzero(x, [0, 1]) # 3 ``` **NOTE** Strings are compared against zero-length empty string `""`. Any string with a size greater than zero is already considered as nonzero. For example: ```python x = tf.constant(["", "a", " ", "b", ""]) tf.count_nonzero(x) # 3, with "a", " ", and "b" as nonzero strings. ``` Args: input: The tensor to reduce. Should be of numeric type, `bool`, or `string`. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input), rank(input))`. keepdims: If true, retains reduced dimensions with length 1. dtype: The output dtype; defaults to `tf.int64`. name: A name for the operation (optional). Returns: The reduced tensor (number of nonzero values). """ if keepdims is None: keepdims = False with ops.name_scope(name, "count_nonzero", [input]): input = ops.convert_to_tensor(input, name="input") # A scalar of 'zero' is enough as `not_equal` will broadcast. zero = array_ops.zeros([], dtype=input.dtype) return cast( reduce_sum( # int64 reduction happens on GPU cast(gen_math_ops.not_equal(input, zero), dtypes.int64), axis=axis, keepdims=keepdims), dtype=dtype) @tf_export(v1=["math.reduce_mean", "reduce_mean"]) def reduce_mean_v1(input_tensor, axis=None, keepdims=None, name=None, reduction_indices=None, keep_dims=None): """Computes the mean of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1., 1.], [2., 2.]]) tf.reduce_mean(x) # 1.5 tf.reduce_mean(x, 0) # [1.5, 1.5] tf.reduce_mean(x, 1) # [1., 2.] ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.mean Please note that `np.mean` has a `dtype` parameter that could be used to specify the output type. By default this is `dtype=float64`. On the other hand, `tf.reduce_mean` has an aggressive type inference from `input_tensor`, for example: ```python x = tf.constant([1, 0, 1, 0]) tf.reduce_mean(x) # 0 y = tf.constant([1., 0., 1., 0.]) tf.reduce_mean(y) # 0.5 ``` @end_compatibility """ axis = deprecation.deprecated_argument_lookup( "axis", axis, "reduction_indices", reduction_indices) keepdims = deprecation.deprecated_argument_lookup("keepdims", keepdims, "keep_dims", keep_dims) return reduce_mean(input_tensor, axis, keepdims, name) @tf_export("math.reduce_mean", "reduce_mean", v1=[]) @dispatch.add_dispatch_support def reduce_mean(input_tensor, axis=None, keepdims=False, name=None): """Computes the mean of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1., 1.], [2., 2.]]) tf.reduce_mean(x) # 1.5 tf.reduce_mean(x, 0) # [1.5, 1.5] tf.reduce_mean(x, 1) # [1., 2.] ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.mean Please note that `np.mean` has a `dtype` parameter that could be used to specify the output type. By default this is `dtype=float64`. On the other hand, `tf.reduce_mean` has an aggressive type inference from `input_tensor`, for example: ```python x = tf.constant([1, 0, 1, 0]) tf.reduce_mean(x) # 0 y = tf.constant([1., 0., 1., 0.]) tf.reduce_mean(y) # 0.5 ``` @end_compatibility """ keepdims = False if keepdims is None else keepdims return _may_reduce_to_scalar( keepdims, axis, gen_math_ops.mean( input_tensor, _ReductionDims(input_tensor, axis), keepdims, name=name)) @tf_export("math.reduce_variance") def reduce_variance(input_tensor, axis=None, keepdims=False, name=None): """Computes the variance of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1., 2.], [3., 4.]]) tf.reduce_variance(x) # 1.25 tf.reduce_variance(x, 0) # [1., 1.] tf.reduce_variance(x, 1) # [0.25, 0.25] ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name scope for the associated operations (optional). Returns: The reduced tensor, of the same dtype as the input_tensor. @compatibility(numpy) Equivalent to np.var Please note that `np.var` has a `dtype` parameter that could be used to specify the output type. By default this is `dtype=float64`. On the other hand, `tf.reduce_variance` has an aggressive type inference from `input_tensor`, @end_compatibility """ name = name if name else "reduce_variance" with ops.name_scope(name): means = reduce_mean(input_tensor, axis=axis, keepdims=True) squared_deviations = gen_math_ops.square(input_tensor - means) return reduce_mean(squared_deviations, axis=axis, keepdims=keepdims) @tf_export("math.reduce_std") def reduce_std(input_tensor, axis=None, keepdims=False, name=None): """Computes the standard deviation of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1., 2.], [3., 4.]]) tf.reduce_std(x) # 1.1180339887498949 tf.reduce_std(x, 0) # [1., 1.] tf.reduce_std(x, 1) # [0.5, 0.5] ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name scope for the associated operations (optional). Returns: The reduced tensor, of the same dtype as the input_tensor. @compatibility(numpy) Equivalent to np.std Please note that `np.std` has a `dtype` parameter that could be used to specify the output type. By default this is `dtype=float64`. On the other hand, `tf.reduce_std` has an aggressive type inference from `input_tensor`, @end_compatibility """ name = name if name else "reduce_std" with ops.name_scope(name): variance = reduce_variance(input_tensor, axis=axis, keepdims=keepdims) return gen_math_ops.sqrt(variance) @tf_export("math.reduce_prod", "reduce_prod", v1=[]) @dispatch.add_dispatch_support def reduce_prod(input_tensor, axis=None, keepdims=False, name=None): """Computes the product of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.prod @end_compatibility """ keepdims = False if keepdims is None else keepdims return _may_reduce_to_scalar( keepdims, axis, gen_math_ops.prod( input_tensor, _ReductionDims(input_tensor, axis), keepdims, name=name)) @tf_export(v1=["math.reduce_prod", "reduce_prod"]) @deprecation.deprecated_args( None, "keep_dims is deprecated, use keepdims instead", "keep_dims") def reduce_prod_v1(input_tensor, axis=None, keepdims=None, name=None, reduction_indices=None, keep_dims=None): """Computes the product of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.prod @end_compatibility """ axis = deprecation.deprecated_argument_lookup( "axis", axis, "reduction_indices", reduction_indices) keepdims = deprecation.deprecated_argument_lookup("keepdims", keepdims, "keep_dims", keep_dims) return reduce_prod(input_tensor, axis, keepdims, name) @tf_export(v1=["math.reduce_min", "reduce_min"]) @deprecation.deprecated_args( None, "keep_dims is deprecated, use keepdims instead", "keep_dims") def reduce_min_v1(input_tensor, axis=None, keepdims=None, name=None, reduction_indices=None, keep_dims=None): """Computes the minimum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have real numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.min @end_compatibility """ axis = deprecation.deprecated_argument_lookup( "axis", axis, "reduction_indices", reduction_indices) keepdims = deprecation.deprecated_argument_lookup("keepdims", keepdims, "keep_dims", keep_dims) return reduce_min(input_tensor, axis, keepdims, name) @tf_export("math.reduce_min", "reduce_min", v1=[]) @dispatch.add_dispatch_support def reduce_min(input_tensor, axis=None, keepdims=False, name=None): """Computes the minimum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have real numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.min @end_compatibility """ keepdims = False if keepdims is None else keepdims return _may_reduce_to_scalar( keepdims, axis, gen_math_ops._min( input_tensor, _ReductionDims(input_tensor, axis), keepdims, name=name)) @tf_export(v1=["math.reduce_max", "reduce_max"]) @deprecation.deprecated_args( None, "keep_dims is deprecated, use keepdims instead", "keep_dims") def reduce_max_v1(input_tensor, axis=None, keepdims=None, name=None, reduction_indices=None, keep_dims=None): """Computes the maximum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have real numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.max @end_compatibility """ axis = deprecation.deprecated_argument_lookup( "axis", axis, "reduction_indices", reduction_indices) keepdims = deprecation.deprecated_argument_lookup("keepdims", keepdims, "keep_dims", keep_dims) return reduce_max(input_tensor, axis, keepdims, name) @tf_export("math.reduce_max", "reduce_max", v1=[]) @dispatch.add_dispatch_support def reduce_max(input_tensor, axis=None, keepdims=False, name=None): """Computes the maximum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have real numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.max @end_compatibility """ keepdims = False if keepdims is None else keepdims return _may_reduce_to_scalar( keepdims, axis, gen_math_ops._max( input_tensor, _ReductionDims(input_tensor, axis), keepdims, name=name)) @tf_export(v1=["math.reduce_all", "reduce_all"]) @deprecation.deprecated_args( None, "keep_dims is deprecated, use keepdims instead", "keep_dims") def reduce_all_v1(input_tensor, axis=None, keepdims=None, name=None, reduction_indices=None, keep_dims=None): """Computes the "logical and" of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[True, True], [False, False]]) tf.reduce_all(x) # False tf.reduce_all(x, 0) # [False, False] tf.reduce_all(x, 1) # [True, False] ``` Args: input_tensor: The boolean tensor to reduce. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.all @end_compatibility """ axis = deprecation.deprecated_argument_lookup( "axis", axis, "reduction_indices", reduction_indices) keepdims = deprecation.deprecated_argument_lookup("keepdims", keepdims, "keep_dims", keep_dims) return reduce_all(input_tensor, axis, keepdims, name) @tf_export("reduce_all", "math.reduce_all", v1=[]) @dispatch.add_dispatch_support def reduce_all(input_tensor, axis=None, keepdims=False, name=None): """Computes the "logical and" of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[True, True], [False, False]]) tf.reduce_all(x) # False tf.reduce_all(x, 0) # [False, False] tf.reduce_all(x, 1) # [True, False] ``` Args: input_tensor: The boolean tensor to reduce. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.all @end_compatibility """ keepdims = False if keepdims is None else keepdims return _may_reduce_to_scalar( keepdims, axis, gen_math_ops._all( input_tensor, _ReductionDims(input_tensor, axis), keepdims, name=name)) @tf_export(v1=["math.reduce_any", "reduce_any"]) @deprecation.deprecated_args( None, "keep_dims is deprecated, use keepdims instead", "keep_dims") def reduce_any_v1(input_tensor, axis=None, keepdims=None, name=None, reduction_indices=None, keep_dims=None): """Computes the "logical or" of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[True, True], [False, False]]) tf.reduce_any(x) # True tf.reduce_any(x, 0) # [True, True] tf.reduce_any(x, 1) # [True, False] ``` Args: input_tensor: The boolean tensor to reduce. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.any @end_compatibility """ axis = deprecation.deprecated_argument_lookup( "axis", axis, "reduction_indices", reduction_indices) keepdims = deprecation.deprecated_argument_lookup("keepdims", keepdims, "keep_dims", keep_dims) return reduce_any(input_tensor, axis, keepdims, name) @tf_export("math.reduce_any", "reduce_any", v1=[]) @dispatch.add_dispatch_support def reduce_any(input_tensor, axis=None, keepdims=False, name=None): """Computes the "logical or" of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[True, True], [False, False]]) tf.reduce_any(x) # True tf.reduce_any(x, 0) # [True, True] tf.reduce_any(x, 1) # [True, False] ``` Args: input_tensor: The boolean tensor to reduce. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.any @end_compatibility """ keepdims = False if keepdims is None else keepdims return _may_reduce_to_scalar( keepdims, axis, gen_math_ops._any( input_tensor, _ReductionDims(input_tensor, axis), keepdims, name=name)) @tf_export(v1=["math.reduce_logsumexp", "reduce_logsumexp"]) @deprecation.deprecated_args( None, "keep_dims is deprecated, use keepdims instead", "keep_dims") def reduce_logsumexp_v1(input_tensor, axis=None, keepdims=None, name=None, reduction_indices=None, keep_dims=None): """Computes log(sum(exp(elements across dimensions of a tensor))). Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` has no entries, all dimensions are reduced, and a tensor with a single element is returned. This function is more numerically stable than log(sum(exp(input))). It avoids overflows caused by taking the exp of large inputs and underflows caused by taking the log of small inputs. For example: ```python x = tf.constant([[0., 0., 0.], [0., 0., 0.]]) tf.reduce_logsumexp(x) # log(6) tf.reduce_logsumexp(x, 0) # [log(2), log(2), log(2)] tf.reduce_logsumexp(x, 1) # [log(3), log(3)] tf.reduce_logsumexp(x, 1, keepdims=True) # [[log(3)], [log(3)]] tf.reduce_logsumexp(x, [0, 1]) # log(6) ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. """ axis = deprecation.deprecated_argument_lookup( "axis", axis, "reduction_indices", reduction_indices) keepdims = deprecation.deprecated_argument_lookup("keepdims", keepdims, "keep_dims", keep_dims) return reduce_logsumexp(input_tensor, axis, keepdims, name) @tf_export("math.reduce_logsumexp", "reduce_logsumexp", v1=[]) def reduce_logsumexp(input_tensor, axis=None, keepdims=False, name=None): """Computes log(sum(exp(elements across dimensions of a tensor))). Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` has no entries, all dimensions are reduced, and a tensor with a single element is returned. This function is more numerically stable than log(sum(exp(input))). It avoids overflows caused by taking the exp of large inputs and underflows caused by taking the log of small inputs. For example: ```python x = tf.constant([[0., 0., 0.], [0., 0., 0.]]) tf.reduce_logsumexp(x) # log(6) tf.reduce_logsumexp(x, 0) # [log(2), log(2), log(2)] tf.reduce_logsumexp(x, 1) # [log(3), log(3)] tf.reduce_logsumexp(x, 1, keepdims=True) # [[log(3)], [log(3)]] tf.reduce_logsumexp(x, [0, 1]) # log(6) ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. """ keepdims = False if keepdims is None else keepdims input_tensor = ops.convert_to_tensor(input_tensor) with ops.name_scope(name, "ReduceLogSumExp", [input_tensor]) as name: raw_max = reduce_max( input_tensor, axis=axis, keepdims=True) my_max = array_ops.stop_gradient( array_ops.where( gen_math_ops.is_finite(raw_max), raw_max, array_ops.zeros_like(raw_max))) result = gen_math_ops.log( reduce_sum( gen_math_ops.exp(gen_math_ops.sub(input_tensor, my_max)), axis, keepdims=keepdims)) if not keepdims: my_max = array_ops.reshape(my_max, array_ops.shape(result)) result = gen_math_ops.add(result, my_max) return _may_reduce_to_scalar(keepdims, axis, result) @tf_export("linalg.trace", v1=["linalg.trace", "trace"]) @deprecation.deprecated_endpoints("trace") def trace(x, name=None): """Compute the trace of a tensor `x`. `trace(x)` returns the sum along the main diagonal of each inner-most matrix in x. If x is of rank `k` with shape `[I, J, K, ..., L, M, N]`, then output is a tensor of rank `k-2` with dimensions `[I, J, K, ..., L]` where `output[i, j, k, ..., l] = trace(x[i, j, i, ..., l, :, :])` For example: ```python x = tf.constant([[1, 2], [3, 4]]) tf.linalg.trace(x) # 5 x = tf.constant([[1, 2, 3], [4, 5, 6], [7, 8, 9]]) tf.linalg.trace(x) # 15 x = tf.constant([[[1, 2, 3], [4, 5, 6], [7, 8, 9]], [[-1, -2, -3], [-4, -5, -6], [-7, -8, -9]]]) tf.linalg.trace(x) # [15, -15] ``` Args: x: tensor. name: A name for the operation (optional). Returns: The trace of input tensor. """ with ops.name_scope(name, "Trace", [x]) as name: x = ops.convert_to_tensor(x, name="x") return reduce_sum(array_ops.matrix_diag_part(x), [-1], name=name) @tf_export("linalg.matmul", "matmul") def matmul(a, b, transpose_a=False, transpose_b=False, adjoint_a=False, adjoint_b=False, a_is_sparse=False, b_is_sparse=False, name=None): """Multiplies matrix `a` by matrix `b`, producing `a` * `b`. The inputs must, following any transpositions, be tensors of rank >= 2 where the inner 2 dimensions specify valid matrix multiplication arguments, and any further outer dimensions match. Both matrices must be of the same type. The supported types are: `float16`, `float32`, `float64`, `int32`, `complex64`, `complex128`. Either matrix can be transposed or adjointed (conjugated and transposed) on the fly by setting one of the corresponding flag to `True`. These are `False` by default. If one or both of the matrices contain a lot of zeros, a more efficient multiplication algorithm can be used by setting the corresponding `a_is_sparse` or `b_is_sparse` flag to `True`. These are `False` by default. This optimization is only available for plain matrices (rank-2 tensors) with datatypes `bfloat16` or `float32`. For example: ```python # 2-D tensor `a` # [[1, 2, 3], # [4, 5, 6]] a = tf.constant([1, 2, 3, 4, 5, 6], shape=[2, 3]) # 2-D tensor `b` # [[ 7, 8], # [ 9, 10], # [11, 12]] b = tf.constant([7, 8, 9, 10, 11, 12], shape=[3, 2]) # `a` * `b` # [[ 58, 64], # [139, 154]] c = tf.matmul(a, b) # 3-D tensor `a` # [[[ 1, 2, 3], # [ 4, 5, 6]], # [[ 7, 8, 9], # [10, 11, 12]]] a = tf.constant(np.arange(1, 13, dtype=np.int32), shape=[2, 2, 3]) # 3-D tensor `b` # [[[13, 14], # [15, 16], # [17, 18]], # [[19, 20], # [21, 22], # [23, 24]]] b = tf.constant(np.arange(13, 25, dtype=np.int32), shape=[2, 3, 2]) # `a` * `b` # [[[ 94, 100], # [229, 244]], # [[508, 532], # [697, 730]]] c = tf.matmul(a, b) # Since python >= 3.5 the @ operator is supported (see PEP 465). # In TensorFlow, it simply calls the `tf.matmul()` function, so the # following lines are equivalent: d = a @ b @ [[10.], [11.]] d = tf.matmul(tf.matmul(a, b), [[10.], [11.]]) ``` Args: a: `Tensor` of type `float16`, `float32`, `float64`, `int32`, `complex64`, `complex128` and rank > 1. b: `Tensor` with same type and rank as `a`. transpose_a: If `True`, `a` is transposed before multiplication. transpose_b: If `True`, `b` is transposed before multiplication. adjoint_a: If `True`, `a` is conjugated and transposed before multiplication. adjoint_b: If `True`, `b` is conjugated and transposed before multiplication. a_is_sparse: If `True`, `a` is treated as a sparse matrix. b_is_sparse: If `True`, `b` is treated as a sparse matrix. name: Name for the operation (optional). Returns: A `Tensor` of the same type as `a` and `b` where each inner-most matrix is the product of the corresponding matrices in `a` and `b`, e.g. if all transpose or adjoint attributes are `False`: `output`[..., i, j] = sum_k (`a`[..., i, k] * `b`[..., k, j]), for all indices i, j. Note: This is matrix product, not element-wise product. Raises: ValueError: If transpose_a and adjoint_a, or transpose_b and adjoint_b are both set to True. """ with ops.name_scope(name, "MatMul", [a, b]) as name: if transpose_a and adjoint_a: raise ValueError("Only one of transpose_a and adjoint_a can be True.") if transpose_b and adjoint_b: raise ValueError("Only one of transpose_b and adjoint_b can be True.") if context.executing_eagerly(): if not isinstance(a, (ops.EagerTensor, _resource_variable_type)): a = ops.convert_to_tensor(a, name="a") if not isinstance(b, (ops.EagerTensor, _resource_variable_type)): b = ops.convert_to_tensor(b, name="b") else: a = ops.convert_to_tensor(a, name="a") b = ops.convert_to_tensor(b, name="b") # TODO(apassos) remove _shape_tuple here when it is not needed. a_shape = a._shape_tuple() # pylint: disable=protected-access b_shape = b._shape_tuple() # pylint: disable=protected-access if (not a_is_sparse and not b_is_sparse) and ((a_shape is None or len(a_shape) > 2) and (b_shape is None or len(b_shape) > 2)): # BatchMatmul does not support transpose, so we conjugate the matrix and # use adjoint instead. Conj() is a noop for real matrices. if transpose_a: a = conj(a) adjoint_a = True if transpose_b: b = conj(b) adjoint_b = True return gen_math_ops.batch_mat_mul( a, b, adj_x=adjoint_a, adj_y=adjoint_b, name=name) # Neither matmul nor sparse_matmul support adjoint, so we conjugate # the matrix and use transpose instead. Conj() is a noop for real # matrices. if adjoint_a: a = conj(a) transpose_a = True if adjoint_b: b = conj(b) transpose_b = True use_sparse_matmul = False if a_is_sparse or b_is_sparse: sparse_matmul_types = [dtypes.bfloat16, dtypes.float32] use_sparse_matmul = ( a.dtype in sparse_matmul_types and b.dtype in sparse_matmul_types) if ((a.dtype == dtypes.bfloat16 or b.dtype == dtypes.bfloat16) and a.dtype != b.dtype): # matmul currently doesn't handle mixed-precision inputs. use_sparse_matmul = True if use_sparse_matmul: ret = sparse_matmul( a, b, transpose_a=transpose_a, transpose_b=transpose_b, a_is_sparse=a_is_sparse, b_is_sparse=b_is_sparse, name=name) # sparse_matmul always returns float32, even with # bfloat16 inputs. This prevents us from configuring bfloat16 training. # casting to bfloat16 also matches non-sparse matmul behavior better. if a.dtype == dtypes.bfloat16 and b.dtype == dtypes.bfloat16: ret = cast(ret, dtypes.bfloat16) return ret else: return gen_math_ops.mat_mul( a, b, transpose_a=transpose_a, transpose_b=transpose_b, name=name) @tf_export("linalg.matvec") def matvec(a, b, transpose_a=False, adjoint_a=False, a_is_sparse=False, b_is_sparse=False, name=None): """Multiplies matrix `a` by vector `b`, producing `a` * `b`. The matrix `a` must, following any transpositions, be a tensor of rank >= 2, and we must have `shape(b) = shape(a)[:-2] + [shape(a)[-1]]`. Both `a` and `b` must be of the same type. The supported types are: `float16`, `float32`, `float64`, `int32`, `complex64`, `complex128`. Matrix `a` can be transposed or adjointed (conjugated and transposed) on the fly by setting one of the corresponding flag to `True`. These are `False` by default. If one or both of the inputs contain a lot of zeros, a more efficient multiplication algorithm can be used by setting the corresponding `a_is_sparse` or `b_is_sparse` flag to `True`. These are `False` by default. This optimization is only available for plain matrices/vectors (rank-2/1 tensors) with datatypes `bfloat16` or `float32`. For example: ```python # 2-D tensor `a` # [[1, 2, 3], # [4, 5, 6]] a = tf.constant([1, 2, 3, 4, 5, 6], shape=[2, 3]) # 1-D tensor `b` # [7, 9, 11] b = tf.constant([7, 9, 11], shape=[3]) # `a` * `b` # [ 58, 64] c = tf.matvec(a, b) # 3-D tensor `a` # [[[ 1, 2, 3], # [ 4, 5, 6]], # [[ 7, 8, 9], # [10, 11, 12]]] a = tf.constant(np.arange(1, 13, dtype=np.int32), shape=[2, 2, 3]) # 2-D tensor `b` # [[13, 14, 15], # [16, 17, 18]] b = tf.constant(np.arange(13, 19, dtype=np.int32), shape=[2, 3]) # `a` * `b` # [[ 86, 212], # [410, 563]] c = tf.matvec(a, b) ``` Args: a: `Tensor` of type `float16`, `float32`, `float64`, `int32`, `complex64`, `complex128` and rank > 1. b: `Tensor` with same type and rank = `rank(a) - 1`. transpose_a: If `True`, `a` is transposed before multiplication. adjoint_a: If `True`, `a` is conjugated and transposed before multiplication. a_is_sparse: If `True`, `a` is treated as a sparse matrix. b_is_sparse: If `True`, `b` is treated as a sparse matrix. name: Name for the operation (optional). Returns: A `Tensor` of the same type as `a` and `b` where each inner-most vector is the product of the corresponding matrices in `a` and vectors in `b`, e.g. if all transpose or adjoint attributes are `False`: `output`[..., i] = sum_k (`a`[..., i, k] * `b`[..., k]), for all indices i. Note: This is matrix-vector product, not element-wise product. Raises: ValueError: If transpose_a and adjoint_a are both set to True. """ with ops.name_scope(name, "MatVec", [a, b]) as name: output = matmul( a, array_ops.expand_dims(b, axis=-1), transpose_a=transpose_a, adjoint_a=adjoint_a, a_is_sparse=a_is_sparse, b_is_sparse=b_is_sparse) return array_ops.squeeze(output, axis=-1) _OverrideBinaryOperatorHelper(matmul, "matmul") sparse_matmul = deprecation.deprecated(None, "Use `tf.linalg.matmul` instead")( gen_math_ops.sparse_mat_mul) tf_export(v1=["sparse_matmul"])(sparse_matmul) @ops.RegisterStatistics("MatMul", "flops") def _calc_mat_mul_flops(graph, node): """Calculates the compute resources needed for MatMul.""" transpose_a = node.attr["transpose_a"].b a_shape = graph_util.tensor_shape_from_node_def_name(graph, node.input[0]) a_shape.assert_is_fully_defined() if transpose_a: k = int(a_shape[0]) else: k = int(a_shape[1]) output_shape = graph_util.tensor_shape_from_node_def_name(graph, node.name) output_shape.assert_is_fully_defined() output_count = np.prod(output_shape.as_list()) return ops.OpStats("flops", (k * output_count * 2)) def _as_indexed_slices(x, optimize=True): """Convert 'x' to IndexedSlices. Convert a dense Tensor to a block-sparse IndexedSlices. Args: x: Either a Tensor object, or an IndexedSlices object. optimize: if true, attempt to optimize the conversion of 'x'. Returns: An IndexedSlices object. Raises: TypeError: If 'x' is not a Tensor or an IndexedSlices object. """ # TODO(touts): op_scope if not isinstance(x, (ops.Tensor, ops.IndexedSlices)): raise TypeError("Not a Tensor or IndexedSlices: %s" % type(x)) if isinstance(x, ops.IndexedSlices): return x x_shape = array_ops.shape_internal(x, optimize=optimize) return ops.IndexedSlices(x, range(0, x_shape[0]), x_shape) def _as_indexed_slices_list(inputs, optimize=True): """Convert all elements of 'inputs' to IndexedSlices. Additionally, homogenize the types of all the indices to either int32 or int64. Args: inputs: List containing either Tensor or IndexedSlices objects. optimize: if true, attempt to optimize the conversion of each input. Returns: A list of IndexedSlices objects. Raises: TypeError: If 'inputs' is not a list or a tuple. """ if not isinstance(inputs, (list, tuple)): raise TypeError("Expected a list or tuple, not a %s" % type(inputs)) outputs = [_as_indexed_slices(i, optimize=optimize) for i in inputs] with_int32_index = [ o.indices for o in outputs if o.indices.dtype == dtypes.int32 ] if not with_int32_index or len(with_int32_index) == len(outputs): return outputs casted_outputs = [] for o in outputs: if o.indices.dtype == dtypes.int32: casted_outputs.append( ops.IndexedSlices(o.values, cast(o.indices, dtypes.int64), o.dense_shape)) else: casted_outputs.append(o) return casted_outputs @tf_export("math.add_n", "add_n") @dispatch.add_dispatch_support def add_n(inputs, name=None): """Adds all input tensors element-wise. Converts `IndexedSlices` objects into dense tensors prior to adding. Args: inputs: A list of `Tensor` or `IndexedSlices` objects, each with same shape and type. name: A name for the operation (optional). Returns: A `Tensor` of same shape and type as the elements of `inputs`. Raises: ValueError: If `inputs` don't all have same shape and dtype or the shape cannot be inferred. """ if not inputs or not isinstance(inputs, (list, tuple)): raise ValueError("inputs must be a list of at least one " "Tensor/IndexedSlices with the same dtype and shape") inputs = ops.convert_n_to_tensor_or_indexed_slices(inputs) if not all(isinstance(x, (ops.Tensor, ops.IndexedSlices)) for x in inputs): raise ValueError("inputs must be a list of at least one " "Tensor/IndexedSlices with the same dtype and shape") if len(inputs) == 1: if isinstance(inputs[0], ops.IndexedSlices): values = ops.convert_to_tensor(inputs[0]) else: values = inputs[0] if name: return array_ops.identity(values, name=name) return values return gen_math_ops.add_n(inputs, name=name) @tf_export("math.accumulate_n", v1=["math.accumulate_n", "accumulate_n"]) @deprecation.deprecated_endpoints("accumulate_n") def accumulate_n(inputs, shape=None, tensor_dtype=None, name=None): """Returns the element-wise sum of a list of tensors. Optionally, pass `shape` and `tensor_dtype` for shape and type checking, otherwise, these are inferred. `tf.math.accumulate_n` performs the same operation as `tf.add_n`, but does not wait for all of its inputs to be ready before beginning to sum. This can save memory if inputs are ready at different times, since minimum temporary storage is proportional to the output size rather than the inputs size. `accumulate_n` is differentiable (but wasn't previous to TensorFlow 1.7). For example: ```python a = tf.constant([[1, 2], [3, 4]]) b = tf.constant([[5, 0], [0, 6]]) tf.math.accumulate_n([a, b, a]) # [[7, 4], [6, 14]] # Explicitly pass shape and type tf.math.accumulate_n([a, b, a], shape=[2, 2], tensor_dtype=tf.int32) # [[7, 4], # [6, 14]] ``` Args: inputs: A list of `Tensor` objects, each with same shape and type. shape: Shape of elements of `inputs`. tensor_dtype: The type of `inputs`. name: A name for the operation (optional). Returns: A `Tensor` of same shape and type as the elements of `inputs`. Raises: ValueError: If `inputs` don't all have same shape and dtype or the shape cannot be inferred. """ def _input_error(): return ValueError("inputs must be a list of at least one Tensor with the " "same dtype and shape") if not inputs or not isinstance(inputs, (list, tuple)): raise _input_error() inputs = ops.convert_n_to_tensor_or_indexed_slices(inputs) if not all(isinstance(x, ops.Tensor) for x in inputs): raise _input_error() if not all(x.dtype == inputs[0].dtype for x in inputs): raise _input_error() if shape is not None: shape = tensor_shape.as_shape(shape) else: shape = tensor_shape.unknown_shape() for input_tensor in inputs: if isinstance(input_tensor, ops.Tensor): shape = shape.merge_with(input_tensor.get_shape()) # tensor_dtype is for safety only; operator's output type computed in C++ if tensor_dtype is not None and tensor_dtype != inputs[0].dtype: raise TypeError("tensor_dtype is {}, but input is of type {}".format( tensor_dtype, inputs[0].dtype)) if len(inputs) == 1 and name is None: return inputs[0] elif len(inputs) == 1 and name is not None: return array_ops.identity(inputs[0], name=name) elif context.executing_eagerly(): # TemporaryVariable not currently supported in eager mode; fall back # onto AddN for now. # TODO(frreiss) remove this once the lifetime of eager variables gets # addressed return add_n(inputs, name=name) else: return gen_math_ops.accumulate_nv2(inputs, name=name, shape=shape) # pylint: disable=protected-access @ops.RegisterGradient("AccumulateNV2") def _accumulate_n_grad(op, grad): """Same as gradient for AddN. Copies the gradient to all inputs.""" # Not broadcasting. return [grad] * len(op.inputs) @tf_export("math.sigmoid", "nn.sigmoid", "sigmoid") def sigmoid(x, name=None): """Computes sigmoid of `x` element-wise. Specifically, `y = 1 / (1 + exp(-x))`. Args: x: A Tensor with type `float16`, `float32`, `float64`, `complex64`, or `complex128`. name: A name for the operation (optional). Returns: A Tensor with the same type as `x`. @compatibility(scipy) Equivalent to scipy.special.expit @end_compatibility """ with ops.name_scope(name, "Sigmoid", [x]) as name: x = ops.convert_to_tensor(x, name="x") return gen_math_ops.sigmoid(x, name=name) @tf_export("math.log_sigmoid", v1=["math.log_sigmoid", "log_sigmoid"]) @dispatch.add_dispatch_support @deprecation.deprecated_endpoints("log_sigmoid") def log_sigmoid(x, name=None): """Computes log sigmoid of `x` element-wise. Specifically, `y = log(1 / (1 + exp(-x)))`. For numerical stability, we use `y = -tf.nn.softplus(-x)`. Args: x: A Tensor with type `float32` or `float64`. name: A name for the operation (optional). Returns: A Tensor with the same type as `x`. """ with ops.name_scope(name, "LogSigmoid", [x]) as name: x = ops.convert_to_tensor(x, name="x") return gen_math_ops.neg(gen_nn_ops.softplus(-x), name=name) @tf_export("math.bincount", v1=[]) def bincount(arr, weights=None, minlength=None, maxlength=None, dtype=dtypes.int32, name=None): """Counts the number of occurrences of each value in an integer array. If `minlength` and `maxlength` are not given, returns a vector with length `tf.reduce_max(arr) + 1` if `arr` is non-empty, and length 0 otherwise. If `weights` are non-None, then index `i` of the output stores the sum of the value in `weights` at each index where the corresponding value in `arr` is `i`. Args: arr: An int32 tensor of non-negative values. weights: If non-None, must be the same shape as arr. For each value in `arr`, the bin will be incremented by the corresponding weight instead of 1. minlength: If given, ensures the output has length at least `minlength`, padding with zeros at the end if necessary. maxlength: If given, skips values in `arr` that are equal or greater than `maxlength`, ensuring that the output has length at most `maxlength`. dtype: If `weights` is None, determines the type of the output bins. name: A name scope for the associated operations (optional). Returns: A vector with the same dtype as `weights` or the given `dtype`. The bin values. """ name = "bincount" if name is None else name with ops.name_scope(name): arr = ops.convert_to_tensor(arr, name="arr", dtype=dtypes.int32) array_is_nonempty = reduce_prod(array_ops.shape(arr)) > 0 output_size = cast(array_is_nonempty, dtypes.int32) * (reduce_max(arr) + 1) if minlength is not None: minlength = ops.convert_to_tensor( minlength, name="minlength", dtype=dtypes.int32) output_size = gen_math_ops.maximum(minlength, output_size) if maxlength is not None: maxlength = ops.convert_to_tensor( maxlength, name="maxlength", dtype=dtypes.int32) output_size = gen_math_ops.minimum(maxlength, output_size) if weights is not None: weights = ops.convert_to_tensor(weights, name="weights") return gen_math_ops.unsorted_segment_sum(weights, arr, output_size) weights = constant_op.constant([], dtype) return gen_math_ops.bincount(arr, output_size, weights) @tf_export(v1=["math.bincount", "bincount"]) @deprecation.deprecated_endpoints("bincount") def bincount_v1(arr, weights=None, minlength=None, maxlength=None, dtype=dtypes.int32): """Counts the number of occurrences of each value in an integer array. If `minlength` and `maxlength` are not given, returns a vector with length `tf.reduce_max(arr) + 1` if `arr` is non-empty, and length 0 otherwise. If `weights` are non-None, then index `i` of the output stores the sum of the value in `weights` at each index where the corresponding value in `arr` is `i`. Args: arr: An int32 tensor of non-negative values. weights: If non-None, must be the same shape as arr. For each value in `arr`, the bin will be incremented by the corresponding weight instead of 1. minlength: If given, ensures the output has length at least `minlength`, padding with zeros at the end if necessary. maxlength: If given, skips values in `arr` that are equal or greater than `maxlength`, ensuring that the output has length at most `maxlength`. dtype: If `weights` is None, determines the type of the output bins. Returns: A vector with the same dtype as `weights` or the given `dtype`. The bin values. """ return bincount(arr, weights, minlength, maxlength, dtype) @tf_export("math.cumsum", "cumsum") def cumsum(x, axis=0, exclusive=False, reverse=False, name=None): """Compute the cumulative sum of the tensor `x` along `axis`. By default, this op performs an inclusive cumsum, which means that the first element of the input is identical to the first element of the output: ```python tf.cumsum([a, b, c]) # [a, a + b, a + b + c] ``` By setting the `exclusive` kwarg to `True`, an exclusive cumsum is performed instead: ```python tf.cumsum([a, b, c], exclusive=True) # [0, a, a + b] ``` By setting the `reverse` kwarg to `True`, the cumsum is performed in the opposite direction: ```python tf.cumsum([a, b, c], reverse=True) # [a + b + c, b + c, c] ``` This is more efficient than using separate `tf.reverse` ops. The `reverse` and `exclusive` kwargs can also be combined: ```python tf.cumsum([a, b, c], exclusive=True, reverse=True) # [b + c, c, 0] ``` Args: x: A `Tensor`. Must be one of the following types: `float32`, `float64`, `int64`, `int32`, `uint8`, `uint16`, `int16`, `int8`, `complex64`, `complex128`, `qint8`, `quint8`, `qint32`, `half`. axis: A `Tensor` of type `int32` (default: 0). Must be in the range `[-rank(x), rank(x))`. exclusive: If `True`, perform exclusive cumsum. reverse: A `bool` (default: False). name: A name for the operation (optional). Returns: A `Tensor`. Has the same type as `x`. """ with ops.name_scope(name, "Cumsum", [x]) as name: x = ops.convert_to_tensor(x, name="x") return gen_math_ops.cumsum( x, axis, exclusive=exclusive, reverse=reverse, name=name) @tf_export("math.cumprod", v1=["math.cumprod", "cumprod"]) @deprecation.deprecated_endpoints("cumprod") def cumprod(x, axis=0, exclusive=False, reverse=False, name=None): """Compute the cumulative product of the tensor `x` along `axis`. By default, this op performs an inclusive cumprod, which means that the first element of the input is identical to the first element of the output: ```python tf.math.cumprod([a, b, c]) # [a, a * b, a * b * c] ``` By setting the `exclusive` kwarg to `True`, an exclusive cumprod is performed instead: ```python tf.math.cumprod([a, b, c], exclusive=True) # [1, a, a * b] ``` By setting the `reverse` kwarg to `True`, the cumprod is performed in the opposite direction: ```python tf.math.cumprod([a, b, c], reverse=True) # [a * b * c, b * c, c] ``` This is more efficient than using separate `tf.reverse` ops. The `reverse` and `exclusive` kwargs can also be combined: ```python tf.math.cumprod([a, b, c], exclusive=True, reverse=True) # [b * c, c, 1] ``` Args: x: A `Tensor`. Must be one of the following types: `float32`, `float64`, `int64`, `int32`, `uint8`, `uint16`, `int16`, `int8`, `complex64`, `complex128`, `qint8`, `quint8`, `qint32`, `half`. axis: A `Tensor` of type `int32` (default: 0). Must be in the range `[-rank(x), rank(x))`. exclusive: If `True`, perform exclusive cumprod. reverse: A `bool` (default: False). name: A name for the operation (optional). Returns: A `Tensor`. Has the same type as `x`. """ with ops.name_scope(name, "Cumprod", [x]) as name: x = ops.convert_to_tensor(x, name="x") return gen_math_ops.cumprod( x, axis, exclusive=exclusive, reverse=reverse, name=name) @tf_export("math.conj", v1=["math.conj", "conj"]) @dispatch.add_dispatch_support @deprecation.deprecated_endpoints("conj") def conj(x, name=None): r"""Returns the complex conjugate of a complex number. Given a tensor `input` of complex numbers, this operation returns a tensor of complex numbers that are the complex conjugate of each element in `input`. The complex numbers in `input` must be of the form \\(a + bj\\), where *a* is the real part and *b* is the imaginary part. The complex conjugate returned by this operation is of the form \\(a - bj\\). For example: # tensor 'input' is [-2.25 + 4.75j, 3.25 + 5.75j] tf.math.conj(input) ==> [-2.25 - 4.75j, 3.25 - 5.75j] If `x` is real, it is returned unchanged. Args: x: `Tensor` to conjugate. Must have numeric or variant type. name: A name for the operation (optional). Returns: A `Tensor` that is the conjugate of `x` (with the same type). Raises: TypeError: If `x` is not a numeric tensor. """ if isinstance(x, ops.Tensor): dt = x.dtype if dt.is_floating or dt.is_integer: return x with ops.name_scope(name, "Conj", [x]) as name: x = ops.convert_to_tensor(x, name="x") if x.dtype.is_complex or x.dtype == dtypes.variant: return gen_math_ops.conj(x, name=name) elif x.dtype.is_floating or x.dtype.is_integer: return x else: raise TypeError( "Expected numeric or variant tensor, got dtype %r" % x.dtype) def _BroadcastShape(op): """Common shape function for binary operators that broadcast their inputs.""" return [ common_shapes.broadcast_shape(op.inputs[0].get_shape(), op.inputs[1].get_shape()) ] def reduced_shape(input_shape, axes): """Helper function for reduction ops. Args: input_shape: 1-D Tensor, the shape of the Tensor being reduced. axes: 1-D Tensor, the reduction axes. Returns: A 1-D Tensor, the output shape as if keepdims were set to True. """ # Example: # cast needed for SparseTensor reductions if context.executing_eagerly(): input_shape = input_shape.numpy() axes = axes.numpy() input_shape[axes] = 1 return input_shape input_shape = cast(input_shape, dtypes.int32) # [2, 3, 5, 7] axes = cast(axes, dtypes.int32) # [1, 2] input_rank = array_ops.size(input_shape) # 4 axes = (axes + input_rank) % input_rank axes_shape = array_ops.shape(axes) # [2] return gen_data_flow_ops.dynamic_stitch( # [2, 1, 1, 7] [ range(input_rank), # [0, 1, 2, 3] axes ], # [1, 2] [ input_shape, # [2, 3, 5, 7] array_ops.fill(axes_shape, 1) ]) # [1, 1] def _unsorted_segment_N(data, segment_ids, num_segments): """ Helper function for unsorted_segment_mean/_sqrtN. Computes the number of segment entries with 0-entries set to 1 to allow division by N. """ # bincount doesn't support negative indices so we use unsorted_segment_sum segment_ids_shape = array_ops.shape_internal(segment_ids) ones_tensor = array_ops.ones(segment_ids_shape, dtype=data.dtype) N = gen_math_ops.unsorted_segment_sum(ones_tensor, segment_ids, num_segments) # add dimensions for all non-reduced axes ndims_output = data.shape.ndims - segment_ids.shape.ndims broadcast_shape = [num_segments] + [1] * ndims_output N = array_ops.reshape(N, broadcast_shape) return gen_math_ops.maximum(N, 1) @tf_export( "math.unsorted_segment_mean", v1=["math.unsorted_segment_mean", "unsorted_segment_mean"]) @deprecation.deprecated_endpoints("unsorted_segment_mean") @dispatch.add_dispatch_support def unsorted_segment_mean(data, segment_ids, num_segments, name=None): r"""Computes the mean along segments of a tensor. Read [the section on segmentation](https://tensorflow.org/api_guides/python/math_ops#segmentation) for an explanation of segments. This operator is similar to the unsorted segment sum operator found [here](../../../api_docs/python/math_ops.md#UnsortedSegmentSum). Instead of computing the sum over segments, it computes the mean of all entries belonging to a segment such that: \\(output_i = 1/N_i \sum_{j...} data[j...]\\) where the sum is over tuples `j...` such that `segment_ids[j...] == i` with \\N_i\\ being the number of occurrences of id \\i\\. If there is no entry for a given segment ID `i`, it outputs 0. If the given segment ID `i` is negative, the value is dropped and will not be added to the sum of the segment. Args: data: A `Tensor` with floating point or complex dtype. segment_ids: An integer tensor whose shape is a prefix of `data.shape`. num_segments: An integer scalar `Tensor`. The number of distinct segment IDs. name: A name for the operation (optional). Returns: A `Tensor`. Has same shape as data, except for the first `segment_ids.rank` dimensions, which are replaced with a single dimension which has size `num_segments`. """ with ops.name_scope(name, "UnsortedSegmentMean"): data = ops.convert_to_tensor(data) segment_ids = ops.convert_to_tensor(segment_ids) N = _unsorted_segment_N(data, segment_ids, num_segments) summed = gen_math_ops.unsorted_segment_sum(data, segment_ids, num_segments) return summed / N @tf_export( "math.unsorted_segment_sqrt_n", v1=["math.unsorted_segment_sqrt_n", "unsorted_segment_sqrt_n"]) @deprecation.deprecated_endpoints("unsorted_segment_sqrt_n") @dispatch.add_dispatch_support def unsorted_segment_sqrt_n(data, segment_ids, num_segments, name=None): r"""Computes the sum along segments of a tensor divided by the sqrt(N). Read [the section on segmentation](https://tensorflow.org/api_guides/python/math_ops#segmentation) for an explanation of segments. This operator is similar to the unsorted segment sum operator found [here](../../../api_docs/python/math_ops.md#UnsortedSegmentSum). Additionally to computing the sum over segments, it divides the results by sqrt(N). \\(output_i = 1/sqrt(N_i) \sum_{j...} data[j...]\\) where the sum is over tuples `j...` such that `segment_ids[j...] == i` with \\N_i\\ being the number of occurrences of id \\i\\. If there is no entry for a given segment ID `i`, it outputs 0. Note that this op only supports floating point and complex dtypes, due to tf.sqrt only supporting these types. If the given segment ID `i` is negative, the value is dropped and will not be added to the sum of the segment. Args: data: A `Tensor` with floating point or complex dtype. segment_ids: An integer tensor whose shape is a prefix of `data.shape`. num_segments: An integer scalar `Tensor`. The number of distinct segment IDs. name: A name for the operation (optional). Returns: A `Tensor`. Has same shape as data, except for the first `segment_ids.rank` dimensions, which are replaced with a single dimension which has size `num_segments`. """ with ops.name_scope(name, "UnsortedSegmentSqrtN"): data = ops.convert_to_tensor(data) segment_ids = ops.convert_to_tensor(segment_ids) N = _unsorted_segment_N(data, segment_ids, num_segments) summed = gen_math_ops.unsorted_segment_sum(data, segment_ids, num_segments) return summed / gen_math_ops.sqrt(N) @tf_export(v1=["sparse.segment_sum", "sparse_segment_sum"]) @deprecation.deprecated_endpoints("sparse_segment_sum") def sparse_segment_sum(data, indices, segment_ids, name=None, num_segments=None): r"""Computes the sum along sparse segments of a tensor. Read [the section on segmentation](https://tensorflow.org/api_guides/python/math_ops#Segmentation) for an explanation of segments. Like `SegmentSum`, but `segment_ids` can have rank less than `data`'s first dimension, selecting a subset of dimension 0, specified by `indices`. `segment_ids` is allowed to have missing ids, in which case the output will be zeros at those indices. In those cases `num_segments` is used to determine the size of the output. For example: ```python c = tf.constant([[1,2,3,4], [-1,-2,-3,-4], [5,6,7,8]]) # Select two rows, one segment. tf.sparse.segment_sum(c, tf.constant([0, 1]), tf.constant([0, 0])) # => [[0 0 0 0]] # Select two rows, two segment. tf.sparse.segment_sum(c, tf.constant([0, 1]), tf.constant([0, 1])) # => [[ 1 2 3 4] # [-1 -2 -3 -4]] # With missing segment ids. tf.sparse.segment_sum(c, tf.constant([0, 1]), tf.constant([0, 2]), num_segments=4) # => [[ 1 2 3 4] # [ 0 0 0 0] # [-1 -2 -3 -4] # [ 0 0 0 0]] # Select all rows, two segments. tf.sparse.segment_sum(c, tf.constant([0, 1, 2]), tf.constant([0, 0, 1])) # => [[0 0 0 0] # [5 6 7 8]] # Which is equivalent to: tf.segment_sum(c, tf.constant([0, 0, 1])) ``` Args: data: A `Tensor` with data that will be assembled in the output. indices: A 1-D `Tensor` with indices into `data`. Has same rank as `segment_ids`. segment_ids: A 1-D `Tensor` with indices into the output `Tensor`. Values should be sorted and can be repeated. name: A name for the operation (optional). num_segments: An optional int32 scalar. Indicates the size of the output `Tensor`. Returns: A `tensor` of the shape as data, except for dimension 0 which has size `k`, the number of segments specified via `num_segments` or inferred for the last element in `segments_ids`. """ if num_segments is not None: return gen_math_ops.sparse_segment_sum_with_num_segments( data=data, indices=indices, segment_ids=segment_ids, num_segments=num_segments, name=name) else: return gen_math_ops.sparse_segment_sum( data=data, indices=indices, segment_ids=segment_ids, name=name) @tf_export("sparse.segment_sum", v1=[]) def sparse_segment_sum_v2(data, indices, segment_ids, num_segments=None, name=None): return sparse_segment_mean( data, indices, segment_ids, name=name, num_segments=num_segments) @tf_export(v1=["sparse.segment_mean", "sparse_segment_mean"]) @deprecation.deprecated_endpoints("sparse_segment_mean") def sparse_segment_mean(data, indices, segment_ids, name=None, num_segments=None): r"""Computes the mean along sparse segments of a tensor. Read [the section on segmentation](https://tensorflow.org/api_guides/python/math_ops#Segmentation) for an explanation of segments. Like `SegmentMean`, but `segment_ids` can have rank less than `data`'s first dimension, selecting a subset of dimension 0, specified by `indices`. `segment_ids` is allowed to have missing ids, in which case the output will be zeros at those indices. In those cases `num_segments` is used to determine the size of the output. Args: data: A `Tensor` with data that will be assembled in the output. indices: A 1-D `Tensor` with indices into `data`. Has same rank as `segment_ids`. segment_ids: A 1-D `Tensor` with indices into the output `Tensor`. Values should be sorted and can be repeated. name: A name for the operation (optional). num_segments: An optional int32 scalar. Indicates the size of the output `Tensor`. Returns: A `tensor` of the shape as data, except for dimension 0 which has size `k`, the number of segments specified via `num_segments` or inferred for the last element in `segments_ids`. """ if num_segments is not None: return gen_math_ops.sparse_segment_mean_with_num_segments( data=data, indices=indices, segment_ids=segment_ids, num_segments=num_segments, name=name) else: return gen_math_ops.sparse_segment_mean( data=data, indices=indices, segment_ids=segment_ids, name=name) @tf_export("sparse.segment_mean", v1=[]) def sparse_segment_mean_v2(data, indices, segment_ids, num_segments=None, name=None): r"""Computes the mean along sparse segments of a tensor. Read [the section on segmentation](https://tensorflow.org/api_guides/python/math_ops#Segmentation) for an explanation of segments. Like `SegmentMean`, but `segment_ids` can have rank less than `data`'s first dimension, selecting a subset of dimension 0, specified by `indices`. `segment_ids` is allowed to have missing ids, in which case the output will be zeros at those indices. In those cases `num_segments` is used to determine the size of the output. Args: data: A `Tensor` with data that will be assembled in the output. indices: A 1-D `Tensor` with indices into `data`. Has same rank as `segment_ids`. segment_ids: A 1-D `Tensor` with indices into the output `Tensor`. Values should be sorted and can be repeated. num_segments: An optional int32 scalar. Indicates the size of the output `Tensor`. name: A name for the operation (optional). Returns: A `tensor` of the shape as data, except for dimension 0 which has size `k`, the number of segments specified via `num_segments` or inferred for the last element in `segments_ids`. """ return sparse_segment_mean( data, indices, segment_ids, name=name, num_segments=num_segments) @tf_export(v1=["sparse.segment_sqrt_n", "sparse_segment_sqrt_n"]) @deprecation.deprecated_endpoints("sparse_segment_sqrt_n") def sparse_segment_sqrt_n(data, indices, segment_ids, name=None, num_segments=None): r"""Computes the sum along sparse segments of a tensor divided by the sqrt(N). `N` is the size of the segment being reduced. Args: data: A `Tensor` with data that will be assembled in the output. indices: A 1-D `Tensor` with indices into `data`. Has same rank as `segment_ids`. segment_ids: A 1-D `Tensor` with indices into the output `Tensor`. Values should be sorted and can be repeated. name: A name for the operation (optional). num_segments: An optional int32 scalar. Indicates the size of the output `Tensor`. Returns: A `tensor` of the shape as data, except for dimension 0 which has size `k`, the number of segments specified via `num_segments` or inferred for the last element in `segments_ids`. """ if num_segments is not None: return gen_math_ops.sparse_segment_sqrt_n_with_num_segments( data=data, indices=indices, segment_ids=segment_ids, num_segments=num_segments, name=name) else: return gen_math_ops.sparse_segment_sqrt_n( data=data, indices=indices, segment_ids=segment_ids, name=name) @tf_export("sparse.segment_sqrt_n", v1=[]) def sparse_segment_sqrt_n_v2(data, indices, segment_ids, num_segments=None, name=None): r"""Computes the sum along sparse segments of a tensor divided by the sqrt(N). `N` is the size of the segment being reduced. Args: data: A `Tensor` with data that will be assembled in the output. indices: A 1-D `Tensor` with indices into `data`. Has same rank as `segment_ids`. segment_ids: A 1-D `Tensor` with indices into the output `Tensor`. Values should be sorted and can be repeated. num_segments: An optional int32 scalar. Indicates the size of the output `Tensor`. name: A name for the operation (optional). Returns: A `tensor` of the shape as data, except for dimension 0 which has size `k`, the number of segments specified via `num_segments` or inferred for the last element in `segments_ids`. """ return sparse_segment_sqrt_n( data, indices, segment_ids, name=name, num_segments=num_segments) @tf_export("tensordot", "linalg.tensordot") def tensordot(a, b, axes, name=None): r"""Tensor contraction of a and b along specified axes. Tensordot (also known as tensor contraction) sums the product of elements from `a` and `b` over the indices specified by `a_axes` and `b_axes`. The lists `a_axes` and `b_axes` specify those pairs of axes along which to contract the tensors. The axis `a_axes[i]` of `a` must have the same dimension as axis `b_axes[i]` of `b` for all `i` in `range(0, len(a_axes))`. The lists `a_axes` and `b_axes` must have identical length and consist of unique integers that specify valid axes for each of the tensors. This operation corresponds to `numpy.tensordot(a, b, axes)`. Example 1: When `a` and `b` are matrices (order 2), the case `axes = 1` is equivalent to matrix multiplication. Example 2: When `a` and `b` are matrices (order 2), the case `axes = [[1], [0]]` is equivalent to matrix multiplication. Example 3: Suppose that \\(a_{ijk}\\) and \\(b_{lmn}\\) represent two tensors of order 3. Then, `contract(a, b, [[0], [2]])` is the order 4 tensor \\(c_{jklm}\\) whose entry corresponding to the indices \\((j,k,l,m)\\) is given by: \\( c_{jklm} = \sum_i a_{ijk} b_{lmi} \\). In general, `order(c) = order(a) + order(b) - 2*len(axes[0])`. Args: a: `Tensor` of type `float32` or `float64`. b: `Tensor` with the same type as `a`. axes: Either a scalar `N`, or a list or an `int32` `Tensor` of shape [2, k]. If axes is a scalar, sum over the last N axes of a and the first N axes of b in order. If axes is a list or `Tensor` the first and second row contain the set of unique integers specifying axes along which the contraction is computed, for `a` and `b`, respectively. The number of axes for `a` and `b` must be equal. name: A name for the operation (optional). Returns: A `Tensor` with the same type as `a`. Raises: ValueError: If the shapes of `a`, `b`, and `axes` are incompatible. IndexError: If the values in axes exceed the rank of the corresponding tensor. """ def _tensordot_reshape(a, axes, flipped=False): """Helper method to perform transpose and reshape for contraction op. This method is helpful in reducing `math_ops.tensordot` to `math_ops.matmul` using `array_ops.transpose` and `array_ops.reshape`. The method takes a tensor and performs the correct transpose and reshape operation for a given set of indices. It returns the reshaped tensor as well as a list of indices necessary to reshape the tensor again after matrix multiplication. Args: a: `Tensor`. axes: List or `int32` `Tensor` of unique indices specifying valid axes of `a`. flipped: An optional `bool`. Defaults to `False`. If `True`, the method assumes that `a` is the second argument in the contraction operation. Returns: A tuple `(reshaped_a, free_dims, free_dims_static)` where `reshaped_a` is the tensor `a` reshaped to allow contraction via `matmul`, `free_dims` is either a list of integers or an `int32` `Tensor`, depending on whether the shape of a is fully specified, and free_dims_static is either a list of integers and None values, or None, representing the inferred static shape of the free dimensions """ if a.get_shape().is_fully_defined() and isinstance(axes, (list, tuple)): shape_a = a.get_shape().as_list() axes = [i if i >= 0 else i + len(shape_a) for i in axes] free = [i for i in xrange(len(shape_a)) if i not in axes] free_dims = [shape_a[i] for i in free] prod_free = int(np.prod([shape_a[i] for i in free])) prod_axes = int(np.prod([shape_a[i] for i in axes])) perm = list(axes) + free if flipped else free + list(axes) new_shape = [prod_axes, prod_free] if flipped else [prod_free, prod_axes] reshaped_a = array_ops.reshape(array_ops.transpose(a, perm), new_shape) return reshaped_a, free_dims, free_dims else: if a.get_shape().ndims is not None and isinstance(axes, (list, tuple)): shape_a = a.get_shape().as_list() axes = [i if i >= 0 else i + len(shape_a) for i in axes] free = [i for i in xrange(len(shape_a)) if i not in axes] axes_dims = [shape_a[i] for i in axes] free_dims = [shape_a[i] for i in free] free_dims_static = free_dims axes = ops.convert_to_tensor(axes, dtype=dtypes.int32, name="axes") free = ops.convert_to_tensor(free, dtype=dtypes.int32, name="free") shape_a = array_ops.shape(a) else: free_dims_static = None shape_a = array_ops.shape(a) rank_a = array_ops.rank(a) axes = ops.convert_to_tensor(axes, dtype=dtypes.int32, name="axes") axes = array_ops.where(axes >= 0, axes, axes + rank_a) free, _ = array_ops.setdiff1d(range(rank_a), axes) free_dims = array_ops.gather(shape_a, free) axes_dims = array_ops.gather(shape_a, axes) prod_free_dims = reduce_prod(free_dims) prod_axes_dims = reduce_prod(axes_dims) if flipped: perm = array_ops.concat([axes, free], 0) new_shape = array_ops.stack([prod_axes_dims, prod_free_dims]) else: perm = array_ops.concat([free, axes], 0) new_shape = array_ops.stack([prod_free_dims, prod_axes_dims]) reshaped_a = array_ops.reshape(array_ops.transpose(a, perm), new_shape) return reshaped_a, free_dims, free_dims_static def _tensordot_axes(a, axes): """Generates two sets of contraction axes for the two tensor arguments.""" a_shape = a.get_shape() if isinstance(axes, compat.integral_types): if axes < 0: raise ValueError("'axes' must be at least 0.") if a_shape.ndims is not None: if axes > a_shape.ndims: raise ValueError("'axes' must not be larger than the number of " "dimensions of tensor %s." % a) return (list(xrange(a_shape.ndims - axes, a_shape.ndims)), list(xrange(axes))) else: rank = array_ops.rank(a) return (range(rank - axes, rank, dtype=dtypes.int32), range(axes, dtype=dtypes.int32)) elif isinstance(axes, (list, tuple)): if len(axes) != 2: raise ValueError("'axes' must be an integer or have length 2.") a_axes = axes[0] b_axes = axes[1] if isinstance(a_axes, compat.integral_types) and \ isinstance(b_axes, compat.integral_types): a_axes = [a_axes] b_axes = [b_axes] if len(a_axes) != len(b_axes): raise ValueError( "Different number of contraction axes 'a' and 'b', %s != %s." % (len(a_axes), len(b_axes))) return a_axes, b_axes else: axes = ops.convert_to_tensor(axes, name="axes", dtype=dtypes.int32) return axes[0], axes[1] with ops.name_scope(name, "Tensordot", [a, b, axes]) as name: a = ops.convert_to_tensor(a, name="a") b = ops.convert_to_tensor(b, name="b") a_axes, b_axes = _tensordot_axes(a, axes) a_reshape, a_free_dims, a_free_dims_static = _tensordot_reshape(a, a_axes) b_reshape, b_free_dims, b_free_dims_static = _tensordot_reshape( b, b_axes, True) ab_matmul = matmul(a_reshape, b_reshape) if isinstance(a_free_dims, list) and isinstance(b_free_dims, list): return array_ops.reshape(ab_matmul, a_free_dims + b_free_dims, name=name) else: a_free_dims = ops.convert_to_tensor(a_free_dims, dtype=dtypes.int32) b_free_dims = ops.convert_to_tensor(b_free_dims, dtype=dtypes.int32) product = array_ops.reshape( ab_matmul, array_ops.concat([a_free_dims, b_free_dims], 0), name=name) if a_free_dims_static is not None and b_free_dims_static is not None: product.set_shape(a_free_dims_static + b_free_dims_static) return product @tf_export("math.polyval") def polyval(coeffs, x, name=None): r"""Computes the elementwise value of a polynomial. If `x` is a tensor and `coeffs` is a list n + 1 tensors, this function returns the value of the n-th order polynomial p(x) = coeffs[n-1] + coeffs[n-2] * x + ... + coeffs[0] * x**(n-1) evaluated using Horner's method, i.e. p(x) = coeffs[n-1] + x * (coeffs[n-2] + ... + x * (coeffs[1] + x * coeffs[0])) Args: coeffs: A list of `Tensor` representing the coefficients of the polynomial. x: A `Tensor` representing the variable of the polynomial. name: A name for the operation (optional). Returns: A `tensor` of the shape as the expression p(x) with usual broadcasting rules for element-wise addition and multiplication applied. @compatibility(numpy) Equivalent to numpy.polyval. @end_compatibility """ with ops.name_scope(name, "polyval", nest.flatten(coeffs) + [x]) as name: x = ops.convert_to_tensor(x, name="x") if len(coeffs) < 1: return array_ops.zeros_like(x, name=name) coeffs = [ ops.convert_to_tensor(coeff, name=("coeff_%d" % index)) for index, coeff in enumerate(coeffs) ] p = coeffs[0] for c in coeffs[1:]: p = c + p * x return p
tensorflow/python/ops/math_ops.py
129,354
Use Python2/Python3 division delegation to implement divide for tensors. Common shape function for binary operators that broadcast their inputs. Register operators with different tensor and scalar versions. If `clazz_object` is `SparseTensor`, assumes `func` takes `(sp_indices, sp_values, sp_shape, dense)` and outputs `(new_sp_values)`. Args: func: the operator op_name: name of the operator being overridden clazz_object: class to override for. Either `Tensor` or `SparseTensor`. Returns range(0, rank(x)) if reduction_indices is None. Construct DivideDelegateWithName. Args: x: Tensor to use as left operand in operator overloads name: The name that is preferred for the op created. Same as gradient for AddN. Copies the gradient to all inputs. Convert 'x' to IndexedSlices. Convert a dense Tensor to a block-sparse IndexedSlices. Args: x: Either a Tensor object, or an IndexedSlices object. optimize: if true, attempt to optimize the conversion of 'x'. Returns: An IndexedSlices object. Raises: TypeError: If 'x' is not a Tensor or an IndexedSlices object. Convert all elements of 'inputs' to IndexedSlices. Additionally, homogenize the types of all the indices to either int32 or int64. Args: inputs: List containing either Tensor or IndexedSlices objects. optimize: if true, attempt to optimize the conversion of each input. Returns: A list of IndexedSlices objects. Raises: TypeError: If 'inputs' is not a list or a tuple. Calculates the compute resources needed for MatMul. Divide two values using Python 2 semantics. Used for Tensor.__div__. Args: x: `Tensor` numerator of real numeric type. y: `Tensor` denominator of real numeric type. name: A name for the operation (optional). Returns: `x / y` returns the quotient of x and y. Set a reduction's output shape to be a scalar if we are certain. Dispatches cwise mul for "Dense*Dense" and "Dense*Sparse". Computes numerical negative value element-wise. I.e., \(y = -x\). Args: x: A `Tensor` or `SparseTensor`. Must be one of the following types: `half`, `float32`, `float64`, `int32`, `int64`, `complex64`, `complex128`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor`, respectively. Has the same type as `x`. Internal helper function for 'sp_t / dense_t'. Generates two sets of contraction axes for the two tensor arguments. Helper method to perform transpose and reshape for contraction op. This method is helpful in reducing `math_ops.tensordot` to `math_ops.matmul` using `array_ops.transpose` and `array_ops.reshape`. The method takes a tensor and performs the correct transpose and reshape operation for a given set of indices. It returns the reshaped tensor as well as a list of indices necessary to reshape the tensor again after matrix multiplication. Args: a: `Tensor`. axes: List or `int32` `Tensor` of unique indices specifying valid axes of `a`. flipped: An optional `bool`. Defaults to `False`. If `True`, the method assumes that `a` is the second argument in the contraction operation. Returns: A tuple `(reshaped_a, free_dims, free_dims_static)` where `reshaped_a` is the tensor `a` reshaped to allow contraction via `matmul`, `free_dims` is either a list of integers or an `int32` `Tensor`, depending on whether the shape of a is fully specified, and free_dims_static is either a list of integers and None values, or None, representing the inferred static shape of the free dimensions Helper function for unsorted_segment_mean/_sqrtN. Computes the number of segment entries with 0-entries set to 1 to allow division by N. Computes the absolute value of a tensor. Given a tensor `x` of complex numbers, this operation returns a tensor of type `float32` or `float64` that is the absolute value of each element in `x`. All elements in `x` must be complex numbers of the form \\(a + bj\\). The absolute value is computed as \\( \sqrt{a^2 + b^2}\\). For example: ```python x = tf.constant([[-2.25 + 4.75j], [-3.25 + 5.75j]]) tf.abs(x) # [5.25594902, 6.60492229] ``` Args: x: A `Tensor` or `SparseTensor` of type `float16`, `float32`, `float64`, `int32`, `int64`, `complex64` or `complex128`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` the same size and type as `x` with absolute values. Note, for `complex64` or `complex128` input, the returned `Tensor` will be of type `float32` or `float64`, respectively. Returns the element-wise sum of a list of tensors. Optionally, pass `shape` and `tensor_dtype` for shape and type checking, otherwise, these are inferred. `tf.math.accumulate_n` performs the same operation as `tf.add_n`, but does not wait for all of its inputs to be ready before beginning to sum. This can save memory if inputs are ready at different times, since minimum temporary storage is proportional to the output size rather than the inputs size. `accumulate_n` is differentiable (but wasn't previous to TensorFlow 1.7). For example: ```python a = tf.constant([[1, 2], [3, 4]]) b = tf.constant([[5, 0], [0, 6]]) tf.math.accumulate_n([a, b, a]) # [[7, 4], [6, 14]] # Explicitly pass shape and type tf.math.accumulate_n([a, b, a], shape=[2, 2], tensor_dtype=tf.int32) # [[7, 4], # [6, 14]] ``` Args: inputs: A list of `Tensor` objects, each with same shape and type. shape: Shape of elements of `inputs`. tensor_dtype: The type of `inputs`. name: A name for the operation (optional). Returns: A `Tensor` of same shape and type as the elements of `inputs`. Raises: ValueError: If `inputs` don't all have same shape and dtype or the shape cannot be inferred. Adds all input tensors element-wise. Converts `IndexedSlices` objects into dense tensors prior to adding. Args: inputs: A list of `Tensor` or `IndexedSlices` objects, each with same shape and type. name: A name for the operation (optional). Returns: A `Tensor` of same shape and type as the elements of `inputs`. Raises: ValueError: If `inputs` don't all have same shape and dtype or the shape cannot be inferred. Returns the element-wise argument of a complex (or real) tensor. Given a tensor `input`, this operation returns a tensor of type `float` that is the argument of each element in `input` considered as a complex number. The elements in `input` are considered to be complex numbers of the form \\(a + bj\\), where *a* is the real part and *b* is the imaginary part. If `input` is real then *b* is zero by definition. The argument returned by this function is of the form \\(atan2(b, a)\\). If `input` is real, a tensor of all zeros is returned. For example: ``` # tensor 'input' is [-2.25 + 4.75j, 3.25 + 5.75j] tf.angle(input) ==> [2.0132, 1.056] ``` Args: input: A `Tensor`. Must be one of the following types: `float`, `double`, `complex64`, `complex128`. name: A name for the operation (optional). Returns: A `Tensor` of type `float32` or `float64`. Returns the index with the largest value across axes of a tensor. Note that in case of ties the identity of the return value is not guaranteed. Args: input: A `Tensor`. Must be one of the following types: `float32`, `float64`, `int32`, `uint8`, `int16`, `int8`, `complex64`, `int64`, `qint8`, `quint8`, `qint32`, `bfloat16`, `uint16`, `complex128`, `half`, `uint32`, `uint64`. axis: A `Tensor`. Must be one of the following types: `int32`, `int64`. int32 or int64, must be in the range `-rank(input), rank(input))`. Describes which axis of the input Tensor to reduce across. For vectors, use axis = 0. output_type: An optional `tf.DType` from: `tf.int32, tf.int64`. Defaults to `tf.int64`. name: A name for the operation (optional). Returns: A `Tensor` of type `output_type`. Returns the index with the smallest value across axes of a tensor. Note that in case of ties the identity of the return value is not guaranteed. Args: input: A `Tensor`. Must be one of the following types: `float32`, `float64`, `int32`, `uint8`, `int16`, `int8`, `complex64`, `int64`, `qint8`, `quint8`, `qint32`, `bfloat16`, `uint16`, `complex128`, `half`, `uint32`, `uint64`. axis: A `Tensor`. Must be one of the following types: `int32`, `int64`. int32 or int64, must be in the range `-rank(input), rank(input))`. Describes which axis of the input Tensor to reduce across. For vectors, use axis = 0. output_type: An optional `tf.DType` from: `tf.int32, tf.int64`. Defaults to `tf.int64`. name: A name for the operation (optional). Returns: A `Tensor` of type `output_type`. Counts the number of occurrences of each value in an integer array. If `minlength` and `maxlength` are not given, returns a vector with length `tf.reduce_max(arr) + 1` if `arr` is non-empty, and length 0 otherwise. If `weights` are non-None, then index `i` of the output stores the sum of the value in `weights` at each index where the corresponding value in `arr` is `i`. Args: arr: An int32 tensor of non-negative values. weights: If non-None, must be the same shape as arr. For each value in `arr`, the bin will be incremented by the corresponding weight instead of 1. minlength: If given, ensures the output has length at least `minlength`, padding with zeros at the end if necessary. maxlength: If given, skips values in `arr` that are equal or greater than `maxlength`, ensuring that the output has length at most `maxlength`. dtype: If `weights` is None, determines the type of the output bins. name: A name scope for the associated operations (optional). Returns: A vector with the same dtype as `weights` or the given `dtype`. The bin values. Counts the number of occurrences of each value in an integer array. If `minlength` and `maxlength` are not given, returns a vector with length `tf.reduce_max(arr) + 1` if `arr` is non-empty, and length 0 otherwise. If `weights` are non-None, then index `i` of the output stores the sum of the value in `weights` at each index where the corresponding value in `arr` is `i`. Args: arr: An int32 tensor of non-negative values. weights: If non-None, must be the same shape as arr. For each value in `arr`, the bin will be incremented by the corresponding weight instead of 1. minlength: If given, ensures the output has length at least `minlength`, padding with zeros at the end if necessary. maxlength: If given, skips values in `arr` that are equal or greater than `maxlength`, ensuring that the output has length at most `maxlength`. dtype: If `weights` is None, determines the type of the output bins. Returns: A vector with the same dtype as `weights` or the given `dtype`. The bin values. Casts a tensor to a new type. The operation casts `x` (in case of `Tensor`) or `x.values` (in case of `SparseTensor` or `IndexedSlices`) to `dtype`. For example: ```python x = tf.constant([1.8, 2.2], dtype=tf.float32) tf.cast(x, tf.int32) # [1, 2], dtype=tf.int32 ``` The operation supports data types (for `x` and `dtype`) of `uint8`, `uint16`, `uint32`, `uint64`, `int8`, `int16`, `int32`, `int64`, `float16`, `float32`, `float64`, `complex64`, `complex128`, `bfloat16`. In case of casting from complex types (`complex64`, `complex128`) to real types, only the real part of `x` is returned. In case of casting from real types to complex types (`complex64`, `complex128`), the imaginary part of the returned value is set to `0`. The handling of complex types here matches the behavior of numpy. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices` of numeric type. It could be `uint8`, `uint16`, `uint32`, `uint64`, `int8`, `int16`, `int32`, `int64`, `float16`, `float32`, `float64`, `complex64`, `complex128`, `bfloat16`. dtype: The destination type. The list of supported dtypes is the same as `x`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` and same type as `dtype`. Raises: TypeError: If `x` cannot be cast to the `dtype`. Converts two real numbers to a complex number. Given a tensor `real` representing the real part of a complex number, and a tensor `imag` representing the imaginary part of a complex number, this operation returns complex numbers elementwise of the form \\(a + bj\\), where *a* represents the `real` part and *b* represents the `imag` part. The input tensors `real` and `imag` must have the same shape. For example: ```python real = tf.constant([2.25, 3.25]) imag = tf.constant([4.75, 5.75]) tf.complex(real, imag) # [[2.25 + 4.75j], [3.25 + 5.75j]] ``` Args: real: A `Tensor`. Must be one of the following types: `float32`, `float64`. imag: A `Tensor`. Must have the same type as `real`. name: A name for the operation (optional). Returns: A `Tensor` of type `complex64` or `complex128`. Returns the complex conjugate of a complex number. Given a tensor `input` of complex numbers, this operation returns a tensor of complex numbers that are the complex conjugate of each element in `input`. The complex numbers in `input` must be of the form \\(a + bj\\), where *a* is the real part and *b* is the imaginary part. The complex conjugate returned by this operation is of the form \\(a - bj\\). For example: # tensor 'input' is [-2.25 + 4.75j, 3.25 + 5.75j] tf.math.conj(input) ==> [-2.25 - 4.75j, 3.25 - 5.75j] If `x` is real, it is returned unchanged. Args: x: `Tensor` to conjugate. Must have numeric or variant type. name: A name for the operation (optional). Returns: A `Tensor` that is the conjugate of `x` (with the same type). Raises: TypeError: If `x` is not a numeric tensor. Computes number of nonzero elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` has no entries, all dimensions are reduced, and a tensor with a single element is returned. **NOTE** Floating point comparison to zero is done by exact floating point equality check. Small values are **not** rounded to zero for purposes of the nonzero check. For example: ```python x = tf.constant([[0, 1, 0], [1, 1, 0]]) tf.count_nonzero(x) # 3 tf.count_nonzero(x, 0) # [1, 2, 0] tf.count_nonzero(x, 1) # [1, 2] tf.count_nonzero(x, 1, keepdims=True) # [[1], [2]] tf.count_nonzero(x, [0, 1]) # 3 ``` **NOTE** Strings are compared against zero-length empty string `""`. Any string with a size greater than zero is already considered as nonzero. For example: ```python x = tf.constant(["", "a", " ", "b", ""]) tf.count_nonzero(x) # 3, with "a", " ", and "b" as nonzero strings. ``` Args: input_tensor: The tensor to reduce. Should be of numeric type, `bool`, or `string`. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. dtype: The output dtype; defaults to `tf.int64`. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor (number of nonzero values). Computes number of nonzero elements across dimensions of a tensor. Reduces `input` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` has no entries, all dimensions are reduced, and a tensor with a single element is returned. **NOTE** Floating point comparison to zero is done by exact floating point equality check. Small values are **not** rounded to zero for purposes of the nonzero check. For example: ```python x = tf.constant([[0, 1, 0], [1, 1, 0]]) tf.count_nonzero(x) # 3 tf.count_nonzero(x, 0) # [1, 2, 0] tf.count_nonzero(x, 1) # [1, 2] tf.count_nonzero(x, 1, keepdims=True) # [[1], [2]] tf.count_nonzero(x, [0, 1]) # 3 ``` **NOTE** Strings are compared against zero-length empty string `""`. Any string with a size greater than zero is already considered as nonzero. For example: ```python x = tf.constant(["", "a", " ", "b", ""]) tf.count_nonzero(x) # 3, with "a", " ", and "b" as nonzero strings. ``` Args: input: The tensor to reduce. Should be of numeric type, `bool`, or `string`. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input), rank(input))`. keepdims: If true, retains reduced dimensions with length 1. dtype: The output dtype; defaults to `tf.int64`. name: A name for the operation (optional). Returns: The reduced tensor (number of nonzero values). Compute the cumulative product of the tensor `x` along `axis`. By default, this op performs an inclusive cumprod, which means that the first element of the input is identical to the first element of the output: ```python tf.math.cumprod([a, b, c]) # [a, a * b, a * b * c] ``` By setting the `exclusive` kwarg to `True`, an exclusive cumprod is performed instead: ```python tf.math.cumprod([a, b, c], exclusive=True) # [1, a, a * b] ``` By setting the `reverse` kwarg to `True`, the cumprod is performed in the opposite direction: ```python tf.math.cumprod([a, b, c], reverse=True) # [a * b * c, b * c, c] ``` This is more efficient than using separate `tf.reverse` ops. The `reverse` and `exclusive` kwargs can also be combined: ```python tf.math.cumprod([a, b, c], exclusive=True, reverse=True) # [b * c, c, 1] ``` Args: x: A `Tensor`. Must be one of the following types: `float32`, `float64`, `int64`, `int32`, `uint8`, `uint16`, `int16`, `int8`, `complex64`, `complex128`, `qint8`, `quint8`, `qint32`, `half`. axis: A `Tensor` of type `int32` (default: 0). Must be in the range `[-rank(x), rank(x))`. exclusive: If `True`, perform exclusive cumprod. reverse: A `bool` (default: False). name: A name for the operation (optional). Returns: A `Tensor`. Has the same type as `x`. Compute the cumulative sum of the tensor `x` along `axis`. By default, this op performs an inclusive cumsum, which means that the first element of the input is identical to the first element of the output: ```python tf.cumsum([a, b, c]) # [a, a + b, a + b + c] ``` By setting the `exclusive` kwarg to `True`, an exclusive cumsum is performed instead: ```python tf.cumsum([a, b, c], exclusive=True) # [0, a, a + b] ``` By setting the `reverse` kwarg to `True`, the cumsum is performed in the opposite direction: ```python tf.cumsum([a, b, c], reverse=True) # [a + b + c, b + c, c] ``` This is more efficient than using separate `tf.reverse` ops. The `reverse` and `exclusive` kwargs can also be combined: ```python tf.cumsum([a, b, c], exclusive=True, reverse=True) # [b + c, c, 0] ``` Args: x: A `Tensor`. Must be one of the following types: `float32`, `float64`, `int64`, `int32`, `uint8`, `uint16`, `int16`, `int8`, `complex64`, `complex128`, `qint8`, `quint8`, `qint32`, `half`. axis: A `Tensor` of type `int32` (default: 0). Must be in the range `[-rank(x), rank(x))`. exclusive: If `True`, perform exclusive cumsum. reverse: A `bool` (default: False). name: A name for the operation (optional). Returns: A `Tensor`. Has the same type as `x`. Divides x / y elementwise (using Python 2 division operator semantics). NOTE: Prefer using the Tensor division operator or tf.divide which obey Python division operator semantics. This function divides `x` and `y`, forcing Python 2.7 semantics. That is, if one of `x` or `y` is a float, then the result will be a float. Otherwise, the output will be an integer type. Flooring semantics are used for integer division. Args: x: `Tensor` numerator of real numeric type. y: `Tensor` denominator of real numeric type. name: A name for the operation (optional). Returns: `x / y` returns the quotient of x and y. Computes an unsafe divide which returns 0 if the y is zero. Args: x: A `Tensor`. Must be one of the following types: `float32`, `float64`. y: A `Tensor` whose dtype is compatible with `x`. name: A name for the operation (optional). Returns: The element-wise value of the x divided by y. Computes Python style division of `x` by `y`. Divides `x / y` elementwise, rounding toward the most negative integer. The same as `tf.div(x,y)` for integers, but uses `tf.floor(tf.div(x,y))` for floating point arguments so that the result is always an integer (though possibly an integer represented as floating point). This op is generated by `x // y` floor division in Python 3 and in Python 2.7 with `from __future__ import division`. `x` and `y` must have the same type, and the result will have the same type as well. Args: x: `Tensor` numerator of real numeric type. y: `Tensor` denominator of real numeric type. name: A name for the operation (optional). Returns: `x / y` rounded down. Raises: TypeError: If the inputs are complex. Returns the imaginary part of a complex (or real) tensor. Given a tensor `input`, this operation returns a tensor of type `float` that is the imaginary part of each element in `input` considered as a complex number. If `input` is real, a tensor of all zeros is returned. For example: ```python x = tf.constant([-2.25 + 4.75j, 3.25 + 5.75j]) tf.imag(x) # [4.75, 5.75] ``` Args: input: A `Tensor`. Must be one of the following types: `float`, `double`, `complex64`, `complex128`. name: A name for the operation (optional). Returns: A `Tensor` of type `float32` or `float64`. Computes log sigmoid of `x` element-wise. Specifically, `y = log(1 / (1 + exp(-x)))`. For numerical stability, we use `y = -tf.nn.softplus(-x)`. Args: x: A Tensor with type `float32` or `float64`. name: A name for the operation (optional). Returns: A Tensor with the same type as `x`. x ^ y = (x | y) & ~(x & y). Multiplies matrix `a` by matrix `b`, producing `a` * `b`. The inputs must, following any transpositions, be tensors of rank >= 2 where the inner 2 dimensions specify valid matrix multiplication arguments, and any further outer dimensions match. Both matrices must be of the same type. The supported types are: `float16`, `float32`, `float64`, `int32`, `complex64`, `complex128`. Either matrix can be transposed or adjointed (conjugated and transposed) on the fly by setting one of the corresponding flag to `True`. These are `False` by default. If one or both of the matrices contain a lot of zeros, a more efficient multiplication algorithm can be used by setting the corresponding `a_is_sparse` or `b_is_sparse` flag to `True`. These are `False` by default. This optimization is only available for plain matrices (rank-2 tensors) with datatypes `bfloat16` or `float32`. For example: ```python # 2-D tensor `a` # [[1, 2, 3], # [4, 5, 6]] a = tf.constant([1, 2, 3, 4, 5, 6], shape=[2, 3]) # 2-D tensor `b` # [[ 7, 8], # [ 9, 10], # [11, 12]] b = tf.constant([7, 8, 9, 10, 11, 12], shape=[3, 2]) # `a` * `b` # [[ 58, 64], # [139, 154]] c = tf.matmul(a, b) # 3-D tensor `a` # [[[ 1, 2, 3], # [ 4, 5, 6]], # [[ 7, 8, 9], # [10, 11, 12]]] a = tf.constant(np.arange(1, 13, dtype=np.int32), shape=[2, 2, 3]) # 3-D tensor `b` # [[[13, 14], # [15, 16], # [17, 18]], # [[19, 20], # [21, 22], # [23, 24]]] b = tf.constant(np.arange(13, 25, dtype=np.int32), shape=[2, 3, 2]) # `a` * `b` # [[[ 94, 100], # [229, 244]], # [[508, 532], # [697, 730]]] c = tf.matmul(a, b) # Since python >= 3.5 the @ operator is supported (see PEP 465). # In TensorFlow, it simply calls the `tf.matmul()` function, so the # following lines are equivalent: d = a @ b @ [[10.], [11.]] d = tf.matmul(tf.matmul(a, b), [[10.], [11.]]) ``` Args: a: `Tensor` of type `float16`, `float32`, `float64`, `int32`, `complex64`, `complex128` and rank > 1. b: `Tensor` with same type and rank as `a`. transpose_a: If `True`, `a` is transposed before multiplication. transpose_b: If `True`, `b` is transposed before multiplication. adjoint_a: If `True`, `a` is conjugated and transposed before multiplication. adjoint_b: If `True`, `b` is conjugated and transposed before multiplication. a_is_sparse: If `True`, `a` is treated as a sparse matrix. b_is_sparse: If `True`, `b` is treated as a sparse matrix. name: Name for the operation (optional). Returns: A `Tensor` of the same type as `a` and `b` where each inner-most matrix is the product of the corresponding matrices in `a` and `b`, e.g. if all transpose or adjoint attributes are `False`: `output`[..., i, j] = sum_k (`a`[..., i, k] * `b`[..., k, j]), for all indices i, j. Note: This is matrix product, not element-wise product. Raises: ValueError: If transpose_a and adjoint_a, or transpose_b and adjoint_b are both set to True. Multiplies matrix `a` by vector `b`, producing `a` * `b`. The matrix `a` must, following any transpositions, be a tensor of rank >= 2, and we must have `shape(b) = shape(a)[:-2] + [shape(a)[-1]]`. Both `a` and `b` must be of the same type. The supported types are: `float16`, `float32`, `float64`, `int32`, `complex64`, `complex128`. Matrix `a` can be transposed or adjointed (conjugated and transposed) on the fly by setting one of the corresponding flag to `True`. These are `False` by default. If one or both of the inputs contain a lot of zeros, a more efficient multiplication algorithm can be used by setting the corresponding `a_is_sparse` or `b_is_sparse` flag to `True`. These are `False` by default. This optimization is only available for plain matrices/vectors (rank-2/1 tensors) with datatypes `bfloat16` or `float32`. For example: ```python # 2-D tensor `a` # [[1, 2, 3], # [4, 5, 6]] a = tf.constant([1, 2, 3, 4, 5, 6], shape=[2, 3]) # 1-D tensor `b` # [7, 9, 11] b = tf.constant([7, 9, 11], shape=[3]) # `a` * `b` # [ 58, 64] c = tf.matvec(a, b) # 3-D tensor `a` # [[[ 1, 2, 3], # [ 4, 5, 6]], # [[ 7, 8, 9], # [10, 11, 12]]] a = tf.constant(np.arange(1, 13, dtype=np.int32), shape=[2, 2, 3]) # 2-D tensor `b` # [[13, 14, 15], # [16, 17, 18]] b = tf.constant(np.arange(13, 19, dtype=np.int32), shape=[2, 3]) # `a` * `b` # [[ 86, 212], # [410, 563]] c = tf.matvec(a, b) ``` Args: a: `Tensor` of type `float16`, `float32`, `float64`, `int32`, `complex64`, `complex128` and rank > 1. b: `Tensor` with same type and rank = `rank(a) - 1`. transpose_a: If `True`, `a` is transposed before multiplication. adjoint_a: If `True`, `a` is conjugated and transposed before multiplication. a_is_sparse: If `True`, `a` is treated as a sparse matrix. b_is_sparse: If `True`, `b` is treated as a sparse matrix. name: Name for the operation (optional). Returns: A `Tensor` of the same type as `a` and `b` where each inner-most vector is the product of the corresponding matrices in `a` and vectors in `b`, e.g. if all transpose or adjoint attributes are `False`: `output`[..., i] = sum_k (`a`[..., i, k] * `b`[..., k]), for all indices i. Note: This is matrix-vector product, not element-wise product. Raises: ValueError: If transpose_a and adjoint_a are both set to True. Computes the elementwise value of a polynomial. If `x` is a tensor and `coeffs` is a list n + 1 tensors, this function returns the value of the n-th order polynomial p(x) = coeffs[n-1] + coeffs[n-2] * x + ... + coeffs[0] * x**(n-1) evaluated using Horner's method, i.e. p(x) = coeffs[n-1] + x * (coeffs[n-2] + ... + x * (coeffs[1] + x * coeffs[0])) Args: coeffs: A list of `Tensor` representing the coefficients of the polynomial. x: A `Tensor` representing the variable of the polynomial. name: A name for the operation (optional). Returns: A `tensor` of the shape as the expression p(x) with usual broadcasting rules for element-wise addition and multiplication applied. @compatibility(numpy) Equivalent to numpy.polyval. @end_compatibility Computes the power of one value to another. Given a tensor `x` and a tensor `y`, this operation computes \\(x^y\\) for corresponding elements in `x` and `y`. For example: ```python x = tf.constant([[2, 2], [3, 3]]) y = tf.constant([[8, 16], [2, 3]]) tf.pow(x, y) # [[256, 65536], [9, 27]] ``` Args: x: A `Tensor` of type `float16`, `float32`, `float64`, `int32`, `int64`, `complex64`, or `complex128`. y: A `Tensor` of type `float16`, `float32`, `float64`, `int32`, `int64`, `complex64`, or `complex128`. name: A name for the operation (optional). Returns: A `Tensor`. Creates a sequence of numbers. Creates a sequence of numbers that begins at `start` and extends by increments of `delta` up to but not including `limit`. The dtype of the resulting tensor is inferred from the inputs unless it is provided explicitly. Like the Python builtin `range`, `start` defaults to 0, so that `range(n) = range(0, n)`. For example: ```python start = 3 limit = 18 delta = 3 tf.range(start, limit, delta) # [3, 6, 9, 12, 15] start = 3 limit = 1 delta = -0.5 tf.range(start, limit, delta) # [3, 2.5, 2, 1.5] limit = 5 tf.range(limit) # [0, 1, 2, 3, 4] ``` Args: start: A 0-D `Tensor` (scalar). Acts as first entry in the range if `limit` is not None; otherwise, acts as range limit and first entry defaults to 0. limit: A 0-D `Tensor` (scalar). Upper limit of sequence, exclusive. If None, defaults to the value of `start` while the first entry of the range defaults to 0. delta: A 0-D `Tensor` (scalar). Number that increments `start`. Defaults to 1. dtype: The type of the elements of the resulting tensor. name: A name for the operation. Defaults to "range". Returns: An 1-D `Tensor` of type `dtype`. @compatibility(numpy) Equivalent to np.arange @end_compatibility Returns the real part of a complex (or real) tensor. Given a tensor `input`, this operation returns a tensor of type `float` that is the real part of each element in `input` considered as a complex number. For example: ```python x = tf.constant([-2.25 + 4.75j, 3.25 + 5.75j]) tf.real(x) # [-2.25, 3.25] ``` If `input` is already real, it is returned unchanged. Args: input: A `Tensor`. Must have numeric type. name: A name for the operation (optional). Returns: A `Tensor` of type `float32` or `float64`. Computes the "logical and" of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[True, True], [False, False]]) tf.reduce_all(x) # False tf.reduce_all(x, 0) # [False, False] tf.reduce_all(x, 1) # [True, False] ``` Args: input_tensor: The boolean tensor to reduce. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.all @end_compatibility Computes the "logical and" of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[True, True], [False, False]]) tf.reduce_all(x) # False tf.reduce_all(x, 0) # [False, False] tf.reduce_all(x, 1) # [True, False] ``` Args: input_tensor: The boolean tensor to reduce. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.all @end_compatibility Computes the "logical or" of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[True, True], [False, False]]) tf.reduce_any(x) # True tf.reduce_any(x, 0) # [True, True] tf.reduce_any(x, 1) # [True, False] ``` Args: input_tensor: The boolean tensor to reduce. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.any @end_compatibility Computes the "logical or" of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[True, True], [False, False]]) tf.reduce_any(x) # True tf.reduce_any(x, 0) # [True, True] tf.reduce_any(x, 1) # [True, False] ``` Args: input_tensor: The boolean tensor to reduce. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.any @end_compatibility Computes log(sum(exp(elements across dimensions of a tensor))). Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` has no entries, all dimensions are reduced, and a tensor with a single element is returned. This function is more numerically stable than log(sum(exp(input))). It avoids overflows caused by taking the exp of large inputs and underflows caused by taking the log of small inputs. For example: ```python x = tf.constant([[0., 0., 0.], [0., 0., 0.]]) tf.reduce_logsumexp(x) # log(6) tf.reduce_logsumexp(x, 0) # [log(2), log(2), log(2)] tf.reduce_logsumexp(x, 1) # [log(3), log(3)] tf.reduce_logsumexp(x, 1, keepdims=True) # [[log(3)], [log(3)]] tf.reduce_logsumexp(x, [0, 1]) # log(6) ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. Computes log(sum(exp(elements across dimensions of a tensor))). Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` has no entries, all dimensions are reduced, and a tensor with a single element is returned. This function is more numerically stable than log(sum(exp(input))). It avoids overflows caused by taking the exp of large inputs and underflows caused by taking the log of small inputs. For example: ```python x = tf.constant([[0., 0., 0.], [0., 0., 0.]]) tf.reduce_logsumexp(x) # log(6) tf.reduce_logsumexp(x, 0) # [log(2), log(2), log(2)] tf.reduce_logsumexp(x, 1) # [log(3), log(3)] tf.reduce_logsumexp(x, 1, keepdims=True) # [[log(3)], [log(3)]] tf.reduce_logsumexp(x, [0, 1]) # log(6) ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. Computes the maximum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have real numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.max @end_compatibility Computes the maximum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have real numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.max @end_compatibility Computes the mean of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1., 1.], [2., 2.]]) tf.reduce_mean(x) # 1.5 tf.reduce_mean(x, 0) # [1.5, 1.5] tf.reduce_mean(x, 1) # [1., 2.] ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.mean Please note that `np.mean` has a `dtype` parameter that could be used to specify the output type. By default this is `dtype=float64`. On the other hand, `tf.reduce_mean` has an aggressive type inference from `input_tensor`, for example: ```python x = tf.constant([1, 0, 1, 0]) tf.reduce_mean(x) # 0 y = tf.constant([1., 0., 1., 0.]) tf.reduce_mean(y) # 0.5 ``` @end_compatibility Computes the mean of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1., 1.], [2., 2.]]) tf.reduce_mean(x) # 1.5 tf.reduce_mean(x, 0) # [1.5, 1.5] tf.reduce_mean(x, 1) # [1., 2.] ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.mean Please note that `np.mean` has a `dtype` parameter that could be used to specify the output type. By default this is `dtype=float64`. On the other hand, `tf.reduce_mean` has an aggressive type inference from `input_tensor`, for example: ```python x = tf.constant([1, 0, 1, 0]) tf.reduce_mean(x) # 0 y = tf.constant([1., 0., 1., 0.]) tf.reduce_mean(y) # 0.5 ``` @end_compatibility Computes the minimum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have real numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.min @end_compatibility Computes the minimum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have real numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.min @end_compatibility Computes the product of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.prod @end_compatibility Computes the product of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor. @compatibility(numpy) Equivalent to np.prod @end_compatibility Computes the standard deviation of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1., 2.], [3., 4.]]) tf.reduce_std(x) # 1.1180339887498949 tf.reduce_std(x, 0) # [1., 1.] tf.reduce_std(x, 1) # [0.5, 0.5] ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name scope for the associated operations (optional). Returns: The reduced tensor, of the same dtype as the input_tensor. @compatibility(numpy) Equivalent to np.std Please note that `np.std` has a `dtype` parameter that could be used to specify the output type. By default this is `dtype=float64`. On the other hand, `tf.reduce_std` has an aggressive type inference from `input_tensor`, @end_compatibility Computes the sum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1, 1, 1], [1, 1, 1]]) tf.reduce_sum(x) # 6 tf.reduce_sum(x, 0) # [2, 2, 2] tf.reduce_sum(x, 1) # [3, 3] tf.reduce_sum(x, 1, keepdims=True) # [[3], [3]] tf.reduce_sum(x, [0, 1]) # 6 ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). Returns: The reduced tensor, of the same dtype as the input_tensor. @compatibility(numpy) Equivalent to np.sum apart the fact that numpy upcast uint8 and int32 to int64 while tensorflow returns the same dtype as the input. @end_compatibility Computes the sum of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1, 1, 1], [1, 1, 1]]) tf.reduce_sum(x) # 6 tf.reduce_sum(x, 0) # [2, 2, 2] tf.reduce_sum(x, 1) # [3, 3] tf.reduce_sum(x, 1, keepdims=True) # [[3], [3]] tf.reduce_sum(x, [0, 1]) # 6 ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name for the operation (optional). reduction_indices: The old (deprecated) name for axis. keep_dims: Deprecated alias for `keepdims`. Returns: The reduced tensor, of the same dtype as the input_tensor. @compatibility(numpy) Equivalent to np.sum apart the fact that numpy upcast uint8 and int32 to int64 while tensorflow returns the same dtype as the input. @end_compatibility Computes the variance of elements across dimensions of a tensor. Reduces `input_tensor` along the dimensions given in `axis`. Unless `keepdims` is true, the rank of the tensor is reduced by 1 for each entry in `axis`. If `keepdims` is true, the reduced dimensions are retained with length 1. If `axis` is None, all dimensions are reduced, and a tensor with a single element is returned. For example: ```python x = tf.constant([[1., 2.], [3., 4.]]) tf.reduce_variance(x) # 1.25 tf.reduce_variance(x, 0) # [1., 1.] tf.reduce_variance(x, 1) # [0.25, 0.25] ``` Args: input_tensor: The tensor to reduce. Should have numeric type. axis: The dimensions to reduce. If `None` (the default), reduces all dimensions. Must be in the range `[-rank(input_tensor), rank(input_tensor))`. keepdims: If true, retains reduced dimensions with length 1. name: A name scope for the associated operations (optional). Returns: The reduced tensor, of the same dtype as the input_tensor. @compatibility(numpy) Equivalent to np.var Please note that `np.var` has a `dtype` parameter that could be used to specify the output type. By default this is `dtype=float64`. On the other hand, `tf.reduce_variance` has an aggressive type inference from `input_tensor`, @end_compatibility Helper function for reduction ops. Args: input_shape: 1-D Tensor, the shape of the Tensor being reduced. axes: 1-D Tensor, the reduction axes. Returns: A 1-D Tensor, the output shape as if keepdims were set to True. Rounds the values of a tensor to the nearest integer, element-wise. Rounds half to even. Also known as bankers rounding. If you want to round according to the current system rounding mode use tf::cint. For example: ```python x = tf.constant([0.9, 2.5, 2.3, 1.5, -4.5]) tf.round(x) # [ 1.0, 2.0, 2.0, 2.0, -4.0 ] ``` Args: x: A `Tensor` of type `float16`, `float32`, `float64`, `int32`, or `int64`. name: A name for the operation (optional). Returns: A `Tensor` of same shape and type as `x`. Performs a safe saturating cast of `value` to `dtype`. This function casts the input to `dtype` without applying any scaling. If there is a danger that values would over or underflow in the cast, this op applies the appropriate clamping before the cast. Args: value: A `Tensor`. dtype: The desired output `DType`. name: A name for the operation (optional). Returns: `value` safely cast to `dtype`. Multiplies a scalar times a `Tensor` or `IndexedSlices` object. Intended for use in gradient code which might deal with `IndexedSlices` objects, which are easy to multiply by a scalar but more expensive to multiply with arbitrary tensors. Args: scalar: A 0-D scalar `Tensor`. Must have known shape. x: A `Tensor` or `IndexedSlices` to be scaled. name: A name for the operation (optional). Returns: `scalar * x` of the same type (`Tensor` or `IndexedSlices`) as `x`. Raises: ValueError: if scalar is not a 0-D `scalar`. Computes sigmoid of `x` element-wise. Specifically, `y = 1 / (1 + exp(-x))`. Args: x: A Tensor with type `float16`, `float32`, `float64`, `complex64`, or `complex128`. name: A name for the operation (optional). Returns: A Tensor with the same type as `x`. @compatibility(scipy) Equivalent to scipy.special.expit @end_compatibility Computes the mean along sparse segments of a tensor. Read [the section on segmentation](https://tensorflow.org/api_guides/python/math_ops#Segmentation) for an explanation of segments. Like `SegmentMean`, but `segment_ids` can have rank less than `data`'s first dimension, selecting a subset of dimension 0, specified by `indices`. `segment_ids` is allowed to have missing ids, in which case the output will be zeros at those indices. In those cases `num_segments` is used to determine the size of the output. Args: data: A `Tensor` with data that will be assembled in the output. indices: A 1-D `Tensor` with indices into `data`. Has same rank as `segment_ids`. segment_ids: A 1-D `Tensor` with indices into the output `Tensor`. Values should be sorted and can be repeated. name: A name for the operation (optional). num_segments: An optional int32 scalar. Indicates the size of the output `Tensor`. Returns: A `tensor` of the shape as data, except for dimension 0 which has size `k`, the number of segments specified via `num_segments` or inferred for the last element in `segments_ids`. Computes the mean along sparse segments of a tensor. Read [the section on segmentation](https://tensorflow.org/api_guides/python/math_ops#Segmentation) for an explanation of segments. Like `SegmentMean`, but `segment_ids` can have rank less than `data`'s first dimension, selecting a subset of dimension 0, specified by `indices`. `segment_ids` is allowed to have missing ids, in which case the output will be zeros at those indices. In those cases `num_segments` is used to determine the size of the output. Args: data: A `Tensor` with data that will be assembled in the output. indices: A 1-D `Tensor` with indices into `data`. Has same rank as `segment_ids`. segment_ids: A 1-D `Tensor` with indices into the output `Tensor`. Values should be sorted and can be repeated. num_segments: An optional int32 scalar. Indicates the size of the output `Tensor`. name: A name for the operation (optional). Returns: A `tensor` of the shape as data, except for dimension 0 which has size `k`, the number of segments specified via `num_segments` or inferred for the last element in `segments_ids`. Computes the sum along sparse segments of a tensor divided by the sqrt(N). `N` is the size of the segment being reduced. Args: data: A `Tensor` with data that will be assembled in the output. indices: A 1-D `Tensor` with indices into `data`. Has same rank as `segment_ids`. segment_ids: A 1-D `Tensor` with indices into the output `Tensor`. Values should be sorted and can be repeated. name: A name for the operation (optional). num_segments: An optional int32 scalar. Indicates the size of the output `Tensor`. Returns: A `tensor` of the shape as data, except for dimension 0 which has size `k`, the number of segments specified via `num_segments` or inferred for the last element in `segments_ids`. Computes the sum along sparse segments of a tensor divided by the sqrt(N). `N` is the size of the segment being reduced. Args: data: A `Tensor` with data that will be assembled in the output. indices: A 1-D `Tensor` with indices into `data`. Has same rank as `segment_ids`. segment_ids: A 1-D `Tensor` with indices into the output `Tensor`. Values should be sorted and can be repeated. num_segments: An optional int32 scalar. Indicates the size of the output `Tensor`. name: A name for the operation (optional). Returns: A `tensor` of the shape as data, except for dimension 0 which has size `k`, the number of segments specified via `num_segments` or inferred for the last element in `segments_ids`. Computes the sum along sparse segments of a tensor. Read [the section on segmentation](https://tensorflow.org/api_guides/python/math_ops#Segmentation) for an explanation of segments. Like `SegmentSum`, but `segment_ids` can have rank less than `data`'s first dimension, selecting a subset of dimension 0, specified by `indices`. `segment_ids` is allowed to have missing ids, in which case the output will be zeros at those indices. In those cases `num_segments` is used to determine the size of the output. For example: ```python c = tf.constant([[1,2,3,4], [-1,-2,-3,-4], [5,6,7,8]]) # Select two rows, one segment. tf.sparse.segment_sum(c, tf.constant([0, 1]), tf.constant([0, 0])) # => [[0 0 0 0]] # Select two rows, two segment. tf.sparse.segment_sum(c, tf.constant([0, 1]), tf.constant([0, 1])) # => [[ 1 2 3 4] # [-1 -2 -3 -4]] # With missing segment ids. tf.sparse.segment_sum(c, tf.constant([0, 1]), tf.constant([0, 2]), num_segments=4) # => [[ 1 2 3 4] # [ 0 0 0 0] # [-1 -2 -3 -4] # [ 0 0 0 0]] # Select all rows, two segments. tf.sparse.segment_sum(c, tf.constant([0, 1, 2]), tf.constant([0, 0, 1])) # => [[0 0 0 0] # [5 6 7 8]] # Which is equivalent to: tf.segment_sum(c, tf.constant([0, 0, 1])) ``` Args: data: A `Tensor` with data that will be assembled in the output. indices: A 1-D `Tensor` with indices into `data`. Has same rank as `segment_ids`. segment_ids: A 1-D `Tensor` with indices into the output `Tensor`. Values should be sorted and can be repeated. name: A name for the operation (optional). num_segments: An optional int32 scalar. Indicates the size of the output `Tensor`. Returns: A `tensor` of the shape as data, except for dimension 0 which has size `k`, the number of segments specified via `num_segments` or inferred for the last element in `segments_ids`. Tensor contraction of a and b along specified axes. Tensordot (also known as tensor contraction) sums the product of elements from `a` and `b` over the indices specified by `a_axes` and `b_axes`. The lists `a_axes` and `b_axes` specify those pairs of axes along which to contract the tensors. The axis `a_axes[i]` of `a` must have the same dimension as axis `b_axes[i]` of `b` for all `i` in `range(0, len(a_axes))`. The lists `a_axes` and `b_axes` must have identical length and consist of unique integers that specify valid axes for each of the tensors. This operation corresponds to `numpy.tensordot(a, b, axes)`. Example 1: When `a` and `b` are matrices (order 2), the case `axes = 1` is equivalent to matrix multiplication. Example 2: When `a` and `b` are matrices (order 2), the case `axes = [[1], [0]]` is equivalent to matrix multiplication. Example 3: Suppose that \\(a_{ijk}\\) and \\(b_{lmn}\\) represent two tensors of order 3. Then, `contract(a, b, [[0], [2]])` is the order 4 tensor \\(c_{jklm}\\) whose entry corresponding to the indices \\((j,k,l,m)\\) is given by: \\( c_{jklm} = \sum_i a_{ijk} b_{lmi} \\). In general, `order(c) = order(a) + order(b) - 2*len(axes[0])`. Args: a: `Tensor` of type `float32` or `float64`. b: `Tensor` with the same type as `a`. axes: Either a scalar `N`, or a list or an `int32` `Tensor` of shape [2, k]. If axes is a scalar, sum over the last N axes of a and the first N axes of b in order. If axes is a list or `Tensor` the first and second row contain the set of unique integers specifying axes along which the contraction is computed, for `a` and `b`, respectively. The number of axes for `a` and `b` must be equal. name: A name for the operation (optional). Returns: A `Tensor` with the same type as `a`. Raises: ValueError: If the shapes of `a`, `b`, and `axes` are incompatible. IndexError: If the values in axes exceed the rank of the corresponding tensor. Casts a tensor to type `bfloat16`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `bfloat16`. Raises: TypeError: If `x` cannot be cast to the `bfloat16`. Casts a tensor to type `complex128`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `complex128`. Raises: TypeError: If `x` cannot be cast to the `complex128`. Casts a tensor to type `complex64`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `complex64`. Raises: TypeError: If `x` cannot be cast to the `complex64`. Casts a tensor to type `float64`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `float64`. Raises: TypeError: If `x` cannot be cast to the `float64`. Casts a tensor to type `float32`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `float32`. Raises: TypeError: If `x` cannot be cast to the `float32`. Casts a tensor to type `int32`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `int32`. Raises: TypeError: If `x` cannot be cast to the `int32`. Casts a tensor to type `int64`. Args: x: A `Tensor` or `SparseTensor` or `IndexedSlices`. name: A name for the operation (optional). Returns: A `Tensor` or `SparseTensor` or `IndexedSlices` with same shape as `x` with type `int64`. Raises: TypeError: If `x` cannot be cast to the `int64`. Compute the trace of a tensor `x`. `trace(x)` returns the sum along the main diagonal of each inner-most matrix in x. If x is of rank `k` with shape `[I, J, K, ..., L, M, N]`, then output is a tensor of rank `k-2` with dimensions `[I, J, K, ..., L]` where `output[i, j, k, ..., l] = trace(x[i, j, i, ..., l, :, :])` For example: ```python x = tf.constant([[1, 2], [3, 4]]) tf.linalg.trace(x) # 5 x = tf.constant([[1, 2, 3], [4, 5, 6], [7, 8, 9]]) tf.linalg.trace(x) # 15 x = tf.constant([[[1, 2, 3], [4, 5, 6], [7, 8, 9]], [[-1, -2, -3], [-4, -5, -6], [-7, -8, -9]]]) tf.linalg.trace(x) # [15, -15] ``` Args: x: tensor. name: A name for the operation (optional). Returns: The trace of input tensor. Divides x / y elementwise (using Python 3 division operator semantics). NOTE: Prefer using the Tensor operator or tf.divide which obey Python division operator semantics. This function forces Python 3 division operator semantics where all integer arguments are cast to floating types first. This op is generated by normal `x / y` division in Python 3 and in Python 2.7 with `from __future__ import division`. If you want integer division that rounds down, use `x // y` or `tf.math.floordiv`. `x` and `y` must have the same numeric type. If the inputs are floating point, the output will have the same type. If the inputs are integral, the inputs are cast to `float32` for `int8` and `int16` and `float64` for `int32` and `int64` (matching the behavior of Numpy). Args: x: `Tensor` numerator of numeric type. y: `Tensor` denominator of numeric type. name: A name for the operation (optional). Returns: `x / y` evaluated in floating point. Raises: TypeError: If `x` and `y` have different dtypes. Computes the mean along segments of a tensor. Read [the section on segmentation](https://tensorflow.org/api_guides/python/math_ops#segmentation) for an explanation of segments. This operator is similar to the unsorted segment sum operator found [here](../../../api_docs/python/math_ops.md#UnsortedSegmentSum). Instead of computing the sum over segments, it computes the mean of all entries belonging to a segment such that: \\(output_i = 1/N_i \sum_{j...} data[j...]\\) where the sum is over tuples `j...` such that `segment_ids[j...] == i` with \\N_i\\ being the number of occurrences of id \\i\\. If there is no entry for a given segment ID `i`, it outputs 0. If the given segment ID `i` is negative, the value is dropped and will not be added to the sum of the segment. Args: data: A `Tensor` with floating point or complex dtype. segment_ids: An integer tensor whose shape is a prefix of `data.shape`. num_segments: An integer scalar `Tensor`. The number of distinct segment IDs. name: A name for the operation (optional). Returns: A `Tensor`. Has same shape as data, except for the first `segment_ids.rank` dimensions, which are replaced with a single dimension which has size `num_segments`. Computes the sum along segments of a tensor divided by the sqrt(N). Read [the section on segmentation](https://tensorflow.org/api_guides/python/math_ops#segmentation) for an explanation of segments. This operator is similar to the unsorted segment sum operator found [here](../../../api_docs/python/math_ops.md#UnsortedSegmentSum). Additionally to computing the sum over segments, it divides the results by sqrt(N). \\(output_i = 1/sqrt(N_i) \sum_{j...} data[j...]\\) where the sum is over tuples `j...` such that `segment_ids[j...] == i` with \\N_i\\ being the number of occurrences of id \\i\\. If there is no entry for a given segment ID `i`, it outputs 0. Note that this op only supports floating point and complex dtypes, due to tf.sqrt only supporting these types. If the given segment ID `i` is negative, the value is dropped and will not be added to the sum of the segment. Args: data: A `Tensor` with floating point or complex dtype. segment_ids: An integer tensor whose shape is a prefix of `data.shape`. num_segments: An integer scalar `Tensor`. The number of distinct segment IDs. name: A name for the operation (optional). Returns: A `Tensor`. Has same shape as data, except for the first `segment_ids.rank` dimensions, which are replaced with a single dimension which has size `num_segments`. Basic arithmetic operators. See the [python/math_ops](python/math_ops) guide. Copyright 2015 The TensorFlow Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ============================================================================== pylint: disable=redefined-builtin go/tf-wildcard-import pylint: disable=wildcard-import pylint: enable=wildcard-import Aliases for some automatically-generated names. pylint: disable=used-before-assignment pylint: disable=used-before-assignment This is set by resource_variable_ops.py. It is included in this way since there is a circular dependency between math_ops and resource_variable_ops pylint: disable=redefined-builtin pylint: enable=redefined-builtin pylint: disable=anomalous-backslash-in-string,protected-access pylint: disable=g-docstring-has-escape pylint: disable=redefined-builtin pylint: enable=g-docstring-has-escape pylint: disable=redefined-builtin pylint: enable=redefined-builtin Cannot use tensors operator overload, because it has no way to track override names. Use a dummy class to track the runtime division behavior TODO(aselle): put deprecation in after another round of global code changes TODO(aselle): put deprecation in after another round of global code changes pylint: disable=g-docstring-has-escape pylint: enable=g-docstring-has-escape pylint: disable=redefined-builtin pylint: disable=redefined-builtin,redefined-outer-name pylint: enable=redefined-outer-name,redefined-builtin pylint: disable=redefined-builtin TODO(josh11b): If x is not already a Tensor, we could return ops.convert_to_tensor(x, dtype=dtype, ...) here, but that allows some conversions that cast() can't do, e.g. casting numbers to strings. When casting to a type with smaller representable range, clamp. Note that this covers casting to unsigned types as well. __invert__ corresponds to the ~ operator. Here we follow the numpy convention ~ marks an elementwise bit-wise inverse. This is only implemented for boolean tensors and will throw a TypeError if used on nonboolean arrays If the RHS is not a tensor, it might be a tensor aware object that can implement the operator with knowledge of itself and the tensor. Propagate func.__doc__ to the wrappers Conversion table for __truediv__. None entries mean no conversion required. NOTE: the support of "sparse (true)div dense" is currently not baked in into "tf.(true_)div()". Until such an API decision is made, the supported usage is to explicitly use the "/" operator to invoke either truediv or div. TODO(aselle): This should be removed TODO(aselle): Deprecate this once all internal functionality uses tf.truncatediv TODO(aselle): Rename this to floordiv when we can. Case: Dense * Sparse. NOTE(aselle): When integer division is added for sparse_dense_cwise, div, truediv, and floordiv should be delegated appropriately for Python sematnics, analogous to dense cwise tensor operations. TODO(alemi) Make this a cwise op if people end up relying on it. pylint: disable=redefined-builtin infer dtype if not explicitly provided Reduction operations pylint: disable=invalid-name TODO(aselle): Remove this after deprecation Fast path: avoid creating Rank and Range ops if ndims is known. sparse.dense_shape is 1-D. Otherwise, we rely on Range and Rank to do the right thing at run-time. pylint: disable=redefined-builtin A scalar of 'zero' is enough as `not_equal` will broadcast. int64 reduction happens on GPU TODO(apassos) remove _shape_tuple here when it is not needed. pylint: disable=protected-access pylint: disable=protected-access BatchMatmul does not support transpose, so we conjugate the matrix and use adjoint instead. Conj() is a noop for real matrices. Neither matmul nor sparse_matmul support adjoint, so we conjugate the matrix and use transpose instead. Conj() is a noop for real matrices. matmul currently doesn't handle mixed-precision inputs. sparse_matmul always returns float32, even with bfloat16 inputs. This prevents us from configuring bfloat16 training. casting to bfloat16 also matches non-sparse matmul behavior better. TODO(touts): op_scope tensor_dtype is for safety only; operator's output type computed in C++ TemporaryVariable not currently supported in eager mode; fall back onto AddN for now. TODO(frreiss) remove this once the lifetime of eager variables gets addressed pylint: disable=protected-access Not broadcasting. Example: cast needed for SparseTensor reductions [2, 3, 5, 7] [1, 2] 4 [2] [2, 1, 1, 7] [0, 1, 2, 3] [1, 2] [2, 3, 5, 7] [1, 1] bincount doesn't support negative indices so we use unsorted_segment_sum add dimensions for all non-reduced axes
71,765
en
0.744922
""" Median and Mean for Cauchy distribution --------------------------------------- This plot shows graphically that mean-based statistics are not robust for the Cauchy distribution. Median-based statistics should be used instead. """ # Author: Jake VanderPlas # License: BSD # The figure produced by this code is published in the textbook # "Statistics, Data Mining, and Machine Learning in Astronomy" (2013) # For more information, see http://astroML.github.com # To report a bug or issue, use the following forum: # https://groups.google.com/forum/#!forum/astroml-general import numpy as np from matplotlib import pyplot as plt from scipy import optimize from scipy.stats import cauchy, norm #---------------------------------------------------------------------- # This function adjusts matplotlib settings for a uniform feel in the textbook. # Note that with usetex=True, fonts are rendered with LaTeX. This may # result in an error if LaTeX is not installed on your system. In that case, # you can set usetex to False. from astroML.plotting import setup_text_plots setup_text_plots(fontsize=8, usetex=True) def robust_mean_mixture(x): """Compute the mean via a mixture of two Gaussians One Gaussian accounts for outliers, and one Gaussian accounts for the true distribution. This cannot be computed analytically, so it uses scipy's function optimization """ if len(x) == 1: return x x = x.ravel() mu_bg = np.mean(x) sig_bg = 3 * np.std(x) likelihood = lambda v: -np.sum(np.log(norm.pdf(x, v[0], v[1]) + norm.pdf(x, mu_bg, sig_bg))) v0 = np.array([0, 30]) v_best = optimize.fmin(likelihood, v0, disp=False) return v_best[0] def robust_mean_iterated(x, sigma_cut=3): """Compute the robust mean iteratively After computing the mean, points further than 3 sigma from the mean are removed and the result is repeated until convergence. """ flag = np.ones(x.shape, dtype=bool) n_to_keep = x.size while True: xf = x[flag] mu = xf.mean() sig = xf.std() if len(xf) == 1: break x_sig = abs((x - mu) / sig) too_far = (x_sig > sigma_cut) flag[too_far] = False n_flag = flag.sum() if n_flag == n_to_keep: break else: n_to_keep = n_flag return mu #------------------------------------------------------------ # Create the distribution and compute means and medians np.random.seed(6) mu = 0 gamma = 2 xi = cauchy(mu, gamma).rvs(100) Nrange = np.arange(1, len(xi) + 1) mean = [np.mean(xi[:N]) for N in Nrange] median = [np.median(xi[:N]) for N in Nrange] mean_mixture = [robust_mean_mixture(xi[:N]) for N in Nrange] mean_iter = [robust_mean_iterated(xi[:N]) for N in Nrange] #------------------------------------------------------------ # Plot the results as a function of number of points fig = plt.figure(figsize=(5, 3.75)) fig.subplots_adjust(hspace=0.05) # first plot the mean ax = fig.add_subplot(211) ax.plot(Nrange, mean, '-.b', label='mean') ax.plot(Nrange, median, '-k', label='median') ax.plot(Nrange, mean_mixture, ':r', label='robust mean (mixture)') ax.plot(Nrange, mean_iter, '--g', label='robust mean (sigma-clip)') ax.plot(Nrange, 0 * Nrange, '-', c='gray', lw=0.5) ax.set_xlim(0, 100) ax.set_ylim(-7, 7) ax.legend(loc=4, ncol=2, frameon=False) ax.set_ylabel('Value') ax.xaxis.set_major_formatter(plt.NullFormatter()) # now plot the median ax = fig.add_subplot(212) ax.scatter(Nrange, xi, lw=0, s=10, c='k') ax.plot(Nrange, 0 * Nrange, '-', c='gray') ax.set_xlim(0, 100) ax.set_ylim(-75, 75) ax.set_xlabel('Sample Size') ax.set_ylabel('Value') plt.show()
book_figures/chapter3/fig_cauchy_median_mean.py
3,753
Compute the robust mean iteratively After computing the mean, points further than 3 sigma from the mean are removed and the result is repeated until convergence. Compute the mean via a mixture of two Gaussians One Gaussian accounts for outliers, and one Gaussian accounts for the true distribution. This cannot be computed analytically, so it uses scipy's function optimization Median and Mean for Cauchy distribution --------------------------------------- This plot shows graphically that mean-based statistics are not robust for the Cauchy distribution. Median-based statistics should be used instead. Author: Jake VanderPlas License: BSD The figure produced by this code is published in the textbook "Statistics, Data Mining, and Machine Learning in Astronomy" (2013) For more information, see http://astroML.github.com To report a bug or issue, use the following forum: https://groups.google.com/forum/!forum/astroml-general---------------------------------------------------------------------- This function adjusts matplotlib settings for a uniform feel in the textbook. Note that with usetex=True, fonts are rendered with LaTeX. This may result in an error if LaTeX is not installed on your system. In that case, you can set usetex to False.------------------------------------------------------------ Create the distribution and compute means and medians------------------------------------------------------------ Plot the results as a function of number of points first plot the mean now plot the median
1,535
en
0.814881
""" Utilities for fast persistence of big data, with optional compression. """ # Author: Gael Varoquaux <gael dot varoquaux at normalesup dot org> # Copyright (c) 2009 Gael Varoquaux # License: BSD Style, 3 clauses. import pickle import traceback import os import zlib import warnings from io import BytesIO from ._compat import _basestring, PY3_OR_LATER if PY3_OR_LATER: Unpickler = pickle._Unpickler Pickler = pickle._Pickler def asbytes(s): if isinstance(s, bytes): return s return s.encode('latin1') else: Unpickler = pickle.Unpickler Pickler = pickle.Pickler asbytes = str def hex_str(an_int): """Converts an int to an hexadecimal string """ return '{0:#x}'.format(an_int) _MEGA = 2 ** 20 # Compressed pickle header format: _ZFILE_PREFIX followed by _MAX_LEN # bytes which contains the length of the zlib compressed data as an # hexadecimal string. For example: 'ZF0x139 ' _ZFILE_PREFIX = asbytes('ZF') _MAX_LEN = len(hex_str(2 ** 64)) ############################################################################### # Compressed file with Zlib def _read_magic(file_handle): """ Utility to check the magic signature of a file identifying it as a Zfile """ magic = file_handle.read(len(_ZFILE_PREFIX)) # Pickling needs file-handles at the beginning of the file file_handle.seek(0) return magic def read_zfile(file_handle): """Read the z-file and return the content as a string Z-files are raw data compressed with zlib used internally by joblib for persistence. Backward compatibility is not guaranteed. Do not use for external purposes. """ file_handle.seek(0) assert _read_magic(file_handle) == _ZFILE_PREFIX, \ "File does not have the right magic" header_length = len(_ZFILE_PREFIX) + _MAX_LEN length = file_handle.read(header_length) length = length[len(_ZFILE_PREFIX):] length = int(length, 16) # With python2 and joblib version <= 0.8.4 compressed pickle header is one # character wider so we need to ignore an additional space if present. # Note: the first byte of the zlib data is guaranteed not to be a # space according to # https://tools.ietf.org/html/rfc6713#section-2.1 next_byte = file_handle.read(1) if next_byte != b' ': # The zlib compressed data has started and we need to go back # one byte file_handle.seek(header_length) # We use the known length of the data to tell Zlib the size of the # buffer to allocate. data = zlib.decompress(file_handle.read(), 15, length) assert len(data) == length, ( "Incorrect data length while decompressing %s." "The file could be corrupted." % file_handle) return data def write_zfile(file_handle, data, compress=1): """Write the data in the given file as a Z-file. Z-files are raw data compressed with zlib used internally by joblib for persistence. Backward compatibility is not guarantied. Do not use for external purposes. """ file_handle.write(_ZFILE_PREFIX) length = hex_str(len(data)) # Store the length of the data file_handle.write(asbytes(length.ljust(_MAX_LEN))) file_handle.write(zlib.compress(asbytes(data), compress)) ############################################################################### # Utility objects for persistence. class NDArrayWrapper(object): """ An object to be persisted instead of numpy arrays. The only thing this object does, is to carry the filename in which the array has been persisted, and the array subclass. """ def __init__(self, filename, subclass, allow_mmap=True): "Store the useful information for later" self.filename = filename self.subclass = subclass self.allow_mmap = allow_mmap def read(self, unpickler): "Reconstruct the array" filename = os.path.join(unpickler._dirname, self.filename) # Load the array from the disk np_ver = [int(x) for x in unpickler.np.__version__.split('.', 2)[:2]] # use getattr instead of self.allow_mmap to ensure backward compat # with NDArrayWrapper instances pickled with joblib < 0.9.0 allow_mmap = getattr(self, 'allow_mmap', True) memmap_kwargs = ({} if not allow_mmap else {'mmap_mode': unpickler.mmap_mode}) array = unpickler.np.load(filename, **memmap_kwargs) # Reconstruct subclasses. This does not work with old # versions of numpy if (hasattr(array, '__array_prepare__') and not self.subclass in (unpickler.np.ndarray, unpickler.np.memmap)): # We need to reconstruct another subclass new_array = unpickler.np.core.multiarray._reconstruct( self.subclass, (0,), 'b') new_array.__array_prepare__(array) array = new_array return array #def __reduce__(self): # return None class ZNDArrayWrapper(NDArrayWrapper): """An object to be persisted instead of numpy arrays. This object store the Zfile filename in which the data array has been persisted, and the meta information to retrieve it. The reason that we store the raw buffer data of the array and the meta information, rather than array representation routine (tostring) is that it enables us to use completely the strided model to avoid memory copies (a and a.T store as fast). In addition saving the heavy information separately can avoid creating large temporary buffers when unpickling data with large arrays. """ def __init__(self, filename, init_args, state): "Store the useful information for later" self.filename = filename self.state = state self.init_args = init_args def read(self, unpickler): "Reconstruct the array from the meta-information and the z-file" # Here we a simply reproducing the unpickling mechanism for numpy # arrays filename = os.path.join(unpickler._dirname, self.filename) array = unpickler.np.core.multiarray._reconstruct(*self.init_args) with open(filename, 'rb') as f: data = read_zfile(f) state = self.state + (data,) array.__setstate__(state) return array ############################################################################### # Pickler classes class NumpyPickler(Pickler): """A pickler to persist of big data efficiently. The main features of this object are: * persistence of numpy arrays in separate .npy files, for which I/O is fast. * optional compression using Zlib, with a special care on avoid temporaries. """ dispatch = Pickler.dispatch.copy() def __init__(self, filename, compress=0, cache_size=10, protocol=None): self._filename = filename self._filenames = [filename, ] self.cache_size = cache_size self.compress = compress if not self.compress: self.file = open(filename, 'wb') else: self.file = BytesIO() # Count the number of npy files that we have created: self._npy_counter = 1 # By default we want a pickle protocol that only changes with # the major python version and not the minor one if protocol is None: protocol = (pickle.DEFAULT_PROTOCOL if PY3_OR_LATER else pickle.HIGHEST_PROTOCOL) Pickler.__init__(self, self.file, protocol=protocol) # delayed import of numpy, to avoid tight coupling try: import numpy as np except ImportError: np = None self.np = np def _write_array(self, array, filename): if not self.compress: self.np.save(filename, array) allow_mmap = not array.dtype.hasobject container = NDArrayWrapper(os.path.basename(filename), type(array), allow_mmap=allow_mmap) else: filename += '.z' # Efficient compressed storage: # The meta data is stored in the container, and the core # numerics in a z-file _, init_args, state = array.__reduce__() # the last entry of 'state' is the data itself with open(filename, 'wb') as zfile: write_zfile(zfile, state[-1], compress=self.compress) state = state[:-1] container = ZNDArrayWrapper(os.path.basename(filename), init_args, state) return container, filename def save(self, obj): """ Subclass the save method, to save ndarray subclasses in npy files, rather than pickling them. Of course, this is a total abuse of the Pickler class. """ if (self.np is not None and type(obj) in (self.np.ndarray, self.np.matrix, self.np.memmap)): size = obj.size * obj.itemsize if self.compress and size < self.cache_size * _MEGA: # When compressing, as we are not writing directly to the # disk, it is more efficient to use standard pickling if type(obj) is self.np.memmap: # Pickling doesn't work with memmaped arrays obj = self.np.asarray(obj) return Pickler.save(self, obj) if not obj.dtype.hasobject: try: filename = '%s_%02i.npy' % (self._filename, self._npy_counter) # This converts the array in a container obj, filename = self._write_array(obj, filename) self._filenames.append(filename) self._npy_counter += 1 except Exception: # XXX: We should have a logging mechanism print('Failed to save %s to .npy file:\n%s' % ( type(obj), traceback.format_exc())) return Pickler.save(self, obj) def close(self): if self.compress: with open(self._filename, 'wb') as zfile: write_zfile(zfile, self.file.getvalue(), self.compress) class NumpyUnpickler(Unpickler): """A subclass of the Unpickler to unpickle our numpy pickles. """ dispatch = Unpickler.dispatch.copy() def __init__(self, filename, file_handle, mmap_mode=None): self._filename = os.path.basename(filename) self._dirname = os.path.dirname(filename) self.mmap_mode = mmap_mode self.file_handle = self._open_pickle(file_handle) Unpickler.__init__(self, self.file_handle) try: import numpy as np except ImportError: np = None self.np = np def _open_pickle(self, file_handle): return file_handle def load_build(self): """ This method is called to set the state of a newly created object. We capture it to replace our place-holder objects, NDArrayWrapper, by the array we are interested in. We replace them directly in the stack of pickler. """ Unpickler.load_build(self) if isinstance(self.stack[-1], NDArrayWrapper): if self.np is None: raise ImportError('Trying to unpickle an ndarray, ' "but numpy didn't import correctly") nd_array_wrapper = self.stack.pop() array = nd_array_wrapper.read(self) self.stack.append(array) # Be careful to register our new method. if PY3_OR_LATER: dispatch[pickle.BUILD[0]] = load_build else: dispatch[pickle.BUILD] = load_build class ZipNumpyUnpickler(NumpyUnpickler): """A subclass of our Unpickler to unpickle on the fly from compressed storage.""" def __init__(self, filename, file_handle): NumpyUnpickler.__init__(self, filename, file_handle, mmap_mode=None) def _open_pickle(self, file_handle): return BytesIO(read_zfile(file_handle)) ############################################################################### # Utility functions def dump(value, filename, compress=0, cache_size=100, protocol=None): """Fast persistence of an arbitrary Python object into one or multiple files, with dedicated storage for numpy arrays. Parameters ----------- value: any Python object The object to store to disk filename: string The name of the file in which it is to be stored compress: integer for 0 to 9, optional Optional compression level for the data. 0 is no compression. Higher means more compression, but also slower read and write times. Using a value of 3 is often a good compromise. See the notes for more details. cache_size: positive number, optional Fixes the order of magnitude (in megabytes) of the cache used for in-memory compression. Note that this is just an order of magnitude estimate and that for big arrays, the code will go over this value at dump and at load time. protocol: positive int Pickle protocol, see pickle.dump documentation for more details. Returns ------- filenames: list of strings The list of file names in which the data is stored. If compress is false, each array is stored in a different file. See Also -------- joblib.load : corresponding loader Notes ----- Memmapping on load cannot be used for compressed files. Thus using compression can significantly slow down loading. In addition, compressed files take extra extra memory during dump and load. """ if compress is True: # By default, if compress is enabled, we want to be using 3 by # default compress = 3 if not isinstance(filename, _basestring): # People keep inverting arguments, and the resulting error is # incomprehensible raise ValueError( 'Second argument should be a filename, %s (type %s) was given' % (filename, type(filename)) ) try: pickler = NumpyPickler(filename, compress=compress, cache_size=cache_size, protocol=protocol) pickler.dump(value) pickler.close() finally: if 'pickler' in locals() and hasattr(pickler, 'file'): pickler.file.flush() pickler.file.close() return pickler._filenames def load(filename, mmap_mode=None): """Reconstruct a Python object from a file persisted with joblib.dump. Parameters ----------- filename: string The name of the file from which to load the object mmap_mode: {None, 'r+', 'r', 'w+', 'c'}, optional If not None, the arrays are memory-mapped from the disk. This mode has no effect for compressed files. Note that in this case the reconstructed object might not longer match exactly the originally pickled object. Returns ------- result: any Python object The object stored in the file. See Also -------- joblib.dump : function to save an object Notes ----- This function can load numpy array files saved separately during the dump. If the mmap_mode argument is given, it is passed to np.load and arrays are loaded as memmaps. As a consequence, the reconstructed object might not match the original pickled object. Note that if the file was saved with compression, the arrays cannot be memmaped. """ with open(filename, 'rb') as file_handle: # We are careful to open the file handle early and keep it open to # avoid race-conditions on renames. That said, if data are stored in # companion files, moving the directory will create a race when # joblib tries to access the companion files. if _read_magic(file_handle) == _ZFILE_PREFIX: if mmap_mode is not None: warnings.warn('file "%(filename)s" appears to be a zip, ' 'ignoring mmap_mode "%(mmap_mode)s" flag passed' % locals(), Warning, stacklevel=2) unpickler = ZipNumpyUnpickler(filename, file_handle=file_handle) else: unpickler = NumpyUnpickler(filename, file_handle=file_handle, mmap_mode=mmap_mode) try: obj = unpickler.load() except UnicodeDecodeError as exc: # More user-friendly error message if PY3_OR_LATER: new_exc = ValueError( 'You may be trying to read with ' 'python 3 a joblib pickle generated with python 2. ' 'This feature is not supported by joblib.') new_exc.__cause__ = exc raise new_exc finally: if hasattr(unpickler, 'file_handle'): unpickler.file_handle.close() return obj
Sklearn_scipy_numpy/source/sklearn/externals/joblib/numpy_pickle.py
17,419
An object to be persisted instead of numpy arrays. The only thing this object does, is to carry the filename in which the array has been persisted, and the array subclass. A pickler to persist of big data efficiently. The main features of this object are: * persistence of numpy arrays in separate .npy files, for which I/O is fast. * optional compression using Zlib, with a special care on avoid temporaries. A subclass of the Unpickler to unpickle our numpy pickles. An object to be persisted instead of numpy arrays. This object store the Zfile filename in which the data array has been persisted, and the meta information to retrieve it. The reason that we store the raw buffer data of the array and the meta information, rather than array representation routine (tostring) is that it enables us to use completely the strided model to avoid memory copies (a and a.T store as fast). In addition saving the heavy information separately can avoid creating large temporary buffers when unpickling data with large arrays. A subclass of our Unpickler to unpickle on the fly from compressed storage. Store the useful information for later Store the useful information for later Utility to check the magic signature of a file identifying it as a Zfile Fast persistence of an arbitrary Python object into one or multiple files, with dedicated storage for numpy arrays. Parameters ----------- value: any Python object The object to store to disk filename: string The name of the file in which it is to be stored compress: integer for 0 to 9, optional Optional compression level for the data. 0 is no compression. Higher means more compression, but also slower read and write times. Using a value of 3 is often a good compromise. See the notes for more details. cache_size: positive number, optional Fixes the order of magnitude (in megabytes) of the cache used for in-memory compression. Note that this is just an order of magnitude estimate and that for big arrays, the code will go over this value at dump and at load time. protocol: positive int Pickle protocol, see pickle.dump documentation for more details. Returns ------- filenames: list of strings The list of file names in which the data is stored. If compress is false, each array is stored in a different file. See Also -------- joblib.load : corresponding loader Notes ----- Memmapping on load cannot be used for compressed files. Thus using compression can significantly slow down loading. In addition, compressed files take extra extra memory during dump and load. Converts an int to an hexadecimal string Reconstruct a Python object from a file persisted with joblib.dump. Parameters ----------- filename: string The name of the file from which to load the object mmap_mode: {None, 'r+', 'r', 'w+', 'c'}, optional If not None, the arrays are memory-mapped from the disk. This mode has no effect for compressed files. Note that in this case the reconstructed object might not longer match exactly the originally pickled object. Returns ------- result: any Python object The object stored in the file. See Also -------- joblib.dump : function to save an object Notes ----- This function can load numpy array files saved separately during the dump. If the mmap_mode argument is given, it is passed to np.load and arrays are loaded as memmaps. As a consequence, the reconstructed object might not match the original pickled object. Note that if the file was saved with compression, the arrays cannot be memmaped. This method is called to set the state of a newly created object. We capture it to replace our place-holder objects, NDArrayWrapper, by the array we are interested in. We replace them directly in the stack of pickler. Reconstruct the array Reconstruct the array from the meta-information and the z-file Read the z-file and return the content as a string Z-files are raw data compressed with zlib used internally by joblib for persistence. Backward compatibility is not guaranteed. Do not use for external purposes. Subclass the save method, to save ndarray subclasses in npy files, rather than pickling them. Of course, this is a total abuse of the Pickler class. Write the data in the given file as a Z-file. Z-files are raw data compressed with zlib used internally by joblib for persistence. Backward compatibility is not guarantied. Do not use for external purposes. Utilities for fast persistence of big data, with optional compression. Author: Gael Varoquaux <gael dot varoquaux at normalesup dot org> Copyright (c) 2009 Gael Varoquaux License: BSD Style, 3 clauses. Compressed pickle header format: _ZFILE_PREFIX followed by _MAX_LEN bytes which contains the length of the zlib compressed data as an hexadecimal string. For example: 'ZF0x139 ' Compressed file with Zlib Pickling needs file-handles at the beginning of the file With python2 and joblib version <= 0.8.4 compressed pickle header is one character wider so we need to ignore an additional space if present. Note: the first byte of the zlib data is guaranteed not to be a space according to https://tools.ietf.org/html/rfc6713section-2.1 The zlib compressed data has started and we need to go back one byte We use the known length of the data to tell Zlib the size of the buffer to allocate. Store the length of the data Utility objects for persistence. Load the array from the disk use getattr instead of self.allow_mmap to ensure backward compat with NDArrayWrapper instances pickled with joblib < 0.9.0 Reconstruct subclasses. This does not work with old versions of numpy We need to reconstruct another subclassdef __reduce__(self): return None Here we a simply reproducing the unpickling mechanism for numpy arrays Pickler classes Count the number of npy files that we have created: By default we want a pickle protocol that only changes with the major python version and not the minor one delayed import of numpy, to avoid tight coupling Efficient compressed storage: The meta data is stored in the container, and the core numerics in a z-file the last entry of 'state' is the data itself When compressing, as we are not writing directly to the disk, it is more efficient to use standard pickling Pickling doesn't work with memmaped arrays This converts the array in a container XXX: We should have a logging mechanism Be careful to register our new method. Utility functions By default, if compress is enabled, we want to be using 3 by default People keep inverting arguments, and the resulting error is incomprehensible We are careful to open the file handle early and keep it open to avoid race-conditions on renames. That said, if data are stored in companion files, moving the directory will create a race when joblib tries to access the companion files. More user-friendly error message
6,865
en
0.869886
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. try: from collections.abc import Iterable except ImportError: from collections import Iterable import contextlib import itertools import logging import re import warnings from typing import Optional, Tuple import numpy as np import torch from fairseq.file_io import PathManager from fairseq import utils import os logger = logging.getLogger(__name__) def infer_language_pair(path): """Infer language pair from filename: <split>.<lang1>-<lang2>.(...).idx""" src, dst = None, None for filename in PathManager.ls(path): parts = filename.split(".") if len(parts) >= 3 and len(parts[1].split("-")) == 2: return parts[1].split("-") return src, dst def collate_tokens( values, pad_idx, eos_idx=None, left_pad=False, move_eos_to_beginning=False, pad_to_length=None, pad_to_multiple=1, pad_to_bsz=None, ): """Convert a list of 1d tensors into a padded 2d tensor.""" size = max(v.size(0) for v in values) size = size if pad_to_length is None else max(size, pad_to_length) if pad_to_multiple != 1 and size % pad_to_multiple != 0: size = int(((size - 0.1) // pad_to_multiple + 1) * pad_to_multiple) batch_size = len(values) if pad_to_bsz is None else max(len(values), pad_to_bsz) res = values[0].new(batch_size, size).fill_(pad_idx) def copy_tensor(src, dst): assert dst.numel() == src.numel() if move_eos_to_beginning: if eos_idx is None: # if no eos_idx is specified, then use the last token in src dst[0] = src[-1] else: dst[0] = eos_idx dst[1:] = src[:-1] else: dst.copy_(src) for i, v in enumerate(values): copy_tensor(v, res[i][size - len(v) :] if left_pad else res[i][: len(v)]) return res def load_indexed_dataset( path, dictionary=None, dataset_impl=None, combine=False, default="cached" ): """A helper function for loading indexed datasets. Args: path (str): path to indexed dataset (e.g., 'data-bin/train') dictionary (~fairseq.data.Dictionary): data dictionary dataset_impl (str, optional): which dataset implementation to use. If not provided, it will be inferred automatically. For legacy indexed data we use the 'cached' implementation by default. combine (bool, optional): automatically load and combine multiple datasets. For example, if *path* is 'data-bin/train', then we will combine 'data-bin/train', 'data-bin/train1', ... and return a single ConcatDataset instance. """ import fairseq.data.indexed_dataset as indexed_dataset from fairseq.data.concat_dataset import ConcatDataset datasets = [] for k in itertools.count(): path_k = path + (str(k) if k > 0 else "") try: path_k = indexed_dataset.get_indexed_dataset_to_local(path_k) except Exception as e: if "StorageException: [404] Path not found" in str(e): logger.warning(f"path_k: {e} not found") else: raise e dataset_impl_k = dataset_impl if dataset_impl_k is None: dataset_impl_k = indexed_dataset.infer_dataset_impl(path_k) dataset = indexed_dataset.make_dataset( path_k, impl=dataset_impl_k or default, fix_lua_indexing=True, dictionary=dictionary, ) if dataset is None: break logger.info("loaded {:,} examples from: {}".format(len(dataset), path_k)) datasets.append(dataset) if not combine: break if len(datasets) == 0: return None elif len(datasets) == 1: return datasets[0] else: return ConcatDataset(datasets) @contextlib.contextmanager def numpy_seed(seed, *addl_seeds): """Context manager which seeds the NumPy PRNG with the specified seed and restores the state afterward""" if seed is None: yield return if len(addl_seeds) > 0: seed = int(hash((seed, *addl_seeds)) % 1e6) state = np.random.get_state() np.random.seed(seed) try: yield finally: np.random.set_state(state) def collect_filtered(function, iterable, filtered): """ Similar to :func:`filter` but collects filtered elements in ``filtered``. Args: function (callable): function that returns ``False`` for elements that should be filtered iterable (iterable): iterable to filter filtered (list): list to store filtered elements """ for el in iterable: if function(el): yield el else: filtered.append(el) def _filter_by_size_dynamic(indices, size_fn, max_positions, raise_exception=False): def compare_leq(a, b): return a <= b if not isinstance(a, tuple) else max(a) <= b def check_size(idx): if isinstance(max_positions, float) or isinstance(max_positions, int): return size_fn(idx) <= max_positions elif isinstance(max_positions, dict): idx_size = size_fn(idx) assert isinstance(idx_size, dict) intersect_keys = set(max_positions.keys()) & set(idx_size.keys()) return all( all( a is None or b is None or a <= b for a, b in zip(idx_size[key], max_positions[key]) ) for key in intersect_keys ) else: # For MultiCorpusSampledDataset, will generalize it later if not isinstance(size_fn(idx), Iterable): return all(size_fn(idx) <= b for b in max_positions) return all( a is None or b is None or a <= b for a, b in zip(size_fn(idx), max_positions) ) ignored = [] itr = collect_filtered(check_size, indices, ignored) indices = np.fromiter(itr, dtype=np.int64, count=-1) return indices, ignored def filter_by_size(indices, dataset, max_positions, raise_exception=False): """ [deprecated] Filter indices based on their size. Use `FairseqDataset::filter_indices_by_size` instead. Args: indices (List[int]): ordered list of dataset indices dataset (FairseqDataset): fairseq dataset instance max_positions (tuple): filter elements larger than this size. Comparisons are done component-wise. raise_exception (bool, optional): if ``True``, raise an exception if any elements are filtered (default: False). """ warnings.warn( "data_utils.filter_by_size is deprecated. " "Use `FairseqDataset::filter_indices_by_size` instead.", stacklevel=2, ) if isinstance(max_positions, float) or isinstance(max_positions, int): if hasattr(dataset, "sizes") and isinstance(dataset.sizes, np.ndarray): ignored = indices[dataset.sizes[indices] > max_positions].tolist() indices = indices[dataset.sizes[indices] <= max_positions] elif ( hasattr(dataset, "sizes") and isinstance(dataset.sizes, list) and len(dataset.sizes) == 1 ): ignored = indices[dataset.sizes[0][indices] > max_positions].tolist() indices = indices[dataset.sizes[0][indices] <= max_positions] else: indices, ignored = _filter_by_size_dynamic( indices, dataset.size, max_positions ) else: indices, ignored = _filter_by_size_dynamic(indices, dataset.size, max_positions) if len(ignored) > 0 and raise_exception: raise Exception( ( "Size of sample #{} is invalid (={}) since max_positions={}, " "skip this example with --skip-invalid-size-inputs-valid-test" ).format(ignored[0], dataset.size(ignored[0]), max_positions) ) if len(ignored) > 0: logger.warning( ( "{} samples have invalid sizes and will be skipped, " "max_positions={}, first few sample ids={}" ).format(len(ignored), max_positions, ignored[:10]) ) return indices def filter_paired_dataset_indices_by_size(src_sizes, tgt_sizes, indices, max_sizes): """Filter a list of sample indices. Remove those that are longer than specified in max_sizes. Args: indices (np.array): original array of sample indices max_sizes (int or list[int] or tuple[int]): max sample size, can be defined separately for src and tgt (then list or tuple) Returns: np.array: filtered sample array list: list of removed indices """ if max_sizes is None: return indices, [] if type(max_sizes) in (int, float): max_src_size, max_tgt_size = max_sizes, max_sizes else: max_src_size, max_tgt_size = max_sizes if tgt_sizes is None: ignored = indices[src_sizes[indices] > max_src_size] else: ignored = indices[ (src_sizes[indices] > max_src_size) | (tgt_sizes[indices] > max_tgt_size) ] if len(ignored) > 0: if tgt_sizes is None: indices = indices[src_sizes[indices] <= max_src_size] else: indices = indices[ (src_sizes[indices] <= max_src_size) & (tgt_sizes[indices] <= max_tgt_size) ] return indices, ignored.tolist() def batch_by_size( indices, num_tokens_fn, num_tokens_vec=None, max_tokens=None, max_sentences=None, required_batch_size_multiple=1, fixed_shapes=None, ): """ Yield mini-batches of indices bucketed by size. Batches may contain sequences of different lengths. Args: indices (List[int]): ordered list of dataset indices num_tokens_fn (callable): function that returns the number of tokens at a given index num_tokens_vec (List[int], optional): precomputed vector of the number of tokens for each index in indices (to enable faster batch generation) max_tokens (int, optional): max number of tokens in each batch (default: None). max_sentences (int, optional): max number of sentences in each batch (default: None). required_batch_size_multiple (int, optional): require batch size to be less than N or a multiple of N (default: 1). fixed_shapes (List[Tuple[int, int]], optional): if given, batches will only be created with the given shapes. *max_sentences* and *required_batch_size_multiple* will be ignored (default: None). """ try: from fairseq.data.data_utils_fast import ( batch_by_size_fn, batch_by_size_vec, batch_fixed_shapes_fast, ) except ImportError: raise ImportError( "Please build Cython components with: " "`python setup.py build_ext --inplace`" ) except ValueError: raise ValueError( "Please build (or rebuild) Cython components with `python setup.py build_ext --inplace`." ) # added int() to avoid TypeError: an integer is required max_tokens = int(max_tokens) if max_tokens is not None else -1 max_sentences = max_sentences if max_sentences is not None else -1 bsz_mult = required_batch_size_multiple if not isinstance(indices, np.ndarray): indices = np.fromiter(indices, dtype=np.int64, count=-1) if num_tokens_vec is not None and not isinstance(num_tokens_vec, np.ndarray): num_tokens_vec = np.fromiter(num_tokens_vec, dtype=np.int64, count=-1) if fixed_shapes is None: if num_tokens_vec is None: return batch_by_size_fn( indices, num_tokens_fn, max_tokens, max_sentences, bsz_mult, ) else: return batch_by_size_vec( indices, num_tokens_vec, max_tokens, max_sentences, bsz_mult, ) else: fixed_shapes = np.array(fixed_shapes, dtype=np.int64) sort_order = np.lexsort( [ fixed_shapes[:, 1].argsort(), # length fixed_shapes[:, 0].argsort(), # bsz ] ) fixed_shapes_sorted = fixed_shapes[sort_order] return batch_fixed_shapes_fast(indices, num_tokens_fn, fixed_shapes_sorted) def post_process(sentence: str, symbol: str): if symbol == "sentencepiece": sentence = sentence.replace(" ", "").replace("\u2581", " ").strip() elif symbol == "wordpiece": sentence = sentence.replace(" ", "").replace("_", " ").strip() elif symbol == "letter": sentence = sentence.replace(" ", "").replace("|", " ").strip() elif symbol == "silence": import re sentence = sentence.replace("<SIL>", "") sentence = re.sub(" +", " ", sentence).strip() elif symbol == "_EOW": sentence = sentence.replace(" ", "").replace("_EOW", " ").strip() elif symbol in {"subword_nmt", "@@ ", "@@"}: if symbol == "subword_nmt": symbol = "@@ " sentence = (sentence + " ").replace(symbol, "").rstrip() elif symbol == "none": pass elif symbol is not None: raise NotImplementedError(f"Unknown post_process option: {symbol}") return sentence def compute_mask_indices( shape: Tuple[int, int], padding_mask: Optional[torch.Tensor], mask_prob: float, mask_length: int, mask_type: str = "static", mask_other: float = 0.0, min_masks: int = 0, no_overlap: bool = False, min_space: int = 0, ) -> np.ndarray: """ Computes random mask spans for a given shape Args: shape: the the shape for which to compute masks. should be of size 2 where first element is batch size and 2nd is timesteps padding_mask: optional padding mask of the same size as shape, which will prevent masking padded elements mask_prob: probability for each token to be chosen as start of the span to be masked. this will be multiplied by number of timesteps divided by length of mask span to mask approximately this percentage of all elements. however due to overlaps, the actual number will be smaller (unless no_overlap is True) mask_type: how to compute mask lengths static = fixed size uniform = sample from uniform distribution [mask_other, mask_length*2] normal = sample from normal distribution with mean mask_length and stdev mask_other. mask is min 1 element poisson = sample from possion distribution with lambda = mask length min_masks: minimum number of masked spans no_overlap: if false, will switch to an alternative recursive algorithm that prevents spans from overlapping min_space: only used if no_overlap is True, this is how many elements to keep unmasked between spans """ bsz, all_sz = shape mask = np.full((bsz, all_sz), False) all_num_mask = int( # add a random number for probabilistic rounding mask_prob * all_sz / float(mask_length) + np.random.rand() ) all_num_mask = max(min_masks, all_num_mask) mask_idcs = [] for i in range(bsz): if padding_mask is not None: sz = all_sz - padding_mask[i].long().sum().item() num_mask = int( # add a random number for probabilistic rounding mask_prob * sz / float(mask_length) + np.random.rand() ) num_mask = max(min_masks, num_mask) else: sz = all_sz num_mask = all_num_mask if mask_type == "static": lengths = np.full(num_mask, mask_length) elif mask_type == "uniform": lengths = np.random.randint(mask_other, mask_length * 2 + 1, size=num_mask) elif mask_type == "normal": lengths = np.random.normal(mask_length, mask_other, size=num_mask) lengths = [max(1, int(round(x))) for x in lengths] elif mask_type == "poisson": lengths = np.random.poisson(mask_length, size=num_mask) lengths = [int(round(x)) for x in lengths] else: raise Exception("unknown mask selection " + mask_type) if sum(lengths) == 0: lengths[0] = min(mask_length, sz - 1) if no_overlap: mask_idc = [] def arrange(s, e, length, keep_length): span_start = np.random.randint(s, e - length) mask_idc.extend(span_start + i for i in range(length)) new_parts = [] if span_start - s - min_space >= keep_length: new_parts.append((s, span_start - min_space + 1)) if e - span_start - keep_length - min_space > keep_length: new_parts.append((span_start + length + min_space, e)) return new_parts parts = [(0, sz)] min_length = min(lengths) for length in sorted(lengths, reverse=True): lens = np.fromiter( (e - s if e - s >= length + min_space else 0 for s, e in parts), np.int, ) l_sum = np.sum(lens) if l_sum == 0: break probs = lens / np.sum(lens) c = np.random.choice(len(parts), p=probs) s, e = parts.pop(c) parts.extend(arrange(s, e, length, min_length)) mask_idc = np.asarray(mask_idc) else: min_len = min(lengths) if sz - min_len <= num_mask: min_len = sz - num_mask - 1 mask_idc = np.random.choice(sz - min_len, num_mask, replace=False) mask_idc = np.asarray( [ mask_idc[j] + offset for j in range(len(mask_idc)) for offset in range(lengths[j]) ] ) mask_idcs.append(np.unique(mask_idc[mask_idc < sz])) min_len = min([len(m) for m in mask_idcs]) for i, mask_idc in enumerate(mask_idcs): if len(mask_idc) > min_len: mask_idc = np.random.choice(mask_idc, min_len, replace=False) mask[i, mask_idc] = True return mask def get_mem_usage(): try: import psutil mb = 1024 * 1024 return f"used={psutil.virtual_memory().used / mb}Mb; avail={psutil.virtual_memory().available / mb}Mb" except ImportError: return "N/A" # lens: torch.LongTensor # returns: torch.BoolTensor def lengths_to_padding_mask(lens): bsz, max_lens = lens.size(0), torch.max(lens).item() mask = torch.arange(max_lens).to(lens.device).view(1, max_lens) mask = mask.expand(bsz, -1) >= lens.view(bsz, 1).expand(-1, max_lens) return mask # lens: torch.LongTensor # returns: torch.BoolTensor def lengths_to_mask(lens): return ~lengths_to_padding_mask(lens) def get_buckets(sizes, num_buckets): buckets = np.unique( np.percentile( sizes, np.linspace(0, 100, num_buckets + 1), interpolation="lower", )[1:] ) return buckets def get_bucketed_sizes(orig_sizes, buckets): sizes = np.copy(orig_sizes) assert np.min(sizes) >= 0 start_val = -1 for end_val in buckets: mask = (sizes > start_val) & (sizes <= end_val) sizes[mask] = end_val start_val = end_val return sizes def _find_extra_valid_paths(dataset_path: str) -> set: paths = utils.split_paths(dataset_path) all_valid_paths = set() for sub_dir in paths: contents = PathManager.ls(sub_dir) valid_paths = [c for c in contents if re.match("valid*[0-9].*", c) is not None] all_valid_paths |= {os.path.basename(p) for p in valid_paths} # Remove .bin, .idx etc roots = {os.path.splitext(p)[0] for p in all_valid_paths} return roots def raise_if_valid_subsets_unintentionally_ignored(train_cfg) -> None: """Raises if there are paths matching 'valid*[0-9].*' which are not combined or ignored.""" if ( train_cfg.dataset.ignore_unused_valid_subsets or train_cfg.dataset.combine_valid_subsets or train_cfg.dataset.disable_validation or not hasattr(train_cfg.task, "data") ): return other_paths = _find_extra_valid_paths(train_cfg.task.data) specified_subsets = train_cfg.dataset.valid_subset.split(",") ignored_paths = [p for p in other_paths if p not in specified_subsets] if ignored_paths: advice = "Set --combine-val to combine them or --ignore-unused-valid-subsets to ignore them." msg = f"Valid paths {ignored_paths} will be ignored. {advice}" raise ValueError(msg)
fairseq/data/data_utils.py
21,283
Yield mini-batches of indices bucketed by size. Batches may contain sequences of different lengths. Args: indices (List[int]): ordered list of dataset indices num_tokens_fn (callable): function that returns the number of tokens at a given index num_tokens_vec (List[int], optional): precomputed vector of the number of tokens for each index in indices (to enable faster batch generation) max_tokens (int, optional): max number of tokens in each batch (default: None). max_sentences (int, optional): max number of sentences in each batch (default: None). required_batch_size_multiple (int, optional): require batch size to be less than N or a multiple of N (default: 1). fixed_shapes (List[Tuple[int, int]], optional): if given, batches will only be created with the given shapes. *max_sentences* and *required_batch_size_multiple* will be ignored (default: None). Convert a list of 1d tensors into a padded 2d tensor. Similar to :func:`filter` but collects filtered elements in ``filtered``. Args: function (callable): function that returns ``False`` for elements that should be filtered iterable (iterable): iterable to filter filtered (list): list to store filtered elements Computes random mask spans for a given shape Args: shape: the the shape for which to compute masks. should be of size 2 where first element is batch size and 2nd is timesteps padding_mask: optional padding mask of the same size as shape, which will prevent masking padded elements mask_prob: probability for each token to be chosen as start of the span to be masked. this will be multiplied by number of timesteps divided by length of mask span to mask approximately this percentage of all elements. however due to overlaps, the actual number will be smaller (unless no_overlap is True) mask_type: how to compute mask lengths static = fixed size uniform = sample from uniform distribution [mask_other, mask_length*2] normal = sample from normal distribution with mean mask_length and stdev mask_other. mask is min 1 element poisson = sample from possion distribution with lambda = mask length min_masks: minimum number of masked spans no_overlap: if false, will switch to an alternative recursive algorithm that prevents spans from overlapping min_space: only used if no_overlap is True, this is how many elements to keep unmasked between spans [deprecated] Filter indices based on their size. Use `FairseqDataset::filter_indices_by_size` instead. Args: indices (List[int]): ordered list of dataset indices dataset (FairseqDataset): fairseq dataset instance max_positions (tuple): filter elements larger than this size. Comparisons are done component-wise. raise_exception (bool, optional): if ``True``, raise an exception if any elements are filtered (default: False). Filter a list of sample indices. Remove those that are longer than specified in max_sizes. Args: indices (np.array): original array of sample indices max_sizes (int or list[int] or tuple[int]): max sample size, can be defined separately for src and tgt (then list or tuple) Returns: np.array: filtered sample array list: list of removed indices Infer language pair from filename: <split>.<lang1>-<lang2>.(...).idx A helper function for loading indexed datasets. Args: path (str): path to indexed dataset (e.g., 'data-bin/train') dictionary (~fairseq.data.Dictionary): data dictionary dataset_impl (str, optional): which dataset implementation to use. If not provided, it will be inferred automatically. For legacy indexed data we use the 'cached' implementation by default. combine (bool, optional): automatically load and combine multiple datasets. For example, if *path* is 'data-bin/train', then we will combine 'data-bin/train', 'data-bin/train1', ... and return a single ConcatDataset instance. Context manager which seeds the NumPy PRNG with the specified seed and restores the state afterward Raises if there are paths matching 'valid*[0-9].*' which are not combined or ignored. Copyright (c) Facebook, Inc. and its affiliates. This source code is licensed under the MIT license found in the LICENSE file in the root directory of this source tree. if no eos_idx is specified, then use the last token in src For MultiCorpusSampledDataset, will generalize it later added int() to avoid TypeError: an integer is required length bsz add a random number for probabilistic rounding add a random number for probabilistic rounding lens: torch.LongTensor returns: torch.BoolTensor lens: torch.LongTensor returns: torch.BoolTensor Remove .bin, .idx etc
4,809
en
0.751023
from flask import Flask from flask_bootstrap import Bootstrap from config import config_options bootstrap = Bootstrap() def create_app(config_name): app = Flask(__name__) # create ap configurations app.config.from_object(config_options[config_name]) # initialize flask extensions bootstrap.init_app(app) # Registering the blueprint from .main import main as main_blueprint app.register_blueprint(main_blueprint) # setting config from .requests import configure_request configure_request(app) return app
app/__init__.py
560
create ap configurations initialize flask extensions Registering the blueprint setting config
93
en
0.404373
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from pathlib import Path import pandas as pd from watchdog.events import FileSystemEventHandler from watchdog.observers import Observer from .processing import process_daily class FileWatcher(FileSystemEventHandler): def __init__(self, path, sheets): super().__init__() self._path = Path(path) self._sheets = sheets self._observer = Observer() self._observer.schedule(self, str(self._path.parent)) def _update_sheets(self): self._sheets.clear() sheets = pd.read_excel(self._path, None) for key, val in sheets.items(): if key[:2] == '20': sheets[key] = process_daily(val) self._sheets.update(sheets) def on_created(self, event): if Path(event.src_path) == self._path: self._update_sheets() def run(self): self._update_sheets() self._observer.start() print('running...') def stop(self): self._observer.stop() self._observer.join() print('stopped')
src/budget/watcher.py
1,241
This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/.
192
en
0.934305
#!/usr/bin/env python # -*- coding: utf-8 -*- from runner.koan import * class AboutTuples(Koan): def test_creating_a_tuple(self): count_of_three = (1, 2, 5) self.assertEqual(5, count_of_three[2]) def test_tuples_are_immutable_so_item_assignment_is_not_possible(self): count_of_three = (1, 2, 5) try: count_of_three[2] = "three" except TypeError as ex: msg = ex.args[0] # Note, assertRegexpMatches() uses regular expression pattern matching, # so you don't have to copy the whole message. self.assertRegexpMatches(msg,"'tuple' object does not support item assignment") def test_tuples_are_immutable_so_appending_is_not_possible(self): count_of_three = (1, 2, 5) with self.assertRaises(AttributeError): count_of_three.append("boom") # Tuples are less flexible than lists, but faster. def test_tuples_can_only_be_changed_through_replacement(self): count_of_three = (1, 2, 5) list_count = list(count_of_three) list_count.append("boom") count_of_three = tuple(list_count) self.assertEqual((1,2,5,"boom"), count_of_three) def test_tuples_of_one_look_peculiar(self): self.assertEqual(int, (1).__class__) self.assertEqual(tuple, (1,).__class__) self.assertEqual(tuple, ("Hello comma!", ).__class__) def test_tuple_constructor_can_be_surprising(self): self.assertEqual(('S', 'u', 'r', 'p', 'r', 'i', 's', 'e', '!'), tuple("Surprise!")) def test_creating_empty_tuples(self): self.assertEqual(tuple(), ()) self.assertEqual((), tuple()) #Sometimes less confusing def test_tuples_can_be_embedded(self): lat = (37, 14, 6, 'N') lon = (115, 48, 40, 'W') place = ('Area 51', lat, lon) self.assertEqual(('Area 51', (37, 14, 6, 'N'), (115, 48, 40, 'W')), place) def test_tuples_are_good_for_representing_records(self): locations = [ ("Illuminati HQ", (38, 52, 15.56, 'N'), (77, 3, 21.46, 'W')), ("Stargate B", (41, 10, 43.92, 'N'), (1, 49, 34.29, 'W')), ] locations.append( ("Cthulu", (26, 40, 1, 'N'), (70, 45, 7, 'W')) ) self.assertEqual("Cthulu", locations[2][0]) self.assertEqual(15.56, locations[0][1][2])
python3/koans/about_tuples.py
2,347
!/usr/bin/env python -*- coding: utf-8 -*- Note, assertRegexpMatches() uses regular expression pattern matching, so you don't have to copy the whole message. Tuples are less flexible than lists, but faster.Sometimes less confusing
230
en
0.813063
# Generated by Django 2.2.1 on 2019-05-17 22:55 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('wells', '0088_fieldsprovided'), ('wells', '0088_remove_well_ems_id'), ] operations = [ ]
app/backend/wells/migrations/0089_merge_20190517_2255.py
265
Generated by Django 2.2.1 on 2019-05-17 22:55
45
en
0.537287
# -*- coding: utf-8 -*- import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models from django.contrib.contenttypes import generic class Migration(SchemaMigration): def forwards(self, orm): # renaming ExperimentACL to 'ObjectACL' db.rename_table('tardis_portal_experimentacl', 'tardis_portal_objectacl') # changing experiment to content_types db.add_column('tardis_portal_objectacl', 'content_type', self.gf('django.db.models.fields.related.ForeignKey')( to=orm['contenttypes.ContentType'], default=1)) db.add_column('tardis_portal_objectacl', 'object_id', self.gf( 'django.db.models.fields.PositiveIntegerField')( default=1)) gfk = generic.GenericForeignKey('content_type', 'object_id') gfk.contribute_to_class(orm.ObjectACL, "content_object") def backwards(self, orm): # renaming ExperimentACL to 'ObjectACL' pass #raise RuntimeError('Cannot reverse this migration.') models = { 'auth.group': { 'Meta': {'object_name': 'Group'}, 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '80'}), 'permissions': ('django.db.models.fields.related.ManyToManyField', [], {'to': "orm['auth.Permission']", 'symmetrical': 'False', 'blank': 'True'}) }, 'auth.permission': { 'Meta': {'ordering': "('content_type__app_label', 'content_type__model', 'codename')", 'unique_together': "(('content_type', 'codename'),)", 'object_name': 'Permission'}, 'codename': ('django.db.models.fields.CharField', [], {'max_length': '100'}), 'content_type': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['contenttypes.ContentType']"}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '50'}) }, 'auth.user': { 'Meta': {'object_name': 'User'}, 'date_joined': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now'}), 'email': ('django.db.models.fields.EmailField', [], {'max_length': '75', 'blank': 'True'}), 'first_name': ('django.db.models.fields.CharField', [], {'max_length': '30', 'blank': 'True'}), 'groups': ('django.db.models.fields.related.ManyToManyField', [], {'to': "orm['auth.Group']", 'symmetrical': 'False', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'is_active': ('django.db.models.fields.BooleanField', [], {'default': 'True'}), 'is_staff': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'is_superuser': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'last_login': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now'}), 'last_name': ('django.db.models.fields.CharField', [], {'max_length': '30', 'blank': 'True'}), 'password': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'user_permissions': ('django.db.models.fields.related.ManyToManyField', [], {'to': "orm['auth.Permission']", 'symmetrical': 'False', 'blank': 'True'}), 'username': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '30'}) }, 'contenttypes.contenttype': { 'Meta': {'ordering': "('name',)", 'unique_together': "(('app_label', 'model'),)", 'object_name': 'ContentType', 'db_table': "'django_content_type'"}, 'app_label': ('django.db.models.fields.CharField', [], {'max_length': '100'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'model': ('django.db.models.fields.CharField', [], {'max_length': '100'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '100'}) }, 'tardis_portal.author_experiment': { 'Meta': {'ordering': "['order']", 'unique_together': "(('experiment', 'author'),)", 'object_name': 'Author_Experiment'}, 'author': ('django.db.models.fields.CharField', [], {'max_length': '255'}), 'experiment': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Experiment']"}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'order': ('django.db.models.fields.PositiveIntegerField', [], {}), 'url': ('django.db.models.fields.URLField', [], {'max_length': '2000', 'blank': 'True'}) }, 'tardis_portal.datafileparameter': { 'Meta': {'ordering': "['name']", 'object_name': 'DatafileParameter'}, 'datetime_value': ('django.db.models.fields.DateTimeField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.ParameterName']"}), 'numerical_value': ('django.db.models.fields.FloatField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}), 'parameterset': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.DatafileParameterSet']"}), 'string_value': ('django.db.models.fields.TextField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}) }, 'tardis_portal.datafileparameterset': { 'Meta': {'ordering': "['id']", 'object_name': 'DatafileParameterSet'}, 'dataset_file': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Dataset_File']"}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'schema': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Schema']"}) }, 'tardis_portal.dataset': { 'Meta': {'ordering': "['-id']", 'object_name': 'Dataset'}, 'description': ('django.db.models.fields.TextField', [], {'blank': 'True'}), 'experiments': ('django.db.models.fields.related.ManyToManyField', [], {'related_name': "'datasets'", 'symmetrical': 'False', 'to': "orm['tardis_portal.Experiment']"}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'immutable': ('django.db.models.fields.BooleanField', [], {'default': 'False'}) }, 'tardis_portal.dataset_file': { 'Meta': {'ordering': "['filename']", 'object_name': 'Dataset_File'}, 'created_time': ('django.db.models.fields.DateTimeField', [], {'null': 'True', 'blank': 'True'}), 'dataset': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Dataset']"}), 'filename': ('django.db.models.fields.CharField', [], {'max_length': '400'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'md5sum': ('django.db.models.fields.CharField', [], {'max_length': '32', 'blank': 'True'}), 'mimetype': ('django.db.models.fields.CharField', [], {'max_length': '80', 'blank': 'True'}), 'modification_time': ('django.db.models.fields.DateTimeField', [], {'null': 'True', 'blank': 'True'}), 'sha512sum': ('django.db.models.fields.CharField', [], {'max_length': '128', 'blank': 'True'}), 'size': ('django.db.models.fields.CharField', [], {'max_length': '400', 'blank': 'True'}) }, 'tardis_portal.datasetparameter': { 'Meta': {'ordering': "['name']", 'object_name': 'DatasetParameter'}, 'datetime_value': ('django.db.models.fields.DateTimeField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.ParameterName']"}), 'numerical_value': ('django.db.models.fields.FloatField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}), 'parameterset': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.DatasetParameterSet']"}), 'string_value': ('django.db.models.fields.TextField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}) }, 'tardis_portal.datasetparameterset': { 'Meta': {'ordering': "['id']", 'object_name': 'DatasetParameterSet'}, 'dataset': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Dataset']"}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'schema': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Schema']"}) }, 'tardis_portal.experiment': { 'Meta': {'object_name': 'Experiment'}, 'approved': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'created_by': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['auth.User']"}), 'created_time': ('django.db.models.fields.DateTimeField', [], {'auto_now_add': 'True', 'blank': 'True'}), 'description': ('django.db.models.fields.TextField', [], {'blank': 'True'}), 'end_time': ('django.db.models.fields.DateTimeField', [], {'null': 'True', 'blank': 'True'}), 'handle': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'institution_name': ('django.db.models.fields.CharField', [], {'default': "'Monash University'", 'max_length': '400'}), 'license': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.License']", 'null': 'True', 'blank': 'True'}), 'locked': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'public_access': ('django.db.models.fields.PositiveSmallIntegerField', [], {'default': '1'}), 'start_time': ('django.db.models.fields.DateTimeField', [], {'null': 'True', 'blank': 'True'}), 'title': ('django.db.models.fields.CharField', [], {'max_length': '400'}), 'update_time': ('django.db.models.fields.DateTimeField', [], {'auto_now': 'True', 'blank': 'True'}), 'url': ('django.db.models.fields.URLField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}) }, 'tardis_portal.experimentparameter': { 'Meta': {'ordering': "['name']", 'object_name': 'ExperimentParameter'}, 'datetime_value': ('django.db.models.fields.DateTimeField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.ParameterName']"}), 'numerical_value': ('django.db.models.fields.FloatField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}), 'parameterset': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.ExperimentParameterSet']"}), 'string_value': ('django.db.models.fields.TextField', [], {'db_index': 'True', 'null': 'True', 'blank': 'True'}) }, 'tardis_portal.experimentparameterset': { 'Meta': {'ordering': "['id']", 'object_name': 'ExperimentParameterSet'}, 'experiment': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Experiment']"}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'schema': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Schema']"}) }, 'tardis_portal.freetextsearchfield': { 'Meta': {'object_name': 'FreeTextSearchField'}, 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'parameter_name': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.ParameterName']"}) }, 'tardis_portal.groupadmin': { 'Meta': {'object_name': 'GroupAdmin'}, 'group': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['auth.Group']"}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'user': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['auth.User']"}) }, 'tardis_portal.license': { 'Meta': {'object_name': 'License'}, 'allows_distribution': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'image_url': ('django.db.models.fields.URLField', [], {'max_length': '2000', 'blank': 'True'}), 'internal_description': ('django.db.models.fields.TextField', [], {}), 'is_active': ('django.db.models.fields.BooleanField', [], {'default': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '400'}), 'url': ('django.db.models.fields.URLField', [], {'unique': 'True', 'max_length': '2000'}) }, 'tardis_portal.location': { 'Meta': {'object_name': 'Location'}, 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'is_available': ('django.db.models.fields.BooleanField', [], {'default': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '10'}), 'priority': ('django.db.models.fields.IntegerField', [], {}), 'transfer_provider': ('django.db.models.fields.CharField', [], {'default': "'local'", 'max_length': '10'}), 'type': ('django.db.models.fields.CharField', [], {'max_length': '10'}), 'url': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '400'}) }, 'tardis_portal.objectacl': { 'Meta': {'ordering': "['content_type', 'object_id']", 'object_name': 'ObjectACL'}, 'aclOwnershipType': ('django.db.models.fields.IntegerField', [], {'default': '1'}), 'canDelete': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'canRead': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'canWrite': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'content_type': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['contenttypes.ContentType']"}), 'effectiveDate': ('django.db.models.fields.DateField', [], {'null': 'True', 'blank': 'True'}), 'entityId': ('django.db.models.fields.CharField', [], {'max_length': '320'}), 'experiment': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Experiment']"}), 'expiryDate': ('django.db.models.fields.DateField', [], {'null': 'True', 'blank': 'True'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'isOwner': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'object_id': ('django.db.models.fields.PositiveIntegerField', [], {}), 'pluginId': ('django.db.models.fields.CharField', [], {'max_length': '30'}) }, 'tardis_portal.parametername': { 'Meta': {'ordering': "('order', 'name')", 'unique_together': "(('schema', 'name'),)", 'object_name': 'ParameterName'}, 'choices': ('django.db.models.fields.CharField', [], {'max_length': '500', 'blank': 'True'}), 'comparison_type': ('django.db.models.fields.IntegerField', [], {'default': '1'}), 'data_type': ('django.db.models.fields.IntegerField', [], {'default': '2'}), 'full_name': ('django.db.models.fields.CharField', [], {'max_length': '60'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'immutable': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'is_searchable': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '60'}), 'order': ('django.db.models.fields.PositiveIntegerField', [], {'default': '9999', 'null': 'True', 'blank': 'True'}), 'schema': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Schema']"}), 'units': ('django.db.models.fields.CharField', [], {'max_length': '60', 'blank': 'True'}) }, 'tardis_portal.providerparameter': { 'Meta': {'unique_together': "(('location', 'name'),)", 'object_name': 'ProviderParameter'}, 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'location': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Location']"}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '10'}), 'value': ('django.db.models.fields.CharField', [], {'max_length': '80', 'blank': 'True'}) }, 'tardis_portal.replica': { 'Meta': {'unique_together': "(('datafile', 'location'),)", 'object_name': 'Replica'}, 'datafile': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Dataset_File']"}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'location': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Location']"}), 'protocol': ('django.db.models.fields.CharField', [], {'max_length': '10', 'blank': 'True'}), 'stay_remote': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'url': ('django.db.models.fields.CharField', [], {'max_length': '400'}), 'verified': ('django.db.models.fields.BooleanField', [], {'default': 'False'}) }, 'tardis_portal.schema': { 'Meta': {'object_name': 'Schema'}, 'hidden': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'immutable': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '50', 'null': 'True', 'blank': 'True'}), 'namespace': ('django.db.models.fields.URLField', [], {'unique': 'True', 'max_length': '255'}), 'subtype': ('django.db.models.fields.CharField', [], {'max_length': '30', 'null': 'True', 'blank': 'True'}), 'type': ('django.db.models.fields.IntegerField', [], {'default': '1'}) }, 'tardis_portal.token': { 'Meta': {'object_name': 'Token'}, 'experiment': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.Experiment']"}), 'expiry_date': ('django.db.models.fields.DateField', [], {'default': 'datetime.datetime(2013, 7, 18, 0, 0)'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'token': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '30'}), 'user': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['auth.User']"}) }, 'tardis_portal.userauthentication': { 'Meta': {'object_name': 'UserAuthentication'}, 'authenticationMethod': ('django.db.models.fields.CharField', [], {'max_length': '30'}), 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'userProfile': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['tardis_portal.UserProfile']"}), 'username': ('django.db.models.fields.CharField', [], {'max_length': '50'}) }, 'tardis_portal.userprofile': { 'Meta': {'object_name': 'UserProfile'}, 'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'isDjangoAccount': ('django.db.models.fields.BooleanField', [], {'default': 'True'}), 'user': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['auth.User']", 'unique': 'True'}) } } complete_apps = ['tardis_portal']
tardis/tardis_portal/south_migrations/0022_object_acl_1.py
21,000
-*- coding: utf-8 -*- renaming ExperimentACL to 'ObjectACL' changing experiment to content_types renaming ExperimentACL to 'ObjectACL'raise RuntimeError('Cannot reverse this migration.')
186
en
0.559006
import discord from discord.ext import commands import asyncio import wolframalpha from aiohttp import ClientSession from html2text import html2text from random import choice, randint from re import sub #setup wolframalpha API client = wolframalpha.Client(open('WA_KEY').readline().rstrip()) class Api(commands.Cog): """Get random cute pics""" def __init__(self, bot): self.bot = bot self.colours = [0x1abc9c, 0x11806a, 0x2ecc71, 0x1f8b4c, 0x3498db, 0x206694, 0x9b59b6, 0x71368a, 0xe91e63, 0xad1457, 0xf1c40f, 0xc27c0e, 0xa84300, 0xe74c3c, 0x992d22, 0x95a5a6, 0x607d8b, 0x979c9f, 0x546e7a] @commands.command(name='ask', description="replies to a query with the short text answer of the wolfram alpha API", brief="wolfram alpha API") async def ask(self, ctx, *, query): res = client.query(query) if res['@success'] == 'false': strRes = "Couldn't find an answer" else: strRes = next(res.results).text embed = discord.Embed( title=query, description=strRes, color=self.bot.embed_color) await ctx.send(embed=embed) @commands.command(name='dog', description="send random dog picture", brief="send dog pic", aliases=['auau']) async def dog(self, ctx): while True: result, error = await get_json('https://random.dog/woof.json') if error: await ctx.send(error) return if result['url'].endswith('.mp4'): pass else: embed = discord.Embed(color=choice(self.colours)) embed.set_image(url=result['url']) await ctx.send(embed=embed) return @commands.command(name='cat', description="send random cat picture", brief="send cat pic", aliases=['antiauau', 'miau']) async def cat(self, ctx): result, error = await get_json('http://aws.random.cat/meow') if error: await ctx.send(error) return embed = discord.Embed(color=choice(self.colours)) embed.set_image(url=result['file']) await ctx.send(embed=embed) @commands.command(name='xkcd', brief="send xkcd comic") async def xkcd(self, ctx, args = None): """ send xkcd comic *xkcd -> sends newest comic *xkcd random -> sends random comic *xkcd [number] -> sends a specific comic """ url = None if not args: url = 'http://xkcd.com/info.0.json' elif args.isdigit(): url = f'http://xkcd.com/{int(args)}/info.0.json' elif args.lower() == 'random': result, error = await get_json('http://xkcd.com/info.0.json') if error: await ctx.send(error) return number = randint(0, result['num']) url = f'http://xkcd.com/{number}/info.0.json' result, error = await get_json(url) if error: await ctx.send(error) return embed = discord.Embed(color=choice(self.colours)) embed.set_image(url=result['img']) await ctx.send(embed=embed) @commands.command(name='lmgtfy', description="give link for let me google that for you", brief="let me google that for you") async def lmgtfy(self, ctx, *query): await ctx.send(f"http://lmgtfy.com/?q={'+'.join(query)}") @commands.command(name='lmddgtfy', description="give link for let me duck duck go that for you", brief="let me duck duck go that for you") async def lmddgtfy(self, ctx, *query): await ctx.send(f"http://lmddgtfy.net/?q={'%20'.join(query)}") @commands.command(name='urban', description="Get a urban defenition of a query", brief="search urban") async def urban(self, ctx, * query : str): url = f"http://api.urbandictionary.com/v0/define?term={'+'.join(query)}" result, error = await get_json(url) if error: await ctx.send(error) return if result["list"]: top_def = result['list'][0] embed = discord.Embed( title=f"Definition of {top_def['word']}", url=top_def['permalink'], description=top_def['definition'], color=self.bot.embed_color) embed.set_thumbnail( url = "http://campbelllawobserver.com/wp-content/uploads/2014/03/Urban-Dictionary-e1372286057646.png") embed.add_field( name="Example", value=top_def['example'], inline=False) embed.add_field( name=":thumbsup:", value=top_def['thumbs_up'], inline=True) embed.add_field( name=":thumbsdown:", value=top_def['thumbs_down'], inline=True) embed.set_footer(text=f"Submited by {top_def['author']}") await ctx.send(embed =embed) else: await ctx.send("Your query gave no results.") @commands.command(name='hoogle', brief="search hoogle") async def hoogle(self, ctx, * query : str): """Searches Hoggle and returns first two options Click title to see full search""" url = f"https://hoogle.haskell.org?mode=json&hoogle={'+'.join(query)}&start=1&count=1" result, error = await get_json(url) if error: await ctx.send(error) return embed = discord.Embed( title=f"Definition of {' '.join(query)}", url=f"https://hoogle.haskell.org/?hoogle={'+'.join(query)}", color=self.bot.embed_color) embed.set_thumbnail( url = "https://upload.wikimedia.org/wikipedia/commons/thumb/c/c3/Lambda-letter-lowercase-symbol-Garamond.svg/1200px-Lambda-letter-lowercase-symbol-Garamond.svg.png") if not result: embed.add_field( name = "No results found", value="*undefined*", inline=False) else: for l in result: val = "*Module:* " + l["module"]["name"] + "\n" val+= sub(r'\n{2,}', '\n\n', sub(r"\n +", "\n" , html2text(l["docs"]))) embed.add_field( name= html2text(l["item"]), value= val, inline=False) embed.set_footer(text="first option in Hoogle (Click title for more)") await ctx.send(embed=embed) async def get_json(url): try: async with ClientSession() as session: async with session.get(url) as response: result = await response.json() return result, None except: return None, "Something unexpected went wrong." def setup(bot): bot.add_cog(Api(bot))
extensions/api.py
7,242
Get random cute pics setup wolframalpha API
44
en
0.258199
""" Copyright 2020 Qiniu Cloud (qiniu.com) Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ import sys from .spl_packet_utils import * from .spl_base_command import SplBaseCommand class SplStreamingBatchCommand(SplBaseCommand): def process_data(self, argv=None, input_stream=sys.stdin.buffer, output_stream=sys.__stdout__.buffer): while True: execute_meta = self.process_protocol_execute(input_stream) resp = self.streaming_handle(self.lines) send_packet(output_stream, execute_meta, resp) self.lines = [] if self.is_finish: break
pdr_python_sdk/pdr_python_sdk/spl/spl_streaming_batch_command.py
1,097
Copyright 2020 Qiniu Cloud (qiniu.com) Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
557
en
0.849154
import argparse import configparser import json import os import sys import time import requests sys.path.append(os.path.join(os.path.dirname(__file__), "..")) from util import logger class ConversationScraper: """Scraper that retrieves, process and stores all messages belonging to a specific Facebook conversation""" REQUEST_WAIT = 10 ERROR_WAIT = 30 CONVERSATION_ENDMARK = "end_of_history" def __init__(self, convID, cookie, fb_dtsg, outDir): self._directory = os.path.join(outDir,str(convID)) self._convID = convID self._cookie = cookie self._fb_dtsg = fb_dtsg """ POST Request full form data (<ids_type> is "thread_fbids" for group conversations, "user_ids" otherwise) "messages[<ids_type>][][offset]": "", "messages[<ids_type>][][timestamp]": "", "messages[<ids_type>][][]": "", "client": "", "__user": "", "__a": "", "__dyn": "", "__req": "", "fb_dtsg": "", "ttstamp": "", "__rev": "" """ def generateRequestData(self, offset, timestamp, chunkSize, isGroupConversation=False): """Generate the data for the POST request. :return: the generated data """ ids_type = "thread_fbids" if isGroupConversation else "user_ids" dataForm = {"messages[{}][{}][offset]".format(ids_type, self._convID) : str(offset), "messages[{}][{}][timestamp]".format(ids_type, self._convID): timestamp, "messages[{}][{}][limit]".format(ids_type, self._convID): str(chunkSize), "client": "web_messenger", "__a": "", "__dyn": "", "__req": "", "fb_dtsg": self._fb_dtsg} return dataForm """ POST Request all header: "Host": "www.facebook.com", "Origin": "https://www.facebook.com", "Referer": "https://www.facebook.com", "accept-encoding": "gzip,deflate", "accept-language": "en-US,en;q=0.8", "cookie": "", "pragma": "no-cache", "user-agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_9_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/37.0.2062.122 Safari/537.36", "content-type": "application/x-www-form-urlencoded", "accept": "*/*", "cache-control": "no-cache" """ def executeRequest(self, requestData): """Executes the POST request and retrieves the correspondent response content. Request headers are generated here :return: the response content """ headers = {"Host": "www.facebook.com", "Origin":"https://www.facebook.com", "Referer":"https://www.facebook.com", "accept-encoding": "gzip,deflate", "accept-language": "en-US,en;q=0.8", "cookie": self._cookie, "pragma": "no-cache", "user-agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_9_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/37.0.2062.122 Safari/537.36", "content-type": "application/x-www-form-urlencoded", "accept": "*/*", "cache-control": "no-cache"} url = "https://www.facebook.com/ajax/mercury/thread_info.php" start = time.time() response = requests.post(url, data=requestData, headers=headers) end = time.time() logger.info("Retrieved in {0:.2f}s".format(end-start)) #Remove additional leading characters msgsData = response.text[9:] return msgsData def writeMessages(self, messages): with open(os.path.join(self._directory,"conversation.json"), 'w') as conv: conv.write(json.dumps(messages)) command = "python -mjson.tool " + os.path.join(self._directory, "conversation.json") + " > " + os.path.join(self._directory, "conversation.pretty.json") os.system(command) def scrapeConversation(self, merge, offset, timestampOffset, chunkSize, limit, isGroupConversation): """Retrieves conversation messages and stores them in a JSON file If merge is specified, the new messages will be merged with the previous version of the conversation, if present """ if merge: if not os.path.exists(os.path.join(self._directory,"conversation.json")): logger.error("Conversation not present. Merge operation not possible") return with open(os.path.join(self._directory,"conversation.json")) as conv: convMessages = json.load(conv) numMergedMsgs = 0 if not os.path.exists(self._directory): os.makedirs(self._directory) logger.info("Starting scraping of conversation {}".format(self._convID)) messages = [] msgsData = "" timestamp = "" if timestampOffset == 0 else str(timestampOffset) while self.CONVERSATION_ENDMARK not in msgsData: requestChunkSize = chunkSize if limit <= 0 else min(chunkSize, limit-len(messages)) reqData = self.generateRequestData(offset, timestamp, requestChunkSize, isGroupConversation) logger.info("Retrieving messages {}-{}".format(offset, requestChunkSize+offset)) msgsData = self.executeRequest(reqData) jsonData = json.loads(msgsData) if jsonData and ('payload' in jsonData) and jsonData['payload']: if ('actions' in jsonData['payload']) and jsonData['payload']['actions']: actions = jsonData['payload']['actions'] #case when the last message already present in the conversation #is older newer than the first one of the current retrieved chunk if merge and convMessages[-1]["timestamp"] > actions[0]["timestamp"]: for i, action in enumerate(actions): if convMessages[-1]["timestamp"] == actions[i]["timestamp"]: numMergedMsgs = len(actions[i+1:-1]) + len(messages) messages = convMessages + actions[i+1:-1] + messages break break #We retrieve one message two times, as the first one of the previous chunk #and as the last one of the new one. So we here remove the duplicate, #but only once we already retrieved at least one chunk if len(messages) == 0: messages = actions else: messages = actions[:-1] + messages #update timestamp timestamp = str(actions[0]["timestamp"]) else: if 'errorSummary' in jsonData: logger.error("Response error: " + jsonData['errorSummary']) else: logger.error("Response error. No messages found") logger.error(msgsData) return else: logger.error("Response error. Empty data or payload") logger.error(msgsData) logger.info("Retrying in {} seconds".format(self.ERROR_WAIT)) time.sleep(self.ERROR_WAIT) continue offset += chunkSize if limit!= 0 and len(messages) >= limit: break time.sleep(self.REQUEST_WAIT) if merge: logger.info("Successfully merged {} new messages".format(numMergedMsgs)) logger.info("Conversation total message count = {}".format(len(messages))) else: logger.info("Conversation scraped successfully. {} messages retrieved".format(len(messages))) self.writeMessages(messages) def main(args=None): parser = argparse.ArgumentParser(description='Conversation Scraper') parser.add_argument('--id', metavar='conversationID', dest='convID', required=True) parser.add_argument('--size', metavar='chunkSize', type=int, dest='chunkSize', default=2000, help="number of messages to retrieve for each request") #TODO not working, the timestamp seems the only relevant parameter parser.add_argument('--off', metavar='offset', type=int, dest='offset', default=0, help="messages number scraping offset") #TODO to test, ??better single var parser.add_argument('--date', metavar='offset', type=int, dest='timestampOffset', default=0, help="messages timestamp scraping offset, has precedence over messages number offset") parser.add_argument('--limit', type=int, dest='limit', default=0, help="number of messages to be retrieved") #Tells the program to try to merge the new messages with the previously scraped conversation #avoid the need to scrape it all from the beginning parser.add_argument('-m', dest='merge', action='store_true', help="merge the new messages with previously scraped conversation") parser.add_argument('-g', dest='isGroupConversation', action='store_true', help="specify if you want to scrape a group conversation") parser.set_defaults(merge=False) baseFolderPath = os.path.abspath(os.path.join(os.path.dirname(__file__), os.pardir, os.pardir)) parser.add_argument('--out', metavar='outputDir', dest='outDir', default=os.path.join(baseFolderPath, 'Messages')) parser.add_argument('--conf', metavar='configFilepath', dest='configFilepath', default=os.path.join(baseFolderPath, 'config.ini')) args = parser.parse_args() convID = args.convID chunkSize = args.chunkSize timestampOffset = args.timestampOffset offset = args.offset limit = args.limit merge = args.merge isGroupConversation = args.isGroupConversation outDir = args.outDir configFilepath = args.configFilepath DATA_SECTION = "User Data" config = configparser.ConfigParser(interpolation=None) config.read(configFilepath) cookie = config.get(DATA_SECTION, "Cookie") fb_dtsg = config.get(DATA_SECTION, "Fb_dtsg") scraper = ConversationScraper(convID, cookie, fb_dtsg, outDir) scraper.scrapeConversation(merge, offset, timestampOffset, chunkSize, limit, isGroupConversation) if __name__ == "__main__": main(sys.argv[1:])
src/util/conversationScraper.py
10,522
Scraper that retrieves, process and stores all messages belonging to a specific Facebook conversation Executes the POST request and retrieves the correspondent response content. Request headers are generated here :return: the response content Generate the data for the POST request. :return: the generated data Retrieves conversation messages and stores them in a JSON file If merge is specified, the new messages will be merged with the previous version of the conversation, if present Remove additional leading characterscase when the last message already present in the conversationis older newer than the first one of the current retrieved chunkWe retrieve one message two times, as the first one of the previous chunkand as the last one of the new one. So we here remove the duplicate,but only once we already retrieved at least one chunkupdate timestampTODO not working, the timestamp seems the only relevant parameterTODO to test, ??better single varTells the program to try to merge the new messages with the previously scraped conversationavoid the need to scrape it all from the beginning
1,099
en
0.848166
# 033 # Ask the user to enter two numbers. Use whole number division to divide # the first number by the second and also work out the remainder and # display the answer in a user-friendly way # (e.g. if they enter 7 and 2 display “7 divided by 2 is 3 # with 1 remaining”). from typing import List def check_num_list(prompt: str, max_length: int = 0, min_length: int = 0) -> List[float]: """Function to check if users input is a number, splitting number by spaces and checking that the correct amount of numbers are entered, returning them in a list""" while True: try: num = input(prompt) num = num.split(' ') if min_length: assert len(num) >= min_length, f'Please enter at least' \ f' {min_length} numbers' if max_length: assert len(num) <= max_length, f'Please enter no more ' \ f'than {max_length} ' \ f'numbers' for index, value in enumerate(num): num[index] = float(value) return num except Exception as e: print(e) def integer_devision_remainder(nums: List[float]) -> str: return f'{nums[0]} / {nums[1]} = {nums[0] // nums[1]} with' \ f' a remainder of {nums[0] % nums[1]}' if __name__ == '__main__': print('This program will tell you floor division and remainder ' 'between two numbers') request = 'Enter two numbers:' print(integer_devision_remainder(check_num_list(request, 2, 2)))
150-Challenges/Challenges 27 - 34/Challenge 33.py
1,668
Function to check if users input is a number, splitting number by spaces and checking that the correct amount of numbers are entered, returning them in a list 033 Ask the user to enter two numbers. Use whole number division to divide the first number by the second and also work out the remainder and display the answer in a user-friendly way (e.g. if they enter 7 and 2 display “7 divided by 2 is 3 with 1 remaining”).
421
en
0.890763
# Code made for Sergio Andrés Díaz Ariza # 23 March 2021 # License MIT # Transport Phenomena: Python Program-Assesment 3.3 import matplotlib.pyplot as plt import seaborn as sns import numpy as np sns.set() class Wire_Param: def Power_from_volt(self, V, R_Km, Dis_insul, Diameter_mm): As = (np.pi * (Diameter_mm / 1000) ** 2) / 4 R = R_Km * Dis_insul * As / 1000 W = (V ** 2) / R # Power generated = Egen return W # [w/m^3] def Param_m(self, h, Kcopper, Diameter_mm): m = np.sqrt((4 * h * 0.2) / (Kcopper * (Diameter_mm / 1000))) return m class Coeff_vector_B: def Eq_1(self, h, Tsurr, Egen, K, L): b1 = h * Tsurr - (Egen * (-L / 2)) + (h * (Egen * ((-L / 2) ** 2)) / 2 * K) return b1 def Eq_2(self, Tsurr): b2 = Tsurr return b2 def Eq_3(self): b3 = 0 return b3 def Eq_4(self): b4 = 0 return b4 class T_profile: def T_z_to_L_2(self, Egen, z1, Kcopper, C1, C2): T = ((-Egen * ((z1) ** 2)) / 2 * Kcopper) + C1 * (z1) + C2 return T def T_z_from_L_2_to_L(self, Tsurr, C3, C4, Param_m, z2): T = Tsurr + (C3 * np.exp((Param_m * z2))) + (C4 * np.exp((-Param_m * z2))) return T Wire = Wire_Param() Vect_B = Coeff_vector_B() Profile = T_profile() Egen = Wire.Power_from_volt(30e-6, 5.32, 0.2, 2.32) # [W/m^3] Param_m = Wire.Param_m(15, 386, 2.32) # Value L--> must be positive fo calculus b1 = Vect_B.Eq_1(15, 25 + 273.15, Egen, 386, -0.2) b2 = Vect_B.Eq_2(25 + 273.15) b3 = Vect_B.Eq_3() b4 = Vect_B.Eq_4() B = np.array([b1, b2, b3, b4]) A = np.array([[-386 + (15 * (-0.2)), 15, 0, 0], [0, 1, -1, -1], [1, 0, -Param_m, Param_m], [0, 0, (np.exp(Param_m * 0.2)) * ((-386 * Param_m) - 15), (np.exp(-Param_m * 0.2)) * ((386 * Param_m) - 15)]]) # A_inv = np.linalg.inv(A) ---> We can find the solution directly C = np.linalg.solve(A, B) # solve for Temperature Profile T_L_2 = Profile.T_z_to_L_2(Egen, np.arange(-0.2, 0.0, 1e-6), 386, C[0], C[1]) T_L = Profile.T_z_from_L_2_to_L(25 + 273.15, C[2], C[3], Param_m, np.arange(0.0, 0.2, 1e-6)) print(T_L_2) print(T_L) plt.figure(1) plt.plot(np.arange(0, 0.2, 1e-6), T_L_2, label='$0<L<0.2$') plt.plot(np.arange(0.2, 0.4, 1e-6), T_L, label='$0.2<L<0.4$') plt.title("Temperature Profile in a Wie\n with Insulated part", fontsize=16) plt.ylabel("Temperature $[K]$", fontsize=14) plt.xlabel("Long of wire $[m]$ ", fontsize=14) plt.legend() plt.show()
Heat_Transfer/3.3-Wire_Isolation_Base__Fin.py
2,514
Code made for Sergio Andrés Díaz Ariza 23 March 2021 License MIT Transport Phenomena: Python Program-Assesment 3.3 Power generated = Egen [w/m^3] [W/m^3] Value L--> must be positive fo calculus A_inv = np.linalg.inv(A) ---> We can find the solution directly solve for Temperature Profile
287
en
0.575088
# -------------------------------------------------------- # Fast R-CNN # Copyright (c) 2015 Microsoft # Licensed under The MIT License [see LICENSE for details] # Written by Ross Girshick # -------------------------------------------------------- """Factory method for easily getting imdbs by name.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function __sets = {} from ..datasets.pascal_voc import pascal_voc from ..datasets.coco import coco from ..datasets.imagenet import imagenet from ..datasets.vg import vg import numpy as np # Set up voc_<year>_<split> for year in ['2007', '2012']: for split in ['train', 'val', 'trainval', 'test']: name = 'voc_{}_{}'.format(year, split) __sets[name] = (lambda split=split, year=year: pascal_voc(split, year)) # Set up coco_2014_<split> for year in ['2014']: for split in ['train', 'val', 'minival', 'valminusminival', 'trainval']: name = 'coco_{}_{}'.format(year, split) __sets[name] = (lambda split=split, year=year: coco(split, year)) # Set up coco_2014_cap_<split> for year in ['2014']: for split in ['train', 'val', 'capval', 'valminuscapval', 'trainval']: name = 'coco_{}_{}'.format(year, split) __sets[name] = (lambda split=split, year=year: coco(split, year)) # Set up coco_2015_<split> for year in ['2015']: for split in ['test', 'test-dev']: name = 'coco_{}_{}'.format(year, split) __sets[name] = (lambda split=split, year=year: coco(split, year)) # Set up vg_<split> # for version in ['1600-400-20']: # for split in ['minitrain', 'train', 'minival', 'val', 'test']: # name = 'vg_{}_{}'.format(version,split) # __sets[name] = (lambda split=split, # version=version: vg(version, split)) for version in ['150-50-20', '150-50-50', '500-150-80', '750-250-150', '1750-700-450', '1600-400-20']: for split in ['minitrain', 'smalltrain', 'train', 'minival', 'smallval', 'val', 'test']: name = 'vg_{}_{}'.format(version, split) __sets[name] = (lambda split=split, version=version: vg(version, split)) # set up image net. for split in ['train', 'val', 'val1', 'val2', 'test']: name = 'imagenet_{}'.format(split) devkit_path = 'data/imagenet/ILSVRC/devkit' data_path = 'data/imagenet/ILSVRC' __sets[name] = (lambda split=split, devkit_path=devkit_path, data_path=data_path: imagenet(split, devkit_path, data_path)) def get_imdb(name): """Get an imdb (image database) by name.""" if name not in __sets: raise KeyError('Unknown dataset: {}'.format(name)) return __sets[name]() def list_imdbs(): """List all registered imdbs.""" return list(__sets.keys())
lib/datasets/factory.py
2,716
Get an imdb (image database) by name. List all registered imdbs. Factory method for easily getting imdbs by name. -------------------------------------------------------- Fast R-CNN Copyright (c) 2015 Microsoft Licensed under The MIT License [see LICENSE for details] Written by Ross Girshick -------------------------------------------------------- Set up voc_<year>_<split> Set up coco_2014_<split> Set up coco_2014_cap_<split> Set up coco_2015_<split> Set up vg_<split> for version in ['1600-400-20']: for split in ['minitrain', 'train', 'minival', 'val', 'test']: name = 'vg_{}_{}'.format(version,split) __sets[name] = (lambda split=split, version=version: vg(version, split)) set up image net.
728
en
0.541779
#!/usr/bin/python # -*- coding: utf-8 -*- """ Script to upload the mappings of Freebase to Wikidata. Can be easily adapted to upload other String identifiers as well This bot needs the dump from https://developers.google.com/freebase/data#freebase-wikidata-mappings The script takes a single parameter: -filename: the filename to read the freebase-wikidata mappings from; default: fb2w.nt.gz """ # # (C) Denny Vrandecic, 2013 # (C) Pywikibot team, 2013-2018 # # Distributed under the terms of the MIT license. # from __future__ import absolute_import, division, unicode_literals import gzip import os import sys import pywikibot class FreebaseMapperRobot(object): """Freebase Mapping bot.""" def __init__(self, filename): """Initializer.""" self.repo = pywikibot.Site('wikidata', 'wikidata').data_repository() self.filename = filename if not os.path.exists(self.filename): pywikibot.output('Cannot find %s. Try providing the absolute path.' % self.filename) sys.exit(1) def run(self): """Run the bot.""" # Set up some items we will use a lot. self.claim = pywikibot.Claim(self.repo, 'P646') # freebase mapping # And sources! self.statedin = pywikibot.Claim(self.repo, 'P248') # stated in # Freebase data dump freebasedumpitem = pywikibot.ItemPage(self.repo, 'Q15241312') self.statedin.setTarget(freebasedumpitem) # date of publication self.dateofpub = pywikibot.Claim(self.repo, 'P577') oct28 = pywikibot.WbTime(site=self.repo, year=2013, month=10, day=28, precision='day') self.dateofpub.setTarget(oct28) for line in gzip.open(self.filename): self.processLine(line.strip()) def processLine(self, line): """Process a single line.""" if not line or line.startswith('#'): return mid, sameas, qid, dot = line.split() if sameas != '<https://www.w3.org/2002/07/owl#sameAs>': return if dot != '.': return if not mid.startswith('<https://rdf.freebase.com/ns/m'): return mid = '/m/' + mid[30:-1] if not qid.startswith('<https://www.wikidata.org/entity/Q'): return qid = 'Q' + qid[33:-1] data = pywikibot.ItemPage(self.repo, qid) data.get() if not data.labels: label = '' elif 'en' in data.labels: label = data.labels['en'] else: # Just pick up the first label label = list(data.labels.values())[0] pywikibot.output('Parsed: {} <--> {}'.format(qid, mid)) pywikibot.output('{} is {}'.format(data.getID(), label)) if data.claims and 'P646' in data.claims: # We assume that there is only one claim. # If there are multiple ones, our logs might be wrong # but the constraint value reports will catch them if mid != data.claims['P646'][0].getTarget(): pywikibot.output('Mismatch: expected {}, has {} instead' .format(mid, data.claims['P646'][0].getTarget())) else: pywikibot.output('Already has mid set, is consistent.') else: # No claim set, lets add it. pywikibot.output('Going to add a new claim.') self.claim.setTarget(mid) data.addClaim(self.claim) self.claim.addSources([self.statedin, self.dateofpub]) pywikibot.output('Claim added!') def main(*args): """ Process command line arguments and invoke bot. If args is an empty list, sys.argv is used. @param args: command line arguments @type args: str """ filename = 'fb2w.nt.gz' # Default filename for arg in pywikibot.handle_args(args): if arg.startswith('-filename'): filename = arg[11:] bot = FreebaseMapperRobot(filename) bot.run() if __name__ == '__main__': main()
scripts/freebasemappingupload.py
4,148
Freebase Mapping bot. Initializer. Process command line arguments and invoke bot. If args is an empty list, sys.argv is used. @param args: command line arguments @type args: str Process a single line. Run the bot. Script to upload the mappings of Freebase to Wikidata. Can be easily adapted to upload other String identifiers as well This bot needs the dump from https://developers.google.com/freebase/data#freebase-wikidata-mappings The script takes a single parameter: -filename: the filename to read the freebase-wikidata mappings from; default: fb2w.nt.gz !/usr/bin/python -*- coding: utf-8 -*- (C) Denny Vrandecic, 2013 (C) Pywikibot team, 2013-2018 Distributed under the terms of the MIT license. Set up some items we will use a lot. freebase mapping And sources! stated in Freebase data dump date of publication Just pick up the first label We assume that there is only one claim. If there are multiple ones, our logs might be wrong but the constraint value reports will catch them No claim set, lets add it. Default filename
1,050
en
0.666926
#!/usr/bin/python3.5 import yaml import os.path import logging logger = logging.getLogger('rsync_backup') config = None cmd = '/usr/bin/rsync' # load the configuration file in def load_config(filepath): global config with open(filepath, "r") as file_descriptor: config = yaml.load(file_descriptor) logger.debug("Config: ", config) # Returns a list of tuplets (of commands src and destination) def get_commands(): logger.debug("Config: ", config) locations = config['backup']['locations'] commands = [] # Test to see if the source and destination locations exist/are accessible _test_locations() # Compile the commands commands = [(_get_cmd(location['src'],location['dest'], 'noexec'),_get_cmd(location['src'],location['dest'],'exec')) for location in locations] _print_commands(commands) return commands def _get_cmd (source, destination, mode='noexec'): params = config['backup']['commands'][mode] noexec_cmd = [cmd] noexec_cmd += params noexec_cmd += (source, destination) #noexec_cmd_str = " ".join(noexec_cmd) return noexec_cmd def _test_locations(): locations = config['backup']['locations'] # Test to see if the source and destination locations exist/are accessible bad_locations = [(os.path.exists(location['src']), location['src']) for location in locations] bad_locations += ([(os.path.exists(location['dest']), location['dest']) for location in locations]) logger.debug(bad_locations) for location in bad_locations: if not location[0]: logger.debug('Error location ' + location[1] + " does not exist") if len([location for location in bad_locations if not location[0]]) > 0: logger.debug("One or more locations are not accessible, existing") #exit() def _print_commands(commands): logger.debug("Commands:") for command_pair in commands: logger.debug("Noexec: %s", command_pair[0]) logger.debug("Exec: %s", command_pair[1]) def get_log_path(): if config == None: exit("You must load the config first") return config['backup']['log'] def get_delete_limit(): if config == None: exit("You must load the config first") return config['backup']['delete_limit'] def get_email_smtp(): if config == None: exit("You must load the config first") return config['backup']['email']['smtp'] def get_email_smtp_port(): if config == None: exit("You must load the config first") return config['backup']['email']['smtp_port'] def get_email_username(): if config == None: exit("You must load the config first") return config['backup']['email']['username'] def get_email_password(): if config == None: exit("You must load the config first") return config['backup']['email']['password'] def get_email_target(): if config == None: exit("You must load the config first") return config['backup']['email']['target']
bin/utils/config.py
3,034
!/usr/bin/python3.5 load the configuration file in Returns a list of tuplets (of commands src and destination) Test to see if the source and destination locations exist/are accessible Compile the commandsnoexec_cmd_str = " ".join(noexec_cmd) Test to see if the source and destination locations exist/are accessibleexit()
320
en
0.637354
""" sphinx.domains.cpp ~~~~~~~~~~~~~~~~~~ The C++ language domain. :copyright: Copyright 2007-2020 by the Sphinx team, see AUTHORS. :license: BSD, see LICENSE for details. """ import re from typing import ( Any, Callable, Dict, Generator, Iterator, List, Tuple, Type, TypeVar, Union ) from docutils import nodes from docutils.nodes import Element, Node, TextElement, system_message from docutils.parsers.rst import directives from sphinx import addnodes from sphinx.addnodes import desc_signature, pending_xref from sphinx.application import Sphinx from sphinx.builders import Builder from sphinx.directives import ObjectDescription from sphinx.domains import Domain, ObjType from sphinx.environment import BuildEnvironment from sphinx.errors import NoUri from sphinx.locale import _, __ from sphinx.roles import SphinxRole, XRefRole from sphinx.transforms import SphinxTransform from sphinx.transforms.post_transforms import ReferencesResolver from sphinx.util import logging from sphinx.util.cfamily import ( NoOldIdError, ASTBaseBase, ASTAttribute, verify_description_mode, StringifyTransform, BaseParser, DefinitionError, UnsupportedMultiCharacterCharLiteral, identifier_re, anon_identifier_re, integer_literal_re, octal_literal_re, hex_literal_re, binary_literal_re, float_literal_re, char_literal_re ) from sphinx.util.docfields import Field, GroupedField from sphinx.util.docutils import SphinxDirective from sphinx.util.nodes import make_refnode logger = logging.getLogger(__name__) T = TypeVar('T') """ Important note on ids ---------------------------------------------------------------------------- Multiple id generation schemes are used due to backwards compatibility. - v1: 1.2.3 <= version < 1.3 The style used before the rewrite. It is not the actual old code, but a replication of the behaviour. - v2: 1.3 <= version < now Standardised mangling scheme from https://itanium-cxx-abi.github.io/cxx-abi/abi.html#mangling though not completely implemented. All versions are generated and attached to elements. The newest is used for the index. All of the versions should work as permalinks. Signature Nodes and Tagnames ---------------------------------------------------------------------------- Each signature is in a desc_signature node, where all children are desc_signature_line nodes. Each of these lines will have the attribute 'sphinx_line_type' set to one of the following (prioritized): - 'declarator', if the line contains the name of the declared object. - 'templateParams', if the line starts a template parameter list, - 'templateParams', if the line has template parameters Note: such lines might get a new tag in the future. - 'templateIntroduction, if the line is on the form 'conceptName{...}' No other desc_signature nodes should exist (so far). Grammar ---------------------------------------------------------------------------- See https://www.nongnu.org/hcb/ for the grammar, and https://github.com/cplusplus/draft/blob/master/source/grammar.tex, and https://github.com/cplusplus/concepts-ts for the newest grammar. common grammar things: template-declaration -> "template" "<" template-parameter-list ">" declaration template-parameter-list -> template-parameter | template-parameter-list "," template-parameter template-parameter -> type-parameter | parameter-declaration # i.e., same as a function argument type-parameter -> "class" "..."[opt] identifier[opt] | "class" identifier[opt] "=" type-id | "typename" "..."[opt] identifier[opt] | "typename" identifier[opt] "=" type-id | "template" "<" template-parameter-list ">" "class" "..."[opt] identifier[opt] | "template" "<" template-parameter-list ">" "class" identifier[opt] "=" id-expression # also, from C++17 we can have "typename" in template templates templateDeclPrefix -> "template" "<" template-parameter-list ">" simple-declaration -> attribute-specifier-seq[opt] decl-specifier-seq[opt] init-declarator-list[opt] ; # Drop the semi-colon. For now: drop the attributes (TODO). # Use at most 1 init-declarator. -> decl-specifier-seq init-declarator -> decl-specifier-seq declarator initializer decl-specifier -> storage-class-specifier -> ( "static" (only for member_object and function_object) | "extern" (only for member_object and function_object) | "register" ) thread_local[opt] (only for member_object) (it can also appear before the others) | type-specifier -> trailing-type-specifier | function-specifier -> "inline" | "virtual" | "explicit" (only for function_object) | "friend" (only for function_object) | "constexpr" (only for member_object and function_object) trailing-type-specifier -> simple-type-specifier | elaborated-type-specifier | typename-specifier | cv-qualifier -> "const" | "volatile" stricter grammar for decl-specifier-seq (with everything, each object uses a subset): visibility storage-class-specifier function-specifier "friend" "constexpr" "volatile" "const" trailing-type-specifier # where trailing-type-specifier can no be cv-qualifier # Inside e.g., template paramters a strict subset is used # (see type-specifier-seq) trailing-type-specifier -> simple-type-specifier -> ::[opt] nested-name-specifier[opt] type-name | ::[opt] nested-name-specifier "template" simple-template-id | "char" | "bool" | ect. | decltype-specifier | elaborated-type-specifier -> class-key attribute-specifier-seq[opt] ::[opt] nested-name-specifier[opt] identifier | class-key ::[opt] nested-name-specifier[opt] template[opt] simple-template-id | "enum" ::[opt] nested-name-specifier[opt] identifier | typename-specifier -> "typename" ::[opt] nested-name-specifier identifier | "typename" ::[opt] nested-name-specifier template[opt] simple-template-id class-key -> "class" | "struct" | "union" type-name ->* identifier | simple-template-id # ignoring attributes and decltype, and then some left-factoring trailing-type-specifier -> rest-of-trailing ("class" | "struct" | "union" | "typename") rest-of-trailing build-in -> "char" | "bool" | ect. decltype-specifier rest-of-trailing -> (with some simplification) "::"[opt] list-of-elements-separated-by-:: element -> "template"[opt] identifier ("<" template-argument-list ">")[opt] template-argument-list -> template-argument "..."[opt] | template-argument-list "," template-argument "..."[opt] template-argument -> constant-expression | type-specifier-seq abstract-declarator | id-expression declarator -> ptr-declarator | noptr-declarator parameters-and-qualifiers trailing-return-type (TODO: for now we don't support trailing-eturn-type) ptr-declarator -> noptr-declarator | ptr-operator ptr-declarator noptr-declarator -> declarator-id attribute-specifier-seq[opt] -> "..."[opt] id-expression | rest-of-trailing | noptr-declarator parameters-and-qualifiers | noptr-declarator "[" constant-expression[opt] "]" attribute-specifier-seq[opt] | "(" ptr-declarator ")" ptr-operator -> "*" attribute-specifier-seq[opt] cv-qualifier-seq[opt] | "& attribute-specifier-seq[opt] | "&&" attribute-specifier-seq[opt] | "::"[opt] nested-name-specifier "*" attribute-specifier-seq[opt] cv-qualifier-seq[opt] # function_object must use a parameters-and-qualifiers, the others may # use it (e.g., function poitners) parameters-and-qualifiers -> "(" parameter-clause ")" attribute-specifier-seq[opt] cv-qualifier-seq[opt] ref-qualifier[opt] exception-specification[opt] ref-qualifier -> "&" | "&&" exception-specification -> "noexcept" ("(" constant-expression ")")[opt] "throw" ("(" type-id-list ")")[opt] # TODO: we don't implement attributes # member functions can have initializers, but we fold them into here memberFunctionInit -> "=" "0" # (note: only "0" is allowed as the value, according to the standard, # right?) enum-head -> enum-key attribute-specifier-seq[opt] nested-name-specifier[opt] identifier enum-base[opt] enum-key -> "enum" | "enum struct" | "enum class" enum-base -> ":" type enumerator-definition -> identifier | identifier "=" constant-expression We additionally add the possibility for specifying the visibility as the first thing. concept_object: goal: just a declaration of the name (for now) grammar: only a single template parameter list, and the nested name may not have any template argument lists "template" "<" template-parameter-list ">" nested-name-specifier type_object: goal: either a single type (e.g., "MyClass:Something_T" or a typedef-like thing (e.g. "Something Something_T" or "int I_arr[]" grammar, single type: based on a type in a function parameter, but without a name: parameter-declaration -> attribute-specifier-seq[opt] decl-specifier-seq abstract-declarator[opt] # Drop the attributes -> decl-specifier-seq abstract-declarator[opt] grammar, typedef-like: no initilizer decl-specifier-seq declarator Can start with a templateDeclPrefix. member_object: goal: as a type_object which must have a declarator, and optionally with a initializer grammar: decl-specifier-seq declarator initializer Can start with a templateDeclPrefix. function_object: goal: a function declaration, TODO: what about templates? for now: skip grammar: no initializer decl-specifier-seq declarator Can start with a templateDeclPrefix. class_object: goal: a class declaration, but with specification of a base class grammar: nested-name "final"[opt] (":" base-specifier-list)[opt] base-specifier-list -> base-specifier "..."[opt] | base-specifier-list, base-specifier "..."[opt] base-specifier -> base-type-specifier | "virtual" access-spe"cifier[opt] base-type-specifier | access-specifier[opt] "virtual"[opt] base-type-specifier Can start with a templateDeclPrefix. enum_object: goal: an unscoped enum or a scoped enum, optionally with the underlying type specified grammar: ("class" | "struct")[opt] visibility[opt] nested-name (":" type)[opt] enumerator_object: goal: an element in a scoped or unscoped enum. The name should be injected according to the scopedness. grammar: nested-name ("=" constant-expression) namespace_object: goal: a directive to put all following declarations in a specific scope grammar: nested-name """ _string_re = re.compile(r"[LuU8]?('([^'\\]*(?:\\.[^'\\]*)*)'" r'|"([^"\\]*(?:\\.[^"\\]*)*)")', re.S) _visibility_re = re.compile(r'\b(public|private|protected)\b') _operator_re = re.compile(r'''(?x) \[\s*\] | \(\s*\) | \+\+ | -- | ->\*? | \, | (<<|>>)=? | && | \|\| | [!<>=/*%+|&^~-]=? | (\b(and|and_eq|bitand|bitor|compl|not|not_eq|or|or_eq|xor|xor_eq)\b) ''') _fold_operator_re = re.compile(r'''(?x) ->\* | \.\* | \, | (<<|>>)=? | && | \|\| | != | [<>=/*%+|&^~-]=? ''') # see https://en.cppreference.com/w/cpp/keyword _keywords = [ 'alignas', 'alignof', 'and', 'and_eq', 'asm', 'auto', 'bitand', 'bitor', 'bool', 'break', 'case', 'catch', 'char', 'char16_t', 'char32_t', 'class', 'compl', 'concept', 'const', 'constexpr', 'const_cast', 'continue', 'decltype', 'default', 'delete', 'do', 'double', 'dynamic_cast', 'else', 'enum', 'explicit', 'export', 'extern', 'false', 'float', 'for', 'friend', 'goto', 'if', 'inline', 'int', 'long', 'mutable', 'namespace', 'new', 'noexcept', 'not', 'not_eq', 'nullptr', 'operator', 'or', 'or_eq', 'private', 'protected', 'public', 'register', 'reinterpret_cast', 'requires', 'return', 'short', 'signed', 'sizeof', 'static', 'static_assert', 'static_cast', 'struct', 'switch', 'template', 'this', 'thread_local', 'throw', 'true', 'try', 'typedef', 'typeid', 'typename', 'union', 'unsigned', 'using', 'virtual', 'void', 'volatile', 'wchar_t', 'while', 'xor', 'xor_eq' ] _max_id = 4 _id_prefix = [None, '', '_CPPv2', '_CPPv3', '_CPPv4'] # Ids are used in lookup keys which are used across pickled files, # so when _max_id changes, make sure to update the ENV_VERSION. # ------------------------------------------------------------------------------ # Id v1 constants # ------------------------------------------------------------------------------ _id_fundamental_v1 = { 'char': 'c', 'signed char': 'c', 'unsigned char': 'C', 'int': 'i', 'signed int': 'i', 'unsigned int': 'U', 'long': 'l', 'signed long': 'l', 'unsigned long': 'L', 'bool': 'b' } _id_shorthands_v1 = { 'std::string': 'ss', 'std::ostream': 'os', 'std::istream': 'is', 'std::iostream': 'ios', 'std::vector': 'v', 'std::map': 'm' } _id_operator_v1 = { 'new': 'new-operator', 'new[]': 'new-array-operator', 'delete': 'delete-operator', 'delete[]': 'delete-array-operator', # the arguments will make the difference between unary and binary # '+(unary)' : 'ps', # '-(unary)' : 'ng', # '&(unary)' : 'ad', # '*(unary)' : 'de', '~': 'inv-operator', '+': 'add-operator', '-': 'sub-operator', '*': 'mul-operator', '/': 'div-operator', '%': 'mod-operator', '&': 'and-operator', '|': 'or-operator', '^': 'xor-operator', '=': 'assign-operator', '+=': 'add-assign-operator', '-=': 'sub-assign-operator', '*=': 'mul-assign-operator', '/=': 'div-assign-operator', '%=': 'mod-assign-operator', '&=': 'and-assign-operator', '|=': 'or-assign-operator', '^=': 'xor-assign-operator', '<<': 'lshift-operator', '>>': 'rshift-operator', '<<=': 'lshift-assign-operator', '>>=': 'rshift-assign-operator', '==': 'eq-operator', '!=': 'neq-operator', '<': 'lt-operator', '>': 'gt-operator', '<=': 'lte-operator', '>=': 'gte-operator', '!': 'not-operator', '&&': 'sand-operator', '||': 'sor-operator', '++': 'inc-operator', '--': 'dec-operator', ',': 'comma-operator', '->*': 'pointer-by-pointer-operator', '->': 'pointer-operator', '()': 'call-operator', '[]': 'subscript-operator' } # ------------------------------------------------------------------------------ # Id v > 1 constants # ------------------------------------------------------------------------------ _id_fundamental_v2 = { # not all of these are actually parsed as fundamental types, TODO: do that 'void': 'v', 'bool': 'b', 'char': 'c', 'signed char': 'a', 'unsigned char': 'h', 'wchar_t': 'w', 'char32_t': 'Di', 'char16_t': 'Ds', 'short': 's', 'short int': 's', 'signed short': 's', 'signed short int': 's', 'unsigned short': 't', 'unsigned short int': 't', 'int': 'i', 'signed': 'i', 'signed int': 'i', 'unsigned': 'j', 'unsigned int': 'j', 'long': 'l', 'long int': 'l', 'signed long': 'l', 'signed long int': 'l', 'unsigned long': 'm', 'unsigned long int': 'm', 'long long': 'x', 'long long int': 'x', 'signed long long': 'x', 'signed long long int': 'x', 'unsigned long long': 'y', 'unsigned long long int': 'y', 'float': 'f', 'double': 'd', 'long double': 'e', 'auto': 'Da', 'decltype(auto)': 'Dc', 'std::nullptr_t': 'Dn' } _id_operator_v2 = { 'new': 'nw', 'new[]': 'na', 'delete': 'dl', 'delete[]': 'da', # the arguments will make the difference between unary and binary # in operator definitions # '+(unary)' : 'ps', # '-(unary)' : 'ng', # '&(unary)' : 'ad', # '*(unary)' : 'de', '~': 'co', 'compl': 'co', '+': 'pl', '-': 'mi', '*': 'ml', '/': 'dv', '%': 'rm', '&': 'an', 'bitand': 'an', '|': 'or', 'bitor': 'or', '^': 'eo', 'xor': 'eo', '=': 'aS', '+=': 'pL', '-=': 'mI', '*=': 'mL', '/=': 'dV', '%=': 'rM', '&=': 'aN', 'and_eq': 'aN', '|=': 'oR', 'or_eq': 'oR', '^=': 'eO', 'xor_eq': 'eO', '<<': 'ls', '>>': 'rs', '<<=': 'lS', '>>=': 'rS', '==': 'eq', '!=': 'ne', 'not_eq': 'ne', '<': 'lt', '>': 'gt', '<=': 'le', '>=': 'ge', '!': 'nt', 'not': 'nt', '&&': 'aa', 'and': 'aa', '||': 'oo', 'or': 'oo', '++': 'pp', '--': 'mm', ',': 'cm', '->*': 'pm', '->': 'pt', '()': 'cl', '[]': 'ix', '.*': 'ds' # this one is not overloadable, but we need it for expressions } _id_operator_unary_v2 = { '++': 'pp_', '--': 'mm_', '*': 'de', '&': 'ad', '+': 'ps', '-': 'ng', '!': 'nt', 'not': 'nt', '~': 'co', 'compl': 'co' } _id_char_from_prefix = { None: 'c', 'u8': 'c', 'u': 'Ds', 'U': 'Di', 'L': 'w' } # type: Dict[Any, str] # these are ordered by preceedence _expression_bin_ops = [ ['||', 'or'], ['&&', 'and'], ['|', 'bitor'], ['^', 'xor'], ['&', 'bitand'], ['==', '!=', 'not_eq'], ['<=', '>=', '<', '>'], ['<<', '>>'], ['+', '-'], ['*', '/', '%'], ['.*', '->*'] ] _expression_unary_ops = ["++", "--", "*", "&", "+", "-", "!", "not", "~", "compl"] _expression_assignment_ops = ["=", "*=", "/=", "%=", "+=", "-=", ">>=", "<<=", "&=", "and_eq", "^=", "|=", "xor_eq", "or_eq"] _id_explicit_cast = { 'dynamic_cast': 'dc', 'static_cast': 'sc', 'const_cast': 'cc', 'reinterpret_cast': 'rc' } class _DuplicateSymbolError(Exception): def __init__(self, symbol: "Symbol", declaration: "ASTDeclaration") -> None: assert symbol assert declaration self.symbol = symbol self.declaration = declaration def __str__(self) -> str: return "Internal C++ duplicate symbol error:\n%s" % self.symbol.dump(0) class ASTBase(ASTBaseBase): pass # Names ################################################################################ class ASTIdentifier(ASTBase): def __init__(self, identifier: str) -> None: assert identifier is not None assert len(identifier) != 0 self.identifier = identifier def is_anon(self) -> bool: return self.identifier[0] == '@' def get_id(self, version: int) -> str: if self.is_anon() and version < 3: raise NoOldIdError() if version == 1: if self.identifier == 'size_t': return 's' else: return self.identifier if self.identifier == "std": return 'St' elif self.identifier[0] == "~": # a destructor, just use an arbitrary version of dtors return 'D0' else: if self.is_anon(): return 'Ut%d_%s' % (len(self.identifier) - 1, self.identifier[1:]) else: return str(len(self.identifier)) + self.identifier # and this is where we finally make a difference between __str__ and the display string def __str__(self) -> str: return self.identifier def get_display_string(self) -> str: return "[anonymous]" if self.is_anon() else self.identifier def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", prefix: str, templateArgs: str, symbol: "Symbol") -> None: verify_description_mode(mode) if mode == 'markType': targetText = prefix + self.identifier + templateArgs pnode = addnodes.pending_xref('', refdomain='cpp', reftype='identifier', reftarget=targetText, modname=None, classname=None) key = symbol.get_lookup_key() pnode['cpp:parent_key'] = key if self.is_anon(): pnode += nodes.strong(text="[anonymous]") else: pnode += nodes.Text(self.identifier) signode += pnode elif mode == 'lastIsName': if self.is_anon(): signode += nodes.strong(text="[anonymous]") else: signode += addnodes.desc_name(self.identifier, self.identifier) elif mode == 'noneIsName': if self.is_anon(): signode += nodes.strong(text="[anonymous]") else: signode += nodes.Text(self.identifier) else: raise Exception('Unknown description mode: %s' % mode) class ASTNestedNameElement(ASTBase): def __init__(self, identOrOp: Union[ASTIdentifier, "ASTOperator"], templateArgs: "ASTTemplateArgs") -> None: self.identOrOp = identOrOp self.templateArgs = templateArgs def is_operator(self) -> bool: return False def get_id(self, version: int) -> str: res = self.identOrOp.get_id(version) if self.templateArgs: res += self.templateArgs.get_id(version) return res def _stringify(self, transform: StringifyTransform) -> str: res = transform(self.identOrOp) if self.templateArgs: res += transform(self.templateArgs) return res def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", prefix: str, symbol: "Symbol") -> None: tArgs = str(self.templateArgs) if self.templateArgs is not None else '' self.identOrOp.describe_signature(signode, mode, env, prefix, tArgs, symbol) if self.templateArgs is not None: self.templateArgs.describe_signature(signode, mode, env, symbol) class ASTNestedName(ASTBase): def __init__(self, names: List[ASTNestedNameElement], templates: List[bool], rooted: bool) -> None: assert len(names) > 0 self.names = names self.templates = templates assert len(self.names) == len(self.templates) self.rooted = rooted @property def name(self) -> "ASTNestedName": return self def num_templates(self) -> int: count = 0 for n in self.names: if n.is_operator(): continue if n.templateArgs: count += 1 return count def get_id(self, version: int, modifiers: str = '') -> str: if version == 1: tt = str(self) if tt in _id_shorthands_v1: return _id_shorthands_v1[tt] else: return '::'.join(n.get_id(version) for n in self.names) res = [] if len(self.names) > 1 or len(modifiers) > 0: res.append('N') res.append(modifiers) for n in self.names: res.append(n.get_id(version)) if len(self.names) > 1 or len(modifiers) > 0: res.append('E') return ''.join(res) def _stringify(self, transform: StringifyTransform) -> str: res = [] if self.rooted: res.append('') for i in range(len(self.names)): n = self.names[i] t = self.templates[i] if t: res.append("template " + transform(n)) else: res.append(transform(n)) return '::'.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) # just print the name part, with template args, not template params if mode == 'noneIsName': signode += nodes.Text(str(self)) elif mode == 'param': name = str(self) signode += nodes.emphasis(name, name) elif mode == 'markType' or mode == 'lastIsName' or mode == 'markName': # Each element should be a pending xref targeting the complete # prefix. however, only the identifier part should be a link, such # that template args can be a link as well. # For 'lastIsName' we should also prepend template parameter lists. templateParams = [] # type: List[Any] if mode == 'lastIsName': assert symbol is not None if symbol.declaration.templatePrefix is not None: templateParams = symbol.declaration.templatePrefix.templates iTemplateParams = 0 templateParamsPrefix = '' prefix = '' first = True names = self.names[:-1] if mode == 'lastIsName' else self.names # If lastIsName, then wrap all of the prefix in a desc_addname, # else append directly to signode. # NOTE: Breathe relies on the prefix being in the desc_addname node, # so it can remove it in inner declarations. dest = signode if mode == 'lastIsName': dest = addnodes.desc_addname() for i in range(len(names)): nne = names[i] template = self.templates[i] if not first: dest += nodes.Text('::') prefix += '::' if template: dest += nodes.Text("template ") first = False txt_nne = str(nne) if txt_nne != '': if nne.templateArgs and iTemplateParams < len(templateParams): templateParamsPrefix += str(templateParams[iTemplateParams]) iTemplateParams += 1 nne.describe_signature(dest, 'markType', env, templateParamsPrefix + prefix, symbol) prefix += txt_nne if mode == 'lastIsName': if len(self.names) > 1: dest += addnodes.desc_addname('::', '::') signode += dest if self.templates[-1]: signode += nodes.Text("template ") self.names[-1].describe_signature(signode, mode, env, '', symbol) else: raise Exception('Unknown description mode: %s' % mode) ################################################################################ # Expressions ################################################################################ class ASTExpression(ASTBase): def get_id(self, version: int) -> str: raise NotImplementedError(repr(self)) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: raise NotImplementedError(repr(self)) # Primary expressions ################################################################################ class ASTLiteral(ASTExpression): pass class ASTPointerLiteral(ASTLiteral): def _stringify(self, transform: StringifyTransform) -> str: return 'nullptr' def get_id(self, version: int) -> str: return 'LDnE' def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('nullptr')) class ASTBooleanLiteral(ASTLiteral): def __init__(self, value: bool) -> None: self.value = value def _stringify(self, transform: StringifyTransform) -> str: if self.value: return 'true' else: return 'false' def get_id(self, version: int) -> str: if self.value: return 'L1E' else: return 'L0E' def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text(str(self))) class ASTNumberLiteral(ASTLiteral): def __init__(self, data: str) -> None: self.data = data def _stringify(self, transform: StringifyTransform) -> str: return self.data def get_id(self, version: int) -> str: return "L%sE" % self.data def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: txt = str(self) signode.append(nodes.Text(txt, txt)) class ASTStringLiteral(ASTLiteral): def __init__(self, data: str) -> None: self.data = data def _stringify(self, transform: StringifyTransform) -> str: return self.data def get_id(self, version: int) -> str: # note: the length is not really correct with escaping return "LA%d_KcE" % (len(self.data) - 2) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: txt = str(self) signode.append(nodes.Text(txt, txt)) class ASTCharLiteral(ASTLiteral): def __init__(self, prefix: str, data: str) -> None: self.prefix = prefix # may be None when no prefix self.data = data assert prefix in _id_char_from_prefix self.type = _id_char_from_prefix[prefix] decoded = data.encode().decode('unicode-escape') if len(decoded) == 1: self.value = ord(decoded) else: raise UnsupportedMultiCharacterCharLiteral(decoded) def _stringify(self, transform: StringifyTransform) -> str: if self.prefix is None: return "'" + self.data + "'" else: return self.prefix + "'" + self.data + "'" def get_id(self, version: int) -> str: return self.type + str(self.value) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: txt = str(self) signode.append(nodes.Text(txt, txt)) class ASTThisLiteral(ASTExpression): def _stringify(self, transform: StringifyTransform) -> str: return "this" def get_id(self, version: int) -> str: return "fpT" def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text("this")) class ASTFoldExpr(ASTExpression): def __init__(self, leftExpr: ASTExpression, op: str, rightExpr: ASTExpression) -> None: assert leftExpr is not None or rightExpr is not None self.leftExpr = leftExpr self.op = op self.rightExpr = rightExpr def _stringify(self, transform: StringifyTransform) -> str: res = ['('] if self.leftExpr: res.append(transform(self.leftExpr)) res.append(' ') res.append(transform(self.op)) res.append(' ') res.append('...') if self.rightExpr: res.append(' ') res.append(transform(self.op)) res.append(' ') res.append(transform(self.rightExpr)) res.append(')') return ''.join(res) def get_id(self, version: int) -> str: assert version >= 3 if version == 3: return str(self) # https://github.com/itanium-cxx-abi/cxx-abi/pull/67 res = [] if self.leftExpr is None: # (... op expr) res.append('fl') elif self.rightExpr is None: # (expr op ...) res.append('fr') else: # (expr op ... op expr) # we don't check where the parameter pack is, # we just always call this a binary left fold res.append('fL') res.append(_id_operator_v2[self.op]) if self.leftExpr: res.append(self.leftExpr.get_id(version)) if self.rightExpr: res.append(self.rightExpr.get_id(version)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('(')) if self.leftExpr: self.leftExpr.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(' ')) signode.append(nodes.Text(self.op)) signode.append(nodes.Text(' ')) signode.append(nodes.Text('...')) if self.rightExpr: signode.append(nodes.Text(' ')) signode.append(nodes.Text(self.op)) signode.append(nodes.Text(' ')) self.rightExpr.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(')')) class ASTParenExpr(ASTExpression): def __init__(self, expr: ASTExpression): self.expr = expr def _stringify(self, transform: StringifyTransform) -> str: return '(' + transform(self.expr) + ')' def get_id(self, version: int) -> str: return self.expr.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('(', '(')) self.expr.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(')', ')')) class ASTIdExpression(ASTExpression): def __init__(self, name: ASTNestedName): # note: this class is basically to cast a nested name as an expression self.name = name def _stringify(self, transform: StringifyTransform) -> str: return transform(self.name) def get_id(self, version: int) -> str: return self.name.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.name.describe_signature(signode, mode, env, symbol) # Postfix expressions ################################################################################ class ASTPostfixOp(ASTBase): def get_id(self, idPrefix: str, version: int) -> str: raise NotImplementedError(repr(self)) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: raise NotImplementedError(repr(self)) class ASTPostfixArray(ASTPostfixOp): def __init__(self, expr: ASTExpression): self.expr = expr def _stringify(self, transform: StringifyTransform) -> str: return '[' + transform(self.expr) + ']' def get_id(self, idPrefix: str, version: int) -> str: return 'ix' + idPrefix + self.expr.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('[')) self.expr.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(']')) class ASTPostfixMember(ASTPostfixOp): def __init__(self, name: ASTNestedName): self.name = name def _stringify(self, transform: StringifyTransform) -> str: return '.' + transform(self.name) def get_id(self, idPrefix: str, version: int) -> str: return 'dt' + idPrefix + self.name.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('.')) self.name.describe_signature(signode, 'noneIsName', env, symbol) class ASTPostfixMemberOfPointer(ASTPostfixOp): def __init__(self, name: ASTNestedName): self.name = name def _stringify(self, transform: StringifyTransform) -> str: return '->' + transform(self.name) def get_id(self, idPrefix: str, version: int) -> str: return 'pt' + idPrefix + self.name.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('->')) self.name.describe_signature(signode, 'noneIsName', env, symbol) class ASTPostfixInc(ASTPostfixOp): def _stringify(self, transform: StringifyTransform) -> str: return '++' def get_id(self, idPrefix: str, version: int) -> str: return 'pp' + idPrefix def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('++')) class ASTPostfixDec(ASTPostfixOp): def _stringify(self, transform: StringifyTransform) -> str: return '--' def get_id(self, idPrefix: str, version: int) -> str: return 'mm' + idPrefix def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('--')) class ASTPostfixCallExpr(ASTPostfixOp): def __init__(self, lst: Union["ASTParenExprList", "ASTBracedInitList"]) -> None: self.lst = lst def _stringify(self, transform: StringifyTransform) -> str: return transform(self.lst) def get_id(self, idPrefix: str, version: int) -> str: res = ['cl', idPrefix] for e in self.lst.exprs: res.append(e.get_id(version)) res.append('E') return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.lst.describe_signature(signode, mode, env, symbol) class ASTPostfixExpr(ASTExpression): def __init__(self, prefix: "ASTType", postFixes: List[ASTPostfixOp]): self.prefix = prefix self.postFixes = postFixes def _stringify(self, transform: StringifyTransform) -> str: res = [transform(self.prefix)] for p in self.postFixes: res.append(transform(p)) return ''.join(res) def get_id(self, version: int) -> str: id = self.prefix.get_id(version) for p in self.postFixes: id = p.get_id(id, version) return id def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.prefix.describe_signature(signode, mode, env, symbol) for p in self.postFixes: p.describe_signature(signode, mode, env, symbol) class ASTExplicitCast(ASTExpression): def __init__(self, cast: str, typ: "ASTType", expr: ASTExpression): assert cast in _id_explicit_cast self.cast = cast self.typ = typ self.expr = expr def _stringify(self, transform: StringifyTransform) -> str: res = [self.cast] res.append('<') res.append(transform(self.typ)) res.append('>(') res.append(transform(self.expr)) res.append(')') return ''.join(res) def get_id(self, version: int) -> str: return (_id_explicit_cast[self.cast] + self.typ.get_id(version) + self.expr.get_id(version)) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text(self.cast)) signode.append(nodes.Text('<')) self.typ.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text('>')) signode.append(nodes.Text('(')) self.expr.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(')')) class ASTTypeId(ASTExpression): def __init__(self, typeOrExpr: Union["ASTType", ASTExpression], isType: bool): self.typeOrExpr = typeOrExpr self.isType = isType def _stringify(self, transform: StringifyTransform) -> str: return 'typeid(' + transform(self.typeOrExpr) + ')' def get_id(self, version: int) -> str: prefix = 'ti' if self.isType else 'te' return prefix + self.typeOrExpr.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('typeid')) signode.append(nodes.Text('(')) self.typeOrExpr.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(')')) # Unary expressions ################################################################################ class ASTUnaryOpExpr(ASTExpression): def __init__(self, op: str, expr: ASTExpression): self.op = op self.expr = expr def _stringify(self, transform: StringifyTransform) -> str: if self.op[0] in 'cn': return transform(self.op) + " " + transform(self.expr) else: return transform(self.op) + transform(self.expr) def get_id(self, version: int) -> str: return _id_operator_unary_v2[self.op] + self.expr.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text(self.op)) if self.op[0] in 'cn': signode.append(nodes.Text(' ')) self.expr.describe_signature(signode, mode, env, symbol) class ASTSizeofParamPack(ASTExpression): def __init__(self, identifier: ASTIdentifier): self.identifier = identifier def _stringify(self, transform: StringifyTransform) -> str: return "sizeof...(" + transform(self.identifier) + ")" def get_id(self, version: int) -> str: return 'sZ' + self.identifier.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('sizeof...(')) self.identifier.describe_signature(signode, mode, env, symbol=symbol, prefix="", templateArgs="") signode.append(nodes.Text(')')) class ASTSizeofType(ASTExpression): def __init__(self, typ: "ASTType"): self.typ = typ def _stringify(self, transform: StringifyTransform) -> str: return "sizeof(" + transform(self.typ) + ")" def get_id(self, version: int) -> str: return 'st' + self.typ.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('sizeof(')) self.typ.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(')')) class ASTSizeofExpr(ASTExpression): def __init__(self, expr: ASTExpression): self.expr = expr def _stringify(self, transform: StringifyTransform) -> str: return "sizeof " + transform(self.expr) def get_id(self, version: int) -> str: return 'sz' + self.expr.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('sizeof ')) self.expr.describe_signature(signode, mode, env, symbol) class ASTAlignofExpr(ASTExpression): def __init__(self, typ: "ASTType"): self.typ = typ def _stringify(self, transform: StringifyTransform) -> str: return "alignof(" + transform(self.typ) + ")" def get_id(self, version: int) -> str: return 'at' + self.typ.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('alignof(')) self.typ.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(')')) class ASTNoexceptExpr(ASTExpression): def __init__(self, expr: ASTExpression): self.expr = expr def _stringify(self, transform: StringifyTransform) -> str: return "noexcept(" + transform(self.expr) + ")" def get_id(self, version: int) -> str: return 'nx' + self.expr.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('noexcept(')) self.expr.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(')')) class ASTNewExpr(ASTExpression): def __init__(self, rooted: bool, isNewTypeId: bool, typ: "ASTType", initList: Union["ASTParenExprList", "ASTBracedInitList"]) -> None: self.rooted = rooted self.isNewTypeId = isNewTypeId self.typ = typ self.initList = initList def _stringify(self, transform: StringifyTransform) -> str: res = [] if self.rooted: res.append('::') res.append('new ') # TODO: placement if self.isNewTypeId: res.append(transform(self.typ)) else: assert False if self.initList is not None: res.append(transform(self.initList)) return ''.join(res) def get_id(self, version: int) -> str: # the array part will be in the type mangling, so na is not used res = ['nw'] # TODO: placement res.append('_') res.append(self.typ.get_id(version)) if self.initList is not None: res.append(self.initList.get_id(version)) else: res.append('E') return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: if self.rooted: signode.append(nodes.Text('::')) signode.append(nodes.Text('new ')) # TODO: placement if self.isNewTypeId: self.typ.describe_signature(signode, mode, env, symbol) else: assert False if self.initList is not None: self.initList.describe_signature(signode, mode, env, symbol) class ASTDeleteExpr(ASTExpression): def __init__(self, rooted: bool, array: bool, expr: ASTExpression): self.rooted = rooted self.array = array self.expr = expr def _stringify(self, transform: StringifyTransform) -> str: res = [] if self.rooted: res.append('::') res.append('delete ') if self.array: res.append('[] ') res.append(transform(self.expr)) return ''.join(res) def get_id(self, version: int) -> str: if self.array: id = "da" else: id = "dl" return id + self.expr.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: if self.rooted: signode.append(nodes.Text('::')) signode.append(nodes.Text('delete ')) if self.array: signode.append(nodes.Text('[] ')) self.expr.describe_signature(signode, mode, env, symbol) # Other expressions ################################################################################ class ASTCastExpr(ASTExpression): def __init__(self, typ: "ASTType", expr: ASTExpression): self.typ = typ self.expr = expr def _stringify(self, transform: StringifyTransform) -> str: res = ['('] res.append(transform(self.typ)) res.append(')') res.append(transform(self.expr)) return ''.join(res) def get_id(self, version: int) -> str: return 'cv' + self.typ.get_id(version) + self.expr.get_id(version) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('(')) self.typ.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(')')) self.expr.describe_signature(signode, mode, env, symbol) class ASTBinOpExpr(ASTExpression): def __init__(self, exprs: List[ASTExpression], ops: List[str]): assert len(exprs) > 0 assert len(exprs) == len(ops) + 1 self.exprs = exprs self.ops = ops def _stringify(self, transform: StringifyTransform) -> str: res = [] res.append(transform(self.exprs[0])) for i in range(1, len(self.exprs)): res.append(' ') res.append(self.ops[i - 1]) res.append(' ') res.append(transform(self.exprs[i])) return ''.join(res) def get_id(self, version: int) -> str: assert version >= 2 res = [] for i in range(len(self.ops)): res.append(_id_operator_v2[self.ops[i]]) res.append(self.exprs[i].get_id(version)) res.append(self.exprs[-1].get_id(version)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.exprs[0].describe_signature(signode, mode, env, symbol) for i in range(1, len(self.exprs)): signode.append(nodes.Text(' ')) signode.append(nodes.Text(self.ops[i - 1])) signode.append(nodes.Text(' ')) self.exprs[i].describe_signature(signode, mode, env, symbol) class ASTBracedInitList(ASTBase): def __init__(self, exprs: List[Union[ASTExpression, "ASTBracedInitList"]], trailingComma: bool) -> None: self.exprs = exprs self.trailingComma = trailingComma def get_id(self, version: int) -> str: return "il%sE" % ''.join(e.get_id(version) for e in self.exprs) def _stringify(self, transform: StringifyTransform) -> str: exprs = [transform(e) for e in self.exprs] trailingComma = ',' if self.trailingComma else '' return '{%s%s}' % (', '.join(exprs), trailingComma) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) signode.append(nodes.Text('{')) first = True for e in self.exprs: if not first: signode.append(nodes.Text(', ')) else: first = False e.describe_signature(signode, mode, env, symbol) if self.trailingComma: signode.append(nodes.Text(',')) signode.append(nodes.Text('}')) class ASTAssignmentExpr(ASTExpression): def __init__(self, exprs: List[Union[ASTExpression, ASTBracedInitList]], ops: List[str]): assert len(exprs) > 0 assert len(exprs) == len(ops) + 1 self.exprs = exprs self.ops = ops def _stringify(self, transform: StringifyTransform) -> str: res = [] res.append(transform(self.exprs[0])) for i in range(1, len(self.exprs)): res.append(' ') res.append(self.ops[i - 1]) res.append(' ') res.append(transform(self.exprs[i])) return ''.join(res) def get_id(self, version: int) -> str: res = [] for i in range(len(self.ops)): res.append(_id_operator_v2[self.ops[i]]) res.append(self.exprs[i].get_id(version)) res.append(self.exprs[-1].get_id(version)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.exprs[0].describe_signature(signode, mode, env, symbol) for i in range(1, len(self.exprs)): signode.append(nodes.Text(' ')) signode.append(nodes.Text(self.ops[i - 1])) signode.append(nodes.Text(' ')) self.exprs[i].describe_signature(signode, mode, env, symbol) class ASTCommaExpr(ASTExpression): def __init__(self, exprs: List[ASTExpression]): assert len(exprs) > 0 self.exprs = exprs def _stringify(self, transform: StringifyTransform) -> str: return ', '.join(transform(e) for e in self.exprs) def get_id(self, version: int) -> str: id_ = _id_operator_v2[','] res = [] for i in range(len(self.exprs) - 1): res.append(id_) res.append(self.exprs[i].get_id(version)) res.append(self.exprs[-1].get_id(version)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.exprs[0].describe_signature(signode, mode, env, symbol) for i in range(1, len(self.exprs)): signode.append(nodes.Text(', ')) self.exprs[i].describe_signature(signode, mode, env, symbol) class ASTFallbackExpr(ASTExpression): def __init__(self, expr: str): self.expr = expr def _stringify(self, transform: StringifyTransform) -> str: return self.expr def get_id(self, version: int) -> str: return str(self.expr) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode += nodes.Text(self.expr) ################################################################################ # Types ################################################################################ # Things for ASTNestedName ################################################################################ class ASTOperator(ASTBase): def is_anon(self) -> bool: return False def is_operator(self) -> bool: return True def get_id(self, version: int) -> str: raise NotImplementedError() def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", prefix: str, templateArgs: str, symbol: "Symbol") -> None: verify_description_mode(mode) identifier = str(self) if mode == 'lastIsName': signode += addnodes.desc_name(identifier, identifier) else: signode += addnodes.desc_addname(identifier, identifier) class ASTOperatorBuildIn(ASTOperator): def __init__(self, op: str) -> None: self.op = op def get_id(self, version: int) -> str: if version == 1: ids = _id_operator_v1 if self.op not in ids: raise NoOldIdError() else: ids = _id_operator_v2 if self.op not in ids: raise Exception('Internal error: Build-in operator "%s" can not ' 'be mapped to an id.' % self.op) return ids[self.op] def _stringify(self, transform: StringifyTransform) -> str: if self.op in ('new', 'new[]', 'delete', 'delete[]') or self.op[0] in "abcnox": return 'operator ' + self.op else: return 'operator' + self.op class ASTOperatorLiteral(ASTOperator): def __init__(self, identifier: ASTIdentifier) -> None: self.identifier = identifier def get_id(self, version: int) -> str: if version == 1: raise NoOldIdError() else: return 'li' + self.identifier.get_id(version) def _stringify(self, transform: StringifyTransform) -> str: return 'operator""' + transform(self.identifier) class ASTOperatorType(ASTOperator): def __init__(self, type: "ASTType") -> None: self.type = type def get_id(self, version: int) -> str: if version == 1: return 'castto-%s-operator' % self.type.get_id(version) else: return 'cv' + self.type.get_id(version) def _stringify(self, transform: StringifyTransform) -> str: return ''.join(['operator ', transform(self.type)]) def get_name_no_template(self) -> str: return str(self) class ASTTemplateArgConstant(ASTBase): def __init__(self, value: ASTExpression) -> None: self.value = value def _stringify(self, transform: StringifyTransform) -> str: return transform(self.value) def get_id(self, version: int) -> str: if version == 1: return str(self).replace(' ', '-') if version == 2: return 'X' + str(self) + 'E' return 'X' + self.value.get_id(version) + 'E' def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) self.value.describe_signature(signode, mode, env, symbol) class ASTTemplateArgs(ASTBase): def __init__(self, args: List[Union["ASTType", ASTTemplateArgConstant]], packExpansion: bool) -> None: assert args is not None self.args = args self.packExpansion = packExpansion def get_id(self, version: int) -> str: if version == 1: res = [] res.append(':') res.append('.'.join(a.get_id(version) for a in self.args)) res.append(':') return ''.join(res) res = [] res.append('I') if len(self.args) > 0: for a in self.args[:-1]: res.append(a.get_id(version)) if self.packExpansion: res.append('J') res.append(self.args[-1].get_id(version)) if self.packExpansion: res.append('E') res.append('E') return ''.join(res) def _stringify(self, transform: StringifyTransform) -> str: res = ', '.join(transform(a) for a in self.args) if self.packExpansion: res += '...' return '<' + res + '>' def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) signode += nodes.Text('<') first = True for a in self.args: if not first: signode += nodes.Text(', ') first = False a.describe_signature(signode, 'markType', env, symbol=symbol) if self.packExpansion: signode += nodes.Text('...') signode += nodes.Text('>') # Main part of declarations ################################################################################ class ASTTrailingTypeSpec(ASTBase): def get_id(self, version: int) -> str: raise NotImplementedError(repr(self)) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: raise NotImplementedError(repr(self)) class ASTTrailingTypeSpecFundamental(ASTTrailingTypeSpec): def __init__(self, name: str) -> None: self.name = name def _stringify(self, transform: StringifyTransform) -> str: return self.name def get_id(self, version: int) -> str: if version == 1: res = [] for a in self.name.split(' '): if a in _id_fundamental_v1: res.append(_id_fundamental_v1[a]) else: res.append(a) return '-'.join(res) if self.name not in _id_fundamental_v2: raise Exception( 'Semi-internal error: Fundamental type "%s" can not be mapped ' 'to an id. Is it a true fundamental type? If not so, the ' 'parser should have rejected it.' % self.name) return _id_fundamental_v2[self.name] def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode += nodes.Text(str(self.name)) class ASTTrailingTypeSpecDecltypeAuto(ASTTrailingTypeSpec): def _stringify(self, transform: StringifyTransform) -> str: return 'decltype(auto)' def get_id(self, version: int) -> str: if version == 1: raise NoOldIdError() return 'Dc' def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text(str(self))) class ASTTrailingTypeSpecDecltype(ASTTrailingTypeSpec): def __init__(self, expr: ASTExpression): self.expr = expr def _stringify(self, transform: StringifyTransform) -> str: return 'decltype(' + transform(self.expr) + ')' def get_id(self, version: int) -> str: if version == 1: raise NoOldIdError() return 'DT' + self.expr.get_id(version) + "E" def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode.append(nodes.Text('decltype(')) self.expr.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(')')) class ASTTrailingTypeSpecName(ASTTrailingTypeSpec): def __init__(self, prefix: str, nestedName: ASTNestedName) -> None: self.prefix = prefix self.nestedName = nestedName @property def name(self) -> ASTNestedName: return self.nestedName def get_id(self, version: int) -> str: return self.nestedName.get_id(version) def _stringify(self, transform: StringifyTransform) -> str: res = [] if self.prefix: res.append(self.prefix) res.append(' ') res.append(transform(self.nestedName)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: if self.prefix: signode += addnodes.desc_annotation(self.prefix, self.prefix) signode += nodes.Text(' ') self.nestedName.describe_signature(signode, mode, env, symbol=symbol) class ASTFunctionParameter(ASTBase): def __init__(self, arg: Union["ASTTypeWithInit", "ASTTemplateParamConstrainedTypeWithInit"], ellipsis: bool = False) -> None: self.arg = arg self.ellipsis = ellipsis def get_id(self, version: int, objectType: str = None, symbol: "Symbol" = None) -> str: # this is not part of the normal name mangling in C++ if symbol: # the anchor will be our parent return symbol.parent.declaration.get_id(version, prefixed=None) # else, do the usual if self.ellipsis: return 'z' else: return self.arg.get_id(version) def _stringify(self, transform: StringifyTransform) -> str: if self.ellipsis: return '...' else: return transform(self.arg) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) if self.ellipsis: signode += nodes.Text('...') else: self.arg.describe_signature(signode, mode, env, symbol=symbol) class ASTParametersQualifiers(ASTBase): def __init__(self, args: List[ASTFunctionParameter], volatile: bool, const: bool, refQual: str, exceptionSpec: str, override: bool, final: bool, initializer: str) -> None: self.args = args self.volatile = volatile self.const = const self.refQual = refQual self.exceptionSpec = exceptionSpec self.override = override self.final = final self.initializer = initializer @property def function_params(self) -> List[ASTFunctionParameter]: return self.args def get_modifiers_id(self, version: int) -> str: res = [] if self.volatile: res.append('V') if self.const: if version == 1: res.append('C') else: res.append('K') if self.refQual == '&&': res.append('O') elif self.refQual == '&': res.append('R') return ''.join(res) def get_param_id(self, version: int) -> str: if version == 1: if len(self.args) == 0: return '' else: return '__' + '.'.join(a.get_id(version) for a in self.args) if len(self.args) == 0: return 'v' else: return ''.join(a.get_id(version) for a in self.args) def _stringify(self, transform: StringifyTransform) -> str: res = [] res.append('(') first = True for a in self.args: if not first: res.append(', ') first = False res.append(str(a)) res.append(')') if self.volatile: res.append(' volatile') if self.const: res.append(' const') if self.refQual: res.append(' ') res.append(self.refQual) if self.exceptionSpec: res.append(' ') res.append(str(self.exceptionSpec)) if self.final: res.append(' final') if self.override: res.append(' override') if self.initializer: res.append(' = ') res.append(self.initializer) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) paramlist = addnodes.desc_parameterlist() for arg in self.args: param = addnodes.desc_parameter('', '', noemph=True) if mode == 'lastIsName': # i.e., outer-function params arg.describe_signature(param, 'param', env, symbol=symbol) else: arg.describe_signature(param, 'markType', env, symbol=symbol) paramlist += param signode += paramlist def _add_anno(signode: TextElement, text: str) -> None: signode += nodes.Text(' ') signode += addnodes.desc_annotation(text, text) def _add_text(signode: TextElement, text: str) -> None: signode += nodes.Text(' ' + text) if self.volatile: _add_anno(signode, 'volatile') if self.const: _add_anno(signode, 'const') if self.refQual: _add_text(signode, self.refQual) if self.exceptionSpec: _add_anno(signode, str(self.exceptionSpec)) if self.final: _add_anno(signode, 'final') if self.override: _add_anno(signode, 'override') if self.initializer: _add_text(signode, '= ' + str(self.initializer)) class ASTDeclSpecsSimple(ASTBase): def __init__(self, storage: str, threadLocal: bool, inline: bool, virtual: bool, explicit: bool, constexpr: bool, volatile: bool, const: bool, friend: bool, attrs: List[ASTAttribute]) -> None: self.storage = storage self.threadLocal = threadLocal self.inline = inline self.virtual = virtual self.explicit = explicit self.constexpr = constexpr self.volatile = volatile self.const = const self.friend = friend self.attrs = attrs def mergeWith(self, other: "ASTDeclSpecsSimple") -> "ASTDeclSpecsSimple": if not other: return self return ASTDeclSpecsSimple(self.storage or other.storage, self.threadLocal or other.threadLocal, self.inline or other.inline, self.virtual or other.virtual, self.explicit or other.explicit, self.constexpr or other.constexpr, self.volatile or other.volatile, self.const or other.const, self.friend or other.friend, self.attrs + other.attrs) def _stringify(self, transform: StringifyTransform) -> str: res = [] # type: List[str] res.extend(transform(attr) for attr in self.attrs) if self.storage: res.append(self.storage) if self.threadLocal: res.append('thread_local') if self.inline: res.append('inline') if self.friend: res.append('friend') if self.virtual: res.append('virtual') if self.explicit: res.append('explicit') if self.constexpr: res.append('constexpr') if self.volatile: res.append('volatile') if self.const: res.append('const') return ' '.join(res) def describe_signature(self, signode: TextElement) -> None: addSpace = False for attr in self.attrs: if addSpace: signode += nodes.Text(' ') addSpace = True attr.describe_signature(signode) def _add(signode: TextElement, text: str) -> bool: if addSpace: signode += nodes.Text(' ') signode += addnodes.desc_annotation(text, text) return True if self.storage: addSpace = _add(signode, self.storage) if self.threadLocal: addSpace = _add(signode, 'thread_local') if self.inline: addSpace = _add(signode, 'inline') if self.friend: addSpace = _add(signode, 'friend') if self.virtual: addSpace = _add(signode, 'virtual') if self.explicit: addSpace = _add(signode, 'explicit') if self.constexpr: addSpace = _add(signode, 'constexpr') if self.volatile: addSpace = _add(signode, 'volatile') if self.const: addSpace = _add(signode, 'const') class ASTDeclSpecs(ASTBase): def __init__(self, outer: str, leftSpecs: ASTDeclSpecsSimple, rightSpecs: ASTDeclSpecsSimple, trailing: ASTTrailingTypeSpec) -> None: # leftSpecs and rightSpecs are used for output # allSpecs are used for id generation self.outer = outer self.leftSpecs = leftSpecs self.rightSpecs = rightSpecs self.allSpecs = self.leftSpecs.mergeWith(self.rightSpecs) self.trailingTypeSpec = trailing def get_id(self, version: int) -> str: if version == 1: res = [] res.append(self.trailingTypeSpec.get_id(version)) if self.allSpecs.volatile: res.append('V') if self.allSpecs.const: res.append('C') return ''.join(res) res = [] if self.allSpecs.volatile: res.append('V') if self.allSpecs.const: res.append('K') if self.trailingTypeSpec is not None: res.append(self.trailingTypeSpec.get_id(version)) return ''.join(res) def _stringify(self, transform: StringifyTransform) -> str: res = [] # type: List[str] l = transform(self.leftSpecs) if len(l) > 0: res.append(l) if self.trailingTypeSpec: if len(res) > 0: res.append(" ") res.append(transform(self.trailingTypeSpec)) r = str(self.rightSpecs) if len(r) > 0: if len(res) > 0: res.append(" ") res.append(r) return "".join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) numChildren = len(signode) self.leftSpecs.describe_signature(signode) addSpace = len(signode) != numChildren if self.trailingTypeSpec: if addSpace: signode += nodes.Text(' ') numChildren = len(signode) self.trailingTypeSpec.describe_signature(signode, mode, env, symbol=symbol) addSpace = len(signode) != numChildren if len(str(self.rightSpecs)) > 0: if addSpace: signode += nodes.Text(' ') self.rightSpecs.describe_signature(signode) # Declarator ################################################################################ class ASTArray(ASTBase): def __init__(self, size: ASTExpression): self.size = size def _stringify(self, transform: StringifyTransform) -> str: if self.size: return '[' + transform(self.size) + ']' else: return '[]' def get_id(self, version: int) -> str: if version == 1: return 'A' if version == 2: if self.size: return 'A' + str(self.size) + '_' else: return 'A_' if self.size: return 'A' + self.size.get_id(version) + '_' else: return 'A_' def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) signode.append(nodes.Text("[")) if self.size: self.size.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text("]")) class ASTDeclarator(ASTBase): @property def name(self) -> ASTNestedName: raise NotImplementedError(repr(self)) @property def isPack(self) -> bool: raise NotImplementedError(repr(self)) @property def function_params(self) -> List[ASTFunctionParameter]: raise NotImplementedError(repr(self)) def require_space_after_declSpecs(self) -> bool: raise NotImplementedError(repr(self)) def get_modifiers_id(self, version: int) -> str: raise NotImplementedError(repr(self)) def get_param_id(self, version: int) -> str: raise NotImplementedError(repr(self)) def get_ptr_suffix_id(self, version: int) -> str: raise NotImplementedError(repr(self)) def get_type_id(self, version: int, returnTypeId: str) -> str: raise NotImplementedError(repr(self)) def is_function_type(self) -> bool: raise NotImplementedError(repr(self)) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: raise NotImplementedError(repr(self)) class ASTDeclaratorNameParamQual(ASTDeclarator): def __init__(self, declId: ASTNestedName, arrayOps: List[ASTArray], paramQual: ASTParametersQualifiers) -> None: self.declId = declId self.arrayOps = arrayOps self.paramQual = paramQual @property def name(self) -> ASTNestedName: return self.declId @property def isPack(self) -> bool: return False @property def function_params(self) -> List[ASTFunctionParameter]: return self.paramQual.function_params # only the modifiers for a function, e.g., def get_modifiers_id(self, version: int) -> str: # cv-qualifiers if self.paramQual: return self.paramQual.get_modifiers_id(version) raise Exception("This should only be called on a function: %s" % self) def get_param_id(self, version: int) -> str: # only the parameters (if any) if self.paramQual: return self.paramQual.get_param_id(version) else: return '' def get_ptr_suffix_id(self, version: int) -> str: # only the array specifiers return ''.join(a.get_id(version) for a in self.arrayOps) def get_type_id(self, version: int, returnTypeId: str) -> str: assert version >= 2 res = [] # TOOD: can we actually have both array ops and paramQual? res.append(self.get_ptr_suffix_id(version)) if self.paramQual: res.append(self.get_modifiers_id(version)) res.append('F') res.append(returnTypeId) res.append(self.get_param_id(version)) res.append('E') else: res.append(returnTypeId) return ''.join(res) # ------------------------------------------------------------------------ def require_space_after_declSpecs(self) -> bool: return self.declId is not None def is_function_type(self) -> bool: return self.paramQual is not None def _stringify(self, transform: StringifyTransform) -> str: res = [] if self.declId: res.append(transform(self.declId)) for op in self.arrayOps: res.append(transform(op)) if self.paramQual: res.append(transform(self.paramQual)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) if self.declId: self.declId.describe_signature(signode, mode, env, symbol) for op in self.arrayOps: op.describe_signature(signode, mode, env, symbol) if self.paramQual: self.paramQual.describe_signature(signode, mode, env, symbol) class ASTDeclaratorNameBitField(ASTDeclarator): def __init__(self, declId: ASTNestedName, size: ASTExpression): self.declId = declId self.size = size @property def name(self) -> ASTNestedName: return self.declId def get_param_id(self, version: int) -> str: # only the parameters (if any) return '' def get_ptr_suffix_id(self, version: int) -> str: # only the array specifiers return '' # ------------------------------------------------------------------------ def require_space_after_declSpecs(self) -> bool: return self.declId is not None def is_function_type(self) -> bool: return False def _stringify(self, transform: StringifyTransform) -> str: res = [] if self.declId: res.append(transform(self.declId)) res.append(" : ") res.append(transform(self.size)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) if self.declId: self.declId.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(' : ', ' : ')) self.size.describe_signature(signode, mode, env, symbol) class ASTDeclaratorPtr(ASTDeclarator): def __init__(self, next: ASTDeclarator, volatile: bool, const: bool, attrs: List[ASTAttribute]) -> None: assert next self.next = next self.volatile = volatile self.const = const self.attrs = attrs @property def name(self) -> ASTNestedName: return self.next.name @property def function_params(self) -> List[ASTFunctionParameter]: return self.next.function_params def require_space_after_declSpecs(self) -> bool: return self.next.require_space_after_declSpecs() def _stringify(self, transform: StringifyTransform) -> str: res = ['*'] for a in self.attrs: res.append(transform(a)) if len(self.attrs) > 0 and (self.volatile or self.const): res.append(' ') if self.volatile: res.append('volatile') if self.const: if self.volatile: res.append(' ') res.append('const') if self.const or self.volatile or len(self.attrs) > 0: if self.next.require_space_after_declSpecs(): res.append(' ') res.append(transform(self.next)) return ''.join(res) def get_modifiers_id(self, version: int) -> str: return self.next.get_modifiers_id(version) def get_param_id(self, version: int) -> str: return self.next.get_param_id(version) def get_ptr_suffix_id(self, version: int) -> str: if version == 1: res = ['P'] if self.volatile: res.append('V') if self.const: res.append('C') res.append(self.next.get_ptr_suffix_id(version)) return ''.join(res) res = [self.next.get_ptr_suffix_id(version)] res.append('P') if self.volatile: res.append('V') if self.const: res.append('C') return ''.join(res) def get_type_id(self, version: int, returnTypeId: str) -> str: # ReturnType *next, so we are part of the return type of 'next res = ['P'] if self.volatile: res.append('V') if self.const: res.append('C') res.append(returnTypeId) return self.next.get_type_id(version, returnTypeId=''.join(res)) def is_function_type(self) -> bool: return self.next.is_function_type() def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) signode += nodes.Text("*") for a in self.attrs: a.describe_signature(signode) if len(self.attrs) > 0 and (self.volatile or self.const): signode += nodes.Text(' ') def _add_anno(signode: TextElement, text: str) -> None: signode += addnodes.desc_annotation(text, text) if self.volatile: _add_anno(signode, 'volatile') if self.const: if self.volatile: signode += nodes.Text(' ') _add_anno(signode, 'const') if self.const or self.volatile or len(self.attrs) > 0: if self.next.require_space_after_declSpecs(): signode += nodes.Text(' ') self.next.describe_signature(signode, mode, env, symbol) class ASTDeclaratorRef(ASTDeclarator): def __init__(self, next: ASTDeclarator, attrs: List[ASTAttribute]) -> None: assert next self.next = next self.attrs = attrs @property def name(self) -> ASTNestedName: return self.next.name @property def isPack(self) -> bool: return True @property def function_params(self) -> List[ASTFunctionParameter]: return self.next.function_params def require_space_after_declSpecs(self) -> bool: return self.next.require_space_after_declSpecs() def _stringify(self, transform: StringifyTransform) -> str: res = ['&'] for a in self.attrs: res.append(transform(a)) if len(self.attrs) > 0 and self.next.require_space_after_declSpecs(): res.append(' ') res.append(transform(self.next)) return ''.join(res) def get_modifiers_id(self, version: int) -> str: return self.next.get_modifiers_id(version) def get_param_id(self, version: int) -> str: # only the parameters (if any) return self.next.get_param_id(version) def get_ptr_suffix_id(self, version: int) -> str: if version == 1: return 'R' + self.next.get_ptr_suffix_id(version) else: return self.next.get_ptr_suffix_id(version) + 'R' def get_type_id(self, version: int, returnTypeId: str) -> str: assert version >= 2 # ReturnType &next, so we are part of the return type of 'next return self.next.get_type_id(version, returnTypeId='R' + returnTypeId) def is_function_type(self) -> bool: return self.next.is_function_type() def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) signode += nodes.Text("&") for a in self.attrs: a.describe_signature(signode) if len(self.attrs) > 0 and self.next.require_space_after_declSpecs(): signode += nodes.Text(' ') self.next.describe_signature(signode, mode, env, symbol) class ASTDeclaratorParamPack(ASTDeclarator): def __init__(self, next: ASTDeclarator) -> None: assert next self.next = next @property def name(self) -> ASTNestedName: return self.next.name @property def function_params(self) -> List[ASTFunctionParameter]: return self.next.function_params def require_space_after_declSpecs(self) -> bool: return False def _stringify(self, transform: StringifyTransform) -> str: res = transform(self.next) if self.next.name: res = ' ' + res return '...' + res def get_modifiers_id(self, version: int) -> str: return self.next.get_modifiers_id(version) def get_param_id(self, version: int) -> str: # only the parameters (if any) return self.next.get_param_id(version) def get_ptr_suffix_id(self, version: int) -> str: if version == 1: return 'Dp' + self.next.get_ptr_suffix_id(version) else: return self.next.get_ptr_suffix_id(version) + 'Dp' def get_type_id(self, version: int, returnTypeId: str) -> str: assert version >= 2 # ReturnType... next, so we are part of the return type of 'next return self.next.get_type_id(version, returnTypeId='Dp' + returnTypeId) def is_function_type(self) -> bool: return self.next.is_function_type() def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) signode += nodes.Text("...") if self.next.name: signode += nodes.Text(' ') self.next.describe_signature(signode, mode, env, symbol) class ASTDeclaratorMemPtr(ASTDeclarator): def __init__(self, className: ASTNestedName, const: bool, volatile: bool, next: ASTDeclarator) -> None: assert className assert next self.className = className self.const = const self.volatile = volatile self.next = next @property def name(self) -> ASTNestedName: return self.next.name @property def function_params(self) -> List[ASTFunctionParameter]: return self.next.function_params def require_space_after_declSpecs(self) -> bool: return True def _stringify(self, transform: StringifyTransform) -> str: res = [] res.append(transform(self.className)) res.append('::*') if self.volatile: res.append('volatile') if self.const: if self.volatile: res.append(' ') res.append('const') if self.next.require_space_after_declSpecs(): res.append(' ') res.append(transform(self.next)) return ''.join(res) def get_modifiers_id(self, version: int) -> str: if version == 1: raise NoOldIdError() else: return self.next.get_modifiers_id(version) def get_param_id(self, version: int) -> str: # only the parameters (if any) if version == 1: raise NoOldIdError() else: return self.next.get_param_id(version) def get_ptr_suffix_id(self, version: int) -> str: if version == 1: raise NoOldIdError() else: raise NotImplementedError() return self.next.get_ptr_suffix_id(version) + 'Dp' def get_type_id(self, version: int, returnTypeId: str) -> str: assert version >= 2 # ReturnType name::* next, so we are part of the return type of next nextReturnTypeId = '' if self.volatile: nextReturnTypeId += 'V' if self.const: nextReturnTypeId += 'K' nextReturnTypeId += 'M' nextReturnTypeId += self.className.get_id(version) nextReturnTypeId += returnTypeId return self.next.get_type_id(version, nextReturnTypeId) def is_function_type(self) -> bool: return self.next.is_function_type() def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) self.className.describe_signature(signode, mode, env, symbol) signode += nodes.Text('::*') def _add_anno(signode: TextElement, text: str) -> None: signode += addnodes.desc_annotation(text, text) if self.volatile: _add_anno(signode, 'volatile') if self.const: if self.volatile: signode += nodes.Text(' ') _add_anno(signode, 'const') if self.next.require_space_after_declSpecs(): signode += nodes.Text(' ') self.next.describe_signature(signode, mode, env, symbol) class ASTDeclaratorParen(ASTDeclarator): def __init__(self, inner: ASTDeclarator, next: ASTDeclarator) -> None: assert inner assert next self.inner = inner self.next = next # TODO: we assume the name, params, and qualifiers are in inner @property def name(self) -> ASTNestedName: return self.inner.name @property def function_params(self) -> List[ASTFunctionParameter]: return self.inner.function_params def require_space_after_declSpecs(self) -> bool: return True def _stringify(self, transform: StringifyTransform) -> str: res = ['('] res.append(transform(self.inner)) res.append(')') res.append(transform(self.next)) return ''.join(res) def get_modifiers_id(self, version: int) -> str: return self.inner.get_modifiers_id(version) def get_param_id(self, version: int) -> str: # only the parameters (if any) return self.inner.get_param_id(version) def get_ptr_suffix_id(self, version: int) -> str: if version == 1: raise NoOldIdError() # TODO: was this implemented before? return self.next.get_ptr_suffix_id(version) + \ self.inner.get_ptr_suffix_id(version) else: return self.inner.get_ptr_suffix_id(version) + \ self.next.get_ptr_suffix_id(version) def get_type_id(self, version: int, returnTypeId: str) -> str: assert version >= 2 # ReturnType (inner)next, so 'inner' returns everything outside nextId = self.next.get_type_id(version, returnTypeId) return self.inner.get_type_id(version, returnTypeId=nextId) def is_function_type(self) -> bool: return self.inner.is_function_type() def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) signode += nodes.Text('(') self.inner.describe_signature(signode, mode, env, symbol) signode += nodes.Text(')') self.next.describe_signature(signode, "noneIsName", env, symbol) # Type and initializer stuff ############################################################################################## class ASTPackExpansionExpr(ASTExpression): def __init__(self, expr: Union[ASTExpression, ASTBracedInitList]): self.expr = expr def _stringify(self, transform: StringifyTransform) -> str: return transform(self.expr) + '...' def get_id(self, version: int) -> str: id = self.expr.get_id(version) return 'sp' + id def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.expr.describe_signature(signode, mode, env, symbol) signode += nodes.Text('...') class ASTParenExprList(ASTBase): def __init__(self, exprs: List[Union[ASTExpression, ASTBracedInitList]]) -> None: self.exprs = exprs def get_id(self, version: int) -> str: return "pi%sE" % ''.join(e.get_id(version) for e in self.exprs) def _stringify(self, transform: StringifyTransform) -> str: exprs = [transform(e) for e in self.exprs] return '(%s)' % ', '.join(exprs) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) signode.append(nodes.Text('(')) first = True for e in self.exprs: if not first: signode.append(nodes.Text(', ')) else: first = False e.describe_signature(signode, mode, env, symbol) signode.append(nodes.Text(')')) class ASTInitializer(ASTBase): def __init__(self, value: Union[ASTExpression, ASTBracedInitList], hasAssign: bool = True) -> None: self.value = value self.hasAssign = hasAssign def _stringify(self, transform: StringifyTransform) -> str: val = transform(self.value) if self.hasAssign: return ' = ' + val else: return val def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) if self.hasAssign: signode.append(nodes.Text(' = ')) self.value.describe_signature(signode, 'markType', env, symbol) class ASTType(ASTBase): def __init__(self, declSpecs: ASTDeclSpecs, decl: ASTDeclarator) -> None: assert declSpecs assert decl self.declSpecs = declSpecs self.decl = decl @property def name(self) -> ASTNestedName: return self.decl.name @property def isPack(self) -> bool: return self.decl.isPack @property def function_params(self) -> List[ASTFunctionParameter]: return self.decl.function_params def get_id(self, version: int, objectType: str = None, symbol: "Symbol" = None) -> str: if version == 1: res = [] if objectType: # needs the name if objectType == 'function': # also modifiers res.append(symbol.get_full_nested_name().get_id(version)) res.append(self.decl.get_param_id(version)) res.append(self.decl.get_modifiers_id(version)) if (self.declSpecs.leftSpecs.constexpr or (self.declSpecs.rightSpecs and self.declSpecs.rightSpecs.constexpr)): res.append('CE') elif objectType == 'type': # just the name res.append(symbol.get_full_nested_name().get_id(version)) else: print(objectType) assert False else: # only type encoding if self.decl.is_function_type(): raise NoOldIdError() res.append(self.declSpecs.get_id(version)) res.append(self.decl.get_ptr_suffix_id(version)) res.append(self.decl.get_param_id(version)) return ''.join(res) # other versions res = [] if objectType: # needs the name if objectType == 'function': # also modifiers modifiers = self.decl.get_modifiers_id(version) res.append(symbol.get_full_nested_name().get_id(version, modifiers)) if version >= 4: # with templates we need to mangle the return type in as well templ = symbol.declaration.templatePrefix if templ is not None: typeId = self.decl.get_ptr_suffix_id(version) returnTypeId = self.declSpecs.get_id(version) res.append(typeId) res.append(returnTypeId) res.append(self.decl.get_param_id(version)) elif objectType == 'type': # just the name res.append(symbol.get_full_nested_name().get_id(version)) else: print(objectType) assert False else: # only type encoding # the 'returnType' of a non-function type is simply just the last # type, i.e., for 'int*' it is 'int' returnTypeId = self.declSpecs.get_id(version) typeId = self.decl.get_type_id(version, returnTypeId) res.append(typeId) return ''.join(res) def _stringify(self, transform: StringifyTransform) -> str: res = [] declSpecs = transform(self.declSpecs) res.append(declSpecs) if self.decl.require_space_after_declSpecs() and len(declSpecs) > 0: res.append(' ') res.append(transform(self.decl)) return ''.join(res) def get_type_declaration_prefix(self) -> str: if self.declSpecs.trailingTypeSpec: return 'typedef' else: return 'type' def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) self.declSpecs.describe_signature(signode, 'markType', env, symbol) if (self.decl.require_space_after_declSpecs() and len(str(self.declSpecs)) > 0): signode += nodes.Text(' ') # for parameters that don't really declare new names we get 'markType', # this should not be propagated, but be 'noneIsName'. if mode == 'markType': mode = 'noneIsName' self.decl.describe_signature(signode, mode, env, symbol) class ASTTemplateParamConstrainedTypeWithInit(ASTBase): def __init__(self, type: ASTType, init: ASTType) -> None: assert type self.type = type self.init = init @property def name(self) -> ASTNestedName: return self.type.name @property def isPack(self) -> bool: return self.type.isPack def get_id(self, version: int, objectType: str = None, symbol: "Symbol" = None) -> str: # this is not part of the normal name mangling in C++ assert version >= 2 if symbol: # the anchor will be our parent return symbol.parent.declaration.get_id(version, prefixed=False) else: return self.type.get_id(version) def _stringify(self, transform: StringifyTransform) -> str: res = transform(self.type) if self.init: res += " = " res += transform(self.init) return res def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.type.describe_signature(signode, mode, env, symbol) if self.init: signode += nodes.Text(" = ") self.init.describe_signature(signode, mode, env, symbol) class ASTTypeWithInit(ASTBase): def __init__(self, type: ASTType, init: ASTInitializer) -> None: self.type = type self.init = init @property def name(self) -> ASTNestedName: return self.type.name @property def isPack(self) -> bool: return self.type.isPack def get_id(self, version: int, objectType: str = None, symbol: "Symbol" = None) -> str: if objectType != 'member': return self.type.get_id(version, objectType) if version == 1: return (symbol.get_full_nested_name().get_id(version) + '__' + self.type.get_id(version)) return symbol.get_full_nested_name().get_id(version) def _stringify(self, transform: StringifyTransform) -> str: res = [] res.append(transform(self.type)) if self.init: res.append(transform(self.init)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) self.type.describe_signature(signode, mode, env, symbol) if self.init: self.init.describe_signature(signode, mode, env, symbol) class ASTTypeUsing(ASTBase): def __init__(self, name: ASTNestedName, type: ASTType) -> None: self.name = name self.type = type def get_id(self, version: int, objectType: str = None, symbol: "Symbol" = None) -> str: if version == 1: raise NoOldIdError() return symbol.get_full_nested_name().get_id(version) def _stringify(self, transform: StringifyTransform) -> str: res = [] res.append(transform(self.name)) if self.type: res.append(' = ') res.append(transform(self.type)) return ''.join(res) def get_type_declaration_prefix(self) -> str: return 'using' def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) self.name.describe_signature(signode, mode, env, symbol=symbol) if self.type: signode += nodes.Text(' = ') self.type.describe_signature(signode, 'markType', env, symbol=symbol) # Other declarations ############################################################################################## class ASTConcept(ASTBase): def __init__(self, nestedName: ASTNestedName, initializer: ASTInitializer) -> None: self.nestedName = nestedName self.initializer = initializer @property def name(self) -> ASTNestedName: return self.nestedName def get_id(self, version: int, objectType: str = None, symbol: "Symbol" = None) -> str: if version == 1: raise NoOldIdError() return symbol.get_full_nested_name().get_id(version) def _stringify(self, transform: StringifyTransform) -> str: res = transform(self.nestedName) if self.initializer: res += transform(self.initializer) return res def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.nestedName.describe_signature(signode, mode, env, symbol) if self.initializer: self.initializer.describe_signature(signode, mode, env, symbol) class ASTBaseClass(ASTBase): def __init__(self, name: ASTNestedName, visibility: str, virtual: bool, pack: bool) -> None: self.name = name self.visibility = visibility self.virtual = virtual self.pack = pack def _stringify(self, transform: StringifyTransform) -> str: res = [] if self.visibility is not None: res.append(self.visibility) res.append(' ') if self.virtual: res.append('virtual ') res.append(transform(self.name)) if self.pack: res.append('...') return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) if self.visibility is not None: signode += addnodes.desc_annotation(self.visibility, self.visibility) signode += nodes.Text(' ') if self.virtual: signode += addnodes.desc_annotation('virtual', 'virtual') signode += nodes.Text(' ') self.name.describe_signature(signode, 'markType', env, symbol=symbol) if self.pack: signode += nodes.Text('...') class ASTClass(ASTBase): def __init__(self, name: ASTNestedName, final: bool, bases: List[ASTBaseClass]) -> None: self.name = name self.final = final self.bases = bases def get_id(self, version: int, objectType: str, symbol: "Symbol") -> str: return symbol.get_full_nested_name().get_id(version) def _stringify(self, transform: StringifyTransform) -> str: res = [] res.append(transform(self.name)) if self.final: res.append(' final') if len(self.bases) > 0: res.append(' : ') first = True for b in self.bases: if not first: res.append(', ') first = False res.append(transform(b)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) self.name.describe_signature(signode, mode, env, symbol=symbol) if self.final: signode += nodes.Text(' ') signode += addnodes.desc_annotation('final', 'final') if len(self.bases) > 0: signode += nodes.Text(' : ') for b in self.bases: b.describe_signature(signode, mode, env, symbol=symbol) signode += nodes.Text(', ') signode.pop() class ASTUnion(ASTBase): def __init__(self, name: ASTNestedName) -> None: self.name = name def get_id(self, version: int, objectType: str, symbol: "Symbol") -> str: if version == 1: raise NoOldIdError() return symbol.get_full_nested_name().get_id(version) def _stringify(self, transform: StringifyTransform) -> str: return transform(self.name) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) self.name.describe_signature(signode, mode, env, symbol=symbol) class ASTEnum(ASTBase): def __init__(self, name: ASTNestedName, scoped: str, underlyingType: ASTType) -> None: self.name = name self.scoped = scoped self.underlyingType = underlyingType def get_id(self, version: int, objectType: str, symbol: "Symbol") -> str: if version == 1: raise NoOldIdError() return symbol.get_full_nested_name().get_id(version) def _stringify(self, transform: StringifyTransform) -> str: res = [] if self.scoped: res.append(self.scoped) res.append(' ') res.append(transform(self.name)) if self.underlyingType: res.append(' : ') res.append(transform(self.underlyingType)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) # self.scoped has been done by the CPPEnumObject self.name.describe_signature(signode, mode, env, symbol=symbol) if self.underlyingType: signode += nodes.Text(' : ') self.underlyingType.describe_signature(signode, 'noneIsName', env, symbol=symbol) class ASTEnumerator(ASTBase): def __init__(self, name: ASTNestedName, init: ASTInitializer) -> None: self.name = name self.init = init def get_id(self, version: int, objectType: str, symbol: "Symbol") -> str: if version == 1: raise NoOldIdError() return symbol.get_full_nested_name().get_id(version) def _stringify(self, transform: StringifyTransform) -> str: res = [] res.append(transform(self.name)) if self.init: res.append(transform(self.init)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: verify_description_mode(mode) self.name.describe_signature(signode, mode, env, symbol) if self.init: self.init.describe_signature(signode, 'markType', env, symbol) ################################################################################ # Templates ################################################################################ # Parameters ################################################################################ class ASTTemplateParam(ASTBase): def get_identifier(self) -> ASTIdentifier: raise NotImplementedError(repr(self)) def get_id(self, version: int) -> str: raise NotImplementedError(repr(self)) def describe_signature(self, parentNode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: raise NotImplementedError(repr(self)) class ASTTemplateKeyParamPackIdDefault(ASTTemplateParam): def __init__(self, key: str, identifier: ASTIdentifier, parameterPack: bool, default: ASTType) -> None: assert key if parameterPack: assert default is None self.key = key self.identifier = identifier self.parameterPack = parameterPack self.default = default def get_identifier(self) -> ASTIdentifier: return self.identifier def get_id(self, version: int) -> str: assert version >= 2 # this is not part of the normal name mangling in C++ res = [] if self.parameterPack: res.append('Dp') else: res.append('0') # we need to put something return ''.join(res) def _stringify(self, transform: StringifyTransform) -> str: res = [self.key] if self.parameterPack: if self.identifier: res.append(' ') res.append('...') if self.identifier: if not self.parameterPack: res.append(' ') res.append(transform(self.identifier)) if self.default: res.append(' = ') res.append(transform(self.default)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode += nodes.Text(self.key) if self.parameterPack: if self.identifier: signode += nodes.Text(' ') signode += nodes.Text('...') if self.identifier: if not self.parameterPack: signode += nodes.Text(' ') self.identifier.describe_signature(signode, mode, env, '', '', symbol) if self.default: signode += nodes.Text(' = ') self.default.describe_signature(signode, 'markType', env, symbol) class ASTTemplateParamType(ASTTemplateParam): def __init__(self, data: ASTTemplateKeyParamPackIdDefault) -> None: assert data self.data = data @property def name(self) -> ASTNestedName: id = self.get_identifier() return ASTNestedName([ASTNestedNameElement(id, None)], [False], rooted=False) @property def isPack(self) -> bool: return self.data.parameterPack def get_identifier(self) -> ASTIdentifier: return self.data.get_identifier() def get_id(self, version: int, objectType: str = None, symbol: "Symbol" = None) -> str: # this is not part of the normal name mangling in C++ assert version >= 2 if symbol: # the anchor will be our parent return symbol.parent.declaration.get_id(version, prefixed=False) else: return self.data.get_id(version) def _stringify(self, transform: StringifyTransform) -> str: return transform(self.data) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.data.describe_signature(signode, mode, env, symbol) class ASTTemplateParamTemplateType(ASTTemplateParam): def __init__(self, nestedParams: "ASTTemplateParams", data: ASTTemplateKeyParamPackIdDefault) -> None: assert nestedParams assert data self.nestedParams = nestedParams self.data = data @property def name(self) -> ASTNestedName: id = self.get_identifier() return ASTNestedName([ASTNestedNameElement(id, None)], [False], rooted=False) @property def isPack(self) -> bool: return self.data.parameterPack def get_identifier(self) -> ASTIdentifier: return self.data.get_identifier() def get_id(self, version: int, objectType: str = None, symbol: "Symbol" = None) -> str: assert version >= 2 # this is not part of the normal name mangling in C++ if symbol: # the anchor will be our parent return symbol.parent.declaration.get_id(version, prefixed=None) else: return self.nestedParams.get_id(version) + self.data.get_id(version) def _stringify(self, transform: StringifyTransform) -> str: return transform(self.nestedParams) + transform(self.data) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.nestedParams.describe_signature(signode, 'noneIsName', env, symbol) signode += nodes.Text(' ') self.data.describe_signature(signode, mode, env, symbol) class ASTTemplateParamNonType(ASTTemplateParam): def __init__(self, param: Union[ASTTypeWithInit, ASTTemplateParamConstrainedTypeWithInit]) -> None: assert param self.param = param @property def name(self) -> ASTNestedName: id = self.get_identifier() return ASTNestedName([ASTNestedNameElement(id, None)], [False], rooted=False) @property def isPack(self) -> bool: return self.param.isPack def get_identifier(self) -> ASTIdentifier: name = self.param.name if name: assert len(name.names) == 1 assert name.names[0].identOrOp assert not name.names[0].templateArgs res = name.names[0].identOrOp assert isinstance(res, ASTIdentifier) return res else: return None def get_id(self, version: int, objectType: str = None, symbol: "Symbol" = None) -> str: assert version >= 2 # this is not part of the normal name mangling in C++ if symbol: # the anchor will be our parent return symbol.parent.declaration.get_id(version, prefixed=None) else: return '_' + self.param.get_id(version) def _stringify(self, transform: StringifyTransform) -> str: return transform(self.param) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: self.param.describe_signature(signode, mode, env, symbol) class ASTTemplateParams(ASTBase): def __init__(self, params: List[ASTTemplateParam]) -> None: assert params is not None self.params = params def get_id(self, version: int) -> str: assert version >= 2 res = [] res.append("I") for param in self.params: res.append(param.get_id(version)) res.append("E") return ''.join(res) def _stringify(self, transform: StringifyTransform) -> str: res = [] res.append("template<") res.append(", ".join(transform(a) for a in self.params)) res.append("> ") return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: signode += nodes.Text("template<") first = True for param in self.params: if not first: signode += nodes.Text(", ") first = False param.describe_signature(signode, mode, env, symbol) signode += nodes.Text(">") def describe_signature_as_introducer( self, parentNode: desc_signature, mode: str, env: "BuildEnvironment", symbol: "Symbol", lineSpec: bool) -> None: def makeLine(parentNode: desc_signature) -> addnodes.desc_signature_line: signode = addnodes.desc_signature_line() parentNode += signode signode.sphinx_line_type = 'templateParams' return signode lineNode = makeLine(parentNode) lineNode += nodes.Text("template<") first = True for param in self.params: if not first: lineNode += nodes.Text(", ") first = False if lineSpec: lineNode = makeLine(parentNode) param.describe_signature(lineNode, mode, env, symbol) if lineSpec and not first: lineNode = makeLine(parentNode) lineNode += nodes.Text(">") # Template introducers ################################################################################ class ASTTemplateIntroductionParameter(ASTBase): def __init__(self, identifier: ASTIdentifier, parameterPack: bool) -> None: self.identifier = identifier self.parameterPack = parameterPack @property def name(self) -> ASTNestedName: id = self.get_identifier() return ASTNestedName([ASTNestedNameElement(id, None)], [False], rooted=False) @property def isPack(self) -> bool: return self.parameterPack def get_identifier(self) -> ASTIdentifier: return self.identifier def get_id(self, version: int, objectType: str = None, symbol: "Symbol" = None) -> str: assert version >= 2 # this is not part of the normal name mangling in C++ if symbol: # the anchor will be our parent return symbol.parent.declaration.get_id(version, prefixed=None) else: if self.parameterPack: return 'Dp' else: return '0' # we need to put something def get_id_as_arg(self, version: int) -> str: assert version >= 2 # used for the implicit requires clause res = self.identifier.get_id(version) if self.parameterPack: return 'sp' + res else: return res def _stringify(self, transform: StringifyTransform) -> str: res = [] if self.parameterPack: res.append('...') res.append(transform(self.identifier)) return ''.join(res) def describe_signature(self, signode: TextElement, mode: str, env: "BuildEnvironment", symbol: "Symbol") -> None: if self.parameterPack: signode += nodes.Text('...') self.identifier.describe_signature(signode, mode, env, '', '', symbol) class ASTTemplateIntroduction(ASTBase): def __init__(self, concept: ASTNestedName, params: List[ASTTemplateIntroductionParameter]) -> None: assert len(params) > 0 self.concept = concept self.params = params def get_id(self, version: int) -> str: assert version >= 2 # first do the same as a normal template parameter list res = [] res.append("I") for param in self.params: res.append(param.get_id(version)) res.append("E") # let's use X expr E, which is otherwise for constant template args res.append("X") res.append(self.concept.get_id(version)) res.append("I") for param in self.params: res.append(param.get_id_as_arg(version)) res.append("E") res.append("E") return ''.join(res) def _stringify(self, transform: StringifyTransform) -> str: res = [] res.append(transform(self.concept)) res.append('{') res.append(', '.join(transform(param) for param in self.params)) res.append('} ') return ''.join(res) def describe_signature_as_introducer( self, parentNode: desc_signature, mode: str, env: "BuildEnvironment", symbol: "Symbol", lineSpec: bool) -> None: # Note: 'lineSpec' has no effect on template introductions. signode = addnodes.desc_signature_line() parentNode += signode signode.sphinx_line_type = 'templateIntroduction' self.concept.describe_signature(signode, 'markType', env, symbol) signode += nodes.Text('{') first = True for param in self.params: if not first: signode += nodes.Text(', ') first = False param.describe_signature(signode, mode, env, symbol) signode += nodes.Text('}') class ASTTemplateDeclarationPrefix(ASTBase): def __init__(self, templates: List[Union[ASTTemplateParams, ASTTemplateIntroduction]]) -> None: # templates is None means it's an explicit instantiation of a variable self.templates = templates def get_id(self, version: int) -> str: assert version >= 2 # this is not part of a normal name mangling system res = [] for t in self.templates: res.append(t.get_id(version)) return ''.join(res) def _stringify(self, transform: StringifyTransform) -> str: res = [] for t in self.templates: res.append(transform(t)) return ''.join(res) def describe_signature(self, signode: desc_signature, mode: str, env: "BuildEnvironment", symbol: "Symbol", lineSpec: bool) -> None: verify_description_mode(mode) for t in self.templates: t.describe_signature_as_introducer(signode, 'lastIsName', env, symbol, lineSpec) ################################################################################ ################################################################################ class ASTDeclaration(ASTBase): def __init__(self, objectType: str, directiveType: str, visibility: str, templatePrefix: ASTTemplateDeclarationPrefix, declaration: Any) -> None: self.objectType = objectType self.directiveType = directiveType self.visibility = visibility self.templatePrefix = templatePrefix self.declaration = declaration self.symbol = None # type: Symbol # set by CPPObject._add_enumerator_to_parent self.enumeratorScopedSymbol = None # type: Symbol def clone(self) -> "ASTDeclaration": if self.templatePrefix: templatePrefixClone = self.templatePrefix.clone() else: templatePrefixClone = None return ASTDeclaration(self.objectType, self.directiveType, self.visibility, templatePrefixClone, self.declaration.clone()) @property def name(self) -> ASTNestedName: return self.declaration.name @property def function_params(self) -> List[ASTFunctionParameter]: if self.objectType != 'function': return None return self.declaration.function_params def get_id(self, version: int, prefixed: bool = True) -> str: if version == 1: if self.templatePrefix: raise NoOldIdError() if self.objectType == 'enumerator' and self.enumeratorScopedSymbol: return self.enumeratorScopedSymbol.declaration.get_id(version) return self.declaration.get_id(version, self.objectType, self.symbol) # version >= 2 if self.objectType == 'enumerator' and self.enumeratorScopedSymbol: return self.enumeratorScopedSymbol.declaration.get_id(version, prefixed) if prefixed: res = [_id_prefix[version]] else: res = [] if self.templatePrefix: res.append(self.templatePrefix.get_id(version)) res.append(self.declaration.get_id(version, self.objectType, self.symbol)) return ''.join(res) def get_newest_id(self) -> str: return self.get_id(_max_id, True) def _stringify(self, transform: StringifyTransform) -> str: res = [] if self.visibility and self.visibility != "public": res.append(self.visibility) res.append(' ') if self.templatePrefix: res.append(transform(self.templatePrefix)) res.append(transform(self.declaration)) return ''.join(res) def describe_signature(self, signode: desc_signature, mode: str, env: "BuildEnvironment", options: Dict) -> None: verify_description_mode(mode) assert self.symbol # The caller of the domain added a desc_signature node. # Always enable multiline: signode['is_multiline'] = True # Put each line in a desc_signature_line node. mainDeclNode = addnodes.desc_signature_line() mainDeclNode.sphinx_line_type = 'declarator' mainDeclNode['add_permalink'] = not self.symbol.isRedeclaration if self.templatePrefix: self.templatePrefix.describe_signature(signode, mode, env, symbol=self.symbol, lineSpec=options.get('tparam-line-spec')) signode += mainDeclNode if self.visibility and self.visibility != "public": mainDeclNode += addnodes.desc_annotation(self.visibility + " ", self.visibility + " ") if self.objectType == 'type': prefix = self.declaration.get_type_declaration_prefix() prefix += ' ' mainDeclNode += addnodes.desc_annotation(prefix, prefix) elif self.objectType == 'concept': mainDeclNode += addnodes.desc_annotation('concept ', 'concept ') elif self.objectType == 'member': pass elif self.objectType == 'function': pass elif self.objectType == 'class': assert self.directiveType in ('class', 'struct') prefix = self.directiveType + ' ' mainDeclNode += addnodes.desc_annotation(prefix, prefix) elif self.objectType == 'union': mainDeclNode += addnodes.desc_annotation('union ', 'union ') elif self.objectType == 'enum': if self.directiveType == 'enum': prefix = 'enum ' elif self.directiveType == 'enum-class': prefix = 'enum class ' elif self.directiveType == 'enum-struct': prefix = 'enum struct ' else: assert False # wrong directiveType used mainDeclNode += addnodes.desc_annotation(prefix, prefix) elif self.objectType == 'enumerator': mainDeclNode += addnodes.desc_annotation('enumerator ', 'enumerator ') else: assert False self.declaration.describe_signature(mainDeclNode, mode, env, self.symbol) class ASTNamespace(ASTBase): def __init__(self, nestedName: ASTNestedName, templatePrefix: ASTTemplateDeclarationPrefix) -> None: self.nestedName = nestedName self.templatePrefix = templatePrefix def _stringify(self, transform: StringifyTransform) -> str: res = [] if self.templatePrefix: res.append(transform(self.templatePrefix)) res.append(transform(self.nestedName)) return ''.join(res) class SymbolLookupResult: def __init__(self, symbols: Iterator["Symbol"], parentSymbol: "Symbol", identOrOp: Union[ASTIdentifier, ASTOperator], templateParams: Any, templateArgs: ASTTemplateArgs) -> None: self.symbols = symbols self.parentSymbol = parentSymbol self.identOrOp = identOrOp self.templateParams = templateParams self.templateArgs = templateArgs class LookupKey: def __init__(self, data: List[Tuple[ASTNestedNameElement, Union[ASTTemplateParams, ASTTemplateIntroduction], str]]) -> None: self.data = data class Symbol: debug_indent = 0 debug_indent_string = " " debug_lookup = False debug_show_tree = False @staticmethod def debug_print(*args: Any) -> None: print(Symbol.debug_indent_string * Symbol.debug_indent, end="") print(*args) def _assert_invariants(self) -> None: if not self.parent: # parent == None means global scope, so declaration means a parent assert not self.identOrOp assert not self.templateParams assert not self.templateArgs assert not self.declaration assert not self.docname else: if self.declaration: assert self.docname def __setattr__(self, key: str, value: Any) -> None: if key == "children": assert False else: return super().__setattr__(key, value) def __init__(self, parent: "Symbol", identOrOp: Union[ASTIdentifier, ASTOperator], templateParams: Union[ASTTemplateParams, ASTTemplateIntroduction], templateArgs: Any, declaration: ASTDeclaration, docname: str) -> None: self.parent = parent # declarations in a single directive are linked together self.siblingAbove = None # type: Symbol self.siblingBelow = None # type: Symbol self.identOrOp = identOrOp self.templateParams = templateParams # template<templateParams> self.templateArgs = templateArgs # identifier<templateArgs> self.declaration = declaration self.docname = docname self.isRedeclaration = False self._assert_invariants() # Remember to modify Symbol.remove if modifications to the parent change. self._children = [] # type: List[Symbol] self._anonChildren = [] # type: List[Symbol] # note: _children includes _anonChildren if self.parent: self.parent._children.append(self) if self.declaration: self.declaration.symbol = self # Do symbol addition after self._children has been initialised. self._add_template_and_function_params() def _fill_empty(self, declaration: ASTDeclaration, docname: str) -> None: self._assert_invariants() assert not self.declaration assert not self.docname assert declaration assert docname self.declaration = declaration self.declaration.symbol = self self.docname = docname self._assert_invariants() # and symbol addition should be done as well self._add_template_and_function_params() def _add_template_and_function_params(self) -> None: if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("_add_template_and_function_params:") # Note: we may be called from _fill_empty, so the symbols we want # to add may actually already be present (as empty symbols). # add symbols for the template params if self.templateParams: for tp in self.templateParams.params: if not tp.get_identifier(): continue # only add a declaration if we our self are from a declaration if self.declaration: decl = ASTDeclaration('templateParam', None, None, None, tp) else: decl = None nne = ASTNestedNameElement(tp.get_identifier(), None) nn = ASTNestedName([nne], [False], rooted=False) self._add_symbols(nn, [], decl, self.docname) # add symbols for function parameters, if any if self.declaration is not None and self.declaration.function_params is not None: for fp in self.declaration.function_params: if fp.arg is None: continue nn = fp.arg.name if nn is None: continue # (comparing to the template params: we have checked that we are a declaration) decl = ASTDeclaration('functionParam', None, None, None, fp) assert not nn.rooted assert len(nn.names) == 1 self._add_symbols(nn, [], decl, self.docname) if Symbol.debug_lookup: Symbol.debug_indent -= 1 def remove(self) -> None: if self.parent is None: return assert self in self.parent._children self.parent._children.remove(self) self.parent = None def clear_doc(self, docname: str) -> None: newChildren = [] # type: List[Symbol] for sChild in self._children: sChild.clear_doc(docname) if sChild.declaration and sChild.docname == docname: sChild.declaration = None sChild.docname = None if sChild.siblingAbove is not None: sChild.siblingAbove.siblingBelow = sChild.siblingBelow if sChild.siblingBelow is not None: sChild.siblingBelow.siblingAbove = sChild.siblingAbove sChild.siblingAbove = None sChild.siblingBelow = None newChildren.append(sChild) self._children = newChildren def get_all_symbols(self) -> Iterator[Any]: yield self for sChild in self._children: for s in sChild.get_all_symbols(): yield s @property def children_recurse_anon(self) -> Generator["Symbol", None, None]: for c in self._children: yield c if not c.identOrOp.is_anon(): continue yield from c.children_recurse_anon def get_lookup_key(self) -> "LookupKey": # The pickle files for the environment and for each document are distinct. # The environment has all the symbols, but the documents has xrefs that # must know their scope. A lookup key is essentially a specification of # how to find a specific symbol. symbols = [] s = self while s.parent: symbols.append(s) s = s.parent symbols.reverse() key = [] for s in symbols: nne = ASTNestedNameElement(s.identOrOp, s.templateArgs) if s.declaration is not None: key.append((nne, s.templateParams, s.declaration.get_newest_id())) else: key.append((nne, s.templateParams, None)) return LookupKey(key) def get_full_nested_name(self) -> ASTNestedName: symbols = [] s = self while s.parent: symbols.append(s) s = s.parent symbols.reverse() names = [] templates = [] for s in symbols: names.append(ASTNestedNameElement(s.identOrOp, s.templateArgs)) templates.append(False) return ASTNestedName(names, templates, rooted=False) def _find_first_named_symbol(self, identOrOp: Union[ASTIdentifier, ASTOperator], templateParams: Any, templateArgs: ASTTemplateArgs, templateShorthand: bool, matchSelf: bool, recurseInAnon: bool, correctPrimaryTemplateArgs: bool ) -> "Symbol": if Symbol.debug_lookup: Symbol.debug_print("_find_first_named_symbol ->") res = self._find_named_symbols(identOrOp, templateParams, templateArgs, templateShorthand, matchSelf, recurseInAnon, correctPrimaryTemplateArgs, searchInSiblings=False) try: return next(res) except StopIteration: return None def _find_named_symbols(self, identOrOp: Union[ASTIdentifier, ASTOperator], templateParams: Any, templateArgs: ASTTemplateArgs, templateShorthand: bool, matchSelf: bool, recurseInAnon: bool, correctPrimaryTemplateArgs: bool, searchInSiblings: bool) -> Iterator["Symbol"]: if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("_find_named_symbols:") Symbol.debug_indent += 1 Symbol.debug_print("self:") print(self.to_string(Symbol.debug_indent + 1), end="") Symbol.debug_print("identOrOp: ", identOrOp) Symbol.debug_print("templateParams: ", templateParams) Symbol.debug_print("templateArgs: ", templateArgs) Symbol.debug_print("templateShorthand: ", templateShorthand) Symbol.debug_print("matchSelf: ", matchSelf) Symbol.debug_print("recurseInAnon: ", recurseInAnon) Symbol.debug_print("correctPrimaryTemplateAargs:", correctPrimaryTemplateArgs) Symbol.debug_print("searchInSiblings: ", searchInSiblings) def isSpecialization() -> bool: # the names of the template parameters must be given exactly as args # and params that are packs must in the args be the name expanded if len(templateParams.params) != len(templateArgs.args): return True # having no template params and no arguments is also a specialization if len(templateParams.params) == 0: return True for i in range(len(templateParams.params)): param = templateParams.params[i] arg = templateArgs.args[i] # TODO: doing this by string manipulation is probably not the most efficient paramName = str(param.name) argTxt = str(arg) isArgPackExpansion = argTxt.endswith('...') if param.isPack != isArgPackExpansion: return True argName = argTxt[:-3] if isArgPackExpansion else argTxt if paramName != argName: return True return False if correctPrimaryTemplateArgs: if templateParams is not None and templateArgs is not None: # If both are given, but it's not a specialization, then do lookup as if # there is no argument list. # For example: template<typename T> int A<T>::var; if not isSpecialization(): templateArgs = None def matches(s: "Symbol") -> bool: if s.identOrOp != identOrOp: return False if (s.templateParams is None) != (templateParams is None): if templateParams is not None: # we query with params, they must match params return False if not templateShorthand: # we don't query with params, and we do care about them return False if templateParams: # TODO: do better comparison if str(s.templateParams) != str(templateParams): return False if (s.templateArgs is None) != (templateArgs is None): return False if s.templateArgs: # TODO: do better comparison if str(s.templateArgs) != str(templateArgs): return False return True def candidates() -> Generator[Symbol, None, None]: s = self if Symbol.debug_lookup: Symbol.debug_print("searching in self:") print(s.to_string(Symbol.debug_indent + 1), end="") while True: if matchSelf: yield s if recurseInAnon: yield from s.children_recurse_anon else: yield from s._children if s.siblingAbove is None: break s = s.siblingAbove if Symbol.debug_lookup: Symbol.debug_print("searching in sibling:") print(s.to_string(Symbol.debug_indent + 1), end="") for s in candidates(): if Symbol.debug_lookup: Symbol.debug_print("candidate:") print(s.to_string(Symbol.debug_indent + 1), end="") if matches(s): if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("matches") Symbol.debug_indent -= 3 yield s if Symbol.debug_lookup: Symbol.debug_indent += 2 if Symbol.debug_lookup: Symbol.debug_indent -= 2 def _symbol_lookup(self, nestedName: ASTNestedName, templateDecls: List[Any], onMissingQualifiedSymbol: Callable[["Symbol", Union[ASTIdentifier, ASTOperator], Any, ASTTemplateArgs], "Symbol"], # NOQA strictTemplateParamArgLists: bool, ancestorLookupType: str, templateShorthand: bool, matchSelf: bool, recurseInAnon: bool, correctPrimaryTemplateArgs: bool, searchInSiblings: bool) -> SymbolLookupResult: # ancestorLookupType: if not None, specifies the target type of the lookup if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("_symbol_lookup:") Symbol.debug_indent += 1 Symbol.debug_print("self:") print(self.to_string(Symbol.debug_indent + 1), end="") Symbol.debug_print("nestedName: ", nestedName) Symbol.debug_print("templateDecls: ", templateDecls) Symbol.debug_print("strictTemplateParamArgLists:", strictTemplateParamArgLists) Symbol.debug_print("ancestorLookupType:", ancestorLookupType) Symbol.debug_print("templateShorthand: ", templateShorthand) Symbol.debug_print("matchSelf: ", matchSelf) Symbol.debug_print("recurseInAnon: ", recurseInAnon) Symbol.debug_print("correctPrimaryTemplateArgs: ", correctPrimaryTemplateArgs) Symbol.debug_print("searchInSiblings: ", searchInSiblings) if strictTemplateParamArgLists: # Each template argument list must have a template parameter list. # But to declare a template there must be an additional template parameter list. assert (nestedName.num_templates() == len(templateDecls) or nestedName.num_templates() + 1 == len(templateDecls)) else: assert len(templateDecls) <= nestedName.num_templates() + 1 names = nestedName.names # find the right starting point for lookup parentSymbol = self if nestedName.rooted: while parentSymbol.parent: parentSymbol = parentSymbol.parent if ancestorLookupType is not None: # walk up until we find the first identifier firstName = names[0] if not firstName.is_operator(): while parentSymbol.parent: if parentSymbol.find_identifier(firstName.identOrOp, matchSelf=matchSelf, recurseInAnon=recurseInAnon, searchInSiblings=searchInSiblings): # if we are in the scope of a constructor but wants to # reference the class we need to walk one extra up if (len(names) == 1 and ancestorLookupType == 'class' and matchSelf and parentSymbol.parent and parentSymbol.parent.identOrOp == firstName.identOrOp): pass else: break parentSymbol = parentSymbol.parent if Symbol.debug_lookup: Symbol.debug_print("starting point:") print(parentSymbol.to_string(Symbol.debug_indent + 1), end="") # and now the actual lookup iTemplateDecl = 0 for name in names[:-1]: identOrOp = name.identOrOp templateArgs = name.templateArgs if strictTemplateParamArgLists: # there must be a parameter list if templateArgs: assert iTemplateDecl < len(templateDecls) templateParams = templateDecls[iTemplateDecl] iTemplateDecl += 1 else: templateParams = None else: # take the next template parameter list if there is one # otherwise it's ok if templateArgs and iTemplateDecl < len(templateDecls): templateParams = templateDecls[iTemplateDecl] iTemplateDecl += 1 else: templateParams = None symbol = parentSymbol._find_first_named_symbol( identOrOp, templateParams, templateArgs, templateShorthand=templateShorthand, matchSelf=matchSelf, recurseInAnon=recurseInAnon, correctPrimaryTemplateArgs=correctPrimaryTemplateArgs) if symbol is None: symbol = onMissingQualifiedSymbol(parentSymbol, identOrOp, templateParams, templateArgs) if symbol is None: if Symbol.debug_lookup: Symbol.debug_indent -= 2 return None # We have now matched part of a nested name, and need to match more # so even if we should matchSelf before, we definitely shouldn't # even more. (see also issue #2666) matchSelf = False parentSymbol = symbol if Symbol.debug_lookup: Symbol.debug_print("handle last name from:") print(parentSymbol.to_string(Symbol.debug_indent + 1), end="") # handle the last name name = names[-1] identOrOp = name.identOrOp templateArgs = name.templateArgs if iTemplateDecl < len(templateDecls): assert iTemplateDecl + 1 == len(templateDecls) templateParams = templateDecls[iTemplateDecl] else: assert iTemplateDecl == len(templateDecls) templateParams = None symbols = parentSymbol._find_named_symbols( identOrOp, templateParams, templateArgs, templateShorthand=templateShorthand, matchSelf=matchSelf, recurseInAnon=recurseInAnon, correctPrimaryTemplateArgs=False, searchInSiblings=searchInSiblings) if Symbol.debug_lookup: symbols = list(symbols) # type: ignore Symbol.debug_indent -= 2 return SymbolLookupResult(symbols, parentSymbol, identOrOp, templateParams, templateArgs) def _add_symbols(self, nestedName: ASTNestedName, templateDecls: List[Any], declaration: ASTDeclaration, docname: str) -> "Symbol": # Used for adding a whole path of symbols, where the last may or may not # be an actual declaration. if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("_add_symbols:") Symbol.debug_indent += 1 Symbol.debug_print("tdecls:", templateDecls) Symbol.debug_print("nn: ", nestedName) Symbol.debug_print("decl: ", declaration) Symbol.debug_print("doc: ", docname) def onMissingQualifiedSymbol(parentSymbol: "Symbol", identOrOp: Union[ASTIdentifier, ASTOperator], templateParams: Any, templateArgs: ASTTemplateArgs ) -> "Symbol": if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("_add_symbols, onMissingQualifiedSymbol:") Symbol.debug_indent += 1 Symbol.debug_print("templateParams:", templateParams) Symbol.debug_print("identOrOp: ", identOrOp) Symbol.debug_print("templateARgs: ", templateArgs) Symbol.debug_indent -= 2 return Symbol(parent=parentSymbol, identOrOp=identOrOp, templateParams=templateParams, templateArgs=templateArgs, declaration=None, docname=None) lookupResult = self._symbol_lookup(nestedName, templateDecls, onMissingQualifiedSymbol, strictTemplateParamArgLists=True, ancestorLookupType=None, templateShorthand=False, matchSelf=False, recurseInAnon=False, correctPrimaryTemplateArgs=True, searchInSiblings=False) assert lookupResult is not None # we create symbols all the way, so that can't happen symbols = list(lookupResult.symbols) if len(symbols) == 0: if Symbol.debug_lookup: Symbol.debug_print("_add_symbols, result, no symbol:") Symbol.debug_indent += 1 Symbol.debug_print("templateParams:", lookupResult.templateParams) Symbol.debug_print("identOrOp: ", lookupResult.identOrOp) Symbol.debug_print("templateArgs: ", lookupResult.templateArgs) Symbol.debug_print("declaration: ", declaration) Symbol.debug_print("docname: ", docname) Symbol.debug_indent -= 1 symbol = Symbol(parent=lookupResult.parentSymbol, identOrOp=lookupResult.identOrOp, templateParams=lookupResult.templateParams, templateArgs=lookupResult.templateArgs, declaration=declaration, docname=docname) if Symbol.debug_lookup: Symbol.debug_indent -= 2 return symbol if Symbol.debug_lookup: Symbol.debug_print("_add_symbols, result, symbols:") Symbol.debug_indent += 1 Symbol.debug_print("number symbols:", len(symbols)) Symbol.debug_indent -= 1 if not declaration: if Symbol.debug_lookup: Symbol.debug_print("no delcaration") Symbol.debug_indent -= 2 # good, just a scope creation # TODO: what if we have more than one symbol? return symbols[0] noDecl = [] withDecl = [] dupDecl = [] for s in symbols: if s.declaration is None: noDecl.append(s) elif s.isRedeclaration: dupDecl.append(s) else: withDecl.append(s) if Symbol.debug_lookup: Symbol.debug_print("#noDecl: ", len(noDecl)) Symbol.debug_print("#withDecl:", len(withDecl)) Symbol.debug_print("#dupDecl: ", len(dupDecl)) # With partial builds we may start with a large symbol tree stripped of declarations. # Essentially any combination of noDecl, withDecl, and dupDecls seems possible. # TODO: make partial builds fully work. What should happen when the primary symbol gets # deleted, and other duplicates exist? The full document should probably be rebuild. # First check if one of those with a declaration matches. # If it's a function, we need to compare IDs, # otherwise there should be only one symbol with a declaration. def makeCandSymbol() -> "Symbol": if Symbol.debug_lookup: Symbol.debug_print("begin: creating candidate symbol") symbol = Symbol(parent=lookupResult.parentSymbol, identOrOp=lookupResult.identOrOp, templateParams=lookupResult.templateParams, templateArgs=lookupResult.templateArgs, declaration=declaration, docname=docname) if Symbol.debug_lookup: Symbol.debug_print("end: creating candidate symbol") return symbol if len(withDecl) == 0: candSymbol = None else: candSymbol = makeCandSymbol() def handleDuplicateDeclaration(symbol: "Symbol", candSymbol: "Symbol") -> None: if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("redeclaration") Symbol.debug_indent -= 1 Symbol.debug_indent -= 2 # Redeclaration of the same symbol. # Let the new one be there, but raise an error to the client # so it can use the real symbol as subscope. # This will probably result in a duplicate id warning. candSymbol.isRedeclaration = True raise _DuplicateSymbolError(symbol, declaration) if declaration.objectType != "function": assert len(withDecl) <= 1 handleDuplicateDeclaration(withDecl[0], candSymbol) # (not reachable) # a function, so compare IDs candId = declaration.get_newest_id() if Symbol.debug_lookup: Symbol.debug_print("candId:", candId) for symbol in withDecl: oldId = symbol.declaration.get_newest_id() if Symbol.debug_lookup: Symbol.debug_print("oldId: ", oldId) if candId == oldId: handleDuplicateDeclaration(symbol, candSymbol) # (not reachable) # no candidate symbol found with matching ID # if there is an empty symbol, fill that one if len(noDecl) == 0: if Symbol.debug_lookup: Symbol.debug_print("no match, no empty, candSybmol is not None?:", candSymbol is not None) # NOQA Symbol.debug_indent -= 2 if candSymbol is not None: return candSymbol else: return makeCandSymbol() else: if Symbol.debug_lookup: Symbol.debug_print("no match, but fill an empty declaration, candSybmol is not None?:", candSymbol is not None) # NOQA Symbol.debug_indent -= 2 if candSymbol is not None: candSymbol.remove() # assert len(noDecl) == 1 # TODO: enable assertion when we at some point find out how to do cleanup # for now, just take the first one, it should work fine ... right? symbol = noDecl[0] # If someone first opened the scope, and then later # declares it, e.g, # .. namespace:: Test # .. namespace:: nullptr # .. class:: Test symbol._fill_empty(declaration, docname) return symbol def merge_with(self, other: "Symbol", docnames: List[str], env: "BuildEnvironment") -> None: if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("merge_with:") assert other is not None def unconditionalAdd(self, otherChild): # TODO: hmm, should we prune by docnames? self._children.append(otherChild) otherChild.parent = self otherChild._assert_invariants() if Symbol.debug_lookup: Symbol.debug_indent += 1 for otherChild in other._children: if Symbol.debug_lookup: Symbol.debug_print("otherChild:\n", otherChild.to_string(Symbol.debug_indent)) Symbol.debug_indent += 1 if otherChild.isRedeclaration: unconditionalAdd(self, otherChild) if Symbol.debug_lookup: Symbol.debug_print("isRedeclaration") Symbol.debug_indent -= 1 continue candiateIter = self._find_named_symbols( identOrOp=otherChild.identOrOp, templateParams=otherChild.templateParams, templateArgs=otherChild.templateArgs, templateShorthand=False, matchSelf=False, recurseInAnon=False, correctPrimaryTemplateArgs=False, searchInSiblings=False) candidates = list(candiateIter) if Symbol.debug_lookup: Symbol.debug_print("raw candidate symbols:", len(candidates)) symbols = [s for s in candidates if not s.isRedeclaration] if Symbol.debug_lookup: Symbol.debug_print("non-duplicate candidate symbols:", len(symbols)) if len(symbols) == 0: unconditionalAdd(self, otherChild) if Symbol.debug_lookup: Symbol.debug_indent -= 1 continue ourChild = None if otherChild.declaration is None: if Symbol.debug_lookup: Symbol.debug_print("no declaration in other child") ourChild = symbols[0] else: queryId = otherChild.declaration.get_newest_id() if Symbol.debug_lookup: Symbol.debug_print("queryId: ", queryId) for symbol in symbols: if symbol.declaration is None: if Symbol.debug_lookup: Symbol.debug_print("empty candidate") # if in the end we have non matching, but have an empty one, # then just continue with that ourChild = symbol continue candId = symbol.declaration.get_newest_id() if Symbol.debug_lookup: Symbol.debug_print("candidate:", candId) if candId == queryId: ourChild = symbol break if Symbol.debug_lookup: Symbol.debug_indent -= 1 if ourChild is None: unconditionalAdd(self, otherChild) continue if otherChild.declaration and otherChild.docname in docnames: if not ourChild.declaration: ourChild._fill_empty(otherChild.declaration, otherChild.docname) elif ourChild.docname != otherChild.docname: name = str(ourChild.declaration) msg = __("Duplicate declaration, also defined in '%s'.\n" "Declaration is '%s'.") msg = msg % (ourChild.docname, name) logger.warning(msg, location=otherChild.docname) else: # Both have declarations, and in the same docname. # This can apparently happen, it should be safe to # just ignore it, right? # Hmm, only on duplicate declarations, right? msg = "Internal C++ domain error during symbol merging.\n" msg += "ourChild:\n" + ourChild.to_string(1) msg += "\notherChild:\n" + otherChild.to_string(1) logger.warning(msg, location=otherChild.docname) ourChild.merge_with(otherChild, docnames, env) if Symbol.debug_lookup: Symbol.debug_indent -= 2 def add_name(self, nestedName: ASTNestedName, templatePrefix: ASTTemplateDeclarationPrefix = None) -> "Symbol": if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("add_name:") if templatePrefix: templateDecls = templatePrefix.templates else: templateDecls = [] res = self._add_symbols(nestedName, templateDecls, declaration=None, docname=None) if Symbol.debug_lookup: Symbol.debug_indent -= 1 return res def add_declaration(self, declaration: ASTDeclaration, docname: str) -> "Symbol": if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("add_declaration:") assert declaration assert docname nestedName = declaration.name if declaration.templatePrefix: templateDecls = declaration.templatePrefix.templates else: templateDecls = [] res = self._add_symbols(nestedName, templateDecls, declaration, docname) if Symbol.debug_lookup: Symbol.debug_indent -= 1 return res def find_identifier(self, identOrOp: Union[ASTIdentifier, ASTOperator], matchSelf: bool, recurseInAnon: bool, searchInSiblings: bool ) -> "Symbol": if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("find_identifier:") Symbol.debug_indent += 1 Symbol.debug_print("identOrOp: ", identOrOp) Symbol.debug_print("matchSelf: ", matchSelf) Symbol.debug_print("recurseInAnon: ", recurseInAnon) Symbol.debug_print("searchInSiblings:", searchInSiblings) print(self.to_string(Symbol.debug_indent + 1), end="") Symbol.debug_indent -= 2 current = self while current is not None: if Symbol.debug_lookup: Symbol.debug_indent += 2 Symbol.debug_print("trying:") print(current.to_string(Symbol.debug_indent + 1), end="") Symbol.debug_indent -= 2 if matchSelf and current.identOrOp == identOrOp: return current children = current.children_recurse_anon if recurseInAnon else current._children for s in children: if s.identOrOp == identOrOp: return s if not searchInSiblings: break current = current.siblingAbove return None def direct_lookup(self, key: "LookupKey") -> "Symbol": if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("direct_lookup:") Symbol.debug_indent += 1 s = self for name, templateParams, id_ in key.data: if id_ is not None: res = None for cand in s._children: if cand.declaration is None: continue if cand.declaration.get_newest_id() == id_: res = cand break s = res else: identOrOp = name.identOrOp templateArgs = name.templateArgs s = s._find_first_named_symbol(identOrOp, templateParams, templateArgs, templateShorthand=False, matchSelf=False, recurseInAnon=False, correctPrimaryTemplateArgs=False) if Symbol.debug_lookup: Symbol.debug_print("name: ", name) Symbol.debug_print("templateParams:", templateParams) Symbol.debug_print("id: ", id_) if s is not None: print(s.to_string(Symbol.debug_indent + 1), end="") else: Symbol.debug_print("not found") if s is None: if Symbol.debug_lookup: Symbol.debug_indent -= 2 return None if Symbol.debug_lookup: Symbol.debug_indent -= 2 return s def find_name(self, nestedName: ASTNestedName, templateDecls: List[Any], typ: str, templateShorthand: bool, matchSelf: bool, recurseInAnon: bool, searchInSiblings: bool) -> Tuple[List["Symbol"], str]: # templateShorthand: missing template parameter lists for templates is ok # If the first component is None, # then the second component _may_ be a string explaining why. if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("find_name:") Symbol.debug_indent += 1 Symbol.debug_print("self:") print(self.to_string(Symbol.debug_indent + 1), end="") Symbol.debug_print("nestedName: ", nestedName) Symbol.debug_print("templateDecls: ", templateDecls) Symbol.debug_print("typ: ", typ) Symbol.debug_print("templateShorthand:", templateShorthand) Symbol.debug_print("matchSelf: ", matchSelf) Symbol.debug_print("recurseInAnon: ", recurseInAnon) Symbol.debug_print("searchInSiblings: ", searchInSiblings) class QualifiedSymbolIsTemplateParam(Exception): pass def onMissingQualifiedSymbol(parentSymbol: "Symbol", identOrOp: Union[ASTIdentifier, ASTOperator], templateParams: Any, templateArgs: ASTTemplateArgs) -> "Symbol": # TODO: Maybe search without template args? # Though, the correctPrimaryTemplateArgs does # that for primary templates. # Is there another case where it would be good? if parentSymbol.declaration is not None: if parentSymbol.declaration.objectType == 'templateParam': raise QualifiedSymbolIsTemplateParam() return None try: lookupResult = self._symbol_lookup(nestedName, templateDecls, onMissingQualifiedSymbol, strictTemplateParamArgLists=False, ancestorLookupType=typ, templateShorthand=templateShorthand, matchSelf=matchSelf, recurseInAnon=recurseInAnon, correctPrimaryTemplateArgs=False, searchInSiblings=searchInSiblings) except QualifiedSymbolIsTemplateParam: return None, "templateParamInQualified" if lookupResult is None: # if it was a part of the qualification that could not be found if Symbol.debug_lookup: Symbol.debug_indent -= 2 return None, None res = list(lookupResult.symbols) if len(res) != 0: if Symbol.debug_lookup: Symbol.debug_indent -= 2 return res, None if lookupResult.parentSymbol.declaration is not None: if lookupResult.parentSymbol.declaration.objectType == 'templateParam': return None, "templateParamInQualified" # try without template params and args symbol = lookupResult.parentSymbol._find_first_named_symbol( lookupResult.identOrOp, None, None, templateShorthand=templateShorthand, matchSelf=matchSelf, recurseInAnon=recurseInAnon, correctPrimaryTemplateArgs=False) if Symbol.debug_lookup: Symbol.debug_indent -= 2 if symbol is not None: return [symbol], None else: return None, None def find_declaration(self, declaration: ASTDeclaration, typ: str, templateShorthand: bool, matchSelf: bool, recurseInAnon: bool) -> "Symbol": # templateShorthand: missing template parameter lists for templates is ok if Symbol.debug_lookup: Symbol.debug_indent += 1 Symbol.debug_print("find_declaration:") nestedName = declaration.name if declaration.templatePrefix: templateDecls = declaration.templatePrefix.templates else: templateDecls = [] def onMissingQualifiedSymbol(parentSymbol: "Symbol", identOrOp: Union[ASTIdentifier, ASTOperator], templateParams: Any, templateArgs: ASTTemplateArgs) -> "Symbol": return None lookupResult = self._symbol_lookup(nestedName, templateDecls, onMissingQualifiedSymbol, strictTemplateParamArgLists=False, ancestorLookupType=typ, templateShorthand=templateShorthand, matchSelf=matchSelf, recurseInAnon=recurseInAnon, correctPrimaryTemplateArgs=False, searchInSiblings=False) if Symbol.debug_lookup: Symbol.debug_indent -= 1 if lookupResult is None: return None symbols = list(lookupResult.symbols) if len(symbols) == 0: return None querySymbol = Symbol(parent=lookupResult.parentSymbol, identOrOp=lookupResult.identOrOp, templateParams=lookupResult.templateParams, templateArgs=lookupResult.templateArgs, declaration=declaration, docname='fakeDocnameForQuery') queryId = declaration.get_newest_id() for symbol in symbols: if symbol.declaration is None: continue candId = symbol.declaration.get_newest_id() if candId == queryId: querySymbol.remove() return symbol querySymbol.remove() return None def to_string(self, indent: int) -> str: res = [Symbol.debug_indent_string * indent] if not self.parent: res.append('::') else: if self.templateParams: res.append(str(self.templateParams)) res.append('\n') res.append(Symbol.debug_indent_string * indent) if self.identOrOp: res.append(str(self.identOrOp)) else: res.append(str(self.declaration)) if self.templateArgs: res.append(str(self.templateArgs)) if self.declaration: res.append(": ") if self.isRedeclaration: res.append('!!duplicate!! ') res.append(str(self.declaration)) if self.docname: res.append('\t(') res.append(self.docname) res.append(')') res.append('\n') return ''.join(res) def dump(self, indent: int) -> str: res = [self.to_string(indent)] for c in self._children: res.append(c.dump(indent + 1)) return ''.join(res) class DefinitionParser(BaseParser): # those without signedness and size modifiers # see https://en.cppreference.com/w/cpp/language/types _simple_fundemental_types = ( 'void', 'bool', 'char', 'wchar_t', 'char16_t', 'char32_t', 'int', 'float', 'double', 'auto' ) _prefix_keys = ('class', 'struct', 'enum', 'union', 'typename') @property def language(self) -> str: return 'C++' @property def id_attributes(self): return self.config.cpp_id_attributes @property def paren_attributes(self): return self.config.cpp_paren_attributes def _parse_string(self) -> str: if self.current_char != '"': return None startPos = self.pos self.pos += 1 escape = False while True: if self.eof: self.fail("Unexpected end during inside string.") elif self.current_char == '"' and not escape: self.pos += 1 break elif self.current_char == '\\': escape = True else: escape = False self.pos += 1 return self.definition[startPos:self.pos] def _parse_literal(self) -> ASTLiteral: # -> integer-literal # | character-literal # | floating-literal # | string-literal # | boolean-literal -> "false" | "true" # | pointer-literal -> "nullptr" # | user-defined-literal self.skip_ws() if self.skip_word('nullptr'): return ASTPointerLiteral() if self.skip_word('true'): return ASTBooleanLiteral(True) if self.skip_word('false'): return ASTBooleanLiteral(False) for regex in [float_literal_re, binary_literal_re, hex_literal_re, integer_literal_re, octal_literal_re]: pos = self.pos if self.match(regex): while self.current_char in 'uUlLfF': self.pos += 1 return ASTNumberLiteral(self.definition[pos:self.pos]) string = self._parse_string() if string is not None: return ASTStringLiteral(string) # character-literal if self.match(char_literal_re): prefix = self.last_match.group(1) # may be None when no prefix data = self.last_match.group(2) try: return ASTCharLiteral(prefix, data) except UnicodeDecodeError as e: self.fail("Can not handle character literal. Internal error was: %s" % e) except UnsupportedMultiCharacterCharLiteral: self.fail("Can not handle character literal" " resulting in multiple decoded characters.") # TODO: user-defined lit return None def _parse_fold_or_paren_expression(self) -> ASTExpression: # "(" expression ")" # fold-expression # -> ( cast-expression fold-operator ... ) # | ( ... fold-operator cast-expression ) # | ( cast-expression fold-operator ... fold-operator cast-expression if self.current_char != '(': return None self.pos += 1 self.skip_ws() if self.skip_string_and_ws("..."): # ( ... fold-operator cast-expression ) if not self.match(_fold_operator_re): self.fail("Expected fold operator after '...' in fold expression.") op = self.matched_text rightExpr = self._parse_cast_expression() if not self.skip_string(')'): self.fail("Expected ')' in end of fold expression.") return ASTFoldExpr(None, op, rightExpr) # try first parsing a unary right fold, or a binary fold pos = self.pos try: self.skip_ws() leftExpr = self._parse_cast_expression() self.skip_ws() if not self.match(_fold_operator_re): self.fail("Expected fold operator after left expression in fold expression.") op = self.matched_text self.skip_ws() if not self.skip_string_and_ws('...'): self.fail("Expected '...' after fold operator in fold expression.") except DefinitionError as eFold: self.pos = pos # fall back to a paren expression try: res = self._parse_expression() self.skip_ws() if not self.skip_string(')'): self.fail("Expected ')' in end of parenthesized expression.") except DefinitionError as eExpr: raise self._make_multi_error([ (eFold, "If fold expression"), (eExpr, "If parenthesized expression") ], "Error in fold expression or parenthesized expression.") return ASTParenExpr(res) # now it definitely is a fold expression if self.skip_string(')'): return ASTFoldExpr(leftExpr, op, None) if not self.match(_fold_operator_re): self.fail("Expected fold operator or ')' after '...' in fold expression.") if op != self.matched_text: self.fail("Operators are different in binary fold: '%s' and '%s'." % (op, self.matched_text)) rightExpr = self._parse_cast_expression() self.skip_ws() if not self.skip_string(')'): self.fail("Expected ')' to end binary fold expression.") return ASTFoldExpr(leftExpr, op, rightExpr) def _parse_primary_expression(self) -> ASTExpression: # literal # "this" # lambda-expression # "(" expression ")" # fold-expression # id-expression -> we parse this with _parse_nested_name self.skip_ws() res = self._parse_literal() # type: ASTExpression if res is not None: return res self.skip_ws() if self.skip_word("this"): return ASTThisLiteral() # TODO: try lambda expression res = self._parse_fold_or_paren_expression() if res is not None: return res nn = self._parse_nested_name() if nn is not None: return ASTIdExpression(nn) return None def _parse_initializer_list(self, name: str, open: str, close: str ) -> Tuple[List[Union[ASTExpression, ASTBracedInitList]], bool]: # Parse open and close with the actual initializer-list inbetween # -> initializer-clause '...'[opt] # | initializer-list ',' initializer-clause '...'[opt] self.skip_ws() if not self.skip_string_and_ws(open): return None, None if self.skip_string(close): return [], False exprs = [] # type: List[Union[ASTExpression, ASTBracedInitList]] trailingComma = False while True: self.skip_ws() expr = self._parse_initializer_clause() self.skip_ws() if self.skip_string('...'): exprs.append(ASTPackExpansionExpr(expr)) else: exprs.append(expr) self.skip_ws() if self.skip_string(close): break if not self.skip_string_and_ws(','): self.fail("Error in %s, expected ',' or '%s'." % (name, close)) if self.current_char == close and close == '}': self.pos += 1 trailingComma = True break return exprs, trailingComma def _parse_paren_expression_list(self) -> ASTParenExprList: # -> '(' expression-list ')' # though, we relax it to also allow empty parens # as it's needed in some cases # # expression-list # -> initializer-list exprs, trailingComma = self._parse_initializer_list("parenthesized expression-list", '(', ')') if exprs is None: return None return ASTParenExprList(exprs) def _parse_initializer_clause(self) -> Union[ASTExpression, ASTBracedInitList]: bracedInitList = self._parse_braced_init_list() if bracedInitList is not None: return bracedInitList return self._parse_assignment_expression(inTemplate=False) def _parse_braced_init_list(self) -> ASTBracedInitList: # -> '{' initializer-list ','[opt] '}' # | '{' '}' exprs, trailingComma = self._parse_initializer_list("braced-init-list", '{', '}') if exprs is None: return None return ASTBracedInitList(exprs, trailingComma) def _parse_expression_list_or_braced_init_list( self ) -> Union[ASTParenExprList, ASTBracedInitList]: paren = self._parse_paren_expression_list() if paren is not None: return paren return self._parse_braced_init_list() def _parse_postfix_expression(self) -> ASTPostfixExpr: # -> primary # | postfix "[" expression "]" # | postfix "[" braced-init-list [opt] "]" # | postfix "(" expression-list [opt] ")" # | postfix "." "template" [opt] id-expression # | postfix "->" "template" [opt] id-expression # | postfix "." pseudo-destructor-name # | postfix "->" pseudo-destructor-name # | postfix "++" # | postfix "--" # | simple-type-specifier "(" expression-list [opt] ")" # | simple-type-specifier braced-init-list # | typename-specifier "(" expression-list [opt] ")" # | typename-specifier braced-init-list # | "dynamic_cast" "<" type-id ">" "(" expression ")" # | "static_cast" "<" type-id ">" "(" expression ")" # | "reinterpret_cast" "<" type-id ">" "(" expression ")" # | "const_cast" "<" type-id ">" "(" expression ")" # | "typeid" "(" expression ")" # | "typeid" "(" type-id ")" prefixType = None prefix = None # type: Any self.skip_ws() cast = None for c in _id_explicit_cast: if self.skip_word_and_ws(c): cast = c break if cast is not None: prefixType = "cast" if not self.skip_string("<"): self.fail("Expected '<' afer '%s'." % cast) typ = self._parse_type(False) self.skip_ws() if not self.skip_string_and_ws(">"): self.fail("Expected '>' after type in '%s'." % cast) if not self.skip_string("("): self.fail("Expected '(' in '%s'." % cast) def parser() -> ASTExpression: return self._parse_expression() expr = self._parse_expression_fallback([')'], parser) self.skip_ws() if not self.skip_string(")"): self.fail("Expected ')' to end '%s'." % cast) prefix = ASTExplicitCast(cast, typ, expr) elif self.skip_word_and_ws("typeid"): prefixType = "typeid" if not self.skip_string_and_ws('('): self.fail("Expected '(' after 'typeid'.") pos = self.pos try: typ = self._parse_type(False) prefix = ASTTypeId(typ, isType=True) if not self.skip_string(')'): self.fail("Expected ')' to end 'typeid' of type.") except DefinitionError as eType: self.pos = pos try: def parser() -> ASTExpression: return self._parse_expression() expr = self._parse_expression_fallback([')'], parser) prefix = ASTTypeId(expr, isType=False) if not self.skip_string(')'): self.fail("Expected ')' to end 'typeid' of expression.") except DefinitionError as eExpr: self.pos = pos header = "Error in 'typeid(...)'." header += " Expected type or expression." errors = [] errors.append((eType, "If type")) errors.append((eExpr, "If expression")) raise self._make_multi_error(errors, header) else: # a primary expression or a type pos = self.pos try: prefix = self._parse_primary_expression() prefixType = 'expr' except DefinitionError as eOuter: self.pos = pos try: # we are potentially casting, so save parens for us # TODO: hmm, would we need to try both with operatorCast and with None? prefix = self._parse_type(False, 'operatorCast') prefixType = 'typeOperatorCast' # | simple-type-specifier "(" expression-list [opt] ")" # | simple-type-specifier braced-init-list # | typename-specifier "(" expression-list [opt] ")" # | typename-specifier braced-init-list self.skip_ws() if self.current_char != '(' and self.current_char != '{': self.fail("Expecting '(' or '{' after type in cast expression.") except DefinitionError as eInner: self.pos = pos header = "Error in postfix expression," header += " expected primary expression or type." errors = [] errors.append((eOuter, "If primary expression")) errors.append((eInner, "If type")) raise self._make_multi_error(errors, header) # and now parse postfixes postFixes = [] # type: List[ASTPostfixOp] while True: self.skip_ws() if prefixType in ['expr', 'cast', 'typeid']: if self.skip_string_and_ws('['): expr = self._parse_expression() self.skip_ws() if not self.skip_string(']'): self.fail("Expected ']' in end of postfix expression.") postFixes.append(ASTPostfixArray(expr)) continue if self.skip_string('.'): if self.skip_string('*'): # don't steal the dot self.pos -= 2 elif self.skip_string('..'): # don't steal the dot self.pos -= 3 else: name = self._parse_nested_name() postFixes.append(ASTPostfixMember(name)) continue if self.skip_string('->'): if self.skip_string('*'): # don't steal the arrow self.pos -= 3 else: name = self._parse_nested_name() postFixes.append(ASTPostfixMemberOfPointer(name)) continue if self.skip_string('++'): postFixes.append(ASTPostfixInc()) continue if self.skip_string('--'): postFixes.append(ASTPostfixDec()) continue lst = self._parse_expression_list_or_braced_init_list() if lst is not None: postFixes.append(ASTPostfixCallExpr(lst)) continue break return ASTPostfixExpr(prefix, postFixes) def _parse_unary_expression(self) -> ASTExpression: # -> postfix # | "++" cast # | "--" cast # | unary-operator cast -> (* | & | + | - | ! | ~) cast # The rest: # | "sizeof" unary # | "sizeof" "(" type-id ")" # | "sizeof" "..." "(" identifier ")" # | "alignof" "(" type-id ")" # | noexcept-expression -> noexcept "(" expression ")" # | new-expression # | delete-expression self.skip_ws() for op in _expression_unary_ops: # TODO: hmm, should we be able to backtrack here? if op[0] in 'cn': res = self.skip_word(op) else: res = self.skip_string(op) if res: expr = self._parse_cast_expression() return ASTUnaryOpExpr(op, expr) if self.skip_word_and_ws('sizeof'): if self.skip_string_and_ws('...'): if not self.skip_string_and_ws('('): self.fail("Expecting '(' after 'sizeof...'.") if not self.match(identifier_re): self.fail("Expecting identifier for 'sizeof...'.") ident = ASTIdentifier(self.matched_text) self.skip_ws() if not self.skip_string(")"): self.fail("Expecting ')' to end 'sizeof...'.") return ASTSizeofParamPack(ident) if self.skip_string_and_ws('('): typ = self._parse_type(named=False) self.skip_ws() if not self.skip_string(')'): self.fail("Expecting ')' to end 'sizeof'.") return ASTSizeofType(typ) expr = self._parse_unary_expression() return ASTSizeofExpr(expr) if self.skip_word_and_ws('alignof'): if not self.skip_string_and_ws('('): self.fail("Expecting '(' after 'alignof'.") typ = self._parse_type(named=False) self.skip_ws() if not self.skip_string(')'): self.fail("Expecting ')' to end 'alignof'.") return ASTAlignofExpr(typ) if self.skip_word_and_ws('noexcept'): if not self.skip_string_and_ws('('): self.fail("Expecting '(' after 'noexcept'.") expr = self._parse_expression() self.skip_ws() if not self.skip_string(')'): self.fail("Expecting ')' to end 'noexcept'.") return ASTNoexceptExpr(expr) # new-expression pos = self.pos rooted = self.skip_string('::') self.skip_ws() if not self.skip_word_and_ws('new'): self.pos = pos else: # new-placement[opt] new-type-id new-initializer[opt] # new-placement[opt] ( type-id ) new-initializer[opt] isNewTypeId = True if self.skip_string_and_ws('('): # either this is a new-placement or it's the second production # without placement, and it's actually the ( type-id ) part self.fail("Sorry, neither new-placement nor parenthesised type-id " "in new-epression is supported yet.") # set isNewTypeId = False if it's (type-id) if isNewTypeId: declSpecs = self._parse_decl_specs(outer=None) decl = self._parse_declarator(named=False, paramMode="new") else: self.fail("Sorry, parenthesised type-id in new expression not yet supported.") lst = self._parse_expression_list_or_braced_init_list() return ASTNewExpr(rooted, isNewTypeId, ASTType(declSpecs, decl), lst) # delete-expression pos = self.pos rooted = self.skip_string('::') self.skip_ws() if not self.skip_word_and_ws('delete'): self.pos = pos else: array = self.skip_string_and_ws('[') if array and not self.skip_string_and_ws(']'): self.fail("Expected ']' in array delete-expression.") expr = self._parse_cast_expression() return ASTDeleteExpr(rooted, array, expr) return self._parse_postfix_expression() def _parse_cast_expression(self) -> ASTExpression: # -> unary | "(" type-id ")" cast pos = self.pos self.skip_ws() if self.skip_string('('): try: typ = self._parse_type(False) if not self.skip_string(')'): self.fail("Expected ')' in cast expression.") expr = self._parse_cast_expression() return ASTCastExpr(typ, expr) except DefinitionError as exCast: self.pos = pos try: return self._parse_unary_expression() except DefinitionError as exUnary: errs = [] errs.append((exCast, "If type cast expression")) errs.append((exUnary, "If unary expression")) raise self._make_multi_error(errs, "Error in cast expression.") else: return self._parse_unary_expression() def _parse_logical_or_expression(self, inTemplate: bool) -> ASTExpression: # logical-or = logical-and || # logical-and = inclusive-or && # inclusive-or = exclusive-or | # exclusive-or = and ^ # and = equality & # equality = relational ==, != # relational = shift <, >, <=, >= # shift = additive <<, >> # additive = multiplicative +, - # multiplicative = pm *, /, % # pm = cast .*, ->* def _parse_bin_op_expr(self: DefinitionParser, opId: int, inTemplate: bool) -> ASTExpression: if opId + 1 == len(_expression_bin_ops): def parser(inTemplate: bool) -> ASTExpression: return self._parse_cast_expression() else: def parser(inTemplate: bool) -> ASTExpression: return _parse_bin_op_expr(self, opId + 1, inTemplate=inTemplate) exprs = [] ops = [] exprs.append(parser(inTemplate=inTemplate)) while True: self.skip_ws() if inTemplate and self.current_char == '>': break pos = self.pos oneMore = False for op in _expression_bin_ops[opId]: if op[0] in 'abcnox': if not self.skip_word(op): continue else: if not self.skip_string(op): continue if op == '&' and self.current_char == '&': # don't split the && 'token' self.pos -= 1 # and btw. && has lower precedence, so we are done break try: expr = parser(inTemplate=inTemplate) exprs.append(expr) ops.append(op) oneMore = True break except DefinitionError: self.pos = pos if not oneMore: break return ASTBinOpExpr(exprs, ops) return _parse_bin_op_expr(self, 0, inTemplate=inTemplate) def _parse_conditional_expression_tail(self, orExprHead: Any) -> None: # -> "?" expression ":" assignment-expression return None def _parse_assignment_expression(self, inTemplate: bool) -> ASTExpression: # -> conditional-expression # | logical-or-expression assignment-operator initializer-clause # | throw-expression # TODO: parse throw-expression: "throw" assignment-expression [opt] # if not a throw expression, then: # -> conditional-expression -> # logical-or-expression # | logical-or-expression "?" expression ":" assignment-expression # | logical-or-expression assignment-operator initializer-clause exprs = [] # type: List[Union[ASTExpression, ASTBracedInitList]] ops = [] orExpr = self._parse_logical_or_expression(inTemplate=inTemplate) exprs.append(orExpr) # TODO: handle ternary with _parse_conditional_expression_tail while True: oneMore = False self.skip_ws() for op in _expression_assignment_ops: if op[0] in 'anox': if not self.skip_word(op): continue else: if not self.skip_string(op): continue expr = self._parse_initializer_clause() exprs.append(expr) ops.append(op) oneMore = True if not oneMore: break if len(ops) == 0: return orExpr else: return ASTAssignmentExpr(exprs, ops) def _parse_constant_expression(self, inTemplate: bool) -> ASTExpression: # -> conditional-expression orExpr = self._parse_logical_or_expression(inTemplate=inTemplate) # TODO: use _parse_conditional_expression_tail return orExpr def _parse_expression(self) -> ASTExpression: # -> assignment-expression # | expression "," assignment-expresion exprs = [self._parse_assignment_expression(inTemplate=False)] while True: self.skip_ws() if not self.skip_string(','): break exprs.append(self._parse_assignment_expression(inTemplate=False)) if len(exprs) == 1: return exprs[0] else: return ASTCommaExpr(exprs) def _parse_expression_fallback(self, end: List[str], parser: Callable[[], ASTExpression], allow: bool = True) -> ASTExpression: # Stupidly "parse" an expression. # 'end' should be a list of characters which ends the expression. # first try to use the provided parser prevPos = self.pos try: return parser() except DefinitionError as e: # some places (e.g., template parameters) we really don't want to use fallback, # and for testing we may want to globally disable it if not allow or not self.allowFallbackExpressionParsing: raise self.warn("Parsing of expression failed. Using fallback parser." " Error was:\n%s" % e) self.pos = prevPos # and then the fallback scanning assert end is not None self.skip_ws() startPos = self.pos if self.match(_string_re): value = self.matched_text else: # TODO: add handling of more bracket-like things, and quote handling brackets = {'(': ')', '{': '}', '[': ']', '<': '>'} symbols = [] # type: List[str] while not self.eof: if (len(symbols) == 0 and self.current_char in end): break if self.current_char in brackets.keys(): symbols.append(brackets[self.current_char]) elif len(symbols) > 0 and self.current_char == symbols[-1]: symbols.pop() self.pos += 1 if len(end) > 0 and self.eof: self.fail("Could not find end of expression starting at %d." % startPos) value = self.definition[startPos:self.pos].strip() return ASTFallbackExpr(value.strip()) # ========================================================================== def _parse_operator(self) -> ASTOperator: self.skip_ws() # adapted from the old code # yay, a regular operator definition if self.match(_operator_re): return ASTOperatorBuildIn(self.matched_text) # new/delete operator? for op in 'new', 'delete': if not self.skip_word(op): continue self.skip_ws() if self.skip_string('['): self.skip_ws() if not self.skip_string(']'): self.fail('Expected "]" after "operator ' + op + '["') op += '[]' return ASTOperatorBuildIn(op) # user-defined literal? if self.skip_string('""'): self.skip_ws() if not self.match(identifier_re): self.fail("Expected user-defined literal suffix.") identifier = ASTIdentifier(self.matched_text) return ASTOperatorLiteral(identifier) # oh well, looks like a cast operator definition. # In that case, eat another type. type = self._parse_type(named=False, outer="operatorCast") return ASTOperatorType(type) def _parse_template_argument_list(self) -> ASTTemplateArgs: # template-argument-list: (but we include the < and > here # template-argument ...[opt] # template-argument-list, template-argument ...[opt] # template-argument: # constant-expression # type-id # id-expression self.skip_ws() if not self.skip_string_and_ws('<'): return None if self.skip_string('>'): return ASTTemplateArgs([], False) prevErrors = [] templateArgs = [] # type: List[Union[ASTType, ASTTemplateArgConstant]] packExpansion = False while 1: pos = self.pos parsedComma = False parsedEnd = False try: type = self._parse_type(named=False) self.skip_ws() if self.skip_string_and_ws('...'): packExpansion = True parsedEnd = True if not self.skip_string('>'): self.fail('Expected ">" after "..." in template argument list.') elif self.skip_string('>'): parsedEnd = True elif self.skip_string(','): parsedComma = True else: self.fail('Expected "...>", ">" or "," in template argument list.') templateArgs.append(type) except DefinitionError as e: prevErrors.append((e, "If type argument")) self.pos = pos try: value = self._parse_constant_expression(inTemplate=True) self.skip_ws() if self.skip_string_and_ws('...'): packExpansion = True parsedEnd = True if not self.skip_string('>'): self.fail('Expected ">" after "..." in template argument list.') elif self.skip_string('>'): parsedEnd = True elif self.skip_string(','): parsedComma = True else: self.fail('Expected "...>", ">" or "," in template argument list.') templateArgs.append(ASTTemplateArgConstant(value)) except DefinitionError as e: self.pos = pos prevErrors.append((e, "If non-type argument")) header = "Error in parsing template argument list." raise self._make_multi_error(prevErrors, header) if parsedEnd: assert not parsedComma break else: assert not packExpansion return ASTTemplateArgs(templateArgs, packExpansion) def _parse_nested_name(self, memberPointer: bool = False) -> ASTNestedName: names = [] # type: List[ASTNestedNameElement] templates = [] # type: List[bool] self.skip_ws() rooted = False if self.skip_string('::'): rooted = True while 1: self.skip_ws() if len(names) > 0: template = self.skip_word_and_ws('template') else: template = False templates.append(template) identOrOp = None # type: Union[ASTIdentifier, ASTOperator] if self.skip_word_and_ws('operator'): identOrOp = self._parse_operator() else: if not self.match(identifier_re): if memberPointer and len(names) > 0: templates.pop() break self.fail("Expected identifier in nested name.") identifier = self.matched_text # make sure there isn't a keyword if identifier in _keywords: self.fail("Expected identifier in nested name, " "got keyword: %s" % identifier) identOrOp = ASTIdentifier(identifier) # try greedily to get template arguments, # but otherwise a < might be because we are in an expression pos = self.pos try: templateArgs = self._parse_template_argument_list() except DefinitionError as ex: self.pos = pos templateArgs = None self.otherErrors.append(ex) names.append(ASTNestedNameElement(identOrOp, templateArgs)) self.skip_ws() if not self.skip_string('::'): if memberPointer: self.fail("Expected '::' in pointer to member (function).") break return ASTNestedName(names, templates, rooted) # ========================================================================== def _parse_trailing_type_spec(self) -> ASTTrailingTypeSpec: # fundemental types self.skip_ws() for t in self._simple_fundemental_types: if self.skip_word(t): return ASTTrailingTypeSpecFundamental(t) # TODO: this could/should be more strict elements = [] if self.skip_word_and_ws('signed'): elements.append('signed') elif self.skip_word_and_ws('unsigned'): elements.append('unsigned') while 1: if self.skip_word_and_ws('short'): elements.append('short') elif self.skip_word_and_ws('long'): elements.append('long') else: break if self.skip_word_and_ws('char'): elements.append('char') elif self.skip_word_and_ws('int'): elements.append('int') elif self.skip_word_and_ws('double'): elements.append('double') if len(elements) > 0: return ASTTrailingTypeSpecFundamental(' '.join(elements)) # decltype self.skip_ws() if self.skip_word_and_ws('decltype'): if not self.skip_string_and_ws('('): self.fail("Expected '(' after 'decltype'.") if self.skip_word_and_ws('auto'): if not self.skip_string(')'): self.fail("Expected ')' after 'decltype(auto'.") return ASTTrailingTypeSpecDecltypeAuto() expr = self._parse_expression() self.skip_ws() if not self.skip_string(')'): self.fail("Expected ')' after 'decltype(<expr>'.") return ASTTrailingTypeSpecDecltype(expr) # prefixed prefix = None self.skip_ws() for k in self._prefix_keys: if self.skip_word_and_ws(k): prefix = k break nestedName = self._parse_nested_name() return ASTTrailingTypeSpecName(prefix, nestedName) def _parse_parameters_and_qualifiers(self, paramMode: str) -> ASTParametersQualifiers: if paramMode == 'new': return None self.skip_ws() if not self.skip_string('('): if paramMode == 'function': self.fail('Expecting "(" in parameters-and-qualifiers.') else: return None args = [] self.skip_ws() if not self.skip_string(')'): while 1: self.skip_ws() if self.skip_string('...'): args.append(ASTFunctionParameter(None, True)) self.skip_ws() if not self.skip_string(')'): self.fail('Expected ")" after "..." in ' 'parameters-and-qualifiers.') break # note: it seems that function arguments can always be named, # even in function pointers and similar. arg = self._parse_type_with_init(outer=None, named='single') # TODO: parse default parameters # TODO: didn't we just do that? args.append(ASTFunctionParameter(arg)) self.skip_ws() if self.skip_string(','): continue elif self.skip_string(')'): break else: self.fail( 'Expecting "," or ")" in parameters-and-qualifiers, ' 'got "%s".' % self.current_char) # TODO: why did we have this bail-out? # does it hurt to parse the extra stuff? # it's needed for pointer to member functions if paramMode != 'function' and False: return ASTParametersQualifiers( args, None, None, None, None, None, None, None) self.skip_ws() const = self.skip_word_and_ws('const') volatile = self.skip_word_and_ws('volatile') if not const: # the can be permuted const = self.skip_word_and_ws('const') refQual = None if self.skip_string('&&'): refQual = '&&' if not refQual and self.skip_string('&'): refQual = '&' exceptionSpec = None override = None final = None initializer = None self.skip_ws() if self.skip_string('noexcept'): exceptionSpec = 'noexcept' self.skip_ws() if self.skip_string('('): self.fail('Parameterised "noexcept" not yet implemented.') self.skip_ws() override = self.skip_word_and_ws('override') final = self.skip_word_and_ws('final') if not override: override = self.skip_word_and_ws( 'override') # they can be permuted self.skip_ws() if self.skip_string('='): self.skip_ws() valid = ('0', 'delete', 'default') for w in valid: if self.skip_word_and_ws(w): initializer = w break if not initializer: self.fail( 'Expected "%s" in initializer-specifier.' % '" or "'.join(valid)) return ASTParametersQualifiers( args, volatile, const, refQual, exceptionSpec, override, final, initializer) def _parse_decl_specs_simple(self, outer: str, typed: bool) -> ASTDeclSpecsSimple: """Just parse the simple ones.""" storage = None threadLocal = None inline = None virtual = None explicit = None constexpr = None volatile = None const = None friend = None attrs = [] while 1: # accept any permutation of a subset of some decl-specs self.skip_ws() if not storage: if outer in ('member', 'function'): if self.skip_word('static'): storage = 'static' continue if self.skip_word('extern'): storage = 'extern' continue if outer == 'member': if self.skip_word('mutable'): storage = 'mutable' continue if self.skip_word('register'): storage = 'register' continue if not threadLocal and outer == 'member': threadLocal = self.skip_word('thread_local') if threadLocal: continue if outer == 'function': # function-specifiers if not inline: inline = self.skip_word('inline') if inline: continue if not friend: friend = self.skip_word('friend') if friend: continue if not virtual: virtual = self.skip_word('virtual') if virtual: continue if not explicit: explicit = self.skip_word('explicit') if explicit: continue if not constexpr and outer in ('member', 'function'): constexpr = self.skip_word("constexpr") if constexpr: continue if not volatile and typed: volatile = self.skip_word('volatile') if volatile: continue if not const and typed: const = self.skip_word('const') if const: continue attr = self._parse_attribute() if attr: attrs.append(attr) continue break return ASTDeclSpecsSimple(storage, threadLocal, inline, virtual, explicit, constexpr, volatile, const, friend, attrs) def _parse_decl_specs(self, outer: str, typed: bool = True) -> ASTDeclSpecs: if outer: if outer not in ('type', 'member', 'function', 'templateParam'): raise Exception('Internal error, unknown outer "%s".' % outer) """ storage-class-specifier function-specifier "constexpr" "volatile" "const" trailing-type-specifier storage-class-specifier -> "static" (only for member_object and function_object) | "register" function-specifier -> "inline" | "virtual" | "explicit" (only for function_object) "constexpr" (only for member_object and function_object) """ leftSpecs = self._parse_decl_specs_simple(outer, typed) rightSpecs = None if typed: trailing = self._parse_trailing_type_spec() rightSpecs = self._parse_decl_specs_simple(outer, typed) else: trailing = None return ASTDeclSpecs(outer, leftSpecs, rightSpecs, trailing) def _parse_declarator_name_suffix( self, named: Union[bool, str], paramMode: str, typed: bool ) -> Union[ASTDeclaratorNameParamQual, ASTDeclaratorNameBitField]: # now we should parse the name, and then suffixes if named == 'maybe': pos = self.pos try: declId = self._parse_nested_name() except DefinitionError: self.pos = pos declId = None elif named == 'single': if self.match(identifier_re): identifier = ASTIdentifier(self.matched_text) nne = ASTNestedNameElement(identifier, None) declId = ASTNestedName([nne], [False], rooted=False) # if it's a member pointer, we may have '::', which should be an error self.skip_ws() if self.current_char == ':': self.fail("Unexpected ':' after identifier.") else: declId = None elif named: declId = self._parse_nested_name() else: declId = None arrayOps = [] while 1: self.skip_ws() if typed and self.skip_string('['): self.skip_ws() if self.skip_string(']'): arrayOps.append(ASTArray(None)) continue def parser() -> ASTExpression: return self._parse_expression() value = self._parse_expression_fallback([']'], parser) if not self.skip_string(']'): self.fail("Expected ']' in end of array operator.") arrayOps.append(ASTArray(value)) continue else: break paramQual = self._parse_parameters_and_qualifiers(paramMode) if paramQual is None and len(arrayOps) == 0: # perhaps a bit-field if named and paramMode == 'type' and typed: self.skip_ws() if self.skip_string(':'): size = self._parse_constant_expression(inTemplate=False) return ASTDeclaratorNameBitField(declId=declId, size=size) return ASTDeclaratorNameParamQual(declId=declId, arrayOps=arrayOps, paramQual=paramQual) def _parse_declarator(self, named: Union[bool, str], paramMode: str, typed: bool = True ) -> ASTDeclarator: # 'typed' here means 'parse return type stuff' if paramMode not in ('type', 'function', 'operatorCast', 'new'): raise Exception( "Internal error, unknown paramMode '%s'." % paramMode) prevErrors = [] self.skip_ws() if typed and self.skip_string('*'): self.skip_ws() volatile = False const = False attrs = [] while 1: if not volatile: volatile = self.skip_word_and_ws('volatile') if volatile: continue if not const: const = self.skip_word_and_ws('const') if const: continue attr = self._parse_attribute() if attr is not None: attrs.append(attr) continue break next = self._parse_declarator(named, paramMode, typed) return ASTDeclaratorPtr(next=next, volatile=volatile, const=const, attrs=attrs) # TODO: shouldn't we parse an R-value ref here first? if typed and self.skip_string("&"): attrs = [] while 1: attr = self._parse_attribute() if attr is None: break attrs.append(attr) next = self._parse_declarator(named, paramMode, typed) return ASTDeclaratorRef(next=next, attrs=attrs) if typed and self.skip_string("..."): next = self._parse_declarator(named, paramMode, False) return ASTDeclaratorParamPack(next=next) if typed and self.current_char == '(': # note: peeking, not skipping if paramMode == "operatorCast": # TODO: we should be able to parse cast operators which return # function pointers. For now, just hax it and ignore. return ASTDeclaratorNameParamQual(declId=None, arrayOps=[], paramQual=None) # maybe this is the beginning of params and quals,try that first, # otherwise assume it's noptr->declarator > ( ptr-declarator ) pos = self.pos try: # assume this is params and quals res = self._parse_declarator_name_suffix(named, paramMode, typed) return res except DefinitionError as exParamQual: prevErrors.append((exParamQual, "If declarator-id with parameters-and-qualifiers")) self.pos = pos try: assert self.current_char == '(' self.skip_string('(') # TODO: hmm, if there is a name, it must be in inner, right? # TODO: hmm, if there must be parameters, they must be # inside, right? inner = self._parse_declarator(named, paramMode, typed) if not self.skip_string(')'): self.fail("Expected ')' in \"( ptr-declarator )\"") next = self._parse_declarator(named=False, paramMode="type", typed=typed) return ASTDeclaratorParen(inner=inner, next=next) except DefinitionError as exNoPtrParen: self.pos = pos prevErrors.append((exNoPtrParen, "If parenthesis in noptr-declarator")) header = "Error in declarator" raise self._make_multi_error(prevErrors, header) if typed: # pointer to member pos = self.pos try: name = self._parse_nested_name(memberPointer=True) self.skip_ws() if not self.skip_string('*'): self.fail("Expected '*' in pointer to member declarator.") self.skip_ws() except DefinitionError as e: self.pos = pos prevErrors.append((e, "If pointer to member declarator")) else: volatile = False const = False while 1: if not volatile: volatile = self.skip_word_and_ws('volatile') if volatile: continue if not const: const = self.skip_word_and_ws('const') if const: continue break next = self._parse_declarator(named, paramMode, typed) return ASTDeclaratorMemPtr(name, const, volatile, next=next) pos = self.pos try: res = self._parse_declarator_name_suffix(named, paramMode, typed) # this is a heuristic for error messages, for when there is a < after a # nested name, but it was not a successful template argument list if self.current_char == '<': self.otherErrors.append(self._make_multi_error(prevErrors, "")) return res except DefinitionError as e: self.pos = pos prevErrors.append((e, "If declarator-id")) header = "Error in declarator or parameters-and-qualifiers" raise self._make_multi_error(prevErrors, header) def _parse_initializer(self, outer: str = None, allowFallback: bool = True ) -> ASTInitializer: # initializer # global vars # -> brace-or-equal-initializer # | '(' expression-list ')' # # brace-or-equal-initializer # member vars # -> '=' initializer-clause # | braced-init-list # # initializer-clause # function params, non-type template params (with '=' in front) # -> assignment-expression # | braced-init-list # # we don't distinguish between global and member vars, so disallow paren: # # -> braced-init-list # var only # | '=' assignment-expression # | '=' braced-init-list self.skip_ws() if outer == 'member': bracedInit = self._parse_braced_init_list() if bracedInit is not None: return ASTInitializer(bracedInit, hasAssign=False) if not self.skip_string('='): return None bracedInit = self._parse_braced_init_list() if bracedInit is not None: return ASTInitializer(bracedInit) if outer == 'member': fallbackEnd = [] # type: List[str] elif outer == 'templateParam': fallbackEnd = [',', '>'] elif outer is None: # function parameter fallbackEnd = [',', ')'] else: self.fail("Internal error, initializer for outer '%s' not " "implemented." % outer) inTemplate = outer == 'templateParam' def parser() -> ASTExpression: return self._parse_assignment_expression(inTemplate=inTemplate) value = self._parse_expression_fallback(fallbackEnd, parser, allow=allowFallback) return ASTInitializer(value) def _parse_type(self, named: Union[bool, str], outer: str = None) -> ASTType: """ named=False|'maybe'|True: 'maybe' is e.g., for function objects which doesn't need to name the arguments outer == operatorCast: annoying case, we should not take the params """ if outer: # always named if outer not in ('type', 'member', 'function', 'operatorCast', 'templateParam'): raise Exception('Internal error, unknown outer "%s".' % outer) if outer != 'operatorCast': assert named if outer in ('type', 'function'): # We allow type objects to just be a name. # Some functions don't have normal return types: constructors, # destrutors, cast operators prevErrors = [] startPos = self.pos # first try without the type try: declSpecs = self._parse_decl_specs(outer=outer, typed=False) decl = self._parse_declarator(named=True, paramMode=outer, typed=False) self.assert_end() except DefinitionError as exUntyped: if outer == 'type': desc = "If just a name" elif outer == 'function': desc = "If the function has no return type" else: assert False prevErrors.append((exUntyped, desc)) self.pos = startPos try: declSpecs = self._parse_decl_specs(outer=outer) decl = self._parse_declarator(named=True, paramMode=outer) except DefinitionError as exTyped: self.pos = startPos if outer == 'type': desc = "If typedef-like declaration" elif outer == 'function': desc = "If the function has a return type" else: assert False prevErrors.append((exTyped, desc)) # Retain the else branch for easier debugging. # TODO: it would be nice to save the previous stacktrace # and output it here. if True: if outer == 'type': header = "Type must be either just a name or a " header += "typedef-like declaration." elif outer == 'function': header = "Error when parsing function declaration." else: assert False raise self._make_multi_error(prevErrors, header) else: # For testing purposes. # do it again to get the proper traceback (how do you # reliably save a traceback when an exception is # constructed?) self.pos = startPos typed = True declSpecs = self._parse_decl_specs(outer=outer, typed=typed) decl = self._parse_declarator(named=True, paramMode=outer, typed=typed) else: paramMode = 'type' if outer == 'member': # i.e., member named = True elif outer == 'operatorCast': paramMode = 'operatorCast' outer = None elif outer == 'templateParam': named = 'single' declSpecs = self._parse_decl_specs(outer=outer) decl = self._parse_declarator(named=named, paramMode=paramMode) return ASTType(declSpecs, decl) def _parse_type_with_init( self, named: Union[bool, str], outer: str) -> Union[ASTTypeWithInit, ASTTemplateParamConstrainedTypeWithInit]: if outer: assert outer in ('type', 'member', 'function', 'templateParam') type = self._parse_type(outer=outer, named=named) if outer != 'templateParam': init = self._parse_initializer(outer=outer) return ASTTypeWithInit(type, init) # it could also be a constrained type parameter, e.g., C T = int& pos = self.pos eExpr = None try: init = self._parse_initializer(outer=outer, allowFallback=False) # note: init may be None if there is no = if init is None: return ASTTypeWithInit(type, None) # we parsed an expression, so we must have a , or a >, # otherwise the expression didn't get everything self.skip_ws() if self.current_char != ',' and self.current_char != '>': # pretend it didn't happen self.pos = pos init = None else: # we assume that it was indeed an expression return ASTTypeWithInit(type, init) except DefinitionError as e: self.pos = pos eExpr = e if not self.skip_string("="): return ASTTypeWithInit(type, None) try: typeInit = self._parse_type(named=False, outer=None) return ASTTemplateParamConstrainedTypeWithInit(type, typeInit) except DefinitionError as eType: if eExpr is None: raise eType errs = [] errs.append((eExpr, "If default template argument is an expression")) errs.append((eType, "If default template argument is a type")) msg = "Error in non-type template parameter" msg += " or constrained template parameter." raise self._make_multi_error(errs, msg) def _parse_type_using(self) -> ASTTypeUsing: name = self._parse_nested_name() self.skip_ws() if not self.skip_string('='): return ASTTypeUsing(name, None) type = self._parse_type(False, None) return ASTTypeUsing(name, type) def _parse_concept(self) -> ASTConcept: nestedName = self._parse_nested_name() self.skip_ws() initializer = self._parse_initializer('member') return ASTConcept(nestedName, initializer) def _parse_class(self) -> ASTClass: name = self._parse_nested_name() self.skip_ws() final = self.skip_word_and_ws('final') bases = [] self.skip_ws() if self.skip_string(':'): while 1: self.skip_ws() visibility = None virtual = False pack = False if self.skip_word_and_ws('virtual'): virtual = True if self.match(_visibility_re): visibility = self.matched_text self.skip_ws() if not virtual and self.skip_word_and_ws('virtual'): virtual = True baseName = self._parse_nested_name() self.skip_ws() pack = self.skip_string('...') bases.append(ASTBaseClass(baseName, visibility, virtual, pack)) self.skip_ws() if self.skip_string(','): continue else: break return ASTClass(name, final, bases) def _parse_union(self) -> ASTUnion: name = self._parse_nested_name() return ASTUnion(name) def _parse_enum(self) -> ASTEnum: scoped = None # is set by CPPEnumObject self.skip_ws() name = self._parse_nested_name() self.skip_ws() underlyingType = None if self.skip_string(':'): underlyingType = self._parse_type(named=False) return ASTEnum(name, scoped, underlyingType) def _parse_enumerator(self) -> ASTEnumerator: name = self._parse_nested_name() self.skip_ws() init = None if self.skip_string('='): self.skip_ws() def parser() -> ASTExpression: return self._parse_constant_expression(inTemplate=False) initVal = self._parse_expression_fallback([], parser) init = ASTInitializer(initVal) return ASTEnumerator(name, init) # ========================================================================== def _parse_template_parameter_list(self) -> ASTTemplateParams: # only: '<' parameter-list '>' # we assume that 'template' has just been parsed templateParams = [] # type: List[ASTTemplateParam] self.skip_ws() if not self.skip_string("<"): self.fail("Expected '<' after 'template'") prevErrors = [] while 1: self.skip_ws() if self.skip_word('template'): # declare a tenplate template parameter nestedParams = self._parse_template_parameter_list() else: nestedParams = None self.skip_ws() key = None if self.skip_word_and_ws('typename'): key = 'typename' elif self.skip_word_and_ws('class'): key = 'class' elif nestedParams: self.fail("Expected 'typename' or 'class' after " "template template parameter list.") if key: # declare a type or template type parameter self.skip_ws() parameterPack = self.skip_string('...') self.skip_ws() if self.match(identifier_re): identifier = ASTIdentifier(self.matched_text) else: identifier = None self.skip_ws() if not parameterPack and self.skip_string('='): default = self._parse_type(named=False, outer=None) else: default = None data = ASTTemplateKeyParamPackIdDefault(key, identifier, parameterPack, default) if nestedParams: # template type templateParams.append( ASTTemplateParamTemplateType(nestedParams, data)) else: # type templateParams.append(ASTTemplateParamType(data)) else: # declare a non-type parameter, or constrained type parameter pos = self.pos try: param = self._parse_type_with_init('maybe', 'templateParam') templateParams.append(ASTTemplateParamNonType(param)) except DefinitionError as e: msg = "If non-type template parameter or constrained template parameter" prevErrors.append((e, msg)) self.pos = pos self.skip_ws() if self.skip_string('>'): return ASTTemplateParams(templateParams) elif self.skip_string(','): prevErrors = [] continue else: header = "Error in template parameter list." try: self.fail('Expected "=", ",", or ">".') except DefinitionError as e: prevErrors.append((e, "")) raise self._make_multi_error(prevErrors, header) def _parse_template_introduction(self) -> ASTTemplateIntroduction: pos = self.pos try: concept = self._parse_nested_name() except Exception: self.pos = pos return None self.skip_ws() if not self.skip_string('{'): self.pos = pos return None # for sure it must be a template introduction now params = [] while 1: self.skip_ws() parameterPack = self.skip_string('...') self.skip_ws() if not self.match(identifier_re): self.fail("Expected identifier in template introduction list.") txt_identifier = self.matched_text # make sure there isn't a keyword if txt_identifier in _keywords: self.fail("Expected identifier in template introduction list, " "got keyword: %s" % txt_identifier) identifier = ASTIdentifier(txt_identifier) params.append(ASTTemplateIntroductionParameter(identifier, parameterPack)) self.skip_ws() if self.skip_string('}'): break elif self.skip_string(','): continue else: self.fail("Error in template introduction list. " 'Expected ",", or "}".') return ASTTemplateIntroduction(concept, params) def _parse_template_declaration_prefix(self, objectType: str ) -> ASTTemplateDeclarationPrefix: templates = [] # type: List[Union[ASTTemplateParams, ASTTemplateIntroduction]] while 1: self.skip_ws() # the saved position is only used to provide a better error message params = None # type: Union[ASTTemplateParams, ASTTemplateIntroduction] pos = self.pos if self.skip_word("template"): try: params = self._parse_template_parameter_list() except DefinitionError as e: if objectType == 'member' and len(templates) == 0: return ASTTemplateDeclarationPrefix(None) else: raise e else: params = self._parse_template_introduction() if not params: break if objectType == 'concept' and len(templates) > 0: self.pos = pos self.fail("More than 1 template parameter list for concept.") templates.append(params) if len(templates) == 0 and objectType == 'concept': self.fail('Missing template parameter list for concept.') if len(templates) == 0: return None else: return ASTTemplateDeclarationPrefix(templates) def _check_template_consistency(self, nestedName: ASTNestedName, templatePrefix: ASTTemplateDeclarationPrefix, fullSpecShorthand: bool, isMember: bool = False ) -> ASTTemplateDeclarationPrefix: numArgs = nestedName.num_templates() isMemberInstantiation = False if not templatePrefix: numParams = 0 else: if isMember and templatePrefix.templates is None: numParams = 0 isMemberInstantiation = True else: numParams = len(templatePrefix.templates) if numArgs + 1 < numParams: self.fail("Too few template argument lists comapred to parameter" " lists. Argument lists: %d, Parameter lists: %d." % (numArgs, numParams)) if numArgs > numParams: numExtra = numArgs - numParams if not fullSpecShorthand and not isMemberInstantiation: msg = "Too many template argument lists compared to parameter" \ " lists. Argument lists: %d, Parameter lists: %d," \ " Extra empty parameters lists prepended: %d." \ % (numArgs, numParams, numExtra) msg += " Declaration:\n\t" if templatePrefix: msg += "%s\n\t" % templatePrefix msg += str(nestedName) self.warn(msg) newTemplates = [] # type: List[Union[ASTTemplateParams, ASTTemplateIntroduction]] for i in range(numExtra): newTemplates.append(ASTTemplateParams([])) if templatePrefix and not isMemberInstantiation: newTemplates.extend(templatePrefix.templates) templatePrefix = ASTTemplateDeclarationPrefix(newTemplates) return templatePrefix def parse_declaration(self, objectType: str, directiveType: str) -> ASTDeclaration: if objectType not in ('class', 'union', 'function', 'member', 'type', 'concept', 'enum', 'enumerator'): raise Exception('Internal error, unknown objectType "%s".' % objectType) if directiveType not in ('class', 'struct', 'union', 'function', 'member', 'var', 'type', 'concept', 'enum', 'enum-struct', 'enum-class', 'enumerator'): raise Exception('Internal error, unknown directiveType "%s".' % directiveType) visibility = None templatePrefix = None declaration = None # type: Any self.skip_ws() if self.match(_visibility_re): visibility = self.matched_text if objectType in ('type', 'concept', 'member', 'function', 'class'): templatePrefix = self._parse_template_declaration_prefix(objectType) if objectType == 'type': prevErrors = [] pos = self.pos try: if not templatePrefix: declaration = self._parse_type(named=True, outer='type') except DefinitionError as e: prevErrors.append((e, "If typedef-like declaration")) self.pos = pos pos = self.pos try: if not declaration: declaration = self._parse_type_using() except DefinitionError as e: self.pos = pos prevErrors.append((e, "If type alias or template alias")) header = "Error in type declaration." raise self._make_multi_error(prevErrors, header) elif objectType == 'concept': declaration = self._parse_concept() elif objectType == 'member': declaration = self._parse_type_with_init(named=True, outer='member') elif objectType == 'function': declaration = self._parse_type(named=True, outer='function') elif objectType == 'class': declaration = self._parse_class() elif objectType == 'union': declaration = self._parse_union() elif objectType == 'enum': declaration = self._parse_enum() elif objectType == 'enumerator': declaration = self._parse_enumerator() else: assert False templatePrefix = self._check_template_consistency(declaration.name, templatePrefix, fullSpecShorthand=False, isMember=objectType == 'member') return ASTDeclaration(objectType, directiveType, visibility, templatePrefix, declaration) def parse_namespace_object(self) -> ASTNamespace: templatePrefix = self._parse_template_declaration_prefix(objectType="namespace") name = self._parse_nested_name() templatePrefix = self._check_template_consistency(name, templatePrefix, fullSpecShorthand=False) res = ASTNamespace(name, templatePrefix) res.objectType = 'namespace' # type: ignore return res def parse_xref_object(self) -> Tuple[Union[ASTNamespace, ASTDeclaration], bool]: pos = self.pos try: templatePrefix = self._parse_template_declaration_prefix(objectType="xref") name = self._parse_nested_name() # if there are '()' left, just skip them self.skip_ws() self.skip_string('()') self.assert_end() templatePrefix = self._check_template_consistency(name, templatePrefix, fullSpecShorthand=True) res1 = ASTNamespace(name, templatePrefix) res1.objectType = 'xref' # type: ignore return res1, True except DefinitionError as e1: try: self.pos = pos res2 = self.parse_declaration('function', 'function') # if there are '()' left, just skip them self.skip_ws() self.skip_string('()') self.assert_end() return res2, False except DefinitionError as e2: errs = [] errs.append((e1, "If shorthand ref")) errs.append((e2, "If full function ref")) msg = "Error in cross-reference." raise self._make_multi_error(errs, msg) def parse_expression(self) -> Union[ASTExpression, ASTType]: pos = self.pos try: expr = self._parse_expression() self.skip_ws() self.assert_end() return expr except DefinitionError as exExpr: self.pos = pos try: typ = self._parse_type(False) self.skip_ws() self.assert_end() return typ except DefinitionError as exType: header = "Error when parsing (type) expression." errs = [] errs.append((exExpr, "If expression")) errs.append((exType, "If type")) raise self._make_multi_error(errs, header) def _make_phony_error_name() -> ASTNestedName: nne = ASTNestedNameElement(ASTIdentifier("PhonyNameDueToError"), None) return ASTNestedName([nne], [False], rooted=False) class CPPObject(ObjectDescription): """Description of a C++ language object.""" doc_field_types = [ GroupedField('parameter', label=_('Parameters'), names=('param', 'parameter', 'arg', 'argument'), can_collapse=True), GroupedField('template parameter', label=_('Template Parameters'), names=('tparam', 'template parameter'), can_collapse=True), GroupedField('exceptions', label=_('Throws'), rolename='cpp:class', names=('throws', 'throw', 'exception'), can_collapse=True), Field('returnvalue', label=_('Returns'), has_arg=False, names=('returns', 'return')), ] option_spec = dict(ObjectDescription.option_spec) option_spec['tparam-line-spec'] = directives.flag def _add_enumerator_to_parent(self, ast: ASTDeclaration) -> None: assert ast.objectType == 'enumerator' # find the parent, if it exists && is an enum # && it's unscoped, # then add the name to the parent scope symbol = ast.symbol assert symbol assert symbol.identOrOp is not None assert symbol.templateParams is None assert symbol.templateArgs is None parentSymbol = symbol.parent assert parentSymbol if parentSymbol.parent is None: # TODO: we could warn, but it is somewhat equivalent to unscoped # enums, without the enum return # no parent parentDecl = parentSymbol.declaration if parentDecl is None: # the parent is not explicitly declared # TODO: we could warn, but it could be a style to just assume # enumerator parents to be scoped return if parentDecl.objectType != 'enum': # TODO: maybe issue a warning, enumerators in non-enums is weird, # but it is somewhat equivalent to unscoped enums, without the enum return if parentDecl.directiveType != 'enum': return targetSymbol = parentSymbol.parent s = targetSymbol.find_identifier(symbol.identOrOp, matchSelf=False, recurseInAnon=True, searchInSiblings=False) if s is not None: # something is already declared with that name return declClone = symbol.declaration.clone() declClone.enumeratorScopedSymbol = symbol Symbol(parent=targetSymbol, identOrOp=symbol.identOrOp, templateParams=None, templateArgs=None, declaration=declClone, docname=self.env.docname) def add_target_and_index(self, ast: ASTDeclaration, sig: str, signode: TextElement) -> None: # general note: name must be lstrip(':')'ed, to remove "::" ids = [] for i in range(1, _max_id + 1): try: id = ast.get_id(version=i) ids.append(id) except NoOldIdError: assert i < _max_id # let's keep the newest first ids = list(reversed(ids)) newestId = ids[0] assert newestId # shouldn't be None if not re.compile(r'^[a-zA-Z0-9_]*$').match(newestId): logger.warning('Index id generation for C++ object "%s" failed, please ' 'report as bug (id=%s).', ast, newestId, location=self.get_source_info()) name = ast.symbol.get_full_nested_name().get_display_string().lstrip(':') # Add index entry, but not if it's a declaration inside a concept isInConcept = False s = ast.symbol.parent while s is not None: decl = s.declaration s = s.parent if decl is None: continue if decl.objectType == 'concept': isInConcept = True break if not isInConcept: strippedName = name for prefix in self.env.config.cpp_index_common_prefix: if name.startswith(prefix): strippedName = strippedName[len(prefix):] break indexText = self.get_index_text(strippedName) self.indexnode['entries'].append(('single', indexText, newestId, '', None)) if newestId not in self.state.document.ids: # if the name is not unique, the first one will win names = self.env.domaindata['cpp']['names'] if name not in names: names[name] = ast.symbol.docname # always add the newest id assert newestId signode['ids'].append(newestId) # only add compatibility ids when there are no conflicts for id in ids[1:]: if not id: # is None when the element didn't exist in that version continue if id not in self.state.document.ids: signode['ids'].append(id) self.state.document.note_explicit_target(signode) @property def object_type(self) -> str: raise NotImplementedError() @property def display_object_type(self) -> str: return self.object_type def get_index_text(self, name: str) -> str: return _('%s (C++ %s)') % (name, self.display_object_type) def parse_definition(self, parser: DefinitionParser) -> ASTDeclaration: return parser.parse_declaration(self.object_type, self.objtype) def describe_signature(self, signode: desc_signature, ast: ASTDeclaration, options: Dict) -> None: ast.describe_signature(signode, 'lastIsName', self.env, options) def run(self) -> List[Node]: env = self.state.document.settings.env # from ObjectDescription.run if 'cpp:parent_symbol' not in env.temp_data: root = env.domaindata['cpp']['root_symbol'] env.temp_data['cpp:parent_symbol'] = root env.ref_context['cpp:parent_key'] = root.get_lookup_key() # The lookup keys assume that no nested scopes exists inside overloaded functions. # (see also #5191) # Example: # .. cpp:function:: void f(int) # .. cpp:function:: void f(double) # # .. cpp:function:: void g() # # :cpp:any:`boom` # # So we disallow any signatures inside functions. parentSymbol = env.temp_data['cpp:parent_symbol'] parentDecl = parentSymbol.declaration if parentDecl is not None and parentDecl.objectType == 'function': logger.warning("C++ declarations inside functions are not supported." + " Parent function is " + str(parentSymbol.get_full_nested_name()), location=self.get_source_info()) name = _make_phony_error_name() symbol = parentSymbol.add_name(name) env.temp_data['cpp:last_symbol'] = symbol return [] # When multiple declarations are made in the same directive # they need to know about each other to provide symbol lookup for function parameters. # We use last_symbol to store the latest added declaration in a directive. env.temp_data['cpp:last_symbol'] = None return super().run() def handle_signature(self, sig: str, signode: desc_signature) -> ASTDeclaration: parentSymbol = self.env.temp_data['cpp:parent_symbol'] parser = DefinitionParser(sig, location=signode, config=self.env.config) try: ast = self.parse_definition(parser) parser.assert_end() except DefinitionError as e: logger.warning(e, location=signode) # It is easier to assume some phony name than handling the error in # the possibly inner declarations. name = _make_phony_error_name() symbol = parentSymbol.add_name(name) self.env.temp_data['cpp:last_symbol'] = symbol raise ValueError try: symbol = parentSymbol.add_declaration(ast, docname=self.env.docname) # append the new declaration to the sibling list assert symbol.siblingAbove is None assert symbol.siblingBelow is None symbol.siblingAbove = self.env.temp_data['cpp:last_symbol'] if symbol.siblingAbove is not None: assert symbol.siblingAbove.siblingBelow is None symbol.siblingAbove.siblingBelow = symbol self.env.temp_data['cpp:last_symbol'] = symbol except _DuplicateSymbolError as e: # Assume we are actually in the old symbol, # instead of the newly created duplicate. self.env.temp_data['cpp:last_symbol'] = e.symbol logger.warning("Duplicate declaration, %s", sig, location=signode) if ast.objectType == 'enumerator': self._add_enumerator_to_parent(ast) # note: handle_signature may be called multiple time per directive, # if it has multiple signatures, so don't mess with the original options. options = dict(self.options) options['tparam-line-spec'] = 'tparam-line-spec' in self.options self.describe_signature(signode, ast, options) return ast def before_content(self) -> None: lastSymbol = self.env.temp_data['cpp:last_symbol'] # type: Symbol assert lastSymbol self.oldParentSymbol = self.env.temp_data['cpp:parent_symbol'] self.oldParentKey = self.env.ref_context['cpp:parent_key'] # type: LookupKey self.env.temp_data['cpp:parent_symbol'] = lastSymbol self.env.ref_context['cpp:parent_key'] = lastSymbol.get_lookup_key() def after_content(self) -> None: self.env.temp_data['cpp:parent_symbol'] = self.oldParentSymbol self.env.ref_context['cpp:parent_key'] = self.oldParentKey class CPPTypeObject(CPPObject): object_type = 'type' class CPPConceptObject(CPPObject): object_type = 'concept' class CPPMemberObject(CPPObject): object_type = 'member' class CPPFunctionObject(CPPObject): object_type = 'function' class CPPClassObject(CPPObject): object_type = 'class' @property def display_object_type(self) -> str: # the distinction between class and struct is only cosmetic assert self.objtype in ('class', 'struct') return self.objtype class CPPUnionObject(CPPObject): object_type = 'union' class CPPEnumObject(CPPObject): object_type = 'enum' class CPPEnumeratorObject(CPPObject): object_type = 'enumerator' class CPPNamespaceObject(SphinxDirective): """ This directive is just to tell Sphinx that we're documenting stuff in namespace foo. """ has_content = False required_arguments = 1 optional_arguments = 0 final_argument_whitespace = True option_spec = {} # type: Dict def run(self) -> List[Node]: rootSymbol = self.env.domaindata['cpp']['root_symbol'] if self.arguments[0].strip() in ('NULL', '0', 'nullptr'): symbol = rootSymbol stack = [] # type: List[Symbol] else: parser = DefinitionParser(self.arguments[0], location=self.get_source_info(), config=self.config) try: ast = parser.parse_namespace_object() parser.assert_end() except DefinitionError as e: logger.warning(e, location=self.get_source_info()) name = _make_phony_error_name() ast = ASTNamespace(name, None) symbol = rootSymbol.add_name(ast.nestedName, ast.templatePrefix) stack = [symbol] self.env.temp_data['cpp:parent_symbol'] = symbol self.env.temp_data['cpp:namespace_stack'] = stack self.env.ref_context['cpp:parent_key'] = symbol.get_lookup_key() return [] class CPPNamespacePushObject(SphinxDirective): has_content = False required_arguments = 1 optional_arguments = 0 final_argument_whitespace = True option_spec = {} # type: Dict def run(self) -> List[Node]: if self.arguments[0].strip() in ('NULL', '0', 'nullptr'): return [] parser = DefinitionParser(self.arguments[0], location=self.get_source_info(), config=self.config) try: ast = parser.parse_namespace_object() parser.assert_end() except DefinitionError as e: logger.warning(e, location=self.get_source_info()) name = _make_phony_error_name() ast = ASTNamespace(name, None) oldParent = self.env.temp_data.get('cpp:parent_symbol', None) if not oldParent: oldParent = self.env.domaindata['cpp']['root_symbol'] symbol = oldParent.add_name(ast.nestedName, ast.templatePrefix) stack = self.env.temp_data.get('cpp:namespace_stack', []) stack.append(symbol) self.env.temp_data['cpp:parent_symbol'] = symbol self.env.temp_data['cpp:namespace_stack'] = stack self.env.ref_context['cpp:parent_key'] = symbol.get_lookup_key() return [] class CPPNamespacePopObject(SphinxDirective): has_content = False required_arguments = 0 optional_arguments = 0 final_argument_whitespace = True option_spec = {} # type: Dict def run(self) -> List[Node]: stack = self.env.temp_data.get('cpp:namespace_stack', None) if not stack or len(stack) == 0: logger.warning("C++ namespace pop on empty stack. Defaulting to gobal scope.", location=self.get_source_info()) stack = [] else: stack.pop() if len(stack) > 0: symbol = stack[-1] else: symbol = self.env.domaindata['cpp']['root_symbol'] self.env.temp_data['cpp:parent_symbol'] = symbol self.env.temp_data['cpp:namespace_stack'] = stack self.env.ref_context['cpp:parent_key'] = symbol.get_lookup_key() return [] class AliasNode(nodes.Element): def __init__(self, sig: str, env: "BuildEnvironment" = None, parentKey: LookupKey = None) -> None: super().__init__() self.sig = sig if env is not None: if 'cpp:parent_symbol' not in env.temp_data: root = env.domaindata['cpp']['root_symbol'] env.temp_data['cpp:parent_symbol'] = root self.parentKey = env.temp_data['cpp:parent_symbol'].get_lookup_key() else: assert parentKey is not None self.parentKey = parentKey def copy(self: T) -> T: return self.__class__(self.sig, env=None, parentKey=self.parentKey) # type: ignore class AliasTransform(SphinxTransform): default_priority = ReferencesResolver.default_priority - 1 def apply(self, **kwargs: Any) -> None: for node in self.document.traverse(AliasNode): sig = node.sig parentKey = node.parentKey try: parser = DefinitionParser(sig, location=node, config=self.env.config) ast, isShorthand = parser.parse_xref_object() parser.assert_end() except DefinitionError as e: logger.warning(e, location=node) ast, isShorthand = None, None if ast is None: # could not be parsed, so stop here signode = addnodes.desc_signature(sig, '') signode.clear() signode += addnodes.desc_name(sig, sig) node.replace_self(signode) continue rootSymbol = self.env.domains['cpp'].data['root_symbol'] # type: Symbol parentSymbol = rootSymbol.direct_lookup(parentKey) # type: Symbol if not parentSymbol: print("Target: ", sig) print("ParentKey: ", parentKey) print(rootSymbol.dump(1)) assert parentSymbol # should be there symbols = [] # type: List[Symbol] if isShorthand: assert isinstance(ast, ASTNamespace) ns = ast name = ns.nestedName if ns.templatePrefix: templateDecls = ns.templatePrefix.templates else: templateDecls = [] symbols, failReason = parentSymbol.find_name( nestedName=name, templateDecls=templateDecls, typ='any', templateShorthand=True, matchSelf=True, recurseInAnon=True, searchInSiblings=False) if symbols is None: symbols = [] else: assert isinstance(ast, ASTDeclaration) decl = ast name = decl.name s = parentSymbol.find_declaration(decl, 'any', templateShorthand=True, matchSelf=True, recurseInAnon=True) if s is not None: symbols.append(s) symbols = [s for s in symbols if s.declaration is not None] if len(symbols) == 0: signode = addnodes.desc_signature(sig, '') node.append(signode) signode.clear() signode += addnodes.desc_name(sig, sig) logger.warning("Could not find C++ declaration for alias '%s'." % ast, location=node) node.replace_self(signode) else: nodes = [] options = dict() options['tparam-line-spec'] = False for s in symbols: signode = addnodes.desc_signature(sig, '') nodes.append(signode) s.declaration.describe_signature(signode, 'markName', self.env, options) node.replace_self(nodes) class CPPAliasObject(ObjectDescription): option_spec = {} # type: Dict def run(self) -> List[Node]: """ On purpose this doesn't call the ObjectDescription version, but is based on it. Each alias signature may expand into multiple real signatures (an overload set). The code is therefore based on the ObjectDescription version. """ if ':' in self.name: self.domain, self.objtype = self.name.split(':', 1) else: self.domain, self.objtype = '', self.name node = addnodes.desc() node.document = self.state.document node['domain'] = self.domain # 'desctype' is a backwards compatible attribute node['objtype'] = node['desctype'] = self.objtype node['noindex'] = True self.names = [] # type: List[str] signatures = self.get_signatures() for i, sig in enumerate(signatures): node.append(AliasNode(sig, env=self.env)) contentnode = addnodes.desc_content() node.append(contentnode) self.before_content() self.state.nested_parse(self.content, self.content_offset, contentnode) self.env.temp_data['object'] = None self.after_content() return [node] class CPPXRefRole(XRefRole): def process_link(self, env: BuildEnvironment, refnode: Element, has_explicit_title: bool, title: str, target: str) -> Tuple[str, str]: refnode.attributes.update(env.ref_context) if not has_explicit_title: # major hax: replace anon names via simple string manipulation. # Can this actually fail? title = anon_identifier_re.sub("[anonymous]", str(title)) if refnode['reftype'] == 'any': # Assume the removal part of fix_parens for :any: refs. # The addition part is done with the reference is resolved. if not has_explicit_title and title.endswith('()'): title = title[:-2] if target.endswith('()'): target = target[:-2] # TODO: should this really be here? if not has_explicit_title: target = target.lstrip('~') # only has a meaning for the title # if the first character is a tilde, don't display the module/class # parts of the contents if title[:1] == '~': title = title[1:] dcolon = title.rfind('::') if dcolon != -1: title = title[dcolon + 2:] return title, target class CPPExprRole(SphinxRole): def __init__(self, asCode: bool) -> None: super().__init__() if asCode: # render the expression as inline code self.class_type = 'cpp-expr' self.node_type = nodes.literal # type: Type[TextElement] else: # render the expression as inline text self.class_type = 'cpp-texpr' self.node_type = nodes.inline def run(self) -> Tuple[List[Node], List[system_message]]: text = self.text.replace('\n', ' ') parser = DefinitionParser(text, location=self.get_source_info(), config=self.config) # attempt to mimic XRefRole classes, except that... classes = ['xref', 'cpp', self.class_type] try: ast = parser.parse_expression() except DefinitionError as ex: logger.warning('Unparseable C++ expression: %r\n%s', text, ex, location=self.get_source_info()) # see below return [self.node_type(text, text, classes=classes)], [] parentSymbol = self.env.temp_data.get('cpp:parent_symbol', None) if parentSymbol is None: parentSymbol = self.env.domaindata['cpp']['root_symbol'] # ...most if not all of these classes should really apply to the individual references, # not the container node signode = self.node_type(classes=classes) ast.describe_signature(signode, 'markType', self.env, parentSymbol) return [signode], [] class CPPDomain(Domain): """C++ language domain. There are two 'object type' attributes being used:: - Each object created from directives gets an assigned .objtype from ObjectDescription.run. This is simply the directive name. - Each declaration (see the distinction in the directives dict below) has a nested .ast of type ASTDeclaration. That object has .objectType which corresponds to the keys in the object_types dict below. They are the core different types of declarations in C++ that one can document. """ name = 'cpp' label = 'C++' object_types = { 'class': ObjType(_('class'), 'class', 'type', 'identifier'), 'union': ObjType(_('union'), 'union', 'type', 'identifier'), 'function': ObjType(_('function'), 'function', 'func', 'type', 'identifier'), 'member': ObjType(_('member'), 'member', 'var'), 'type': ObjType(_('type'), 'type', 'identifier'), 'concept': ObjType(_('concept'), 'concept', 'identifier'), 'enum': ObjType(_('enum'), 'enum', 'type', 'identifier'), 'enumerator': ObjType(_('enumerator'), 'enumerator') } directives = { # declarations 'class': CPPClassObject, 'struct': CPPClassObject, 'union': CPPUnionObject, 'function': CPPFunctionObject, 'member': CPPMemberObject, 'var': CPPMemberObject, 'type': CPPTypeObject, 'concept': CPPConceptObject, 'enum': CPPEnumObject, 'enum-struct': CPPEnumObject, 'enum-class': CPPEnumObject, 'enumerator': CPPEnumeratorObject, # scope control 'namespace': CPPNamespaceObject, 'namespace-push': CPPNamespacePushObject, 'namespace-pop': CPPNamespacePopObject, # other 'alias': CPPAliasObject } roles = { 'any': CPPXRefRole(), 'class': CPPXRefRole(), 'struct': CPPXRefRole(), 'union': CPPXRefRole(), 'func': CPPXRefRole(fix_parens=True), 'member': CPPXRefRole(), 'var': CPPXRefRole(), 'type': CPPXRefRole(), 'concept': CPPXRefRole(), 'enum': CPPXRefRole(), 'enumerator': CPPXRefRole(), 'expr': CPPExprRole(asCode=True), 'texpr': CPPExprRole(asCode=False) } initial_data = { 'root_symbol': Symbol(None, None, None, None, None, None), 'names': {} # full name for indexing -> docname } def clear_doc(self, docname: str) -> None: if Symbol.debug_show_tree: print("clear_doc:", docname) print("\tbefore:") print(self.data['root_symbol'].dump(1)) print("\tbefore end") rootSymbol = self.data['root_symbol'] rootSymbol.clear_doc(docname) if Symbol.debug_show_tree: print("\tafter:") print(self.data['root_symbol'].dump(1)) print("\tafter end") print("clear_doc end:", docname) for name, nDocname in list(self.data['names'].items()): if nDocname == docname: del self.data['names'][name] def process_doc(self, env: BuildEnvironment, docname: str, document: nodes.document) -> None: if Symbol.debug_show_tree: print("process_doc:", docname) print(self.data['root_symbol'].dump(0)) print("process_doc end:", docname) def process_field_xref(self, pnode: pending_xref) -> None: pnode.attributes.update(self.env.ref_context) def merge_domaindata(self, docnames: List[str], otherdata: Dict) -> None: if Symbol.debug_show_tree: print("merge_domaindata:") print("\tself:") print(self.data['root_symbol'].dump(1)) print("\tself end") print("\tother:") print(otherdata['root_symbol'].dump(1)) print("\tother end") self.data['root_symbol'].merge_with(otherdata['root_symbol'], docnames, self.env) ourNames = self.data['names'] for name, docname in otherdata['names'].items(): if docname in docnames: if name in ourNames: msg = __("Duplicate declaration, also defined in '%s'.\n" "Name of declaration is '%s'.") msg = msg % (ourNames[name], name) logger.warning(msg, location=docname) else: ourNames[name] = docname if Symbol.debug_show_tree: print("\tresult:") print(self.data['root_symbol'].dump(1)) print("\tresult end") print("merge_domaindata end") def _resolve_xref_inner(self, env: BuildEnvironment, fromdocname: str, builder: Builder, typ: str, target: str, node: pending_xref, contnode: Element) -> Tuple[Element, str]: # add parens again for those that could be functions if typ == 'any' or typ == 'func': target += '()' parser = DefinitionParser(target, location=node, config=env.config) try: ast, isShorthand = parser.parse_xref_object() except DefinitionError as e: # as arg to stop flake8 from complaining def findWarning(e: Exception) -> Tuple[str, Exception]: if typ != 'any' and typ != 'func': return target, e # hax on top of the paren hax to try to get correct errors parser2 = DefinitionParser(target[:-2], location=node, config=env.config) try: parser2.parse_xref_object() except DefinitionError as e2: return target[:-2], e2 # strange, that we don't get the error now, use the original return target, e t, ex = findWarning(e) logger.warning('Unparseable C++ cross-reference: %r\n%s', t, ex, location=node) return None, None parentKey = node.get("cpp:parent_key", None) # type: LookupKey rootSymbol = self.data['root_symbol'] if parentKey: parentSymbol = rootSymbol.direct_lookup(parentKey) # type: Symbol if not parentSymbol: print("Target: ", target) print("ParentKey: ", parentKey.data) print(rootSymbol.dump(1)) assert parentSymbol # should be there else: parentSymbol = rootSymbol if isShorthand: assert isinstance(ast, ASTNamespace) ns = ast name = ns.nestedName if ns.templatePrefix: templateDecls = ns.templatePrefix.templates else: templateDecls = [] # let's be conservative with the sibling lookup for now searchInSiblings = (not name.rooted) and len(name.names) == 1 symbols, failReason = parentSymbol.find_name( name, templateDecls, typ, templateShorthand=True, matchSelf=True, recurseInAnon=True, searchInSiblings=searchInSiblings) if symbols is None: if typ == 'identifier': if failReason == 'templateParamInQualified': # this is an xref we created as part of a signature, # so don't warn for names nested in template parameters raise NoUri(str(name), typ) s = None else: # just refer to the arbitrarily first symbol s = symbols[0] else: assert isinstance(ast, ASTDeclaration) decl = ast name = decl.name s = parentSymbol.find_declaration(decl, typ, templateShorthand=True, matchSelf=True, recurseInAnon=True) if s is None or s.declaration is None: txtName = str(name) if txtName.startswith('std::') or txtName == 'std': raise NoUri(txtName, typ) return None, None if typ.startswith('cpp:'): typ = typ[4:] origTyp = typ if typ == 'func': typ = 'function' if typ == 'struct': typ = 'class' declTyp = s.declaration.objectType def checkType() -> bool: if typ == 'any' or typ == 'identifier': return True if declTyp == 'templateParam': # TODO: perhaps this should be strengthened one day return True if declTyp == 'functionParam': if typ == 'var' or typ == 'member': return True objtypes = self.objtypes_for_role(typ) if objtypes: return declTyp in objtypes print("Type is %s (originally: %s), declType is %s" % (typ, origTyp, declTyp)) assert False if not checkType(): logger.warning("cpp:%s targets a %s (%s).", origTyp, s.declaration.objectType, s.get_full_nested_name(), location=node) declaration = s.declaration if isShorthand: fullNestedName = s.get_full_nested_name() displayName = fullNestedName.get_display_string().lstrip(':') else: displayName = decl.get_display_string() docname = s.docname assert docname # the non-identifier refs are cross-references, which should be processed: # - fix parenthesis due to operator() and add_function_parentheses if typ != "identifier": title = contnode.pop(0).astext() # If it's operator(), we need to add '()' if explicit function parens # are requested. Then the Sphinx machinery will add another pair. # Also, if it's an 'any' ref that resolves to a function, we need to add # parens as well. # However, if it's a non-shorthand function ref, for a function that # takes no arguments, then we may need to add parens again as well. addParen = 0 if not node.get('refexplicit', False) and declaration.objectType == 'function': if isShorthand: # this is just the normal haxing for 'any' roles if env.config.add_function_parentheses and typ == 'any': addParen += 1 # and now this stuff for operator() if (env.config.add_function_parentheses and typ == 'function' and title.endswith('operator()')): addParen += 1 if ((typ == 'any' or typ == 'function') and title.endswith('operator') and displayName.endswith('operator()')): addParen += 1 else: # our job here is to essentially nullify add_function_parentheses if env.config.add_function_parentheses: if typ == 'any' and displayName.endswith('()'): addParen += 1 elif typ == 'function': if title.endswith('()') and not displayName.endswith('()'): title = title[:-2] else: if displayName.endswith('()'): addParen += 1 if addParen > 0: title += '()' * addParen # and reconstruct the title again contnode += nodes.Text(title) return make_refnode(builder, fromdocname, docname, declaration.get_newest_id(), contnode, displayName ), declaration.objectType def resolve_xref(self, env: BuildEnvironment, fromdocname: str, builder: Builder, typ: str, target: str, node: pending_xref, contnode: Element ) -> Element: return self._resolve_xref_inner(env, fromdocname, builder, typ, target, node, contnode)[0] def resolve_any_xref(self, env: BuildEnvironment, fromdocname: str, builder: Builder, target: str, node: pending_xref, contnode: Element ) -> List[Tuple[str, Element]]: with logging.suppress_logging(): retnode, objtype = self._resolve_xref_inner(env, fromdocname, builder, 'any', target, node, contnode) if retnode: if objtype == 'templateParam': return [('cpp:templateParam', retnode)] else: return [('cpp:' + self.role_for_objtype(objtype), retnode)] return [] def get_objects(self) -> Iterator[Tuple[str, str, str, str, str, int]]: rootSymbol = self.data['root_symbol'] for symbol in rootSymbol.get_all_symbols(): if symbol.declaration is None: continue assert symbol.docname fullNestedName = symbol.get_full_nested_name() name = str(fullNestedName).lstrip(':') dispname = fullNestedName.get_display_string().lstrip(':') objectType = symbol.declaration.objectType docname = symbol.docname newestId = symbol.declaration.get_newest_id() yield (name, dispname, objectType, docname, newestId, 1) def get_full_qualified_name(self, node: Element) -> str: target = node.get('reftarget', None) if target is None: return None parentKey = node.get("cpp:parent_key", None) # type: LookupKey if parentKey is None or len(parentKey.data) <= 0: return None rootSymbol = self.data['root_symbol'] parentSymbol = rootSymbol.direct_lookup(parentKey) parentName = parentSymbol.get_full_nested_name() return '::'.join([str(parentName), target]) def setup(app: Sphinx) -> Dict[str, Any]: app.add_domain(CPPDomain) app.add_config_value("cpp_index_common_prefix", [], 'env') app.add_config_value("cpp_id_attributes", [], 'env') app.add_config_value("cpp_paren_attributes", [], 'env') app.add_post_transform(AliasTransform) return { 'version': 'builtin', 'env_version': 2, 'parallel_read_safe': True, 'parallel_write_safe': True, }
sphinx/domains/cpp.py
288,143
C++ language domain. There are two 'object type' attributes being used:: - Each object created from directives gets an assigned .objtype from ObjectDescription.run. This is simply the directive name. - Each declaration (see the distinction in the directives dict below) has a nested .ast of type ASTDeclaration. That object has .objectType which corresponds to the keys in the object_types dict below. They are the core different types of declarations in C++ that one can document. This directive is just to tell Sphinx that we're documenting stuff in namespace foo. Description of a C++ language object. Just parse the simple ones. named=False|'maybe'|True: 'maybe' is e.g., for function objects which doesn't need to name the arguments outer == operatorCast: annoying case, we should not take the params On purpose this doesn't call the ObjectDescription version, but is based on it. Each alias signature may expand into multiple real signatures (an overload set). The code is therefore based on the ObjectDescription version. sphinx.domains.cpp ~~~~~~~~~~~~~~~~~~ The C++ language domain. :copyright: Copyright 2007-2020 by the Sphinx team, see AUTHORS. :license: BSD, see LICENSE for details. see https://en.cppreference.com/w/cpp/keyword Ids are used in lookup keys which are used across pickled files, so when _max_id changes, make sure to update the ENV_VERSION. ------------------------------------------------------------------------------ Id v1 constants ------------------------------------------------------------------------------ the arguments will make the difference between unary and binary '+(unary)' : 'ps', '-(unary)' : 'ng', '&(unary)' : 'ad', '*(unary)' : 'de', ------------------------------------------------------------------------------ Id v > 1 constants ------------------------------------------------------------------------------ not all of these are actually parsed as fundamental types, TODO: do that the arguments will make the difference between unary and binary in operator definitions '+(unary)' : 'ps', '-(unary)' : 'ng', '&(unary)' : 'ad', '*(unary)' : 'de', this one is not overloadable, but we need it for expressions type: Dict[Any, str] these are ordered by preceedence Names a destructor, just use an arbitrary version of dtors and this is where we finally make a difference between __str__ and the display string just print the name part, with template args, not template params Each element should be a pending xref targeting the complete prefix. however, only the identifier part should be a link, such that template args can be a link as well. For 'lastIsName' we should also prepend template parameter lists. type: List[Any] If lastIsName, then wrap all of the prefix in a desc_addname, else append directly to signode. NOTE: Breathe relies on the prefix being in the desc_addname node, so it can remove it in inner declarations. Expressions Primary expressions note: the length is not really correct with escaping may be None when no prefix https://github.com/itanium-cxx-abi/cxx-abi/pull/67 (... op expr) (expr op ...) (expr op ... op expr) we don't check where the parameter pack is, we just always call this a binary left fold note: this class is basically to cast a nested name as an expression Postfix expressions Unary expressions TODO: placement the array part will be in the type mangling, so na is not used TODO: placement TODO: placement Other expressions Types Things for ASTNestedName Main part of declarations this is not part of the normal name mangling in C++ the anchor will be our parent else, do the usual i.e., outer-function params type: List[str] leftSpecs and rightSpecs are used for output allSpecs are used for id generation type: List[str] Declarator only the modifiers for a function, e.g., cv-qualifiers only the parameters (if any) only the array specifiers TOOD: can we actually have both array ops and paramQual? ------------------------------------------------------------------------ only the parameters (if any) only the array specifiers ------------------------------------------------------------------------ ReturnType *next, so we are part of the return type of 'next only the parameters (if any) ReturnType &next, so we are part of the return type of 'next only the parameters (if any) ReturnType... next, so we are part of the return type of 'next only the parameters (if any) ReturnType name::* next, so we are part of the return type of next TODO: we assume the name, params, and qualifiers are in inner only the parameters (if any) TODO: was this implemented before? ReturnType (inner)next, so 'inner' returns everything outside Type and initializer stuff needs the name also modifiers just the name only type encoding other versions needs the name also modifiers with templates we need to mangle the return type in as well just the name only type encoding the 'returnType' of a non-function type is simply just the last type, i.e., for 'int*' it is 'int' for parameters that don't really declare new names we get 'markType', this should not be propagated, but be 'noneIsName'. this is not part of the normal name mangling in C++ the anchor will be our parent Other declarations self.scoped has been done by the CPPEnumObject Templates Parameters this is not part of the normal name mangling in C++ we need to put something this is not part of the normal name mangling in C++ the anchor will be our parent this is not part of the normal name mangling in C++ the anchor will be our parent this is not part of the normal name mangling in C++ the anchor will be our parent Template introducers this is not part of the normal name mangling in C++ the anchor will be our parent we need to put something used for the implicit requires clause first do the same as a normal template parameter list let's use X expr E, which is otherwise for constant template args Note: 'lineSpec' has no effect on template introductions. templates is None means it's an explicit instantiation of a variable this is not part of a normal name mangling system type: Symbol set by CPPObject._add_enumerator_to_parent type: Symbol version >= 2 The caller of the domain added a desc_signature node. Always enable multiline: Put each line in a desc_signature_line node. wrong directiveType used parent == None means global scope, so declaration means a parent declarations in a single directive are linked together type: Symbol type: Symbol template<templateParams> identifier<templateArgs> Remember to modify Symbol.remove if modifications to the parent change. type: List[Symbol] type: List[Symbol] note: _children includes _anonChildren Do symbol addition after self._children has been initialised. and symbol addition should be done as well Note: we may be called from _fill_empty, so the symbols we want to add may actually already be present (as empty symbols). add symbols for the template params only add a declaration if we our self are from a declaration add symbols for function parameters, if any (comparing to the template params: we have checked that we are a declaration) type: List[Symbol] The pickle files for the environment and for each document are distinct. The environment has all the symbols, but the documents has xrefs that must know their scope. A lookup key is essentially a specification of how to find a specific symbol. the names of the template parameters must be given exactly as args and params that are packs must in the args be the name expanded having no template params and no arguments is also a specialization TODO: doing this by string manipulation is probably not the most efficient If both are given, but it's not a specialization, then do lookup as if there is no argument list. For example: template<typename T> int A<T>::var; we query with params, they must match params we don't query with params, and we do care about them TODO: do better comparison TODO: do better comparison NOQA ancestorLookupType: if not None, specifies the target type of the lookup Each template argument list must have a template parameter list. But to declare a template there must be an additional template parameter list. find the right starting point for lookup walk up until we find the first identifier if we are in the scope of a constructor but wants to reference the class we need to walk one extra up and now the actual lookup there must be a parameter list take the next template parameter list if there is one otherwise it's ok We have now matched part of a nested name, and need to match more so even if we should matchSelf before, we definitely shouldn't even more. (see also issue 2666) handle the last name type: ignore Used for adding a whole path of symbols, where the last may or may not be an actual declaration. we create symbols all the way, so that can't happen good, just a scope creation TODO: what if we have more than one symbol? With partial builds we may start with a large symbol tree stripped of declarations. Essentially any combination of noDecl, withDecl, and dupDecls seems possible. TODO: make partial builds fully work. What should happen when the primary symbol gets deleted, and other duplicates exist? The full document should probably be rebuild. First check if one of those with a declaration matches. If it's a function, we need to compare IDs, otherwise there should be only one symbol with a declaration. Redeclaration of the same symbol. Let the new one be there, but raise an error to the client so it can use the real symbol as subscope. This will probably result in a duplicate id warning. (not reachable) a function, so compare IDs (not reachable) no candidate symbol found with matching ID if there is an empty symbol, fill that one NOQA NOQA assert len(noDecl) == 1 TODO: enable assertion when we at some point find out how to do cleanup for now, just take the first one, it should work fine ... right? If someone first opened the scope, and then later declares it, e.g, .. namespace:: Test .. namespace:: nullptr .. class:: Test TODO: hmm, should we prune by docnames? if in the end we have non matching, but have an empty one, then just continue with that Both have declarations, and in the same docname. This can apparently happen, it should be safe to just ignore it, right? Hmm, only on duplicate declarations, right? templateShorthand: missing template parameter lists for templates is ok If the first component is None, then the second component _may_ be a string explaining why. TODO: Maybe search without template args? Though, the correctPrimaryTemplateArgs does that for primary templates. Is there another case where it would be good? if it was a part of the qualification that could not be found try without template params and args templateShorthand: missing template parameter lists for templates is ok those without signedness and size modifiers see https://en.cppreference.com/w/cpp/language/types -> integer-literal | character-literal | floating-literal | string-literal | boolean-literal -> "false" | "true" | pointer-literal -> "nullptr" | user-defined-literal character-literal may be None when no prefix TODO: user-defined lit "(" expression ")" fold-expression -> ( cast-expression fold-operator ... ) | ( ... fold-operator cast-expression ) | ( cast-expression fold-operator ... fold-operator cast-expression ( ... fold-operator cast-expression ) try first parsing a unary right fold, or a binary fold fall back to a paren expression now it definitely is a fold expression literal "this" lambda-expression "(" expression ")" fold-expression id-expression -> we parse this with _parse_nested_name type: ASTExpression TODO: try lambda expression Parse open and close with the actual initializer-list inbetween -> initializer-clause '...'[opt] | initializer-list ',' initializer-clause '...'[opt] type: List[Union[ASTExpression, ASTBracedInitList]] -> '(' expression-list ')' though, we relax it to also allow empty parens as it's needed in some cases expression-list -> initializer-list -> '{' initializer-list ','[opt] '}' | '{' '}' -> primary | postfix "[" expression "]" | postfix "[" braced-init-list [opt] "]" | postfix "(" expression-list [opt] ")" | postfix "." "template" [opt] id-expression | postfix "->" "template" [opt] id-expression | postfix "." pseudo-destructor-name | postfix "->" pseudo-destructor-name | postfix "++" | postfix "--" | simple-type-specifier "(" expression-list [opt] ")" | simple-type-specifier braced-init-list | typename-specifier "(" expression-list [opt] ")" | typename-specifier braced-init-list | "dynamic_cast" "<" type-id ">" "(" expression ")" | "static_cast" "<" type-id ">" "(" expression ")" | "reinterpret_cast" "<" type-id ">" "(" expression ")" | "const_cast" "<" type-id ">" "(" expression ")" | "typeid" "(" expression ")" | "typeid" "(" type-id ")" type: Any a primary expression or a type we are potentially casting, so save parens for us TODO: hmm, would we need to try both with operatorCast and with None? | simple-type-specifier "(" expression-list [opt] ")" | simple-type-specifier braced-init-list | typename-specifier "(" expression-list [opt] ")" | typename-specifier braced-init-list and now parse postfixes type: List[ASTPostfixOp] don't steal the dot don't steal the dot don't steal the arrow -> postfix | "++" cast | "--" cast | unary-operator cast -> (* | & | + | - | ! | ~) cast The rest: | "sizeof" unary | "sizeof" "(" type-id ")" | "sizeof" "..." "(" identifier ")" | "alignof" "(" type-id ")" | noexcept-expression -> noexcept "(" expression ")" | new-expression | delete-expression TODO: hmm, should we be able to backtrack here? new-expression new-placement[opt] new-type-id new-initializer[opt] new-placement[opt] ( type-id ) new-initializer[opt] either this is a new-placement or it's the second production without placement, and it's actually the ( type-id ) part set isNewTypeId = False if it's (type-id) delete-expression -> unary | "(" type-id ")" cast logical-or = logical-and || logical-and = inclusive-or && inclusive-or = exclusive-or | exclusive-or = and ^ and = equality & equality = relational ==, != relational = shift <, >, <=, >= shift = additive <<, >> additive = multiplicative +, - multiplicative = pm *, /, % pm = cast .*, ->* don't split the && 'token' and btw. && has lower precedence, so we are done -> "?" expression ":" assignment-expression -> conditional-expression | logical-or-expression assignment-operator initializer-clause | throw-expression TODO: parse throw-expression: "throw" assignment-expression [opt] if not a throw expression, then: -> conditional-expression -> logical-or-expression | logical-or-expression "?" expression ":" assignment-expression | logical-or-expression assignment-operator initializer-clause type: List[Union[ASTExpression, ASTBracedInitList]] TODO: handle ternary with _parse_conditional_expression_tail -> conditional-expression TODO: use _parse_conditional_expression_tail -> assignment-expression | expression "," assignment-expresion Stupidly "parse" an expression. 'end' should be a list of characters which ends the expression. first try to use the provided parser some places (e.g., template parameters) we really don't want to use fallback, and for testing we may want to globally disable it and then the fallback scanning TODO: add handling of more bracket-like things, and quote handling type: List[str] ========================================================================== adapted from the old code yay, a regular operator definition new/delete operator? user-defined literal? oh well, looks like a cast operator definition. In that case, eat another type. template-argument-list: (but we include the < and > here template-argument ...[opt] template-argument-list, template-argument ...[opt] template-argument: constant-expression type-id id-expression type: List[Union[ASTType, ASTTemplateArgConstant]] type: List[ASTNestedNameElement] type: List[bool] type: Union[ASTIdentifier, ASTOperator] make sure there isn't a keyword try greedily to get template arguments, but otherwise a < might be because we are in an expression ========================================================================== fundemental types TODO: this could/should be more strict decltype prefixed note: it seems that function arguments can always be named, even in function pointers and similar. TODO: parse default parameters TODO: didn't we just do that? TODO: why did we have this bail-out? does it hurt to parse the extra stuff? it's needed for pointer to member functions the can be permuted they can be permuted accept any permutation of a subset of some decl-specs function-specifiers now we should parse the name, and then suffixes if it's a member pointer, we may have '::', which should be an error perhaps a bit-field 'typed' here means 'parse return type stuff' TODO: shouldn't we parse an R-value ref here first? note: peeking, not skipping TODO: we should be able to parse cast operators which return function pointers. For now, just hax it and ignore. maybe this is the beginning of params and quals,try that first, otherwise assume it's noptr->declarator > ( ptr-declarator ) assume this is params and quals TODO: hmm, if there is a name, it must be in inner, right? TODO: hmm, if there must be parameters, they must be inside, right? pointer to member this is a heuristic for error messages, for when there is a < after a nested name, but it was not a successful template argument list initializer global vars -> brace-or-equal-initializer | '(' expression-list ')' brace-or-equal-initializer member vars -> '=' initializer-clause | braced-init-list initializer-clause function params, non-type template params (with '=' in front) -> assignment-expression | braced-init-list we don't distinguish between global and member vars, so disallow paren: -> braced-init-list var only | '=' assignment-expression | '=' braced-init-list type: List[str] function parameter always named We allow type objects to just be a name. Some functions don't have normal return types: constructors, destrutors, cast operators first try without the type Retain the else branch for easier debugging. TODO: it would be nice to save the previous stacktrace and output it here. For testing purposes. do it again to get the proper traceback (how do you reliably save a traceback when an exception is constructed?) i.e., member it could also be a constrained type parameter, e.g., C T = int& note: init may be None if there is no = we parsed an expression, so we must have a , or a >, otherwise the expression didn't get everything pretend it didn't happen we assume that it was indeed an expression is set by CPPEnumObject ========================================================================== only: '<' parameter-list '>' we assume that 'template' has just been parsed type: List[ASTTemplateParam] declare a tenplate template parameter declare a type or template type parameter template type type declare a non-type parameter, or constrained type parameter for sure it must be a template introduction now make sure there isn't a keyword type: List[Union[ASTTemplateParams, ASTTemplateIntroduction]] the saved position is only used to provide a better error message type: Union[ASTTemplateParams, ASTTemplateIntroduction] type: List[Union[ASTTemplateParams, ASTTemplateIntroduction]] type: Any type: ignore if there are '()' left, just skip them type: ignore if there are '()' left, just skip them find the parent, if it exists && is an enum && it's unscoped, then add the name to the parent scope TODO: we could warn, but it is somewhat equivalent to unscoped enums, without the enum no parent the parent is not explicitly declared TODO: we could warn, but it could be a style to just assume enumerator parents to be scoped TODO: maybe issue a warning, enumerators in non-enums is weird, but it is somewhat equivalent to unscoped enums, without the enum something is already declared with that name general note: name must be lstrip(':')'ed, to remove "::" let's keep the newest first shouldn't be None Add index entry, but not if it's a declaration inside a concept if the name is not unique, the first one will win always add the newest id only add compatibility ids when there are no conflicts is None when the element didn't exist in that version from ObjectDescription.run The lookup keys assume that no nested scopes exists inside overloaded functions. (see also 5191) Example: .. cpp:function:: void f(int) .. cpp:function:: void f(double) .. cpp:function:: void g() :cpp:any:`boom` So we disallow any signatures inside functions. When multiple declarations are made in the same directive they need to know about each other to provide symbol lookup for function parameters. We use last_symbol to store the latest added declaration in a directive. It is easier to assume some phony name than handling the error in the possibly inner declarations. append the new declaration to the sibling list Assume we are actually in the old symbol, instead of the newly created duplicate. note: handle_signature may be called multiple time per directive, if it has multiple signatures, so don't mess with the original options. type: Symbol type: LookupKey the distinction between class and struct is only cosmetic type: Dict type: List[Symbol] type: Dict type: Dict type: ignore could not be parsed, so stop here type: Symbol type: Symbol should be there type: List[Symbol] type: Dict 'desctype' is a backwards compatible attribute type: List[str] major hax: replace anon names via simple string manipulation. Can this actually fail? Assume the removal part of fix_parens for :any: refs. The addition part is done with the reference is resolved. TODO: should this really be here? only has a meaning for the title if the first character is a tilde, don't display the module/class parts of the contents render the expression as inline code type: Type[TextElement] render the expression as inline text attempt to mimic XRefRole classes, except that... see below ...most if not all of these classes should really apply to the individual references, not the container node declarations scope control other full name for indexing -> docname add parens again for those that could be functions as arg to stop flake8 from complaining hax on top of the paren hax to try to get correct errors strange, that we don't get the error now, use the original type: LookupKey type: Symbol should be there let's be conservative with the sibling lookup for now this is an xref we created as part of a signature, so don't warn for names nested in template parameters just refer to the arbitrarily first symbol TODO: perhaps this should be strengthened one day the non-identifier refs are cross-references, which should be processed: - fix parenthesis due to operator() and add_function_parentheses If it's operator(), we need to add '()' if explicit function parens are requested. Then the Sphinx machinery will add another pair. Also, if it's an 'any' ref that resolves to a function, we need to add parens as well. However, if it's a non-shorthand function ref, for a function that takes no arguments, then we may need to add parens again as well. this is just the normal haxing for 'any' roles and now this stuff for operator() our job here is to essentially nullify add_function_parentheses and reconstruct the title again type: LookupKey
23,673
en
0.795698
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi SDK Generator. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union from ... import _utilities, _tables from . import outputs __all__ = ['WebAppSitePushSettingsSlot'] class WebAppSitePushSettingsSlot(pulumi.CustomResource): def __init__(__self__, resource_name: str, opts: Optional[pulumi.ResourceOptions] = None, dynamic_tags_json: Optional[pulumi.Input[str]] = None, is_push_enabled: Optional[pulumi.Input[bool]] = None, kind: Optional[pulumi.Input[str]] = None, name: Optional[pulumi.Input[str]] = None, resource_group_name: Optional[pulumi.Input[str]] = None, slot: Optional[pulumi.Input[str]] = None, tag_whitelist_json: Optional[pulumi.Input[str]] = None, tags_requiring_auth: Optional[pulumi.Input[str]] = None, __props__=None, __name__=None, __opts__=None): """ Push settings for the App. :param str resource_name: The name of the resource. :param pulumi.ResourceOptions opts: Options for the resource. :param pulumi.Input[str] dynamic_tags_json: Gets or sets a JSON string containing a list of dynamic tags that will be evaluated from user claims in the push registration endpoint. :param pulumi.Input[bool] is_push_enabled: Gets or sets a flag indicating whether the Push endpoint is enabled. :param pulumi.Input[str] kind: Kind of resource. :param pulumi.Input[str] name: Name of web app. :param pulumi.Input[str] resource_group_name: Name of the resource group to which the resource belongs. :param pulumi.Input[str] slot: Name of web app slot. If not specified then will default to production slot. :param pulumi.Input[str] tag_whitelist_json: Gets or sets a JSON string containing a list of tags that are whitelisted for use by the push registration endpoint. :param pulumi.Input[str] tags_requiring_auth: Gets or sets a JSON string containing a list of tags that require user authentication to be used in the push registration endpoint. Tags can consist of alphanumeric characters and the following: '_', '@', '#', '.', ':', '-'. Validation should be performed at the PushRequestHandler. """ if __name__ is not None: warnings.warn("explicit use of __name__ is deprecated", DeprecationWarning) resource_name = __name__ if __opts__ is not None: warnings.warn("explicit use of __opts__ is deprecated, use 'opts' instead", DeprecationWarning) opts = __opts__ if opts is None: opts = pulumi.ResourceOptions() if not isinstance(opts, pulumi.ResourceOptions): raise TypeError('Expected resource options to be a ResourceOptions instance') if opts.version is None: opts.version = _utilities.get_version() if opts.id is None: if __props__ is not None: raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = dict() __props__['dynamic_tags_json'] = dynamic_tags_json if is_push_enabled is None and not opts.urn: raise TypeError("Missing required property 'is_push_enabled'") __props__['is_push_enabled'] = is_push_enabled __props__['kind'] = kind if name is None and not opts.urn: raise TypeError("Missing required property 'name'") __props__['name'] = name if resource_group_name is None and not opts.urn: raise TypeError("Missing required property 'resource_group_name'") __props__['resource_group_name'] = resource_group_name if slot is None and not opts.urn: raise TypeError("Missing required property 'slot'") __props__['slot'] = slot __props__['tag_whitelist_json'] = tag_whitelist_json __props__['tags_requiring_auth'] = tags_requiring_auth __props__['system_data'] = None __props__['type'] = None alias_opts = pulumi.ResourceOptions(aliases=[pulumi.Alias(type_="azure-nextgen:web/v20200901:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-native:web:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-native:web/latest:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/latest:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20160801:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20160801:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20180201:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20180201:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20181101:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20181101:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20190801:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20190801:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20200601:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20200601:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-native:web/v20201001:WebAppSitePushSettingsSlot"), pulumi.Alias(type_="azure-nextgen:web/v20201001:WebAppSitePushSettingsSlot")]) opts = pulumi.ResourceOptions.merge(opts, alias_opts) super(WebAppSitePushSettingsSlot, __self__).__init__( 'azure-native:web/v20200901:WebAppSitePushSettingsSlot', resource_name, __props__, opts) @staticmethod def get(resource_name: str, id: pulumi.Input[str], opts: Optional[pulumi.ResourceOptions] = None) -> 'WebAppSitePushSettingsSlot': """ Get an existing WebAppSitePushSettingsSlot resource's state with the given name, id, and optional extra properties used to qualify the lookup. :param str resource_name: The unique name of the resulting resource. :param pulumi.Input[str] id: The unique provider ID of the resource to lookup. :param pulumi.ResourceOptions opts: Options for the resource. """ opts = pulumi.ResourceOptions.merge(opts, pulumi.ResourceOptions(id=id)) __props__ = dict() __props__["dynamic_tags_json"] = None __props__["is_push_enabled"] = None __props__["kind"] = None __props__["name"] = None __props__["system_data"] = None __props__["tag_whitelist_json"] = None __props__["tags_requiring_auth"] = None __props__["type"] = None return WebAppSitePushSettingsSlot(resource_name, opts=opts, __props__=__props__) @property @pulumi.getter(name="dynamicTagsJson") def dynamic_tags_json(self) -> pulumi.Output[Optional[str]]: """ Gets or sets a JSON string containing a list of dynamic tags that will be evaluated from user claims in the push registration endpoint. """ return pulumi.get(self, "dynamic_tags_json") @property @pulumi.getter(name="isPushEnabled") def is_push_enabled(self) -> pulumi.Output[bool]: """ Gets or sets a flag indicating whether the Push endpoint is enabled. """ return pulumi.get(self, "is_push_enabled") @property @pulumi.getter def kind(self) -> pulumi.Output[Optional[str]]: """ Kind of resource. """ return pulumi.get(self, "kind") @property @pulumi.getter def name(self) -> pulumi.Output[str]: """ Resource Name. """ return pulumi.get(self, "name") @property @pulumi.getter(name="systemData") def system_data(self) -> pulumi.Output['outputs.SystemDataResponse']: """ The system metadata relating to this resource. """ return pulumi.get(self, "system_data") @property @pulumi.getter(name="tagWhitelistJson") def tag_whitelist_json(self) -> pulumi.Output[Optional[str]]: """ Gets or sets a JSON string containing a list of tags that are whitelisted for use by the push registration endpoint. """ return pulumi.get(self, "tag_whitelist_json") @property @pulumi.getter(name="tagsRequiringAuth") def tags_requiring_auth(self) -> pulumi.Output[Optional[str]]: """ Gets or sets a JSON string containing a list of tags that require user authentication to be used in the push registration endpoint. Tags can consist of alphanumeric characters and the following: '_', '@', '#', '.', ':', '-'. Validation should be performed at the PushRequestHandler. """ return pulumi.get(self, "tags_requiring_auth") @property @pulumi.getter def type(self) -> pulumi.Output[str]: """ Resource type. """ return pulumi.get(self, "type") def translate_output_property(self, prop): return _tables.CAMEL_TO_SNAKE_CASE_TABLE.get(prop) or prop def translate_input_property(self, prop): return _tables.SNAKE_TO_CAMEL_CASE_TABLE.get(prop) or prop
sdk/python/pulumi_azure_native/web/v20200901/web_app_site_push_settings_slot.py
9,647
Push settings for the App. :param str resource_name: The name of the resource. :param pulumi.ResourceOptions opts: Options for the resource. :param pulumi.Input[str] dynamic_tags_json: Gets or sets a JSON string containing a list of dynamic tags that will be evaluated from user claims in the push registration endpoint. :param pulumi.Input[bool] is_push_enabled: Gets or sets a flag indicating whether the Push endpoint is enabled. :param pulumi.Input[str] kind: Kind of resource. :param pulumi.Input[str] name: Name of web app. :param pulumi.Input[str] resource_group_name: Name of the resource group to which the resource belongs. :param pulumi.Input[str] slot: Name of web app slot. If not specified then will default to production slot. :param pulumi.Input[str] tag_whitelist_json: Gets or sets a JSON string containing a list of tags that are whitelisted for use by the push registration endpoint. :param pulumi.Input[str] tags_requiring_auth: Gets or sets a JSON string containing a list of tags that require user authentication to be used in the push registration endpoint. Tags can consist of alphanumeric characters and the following: '_', '@', '#', '.', ':', '-'. Validation should be performed at the PushRequestHandler. Gets or sets a JSON string containing a list of dynamic tags that will be evaluated from user claims in the push registration endpoint. Get an existing WebAppSitePushSettingsSlot resource's state with the given name, id, and optional extra properties used to qualify the lookup. :param str resource_name: The unique name of the resulting resource. :param pulumi.Input[str] id: The unique provider ID of the resource to lookup. :param pulumi.ResourceOptions opts: Options for the resource. Gets or sets a flag indicating whether the Push endpoint is enabled. Kind of resource. Resource Name. The system metadata relating to this resource. Gets or sets a JSON string containing a list of tags that are whitelisted for use by the push registration endpoint. Gets or sets a JSON string containing a list of tags that require user authentication to be used in the push registration endpoint. Tags can consist of alphanumeric characters and the following: '_', '@', '#', '.', ':', '-'. Validation should be performed at the PushRequestHandler. Resource type. coding=utf-8 *** WARNING: this file was generated by the Pulumi SDK Generator. *** *** Do not edit by hand unless you're certain you know what you are doing! ***
2,474
en
0.659273
""" This software was developed by the University of Tennessee as part of the Distributed Data Analysis of Neutron Scattering Experiments (DANSE) project funded by the US National Science Foundation. See the license text in license.txt copyright 2008, 2009, University of Tennessee """ import wx import sys from sas.sasgui.guiframe.panel_base import PanelBase from sas.sascalc.calculator.kiessig_calculator import KiessigThicknessCalculator from calculator_widgets import OutputTextCtrl from calculator_widgets import InputTextCtrl from sas.sasgui.perspectives.calculator import calculator_widgets as widget from sas.sasgui.guiframe.documentation_window import DocumentationWindow _BOX_WIDTH = 77 #Slit length panel size if sys.platform.count("win32") > 0: PANEL_TOP = 0 PANEL_WIDTH = 500 PANEL_HEIGHT = 230 FONT_VARIANT = 0 else: PANEL_TOP = 60 PANEL_WIDTH = 560 PANEL_HEIGHT = 230 FONT_VARIANT = 1 class KiessigThicknessCalculatorPanel(wx.Panel, PanelBase): """ Provides the Kiessig thickness calculator GUI. """ ## Internal nickname for the window, used by the AUI manager window_name = "Kiessig Thickness Calculator" ## Name to appear on the window title bar window_caption = "Kiessig Thickness Calculator" ## Flag to tell the AUI manager to put this panel in the center pane CENTER_PANE = True def __init__(self, parent, *args, **kwds): wx.Panel.__init__(self, parent, *args, **kwds) PanelBase.__init__(self) #Font size self.SetWindowVariant(variant=FONT_VARIANT) # Object that receive status event self.parent = parent self.kiessig = KiessigThicknessCalculator() #layout attribute self.hint_sizer = None self._do_layout() def _define_structure(self): """ Define the main sizers building to build this application. """ self.main_sizer = wx.BoxSizer(wx.VERTICAL) self.box_source = wx.StaticBox(self, -1, str("Kiessig Thickness Calculator")) self.boxsizer_source = wx.StaticBoxSizer(self.box_source, wx.VERTICAL) self.dq_name_sizer = wx.BoxSizer(wx.HORIZONTAL) self.thickness_size_sizer = wx.BoxSizer(wx.HORIZONTAL) self.hint_sizer = wx.BoxSizer(wx.HORIZONTAL) self.button_sizer = wx.BoxSizer(wx.HORIZONTAL) def _layout_dq_name(self): """ Fill the sizer containing dq name """ # get the default dq dq_value = str(self.kiessig.get_deltaq()) dq_unit_txt = wx.StaticText(self, -1, '[1/A]') dq_name_txt = wx.StaticText(self, -1, 'Kiessig Fringe Width (Delta Q): ') self.dq_name_tcl = InputTextCtrl(self, -1, size=(_BOX_WIDTH,-1)) dq_hint = "Type the Kiessig Fringe Width (Delta Q)" self.dq_name_tcl.SetValue(dq_value) self.dq_name_tcl.SetToolTipString(dq_hint) #control that triggers importing data id = wx.NewId() self.compute_button = wx.Button(self, id, "Compute") hint_on_compute = "Compute the diameter/thickness in the real space." self.compute_button.SetToolTipString(hint_on_compute) self.Bind(wx.EVT_BUTTON, self.on_compute, id=id) self.dq_name_sizer.AddMany([(dq_name_txt, 0, wx.LEFT, 15), (self.dq_name_tcl, 0, wx.LEFT, 15), (dq_unit_txt, 0, wx.LEFT, 10), (self.compute_button, 0, wx.LEFT, 30)]) def _layout_thickness_size(self): """ Fill the sizer containing thickness information """ thick_unit = '['+self.kiessig.get_thickness_unit() +']' thickness_size_txt = wx.StaticText(self, -1, 'Thickness (or Diameter): ') self.thickness_size_tcl = OutputTextCtrl(self, -1, size=(_BOX_WIDTH,-1)) thickness_size_hint = " Estimated Size in Real Space" self.thickness_size_tcl.SetToolTipString(thickness_size_hint) thickness_size_unit_txt = wx.StaticText(self, -1, thick_unit) self.thickness_size_sizer.AddMany([(thickness_size_txt, 0, wx.LEFT, 15), (self.thickness_size_tcl, 0, wx.LEFT, 15), (thickness_size_unit_txt, 0, wx.LEFT, 10)]) def _layout_hint(self): """ Fill the sizer containing hint """ hint_msg = "This tool is to approximately estimate " hint_msg += "the thickness of a layer" hint_msg += " or the diameter of particles\n " hint_msg += "from the Kiessig fringe width in SAS/NR data." hint_msg += "" self.hint_txt = wx.StaticText(self, -1, hint_msg) self.hint_sizer.AddMany([(self.hint_txt, 0, wx.LEFT, 15)]) def _layout_button(self): """ Do the layout for the button widgets """ id = wx.NewId() self.bt_help = wx.Button(self, id, 'HELP') self.bt_help.Bind(wx.EVT_BUTTON, self.on_help) self.bt_help.SetToolTipString("Help using the Kiessig fringe calculator.") self.bt_close = wx.Button(self, wx.ID_CANCEL, 'Close') self.bt_close.Bind(wx.EVT_BUTTON, self.on_close) self.bt_close.SetToolTipString("Close this window.") self.button_sizer.AddMany([(self.bt_help, 0, wx.LEFT, 260), (self.bt_close, 0, wx.LEFT, 20)]) def _do_layout(self): """ Draw window content """ self._define_structure() self._layout_dq_name() self._layout_thickness_size() self._layout_hint() self._layout_button() self.boxsizer_source.AddMany([(self.dq_name_sizer, 0, wx.EXPAND|wx.TOP|wx.BOTTOM, 5), (self.thickness_size_sizer, 0, wx.EXPAND|wx.TOP|wx.BOTTOM, 5), (self.hint_sizer, 0, wx.EXPAND|wx.TOP|wx.BOTTOM, 5)]) self.main_sizer.AddMany([(self.boxsizer_source, 0, wx.ALL, 10), (self.button_sizer, 0, wx.EXPAND|wx.TOP|wx.BOTTOM, 5)]) self.SetSizer(self.main_sizer) self.SetAutoLayout(True) def on_help(self, event): """ Bring up the Kiessig fringe calculator Documentation whenever the HELP button is clicked. Calls DocumentationWindow with the path of the location within the documentation tree (after /doc/ ....". Note that when using old versions of Wx (before 2.9) and thus not the release version of installers, the help comes up at the top level of the file as webbrowser does not pass anything past the # to the browser when it is running "file:///...." :param evt: Triggers on clicking the help button """ _TreeLocation = "user/sasgui/perspectives/calculator/" _TreeLocation += "kiessig_calculator_help.html" _doc_viewer = DocumentationWindow(self, -1, _TreeLocation, "", "Density/Volume Calculator Help") def on_close(self, event): """ close the window containing this panel """ self.parent.Close() if event is not None: event.Skip() def on_compute(self, event): """ Execute the computation of thickness """ # skip for another event if event is not None: event.Skip() dq = self.dq_name_tcl.GetValue() self.kiessig.set_deltaq(dq) # calculate the thickness output = self.kiessig.compute_thickness() thickness = self.format_number(output) # set tcl self.thickness_size_tcl.SetValue(str(thickness)) def format_number(self, value=None): """ Return a float in a standardized, human-readable formatted string """ try: value = float(value) except: output = None return output output = "%-7.4g" % value return output.lstrip().rstrip() def _onparamEnter(self, event = None): """ On Text_enter_callback, perform compute """ self.on_compute(event) class KiessigWindow(widget.CHILD_FRAME): def __init__(self, parent=None, manager=None, title="Kiessig Thickness Calculator", size=(PANEL_WIDTH,PANEL_HEIGHT), *args, **kwds): kwds['title'] = title kwds['size'] = size widget.CHILD_FRAME.__init__(self, parent, *args, **kwds) self.parent = parent self.manager = manager self.panel = KiessigThicknessCalculatorPanel(parent=self) self.Bind(wx.EVT_CLOSE, self.on_close) self.SetPosition((wx.LEFT, PANEL_TOP)) self.Show(True) def on_close(self, event): """ Close event """ if self.manager is not None: self.manager.kiessig_frame = None self.Destroy() if __name__ == "__main__": app = wx.PySimpleApp() widget.CHILD_FRAME = wx.Frame frame = KiessigWindow() frame.Show(True) app.MainLoop()
src/sas/sasgui/perspectives/calculator/kiessig_calculator_panel.py
9,529
Provides the Kiessig thickness calculator GUI. Define the main sizers building to build this application. Draw window content Do the layout for the button widgets Fill the sizer containing dq name Fill the sizer containing hint Fill the sizer containing thickness information On Text_enter_callback, perform compute Return a float in a standardized, human-readable formatted string close the window containing this panel Close event Execute the computation of thickness Bring up the Kiessig fringe calculator Documentation whenever the HELP button is clicked. Calls DocumentationWindow with the path of the location within the documentation tree (after /doc/ ....". Note that when using old versions of Wx (before 2.9) and thus not the release version of installers, the help comes up at the top level of the file as webbrowser does not pass anything past the # to the browser when it is running "file:///...." :param evt: Triggers on clicking the help button This software was developed by the University of Tennessee as part of the Distributed Data Analysis of Neutron Scattering Experiments (DANSE) project funded by the US National Science Foundation. See the license text in license.txt copyright 2008, 2009, University of Tennessee Slit length panel size Internal nickname for the window, used by the AUI manager Name to appear on the window title bar Flag to tell the AUI manager to put this panel in the center paneFont size Object that receive status eventlayout attribute get the default dqcontrol that triggers importing data skip for another event calculate the thickness set tcl
1,630
en
0.809921
import unittest from typing import Optional from pathlib import Path from time import strftime, gmtime from github import Github import json import yaml from .. import get_repo_meta, clone_and_archive from ..get_github import dump_list from pathlib import Path class TestGetGithub(unittest.TestCase): """ Test coverage for the get_github module """ _current_dir = Path(__file__).resolve().parent #BORROWED from https://github.com/NOAA-OWP/DMOD/blob/master/python/lib/scheduler/dmod/test/it_redisManager.py @classmethod def find_project_root_directory(cls, current_directory: Optional[Path]) -> Optional[Path]: """ Given a directory (with ``None`` implying the current directory) assumed to be at or under this project's root, find the project root directory. This implementation attempts to find a directory having both a ``.git/`` child directory and a ``.env`` file. Parameters ---------- current_directory Returns ------- Optional[Path] The project root directory, or ``None`` if it fails to find it. """ if not current_directory: current_directory = TestGetGithub._current_dir abs_root = Path(current_directory.absolute().root) while current_directory.absolute() != abs_root: if not current_directory.is_dir(): current_directory = current_directory.parent continue git_sub_dir = current_directory.joinpath('.git') child_env_file = current_directory.joinpath('config.yaml') if git_sub_dir.exists() and git_sub_dir.is_dir() and child_env_file.exists() and child_env_file.is_file(): return current_directory current_directory = current_directory.parent return None @classmethod def load_token(cls): """ Read an API token from a configuration file, if none found, use '' for no auth """ token = '' root_dir = cls.find_project_root_directory(None) if not root_dir: return token config_file = root_dir/'config.yaml' if config_file.exists(): with open(config_file) as file: config = yaml.load(file, Loader=yaml.FullLoader) try: token = config['token'] except: print("Unable to load api-token from project root directory config.yaml") return token def setUp(self): self.token = self.load_token() if self.token: #Token auth github, higher rate limit self.github = Github(self.token) else: #Anonimous github, severly rate limited API self.github = Github() self.org_string = 'NOAA-OWP' self.repo_string = 'owp-open-source-project-template' self.repo_w_wiki = 'DMOD' self.org = self.github.get_organization(self.org_string) self.wiki_repo = self.org.get_repo(self.repo_w_wiki) self.repo = self.org.get_repo(self.repo_string) self.time = strftime("%Y-%m-%d_%H:%M:%S", gmtime()) def tearDown(self): for p in Path(TestGetGithub._current_dir).glob("*.json"): if p.is_file(): p.unlink() for p in Path(TestGetGithub._current_dir).glob("*.tar.gz"): if p.is_file(): p.unlink() def test_get_repo_meta(self): """ Test the archive_repo function to ensure all meta data is properly captured """ meta = get_repo_meta(self.repo, self.time, TestGetGithub._current_dir) self.assertIsNotNone(meta) self.assertTrue(len(meta), 6) #defer name substitution pattern = "{repo}_{name}_{time}.json".format(repo=self.repo_string, name="{name}", time=self.time) self.assertEqual(meta[0].name, pattern.format(name='comments')) self.assertEqual(meta[1].name, pattern.format(name='issues')) self.assertEqual(meta[2].name, pattern.format(name='issue_comments')) self.assertEqual(meta[3].name, pattern.format(name='pulls')) self.assertEqual(meta[4].name, pattern.format(name='pulls_comments')) self.assertEqual(meta[5].name, pattern.format(name='pulls_review_comments')) self.assertTrue((TestGetGithub._current_dir/pattern.format(name='comments')).exists()) self.assertTrue((TestGetGithub._current_dir/pattern.format(name='issues')).exists()) self.assertTrue((TestGetGithub._current_dir/pattern.format(name='issue_comments')).exists()) self.assertTrue((TestGetGithub._current_dir/pattern.format(name='pulls')).exists()) self.assertTrue((TestGetGithub._current_dir/pattern.format(name='pulls_comments')).exists()) self.assertTrue((TestGetGithub._current_dir/pattern.format(name='pulls_review_comments')).exists()) def test_clone_and_archive(self): """ Test the clone functionality """ #Sorta hackily testing the archive_repo logic here...FIXME later #FIXME has_wiki is broke!!! self.assertFalse(self.repo.has_wiki) clone_url = self.repo.clone_url archive_name = clone_and_archive(self.repo_string, clone_url, self.time, TestGetGithub._current_dir, []) name = '{repo}_github_archive_{time}.tar.gz'.format(repo=self.repo_string, time=self.time) self.assertEqual(archive_name.name, name) self.assertTrue((TestGetGithub._current_dir/name).exists()) #TODO test existence of repo in archive def test_clone_and_archive_1(self): """ Test cloning a repo with a wiki """ #Sorta hackily testing the archive_repo logic here...FIXME later self.assertTrue(self.wiki_repo.has_wiki) wiki_url = self.wiki_repo.clone_url[:-3]+'wiki.git' #finally get the repo code itself clone_url = self.wiki_repo.clone_url archive_name = clone_and_archive(self.repo_w_wiki, clone_url, self.time, TestGetGithub._current_dir, [], wiki_url) name = '{repo}_github_archive_{time}.tar.gz'.format(repo=self.repo_w_wiki, time=self.time) self.assertEqual(archive_name.name, name) self.assertTrue((TestGetGithub._current_dir/name).exists()) #TODO test existense of wiki in archive @unittest.skip("Incomplete mock implementation for dumped item") def test_dump_list(self): pass #Quick json format test for list objects outfile = dump_list("test_repo", self.time, TestGetGithub._current_dir, "test_key", [ '{json=dict, test=values}', 'json=dict2, test=values2}']) with open(outfile) as fp: data = json.load(fp) self.assertTrue(data['json'], 'dict')
github_archive/test/test_get_github.py
6,815
Test coverage for the get_github module Given a directory (with ``None`` implying the current directory) assumed to be at or under this project's root, find the project root directory. This implementation attempts to find a directory having both a ``.git/`` child directory and a ``.env`` file. Parameters ---------- current_directory Returns ------- Optional[Path] The project root directory, or ``None`` if it fails to find it. Read an API token from a configuration file, if none found, use '' for no auth Test the clone functionality Test cloning a repo with a wiki Test the archive_repo function to ensure all meta data is properly captured BORROWED from https://github.com/NOAA-OWP/DMOD/blob/master/python/lib/scheduler/dmod/test/it_redisManager.pyToken auth github, higher rate limitAnonimous github, severly rate limited APIdefer name substitutionSorta hackily testing the archive_repo logic here...FIXME laterFIXME has_wiki is broke!!!TODO test existence of repo in archiveSorta hackily testing the archive_repo logic here...FIXME laterfinally get the repo code itselfTODO test existense of wiki in archiveQuick json format test for list objects
1,159
en
0.647168
# Copyright 2020 Google Research. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """AutoAugment. [1] Barret, et al. Learning Data Augmentation Strategies for Object Detection. Arxiv: https://arxiv.org/abs/1906.11172 """ import inspect import math from absl import logging import tensorflow.compat.v1 as tf from tensorflow_addons import image as image_ops import hparams_config # This signifies the max integer that the controller RNN could predict for the # augmentation scheme. _MAX_LEVEL = 10. # Represents an invalid bounding box that is used for checking for padding # lists of bounding box coordinates for a few augmentation operations _INVALID_BOX = [[-1.0, -1.0, -1.0, -1.0]] def policy_v0(): """Autoaugment policy that was used in AutoAugment Detection Paper.""" # Each tuple is an augmentation operation of the form # (operation, probability, magnitude). Each element in policy is a # sub-policy that will be applied sequentially on the image. policy = [ [('TranslateX_BBox', 0.6, 4), ('Equalize', 0.8, 10)], [('TranslateY_Only_BBoxes', 0.2, 2), ('Cutout', 0.8, 8)], [('Sharpness', 0.0, 8), ('ShearX_BBox', 0.4, 0)], [('ShearY_BBox', 1.0, 2), ('TranslateY_Only_BBoxes', 0.6, 6)], [('Rotate_BBox', 0.6, 10), ('Color', 1.0, 6)], ] return policy def policy_v1(): """Autoaugment policy that was used in AutoAugment Detection Paper.""" # Each tuple is an augmentation operation of the form # (operation, probability, magnitude). Each element in policy is a # sub-policy that will be applied sequentially on the image. policy = [ [('TranslateX_BBox', 0.6, 4), ('Equalize', 0.8, 10)], [('TranslateY_Only_BBoxes', 0.2, 2), ('Cutout', 0.8, 8)], [('Sharpness', 0.0, 8), ('ShearX_BBox', 0.4, 0)], [('ShearY_BBox', 1.0, 2), ('TranslateY_Only_BBoxes', 0.6, 6)], [('Rotate_BBox', 0.6, 10), ('Color', 1.0, 6)], [('Color', 0.0, 0), ('ShearX_Only_BBoxes', 0.8, 4)], [('ShearY_Only_BBoxes', 0.8, 2), ('Flip_Only_BBoxes', 0.0, 10)], [('Equalize', 0.6, 10), ('TranslateX_BBox', 0.2, 2)], [('Color', 1.0, 10), ('TranslateY_Only_BBoxes', 0.4, 6)], [('Rotate_BBox', 0.8, 10), ('Contrast', 0.0, 10)], [('Cutout', 0.2, 2), ('Brightness', 0.8, 10)], [('Color', 1.0, 6), ('Equalize', 1.0, 2)], [('Cutout_Only_BBoxes', 0.4, 6), ('TranslateY_Only_BBoxes', 0.8, 2)], [('Color', 0.2, 8), ('Rotate_BBox', 0.8, 10)], [('Sharpness', 0.4, 4), ('TranslateY_Only_BBoxes', 0.0, 4)], [('Sharpness', 1.0, 4), ('SolarizeAdd', 0.4, 4)], [('Rotate_BBox', 1.0, 8), ('Sharpness', 0.2, 8)], [('ShearY_BBox', 0.6, 10), ('Equalize_Only_BBoxes', 0.6, 8)], [('ShearX_BBox', 0.2, 6), ('TranslateY_Only_BBoxes', 0.2, 10)], [('SolarizeAdd', 0.6, 8), ('Brightness', 0.8, 10)], ] return policy def policy_vtest(): """Autoaugment test policy for debugging.""" # Each tuple is an augmentation operation of the form # (operation, probability, magnitude). Each element in policy is a # sub-policy that will be applied sequentially on the image. policy = [ [('TranslateX_BBox', 1.0, 4), ('Equalize', 1.0, 10)], ] return policy def policy_v2(): """Additional policy that performs well on object detection.""" # Each tuple is an augmentation operation of the form # (operation, probability, magnitude). Each element in policy is a # sub-policy that will be applied sequentially on the image. policy = [ [('Color', 0.0, 6), ('Cutout', 0.6, 8), ('Sharpness', 0.4, 8)], [('Rotate_BBox', 0.4, 8), ('Sharpness', 0.4, 2), ('Rotate_BBox', 0.8, 10)], [('TranslateY_BBox', 1.0, 8), ('AutoContrast', 0.8, 2)], [('AutoContrast', 0.4, 6), ('ShearX_BBox', 0.8, 8), ('Brightness', 0.0, 10)], [('SolarizeAdd', 0.2, 6), ('Contrast', 0.0, 10), ('AutoContrast', 0.6, 0)], [('Cutout', 0.2, 0), ('Solarize', 0.8, 8), ('Color', 1.0, 4)], [('TranslateY_BBox', 0.0, 4), ('Equalize', 0.6, 8), ('Solarize', 0.0, 10)], [('TranslateY_BBox', 0.2, 2), ('ShearY_BBox', 0.8, 8), ('Rotate_BBox', 0.8, 8)], [('Cutout', 0.8, 8), ('Brightness', 0.8, 8), ('Cutout', 0.2, 2)], [('Color', 0.8, 4), ('TranslateY_BBox', 1.0, 6), ('Rotate_BBox', 0.6, 6)], [('Rotate_BBox', 0.6, 10), ('BBox_Cutout', 1.0, 4), ('Cutout', 0.2, 8)], [('Rotate_BBox', 0.0, 0), ('Equalize', 0.6, 6), ('ShearY_BBox', 0.6, 8)], [('Brightness', 0.8, 8), ('AutoContrast', 0.4, 2), ('Brightness', 0.2, 2)], [('TranslateY_BBox', 0.4, 8), ('Solarize', 0.4, 6), ('SolarizeAdd', 0.2, 10)], [('Contrast', 1.0, 10), ('SolarizeAdd', 0.2, 8), ('Equalize', 0.2, 4)], ] return policy def policy_v3(): """"Additional policy that performs well on object detection.""" # Each tuple is an augmentation operation of the form # (operation, probability, magnitude). Each element in policy is a # sub-policy that will be applied sequentially on the image. policy = [ [('Posterize', 0.8, 2), ('TranslateX_BBox', 1.0, 8)], [('BBox_Cutout', 0.2, 10), ('Sharpness', 1.0, 8)], [('Rotate_BBox', 0.6, 8), ('Rotate_BBox', 0.8, 10)], [('Equalize', 0.8, 10), ('AutoContrast', 0.2, 10)], [('SolarizeAdd', 0.2, 2), ('TranslateY_BBox', 0.2, 8)], [('Sharpness', 0.0, 2), ('Color', 0.4, 8)], [('Equalize', 1.0, 8), ('TranslateY_BBox', 1.0, 8)], [('Posterize', 0.6, 2), ('Rotate_BBox', 0.0, 10)], [('AutoContrast', 0.6, 0), ('Rotate_BBox', 1.0, 6)], # [('Equalize', 0.0, 4), ('Cutout', 0.8, 10)], [('Brightness', 1.0, 2), ('TranslateY_BBox', 1.0, 6)], [('Contrast', 0.0, 2), ('ShearY_BBox', 0.8, 0)], [('AutoContrast', 0.8, 10), ('Contrast', 0.2, 10)], # [('Rotate_BBox', 1.0, 10), ('Cutout', 1.0, 10)], [('SolarizeAdd', 0.8, 6), ('Equalize', 0.8, 8)], ] return policy def blend(image1, image2, factor): """Blend image1 and image2 using 'factor'. Factor can be above 0.0. A value of 0.0 means only image1 is used. A value of 1.0 means only image2 is used. A value between 0.0 and 1.0 means we linearly interpolate the pixel values between the two images. A value greater than 1.0 "extrapolates" the difference between the two pixel values, and we clip the results to values between 0 and 255. Args: image1: An image Tensor of type uint8. image2: An image Tensor of type uint8. factor: A floating point value above 0.0. Returns: A blended image Tensor of type uint8. """ if factor == 0.0: return tf.convert_to_tensor(image1) if factor == 1.0: return tf.convert_to_tensor(image2) image1 = tf.to_float(image1) image2 = tf.to_float(image2) difference = image2 - image1 scaled = factor * difference # Do addition in float. temp = tf.to_float(image1) + scaled # Interpolate if factor > 0.0 and factor < 1.0: # Interpolation means we always stay within 0 and 255. return tf.cast(temp, tf.uint8) # Extrapolate: # # We need to clip and then cast. return tf.cast(tf.clip_by_value(temp, 0.0, 255.0), tf.uint8) def cutout(image, pad_size, replace=0): """Apply cutout (https://arxiv.org/abs/1708.04552) to image. This operation applies a (2*pad_size x 2*pad_size) mask of zeros to a random location within `img`. The pixel values filled in will be of the value `replace`. The located where the mask will be applied is randomly chosen uniformly over the whole image. Args: image: An image Tensor of type uint8. pad_size: Specifies how big the zero mask that will be generated is that is applied to the image. The mask will be of size (2*pad_size x 2*pad_size). replace: What pixel value to fill in the image in the area that has the cutout mask applied to it. Returns: An image Tensor that is of type uint8. """ image_height = tf.maximum(tf.shape(image)[0], 10) image_width = tf.maximum(tf.shape(image)[1], 10) # Sample the center location in the image where the zero mask will be applied. cutout_center_height = tf.random_uniform( shape=[], minval=0, maxval=image_height, dtype=tf.int32) cutout_center_width = tf.random_uniform( shape=[], minval=0, maxval=image_width, dtype=tf.int32) lower_pad = tf.maximum(0, cutout_center_height - pad_size) upper_pad = tf.maximum(0, image_height - cutout_center_height - pad_size) left_pad = tf.maximum(0, cutout_center_width - pad_size) right_pad = tf.maximum(0, image_width - cutout_center_width - pad_size) cutout_shape = [image_height - (lower_pad + upper_pad), image_width - (left_pad + right_pad)] padding_dims = [[lower_pad, upper_pad], [left_pad, right_pad]] mask = tf.pad( tf.zeros(cutout_shape, dtype=image.dtype), padding_dims, constant_values=1) mask = tf.expand_dims(mask, -1) mask = tf.tile(mask, [1, 1, 3]) image = tf.where( tf.equal(mask, 0), tf.ones_like(image, dtype=image.dtype) * replace, image) return image def solarize(image, threshold=128): # For each pixel in the image, select the pixel # if the value is less than the threshold. # Otherwise, subtract 255 from the pixel. return tf.where(image < threshold, image, 255 - image) def solarize_add(image, addition=0, threshold=128): # For each pixel in the image less than threshold # we add 'addition' amount to it and then clip the # pixel value to be between 0 and 255. The value # of 'addition' is between -128 and 128. added_image = tf.cast(image, tf.int64) + addition added_image = tf.cast(tf.clip_by_value(added_image, 0, 255), tf.uint8) return tf.where(image < threshold, added_image, image) def color(image, factor): """Equivalent of PIL Color.""" degenerate = tf.image.grayscale_to_rgb(tf.image.rgb_to_grayscale(image)) return blend(degenerate, image, factor) def contrast(image, factor): """Equivalent of PIL Contrast.""" degenerate = tf.image.rgb_to_grayscale(image) # Cast before calling tf.histogram. degenerate = tf.cast(degenerate, tf.int32) # Compute the grayscale histogram, then compute the mean pixel value, # and create a constant image size of that value. Use that as the # blending degenerate target of the original image. mean = tf.reduce_mean(tf.cast(degenerate, tf.float32)) degenerate = tf.ones_like(degenerate, dtype=tf.float32) * mean degenerate = tf.clip_by_value(degenerate, 0.0, 255.0) degenerate = tf.image.grayscale_to_rgb(tf.cast(degenerate, tf.uint8)) return blend(degenerate, image, factor) def brightness(image, factor): """Equivalent of PIL Brightness.""" degenerate = tf.zeros_like(image) return blend(degenerate, image, factor) def posterize(image, bits): """Equivalent of PIL Posterize.""" shift = 8 - bits return tf.bitwise.left_shift(tf.bitwise.right_shift(image, shift), shift) def rotate(image, degrees, replace): """Rotates the image by degrees either clockwise or counterclockwise. Args: image: An image Tensor of type uint8. degrees: Float, a scalar angle in degrees to rotate all images by. If degrees is positive the image will be rotated clockwise otherwise it will be rotated counterclockwise. replace: A one or three value 1D tensor to fill empty pixels caused by the rotate operation. Returns: The rotated version of image. """ # Convert from degrees to radians. degrees_to_radians = math.pi / 180.0 radians = degrees * degrees_to_radians # In practice, we should randomize the rotation degrees by flipping # it negatively half the time, but that's done on 'degrees' outside # of the function. image = image_ops.rotate(wrap(image), radians) return unwrap(image, replace) def random_shift_bbox(image, bbox, pixel_scaling, replace, new_min_bbox_coords=None): """Move the bbox and the image content to a slightly new random location. Args: image: 3D uint8 Tensor. bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. The potential values for the new min corner of the bbox will be between [old_min - pixel_scaling * bbox_height/2, old_min - pixel_scaling * bbox_height/2]. pixel_scaling: A float between 0 and 1 that specifies the pixel range that the new bbox location will be sampled from. replace: A one or three value 1D tensor to fill empty pixels. new_min_bbox_coords: If not None, then this is a tuple that specifies the (min_y, min_x) coordinates of the new bbox. Normally this is randomly specified, but this allows it to be manually set. The coordinates are the absolute coordinates between 0 and image height/width and are int32. Returns: The new image that will have the shifted bbox location in it along with the new bbox that contains the new coordinates. """ # Obtains image height and width and create helper clip functions. image_height = tf.to_float(tf.maximum(tf.shape(image)[0], 10)) image_width = tf.to_float(tf.maximum(tf.shape(image)[1], 10)) def clip_y(val): return tf.clip_by_value(val, 0, tf.to_int32(image_height) - 1) def clip_x(val): return tf.clip_by_value(val, 0, tf.to_int32(image_width) - 1) # Convert bbox to pixel coordinates. min_y = tf.to_int32(image_height * bbox[0]) min_x = tf.to_int32(image_width * bbox[1]) max_y = clip_y(tf.to_int32(image_height * bbox[2])) max_x = clip_x(tf.to_int32(image_width * bbox[3])) bbox_height, bbox_width = (max_y - min_y + 1, max_x - min_x + 1) image_height = tf.to_int32(image_height) image_width = tf.to_int32(image_width) # Select the new min/max bbox ranges that are used for sampling the # new min x/y coordinates of the shifted bbox. minval_y = clip_y( min_y - tf.to_int32(pixel_scaling * tf.to_float(bbox_height) / 2.0)) maxval_y = clip_y( min_y + tf.to_int32(pixel_scaling * tf.to_float(bbox_height) / 2.0)) minval_x = clip_x( min_x - tf.to_int32(pixel_scaling * tf.to_float(bbox_width) / 2.0)) maxval_x = clip_x( min_x + tf.to_int32(pixel_scaling * tf.to_float(bbox_width) / 2.0)) # Sample and calculate the new unclipped min/max coordinates of the new bbox. if new_min_bbox_coords is None: unclipped_new_min_y = tf.random_uniform( shape=[], minval=minval_y, maxval=maxval_y, dtype=tf.int32) unclipped_new_min_x = tf.random_uniform( shape=[], minval=minval_x, maxval=maxval_x, dtype=tf.int32) else: unclipped_new_min_y, unclipped_new_min_x = ( clip_y(new_min_bbox_coords[0]), clip_x(new_min_bbox_coords[1])) unclipped_new_max_y = unclipped_new_min_y + bbox_height - 1 unclipped_new_max_x = unclipped_new_min_x + bbox_width - 1 # Determine if any of the new bbox was shifted outside the current image. # This is used for determining if any of the original bbox content should be # discarded. new_min_y, new_min_x, new_max_y, new_max_x = ( clip_y(unclipped_new_min_y), clip_x(unclipped_new_min_x), clip_y(unclipped_new_max_y), clip_x(unclipped_new_max_x)) shifted_min_y = (new_min_y - unclipped_new_min_y) + min_y shifted_max_y = max_y - (unclipped_new_max_y - new_max_y) shifted_min_x = (new_min_x - unclipped_new_min_x) + min_x shifted_max_x = max_x - (unclipped_new_max_x - new_max_x) # Create the new bbox tensor by converting pixel integer values to floats. new_bbox = tf.stack([ tf.to_float(new_min_y) / tf.to_float(image_height), tf.to_float(new_min_x) / tf.to_float(image_width), tf.to_float(new_max_y) / tf.to_float(image_height), tf.to_float(new_max_x) / tf.to_float(image_width)]) # Copy the contents in the bbox and fill the old bbox location # with gray (128). bbox_content = image[shifted_min_y:shifted_max_y + 1, shifted_min_x:shifted_max_x + 1, :] def mask_and_add_image( min_y_, min_x_, max_y_, max_x_, mask, content_tensor, image_): """Applies mask to bbox region in image then adds content_tensor to it.""" mask = tf.pad(mask, [[min_y_, (image_height - 1) - max_y_], [min_x_, (image_width - 1) - max_x_], [0, 0]], constant_values=1) content_tensor = tf.pad(content_tensor, [[min_y_, (image_height - 1) - max_y_], [min_x_, (image_width - 1) - max_x_], [0, 0]], constant_values=0) return image_ * mask + content_tensor # Zero out original bbox location. mask = tf.zeros_like(image)[min_y:max_y+1, min_x:max_x+1, :] grey_tensor = tf.zeros_like(mask) + replace[0] image = mask_and_add_image(min_y, min_x, max_y, max_x, mask, grey_tensor, image) # Fill in bbox content to new bbox location. mask = tf.zeros_like(bbox_content) image = mask_and_add_image(new_min_y, new_min_x, new_max_y, new_max_x, mask, bbox_content, image) return image, new_bbox def _clip_bbox(min_y, min_x, max_y, max_x): """Clip bounding box coordinates between 0 and 1. Args: min_y: Normalized bbox coordinate of type float between 0 and 1. min_x: Normalized bbox coordinate of type float between 0 and 1. max_y: Normalized bbox coordinate of type float between 0 and 1. max_x: Normalized bbox coordinate of type float between 0 and 1. Returns: Clipped coordinate values between 0 and 1. """ min_y = tf.clip_by_value(min_y, 0.0, 1.0) min_x = tf.clip_by_value(min_x, 0.0, 1.0) max_y = tf.clip_by_value(max_y, 0.0, 1.0) max_x = tf.clip_by_value(max_x, 0.0, 1.0) return min_y, min_x, max_y, max_x def _check_bbox_area(min_y, min_x, max_y, max_x, delta=0.05): """Adjusts bbox coordinates to make sure the area is > 0. Args: min_y: Normalized bbox coordinate of type float between 0 and 1. min_x: Normalized bbox coordinate of type float between 0 and 1. max_y: Normalized bbox coordinate of type float between 0 and 1. max_x: Normalized bbox coordinate of type float between 0 and 1. delta: Float, this is used to create a gap of size 2 * delta between bbox min/max coordinates that are the same on the boundary. This prevents the bbox from having an area of zero. Returns: Tuple of new bbox coordinates between 0 and 1 that will now have a guaranteed area > 0. """ height = max_y - min_y width = max_x - min_x def _adjust_bbox_boundaries(min_coord, max_coord): # Make sure max is never 0 and min is never 1. max_coord = tf.maximum(max_coord, 0.0 + delta) min_coord = tf.minimum(min_coord, 1.0 - delta) return min_coord, max_coord min_y, max_y = tf.cond(tf.equal(height, 0.0), lambda: _adjust_bbox_boundaries(min_y, max_y), lambda: (min_y, max_y)) min_x, max_x = tf.cond(tf.equal(width, 0.0), lambda: _adjust_bbox_boundaries(min_x, max_x), lambda: (min_x, max_x)) return min_y, min_x, max_y, max_x def _scale_bbox_only_op_probability(prob): """Reduce the probability of the bbox-only operation. Probability is reduced so that we do not distort the content of too many bounding boxes that are close to each other. The value of 3.0 was a chosen hyper parameter when designing the autoaugment algorithm that we found empirically to work well. Args: prob: Float that is the probability of applying the bbox-only operation. Returns: Reduced probability. """ return prob / 3.0 def _apply_bbox_augmentation(image, bbox, augmentation_func, *args): """Applies augmentation_func to the subsection of image indicated by bbox. Args: image: 3D uint8 Tensor. bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. augmentation_func: Augmentation function that will be applied to the subsection of image. *args: Additional parameters that will be passed into augmentation_func when it is called. Returns: A modified version of image, where the bbox location in the image will have `ugmentation_func applied to it. """ image_height = tf.to_float(tf.maximum(tf.shape(image)[0], 10)) image_width = tf.to_float(tf.maximum(tf.shape(image)[1], 10)) min_y = tf.to_int32(image_height * bbox[0]) min_x = tf.to_int32(image_width * bbox[1]) max_y = tf.to_int32(image_height * bbox[2]) max_x = tf.to_int32(image_width * bbox[3]) image_height = tf.to_int32(image_height) image_width = tf.to_int32(image_width) # Clip to be sure the max values do not fall out of range. max_y = tf.minimum(max_y, image_height - 1) max_x = tf.minimum(max_x, image_width - 1) # Get the sub-tensor that is the image within the bounding box region. bbox_content = image[min_y:max_y + 1, min_x:max_x + 1, :] # Apply the augmentation function to the bbox portion of the image. augmented_bbox_content = augmentation_func(bbox_content, *args) # Pad the augmented_bbox_content and the mask to match the shape of original # image. augmented_bbox_content = tf.pad(augmented_bbox_content, [[min_y, (image_height - 1) - max_y], [min_x, (image_width - 1) - max_x], [0, 0]]) # Create a mask that will be used to zero out a part of the original image. mask_tensor = tf.zeros_like(bbox_content) mask_tensor = tf.pad(mask_tensor, [[min_y, (image_height - 1) - max_y], [min_x, (image_width - 1) - max_x], [0, 0]], constant_values=1) # Replace the old bbox content with the new augmented content. image = image * mask_tensor + augmented_bbox_content return image def _concat_bbox(bbox, bboxes): """Helper function that concats bbox to bboxes along the first dimension.""" # Note if all elements in bboxes are -1 (_INVALID_BOX), then this means # we discard bboxes and start the bboxes Tensor with the current bbox. bboxes_sum_check = tf.reduce_sum(bboxes) bbox = tf.expand_dims(bbox, 0) # This check will be true when it is an _INVALID_BOX bboxes = tf.cond(tf.equal(bboxes_sum_check, -4.0), lambda: bbox, lambda: tf.concat([bboxes, bbox], 0)) return bboxes def _apply_bbox_augmentation_wrapper(image, bbox, new_bboxes, prob, augmentation_func, func_changes_bbox, *args): """Applies _apply_bbox_augmentation with probability prob. Args: image: 3D uint8 Tensor. bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. new_bboxes: 2D Tensor that is a list of the bboxes in the image after they have been altered by aug_func. These will only be changed when func_changes_bbox is set to true. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float that are the normalized bbox coordinates between 0 and 1. prob: Float that is the probability of applying _apply_bbox_augmentation. augmentation_func: Augmentation function that will be applied to the subsection of image. func_changes_bbox: Boolean. Does augmentation_func return bbox in addition to image. *args: Additional parameters that will be passed into augmentation_func when it is called. Returns: A tuple. Fist element is a modified version of image, where the bbox location in the image will have augmentation_func applied to it if it is chosen to be called with probability `prob`. The second element is a Tensor of Tensors of length 4 that will contain the altered bbox after applying augmentation_func. """ should_apply_op = tf.cast( tf.floor(tf.random_uniform([], dtype=tf.float32) + prob), tf.bool) if func_changes_bbox: augmented_image, bbox = tf.cond( should_apply_op, lambda: augmentation_func(image, bbox, *args), lambda: (image, bbox)) else: augmented_image = tf.cond( should_apply_op, lambda: _apply_bbox_augmentation(image, bbox, augmentation_func, *args), lambda: image) new_bboxes = _concat_bbox(bbox, new_bboxes) return augmented_image, new_bboxes def _apply_multi_bbox_augmentation(image, bboxes, prob, aug_func, func_changes_bbox, *args): """Applies aug_func to the image for each bbox in bboxes. Args: image: 3D uint8 Tensor. bboxes: 2D Tensor that is a list of the bboxes in the image. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float. prob: Float that is the probability of applying aug_func to a specific bounding box within the image. aug_func: Augmentation function that will be applied to the subsections of image indicated by the bbox values in bboxes. func_changes_bbox: Boolean. Does augmentation_func return bbox in addition to image. *args: Additional parameters that will be passed into augmentation_func when it is called. Returns: A modified version of image, where each bbox location in the image will have augmentation_func applied to it if it is chosen to be called with probability prob independently across all bboxes. Also the final bboxes are returned that will be unchanged if func_changes_bbox is set to false and if true, the new altered ones will be returned. """ # Will keep track of the new altered bboxes after aug_func is repeatedly # applied. The -1 values are a dummy value and this first Tensor will be # removed upon appending the first real bbox. new_bboxes = tf.constant(_INVALID_BOX) # If the bboxes are empty, then just give it _INVALID_BOX. The result # will be thrown away. bboxes = tf.cond(tf.equal(tf.shape(bboxes)[0], 0), lambda: tf.constant(_INVALID_BOX), lambda: bboxes) bboxes = tf.ensure_shape(bboxes, (None, 4)) # pylint:disable=g-long-lambda # pylint:disable=line-too-long wrapped_aug_func = lambda _image, bbox, _new_bboxes: _apply_bbox_augmentation_wrapper( _image, bbox, _new_bboxes, prob, aug_func, func_changes_bbox, *args) # pylint:enable=g-long-lambda # pylint:enable=line-too-long # Setup the while_loop. num_bboxes = tf.shape(bboxes)[0] # We loop until we go over all bboxes. idx = tf.constant(0) # Counter for the while loop. # Conditional function when to end the loop once we go over all bboxes # images_and_bboxes contain (_image, _new_bboxes) cond = lambda _idx, _images_and_bboxes: tf.less(_idx, num_bboxes) # Shuffle the bboxes so that the augmentation order is not deterministic if # we are not changing the bboxes with aug_func. if not func_changes_bbox: loop_bboxes = tf.random.shuffle(bboxes) else: loop_bboxes = bboxes # Main function of while_loop where we repeatedly apply augmentation on the # bboxes in the image. # pylint:disable=g-long-lambda body = lambda _idx, _images_and_bboxes: [ _idx + 1, wrapped_aug_func(_images_and_bboxes[0], loop_bboxes[_idx], _images_and_bboxes[1])] # pylint:enable=g-long-lambda _, (image, new_bboxes) = tf.while_loop( cond, body, [idx, (image, new_bboxes)], shape_invariants=[idx.get_shape(), (image.get_shape(), tf.TensorShape([None, 4]))]) # Either return the altered bboxes or the original ones depending on if # we altered them in anyway. if func_changes_bbox: final_bboxes = new_bboxes else: final_bboxes = bboxes return image, final_bboxes def _apply_multi_bbox_augmentation_wrapper(image, bboxes, prob, aug_func, func_changes_bbox, *args): """Checks to be sure num bboxes > 0 before calling inner function.""" num_bboxes = tf.shape(bboxes)[0] image, bboxes = tf.cond( tf.equal(num_bboxes, 0), lambda: (image, bboxes), # pylint:disable=g-long-lambda lambda: _apply_multi_bbox_augmentation( image, bboxes, prob, aug_func, func_changes_bbox, *args)) # pylint:enable=g-long-lambda return image, bboxes def rotate_only_bboxes(image, bboxes, prob, degrees, replace): """Apply rotate to each bbox in the image with probability prob.""" func_changes_bbox = False prob = _scale_bbox_only_op_probability(prob) return _apply_multi_bbox_augmentation_wrapper( image, bboxes, prob, rotate, func_changes_bbox, degrees, replace) def shear_x_only_bboxes(image, bboxes, prob, level, replace): """Apply shear_x to each bbox in the image with probability prob.""" func_changes_bbox = False prob = _scale_bbox_only_op_probability(prob) return _apply_multi_bbox_augmentation_wrapper( image, bboxes, prob, shear_x, func_changes_bbox, level, replace) def shear_y_only_bboxes(image, bboxes, prob, level, replace): """Apply shear_y to each bbox in the image with probability prob.""" func_changes_bbox = False prob = _scale_bbox_only_op_probability(prob) return _apply_multi_bbox_augmentation_wrapper( image, bboxes, prob, shear_y, func_changes_bbox, level, replace) def translate_x_only_bboxes(image, bboxes, prob, pixels, replace): """Apply translate_x to each bbox in the image with probability prob.""" func_changes_bbox = False prob = _scale_bbox_only_op_probability(prob) return _apply_multi_bbox_augmentation_wrapper( image, bboxes, prob, translate_x, func_changes_bbox, pixels, replace) def translate_y_only_bboxes(image, bboxes, prob, pixels, replace): """Apply translate_y to each bbox in the image with probability prob.""" func_changes_bbox = False prob = _scale_bbox_only_op_probability(prob) return _apply_multi_bbox_augmentation_wrapper( image, bboxes, prob, translate_y, func_changes_bbox, pixels, replace) def flip_only_bboxes(image, bboxes, prob): """Apply flip_lr to each bbox in the image with probability prob.""" func_changes_bbox = False prob = _scale_bbox_only_op_probability(prob) return _apply_multi_bbox_augmentation_wrapper( image, bboxes, prob, tf.image.flip_left_right, func_changes_bbox) def solarize_only_bboxes(image, bboxes, prob, threshold): """Apply solarize to each bbox in the image with probability prob.""" func_changes_bbox = False prob = _scale_bbox_only_op_probability(prob) return _apply_multi_bbox_augmentation_wrapper( image, bboxes, prob, solarize, func_changes_bbox, threshold) def equalize_only_bboxes(image, bboxes, prob): """Apply equalize to each bbox in the image with probability prob.""" func_changes_bbox = False prob = _scale_bbox_only_op_probability(prob) return _apply_multi_bbox_augmentation_wrapper( image, bboxes, prob, equalize, func_changes_bbox) def cutout_only_bboxes(image, bboxes, prob, pad_size, replace): """Apply cutout to each bbox in the image with probability prob.""" func_changes_bbox = False prob = _scale_bbox_only_op_probability(prob) return _apply_multi_bbox_augmentation_wrapper( image, bboxes, prob, cutout, func_changes_bbox, pad_size, replace) def _rotate_bbox(bbox, image_height, image_width, degrees): """Rotates the bbox coordinated by degrees. Args: bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. image_height: Int, height of the image. image_width: Int, height of the image. degrees: Float, a scalar angle in degrees to rotate all images by. If degrees is positive the image will be rotated clockwise otherwise it will be rotated counterclockwise. Returns: A tensor of the same shape as bbox, but now with the rotated coordinates. """ image_height, image_width = ( tf.to_float(image_height), tf.to_float(image_width)) # Convert from degrees to radians. degrees_to_radians = math.pi / 180.0 radians = degrees * degrees_to_radians # Translate the bbox to the center of the image and turn the normalized 0-1 # coordinates to absolute pixel locations. # Y coordinates are made negative as the y axis of images goes down with # increasing pixel values, so we negate to make sure x axis and y axis points # are in the traditionally positive direction. min_y = -tf.to_int32(image_height * (bbox[0] - 0.5)) min_x = tf.to_int32(image_width * (bbox[1] - 0.5)) max_y = -tf.to_int32(image_height * (bbox[2] - 0.5)) max_x = tf.to_int32(image_width * (bbox[3] - 0.5)) coordinates = tf.stack( [[min_y, min_x], [min_y, max_x], [max_y, min_x], [max_y, max_x]]) coordinates = tf.cast(coordinates, tf.float32) # Rotate the coordinates according to the rotation matrix clockwise if # radians is positive, else negative rotation_matrix = tf.stack( [[tf.cos(radians), tf.sin(radians)], [-tf.sin(radians), tf.cos(radians)]]) new_coords = tf.cast( tf.matmul(rotation_matrix, tf.transpose(coordinates)), tf.int32) # Find min/max values and convert them back to normalized 0-1 floats. min_y = -(tf.to_float(tf.reduce_max(new_coords[0, :])) / image_height - 0.5) min_x = tf.to_float(tf.reduce_min(new_coords[1, :])) / image_width + 0.5 max_y = -(tf.to_float(tf.reduce_min(new_coords[0, :])) / image_height - 0.5) max_x = tf.to_float(tf.reduce_max(new_coords[1, :])) / image_width + 0.5 # Clip the bboxes to be sure the fall between [0, 1]. min_y, min_x, max_y, max_x = _clip_bbox(min_y, min_x, max_y, max_x) min_y, min_x, max_y, max_x = _check_bbox_area(min_y, min_x, max_y, max_x) return tf.stack([min_y, min_x, max_y, max_x]) def rotate_with_bboxes(image, bboxes, degrees, replace): """Equivalent of PIL Rotate that rotates the image and bbox. Args: image: 3D uint8 Tensor. bboxes: 2D Tensor that is a list of the bboxes in the image. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float. degrees: Float, a scalar angle in degrees to rotate all images by. If degrees is positive the image will be rotated clockwise otherwise it will be rotated counterclockwise. replace: A one or three value 1D tensor to fill empty pixels. Returns: A tuple containing a 3D uint8 Tensor that will be the result of rotating image by degrees. The second element of the tuple is bboxes, where now the coordinates will be shifted to reflect the rotated image. """ # Rotate the image. image = rotate(image, degrees, replace) # Convert bbox coordinates to pixel values. image_height = tf.shape(image)[0] image_width = tf.shape(image)[1] # pylint:disable=g-long-lambda wrapped_rotate_bbox = lambda bbox: _rotate_bbox( bbox, image_height, image_width, degrees) # pylint:enable=g-long-lambda bboxes = tf.map_fn(wrapped_rotate_bbox, bboxes) return image, bboxes def translate_x(image, pixels, replace): """Equivalent of PIL Translate in X dimension.""" image = image_ops.translate(wrap(image), [-pixels, 0]) return unwrap(image, replace) def translate_y(image, pixels, replace): """Equivalent of PIL Translate in Y dimension.""" image = image_ops.translate(wrap(image), [0, -pixels]) return unwrap(image, replace) def _shift_bbox(bbox, image_height, image_width, pixels, shift_horizontal): """Shifts the bbox coordinates by pixels. Args: bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. image_height: Int, height of the image. image_width: Int, width of the image. pixels: An int. How many pixels to shift the bbox. shift_horizontal: Boolean. If true then shift in X dimension else shift in Y dimension. Returns: A tensor of the same shape as bbox, but now with the shifted coordinates. """ pixels = tf.to_int32(pixels) # Convert bbox to integer pixel locations. min_y = tf.to_int32(tf.to_float(image_height) * bbox[0]) min_x = tf.to_int32(tf.to_float(image_width) * bbox[1]) max_y = tf.to_int32(tf.to_float(image_height) * bbox[2]) max_x = tf.to_int32(tf.to_float(image_width) * bbox[3]) if shift_horizontal: min_x = tf.maximum(0, min_x - pixels) max_x = tf.minimum(image_width, max_x - pixels) else: min_y = tf.maximum(0, min_y - pixels) max_y = tf.minimum(image_height, max_y - pixels) # Convert bbox back to floats. min_y = tf.to_float(min_y) / tf.to_float(image_height) min_x = tf.to_float(min_x) / tf.to_float(image_width) max_y = tf.to_float(max_y) / tf.to_float(image_height) max_x = tf.to_float(max_x) / tf.to_float(image_width) # Clip the bboxes to be sure the fall between [0, 1]. min_y, min_x, max_y, max_x = _clip_bbox(min_y, min_x, max_y, max_x) min_y, min_x, max_y, max_x = _check_bbox_area(min_y, min_x, max_y, max_x) return tf.stack([min_y, min_x, max_y, max_x]) def translate_bbox(image, bboxes, pixels, replace, shift_horizontal): """Equivalent of PIL Translate in X/Y dimension that shifts image and bbox. Args: image: 3D uint8 Tensor. bboxes: 2D Tensor that is a list of the bboxes in the image. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float with values between [0, 1]. pixels: An int. How many pixels to shift the image and bboxes replace: A one or three value 1D tensor to fill empty pixels. shift_horizontal: Boolean. If true then shift in X dimension else shift in Y dimension. Returns: A tuple containing a 3D uint8 Tensor that will be the result of translating image by pixels. The second element of the tuple is bboxes, where now the coordinates will be shifted to reflect the shifted image. """ if shift_horizontal: image = translate_x(image, pixels, replace) else: image = translate_y(image, pixels, replace) # Convert bbox coordinates to pixel values. image_height = tf.shape(image)[0] image_width = tf.shape(image)[1] # pylint:disable=g-long-lambda wrapped_shift_bbox = lambda bbox: _shift_bbox( bbox, image_height, image_width, pixels, shift_horizontal) # pylint:enable=g-long-lambda bboxes = tf.map_fn(wrapped_shift_bbox, bboxes) return image, bboxes def shear_x(image, level, replace): """Equivalent of PIL Shearing in X dimension.""" # Shear parallel to x axis is a projective transform # with a matrix form of: # [1 level # 0 1]. image = image_ops.transform( wrap(image), [1., level, 0., 0., 1., 0., 0., 0.]) return unwrap(image, replace) def shear_y(image, level, replace): """Equivalent of PIL Shearing in Y dimension.""" # Shear parallel to y axis is a projective transform # with a matrix form of: # [1 0 # level 1]. image = image_ops.transform( wrap(image), [1., 0., 0., level, 1., 0., 0., 0.]) return unwrap(image, replace) def _shear_bbox(bbox, image_height, image_width, level, shear_horizontal): """Shifts the bbox according to how the image was sheared. Args: bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. image_height: Int, height of the image. image_width: Int, height of the image. level: Float. How much to shear the image. shear_horizontal: If true then shear in X dimension else shear in the Y dimension. Returns: A tensor of the same shape as bbox, but now with the shifted coordinates. """ image_height, image_width = ( tf.to_float(image_height), tf.to_float(image_width)) # Change bbox coordinates to be pixels. min_y = tf.to_int32(image_height * bbox[0]) min_x = tf.to_int32(image_width * bbox[1]) max_y = tf.to_int32(image_height * bbox[2]) max_x = tf.to_int32(image_width * bbox[3]) coordinates = tf.stack( [[min_y, min_x], [min_y, max_x], [max_y, min_x], [max_y, max_x]]) coordinates = tf.cast(coordinates, tf.float32) # Shear the coordinates according to the translation matrix. if shear_horizontal: translation_matrix = tf.stack( [[1, 0], [-level, 1]]) else: translation_matrix = tf.stack( [[1, -level], [0, 1]]) translation_matrix = tf.cast(translation_matrix, tf.float32) new_coords = tf.cast( tf.matmul(translation_matrix, tf.transpose(coordinates)), tf.int32) # Find min/max values and convert them back to floats. min_y = tf.to_float(tf.reduce_min(new_coords[0, :])) / image_height min_x = tf.to_float(tf.reduce_min(new_coords[1, :])) / image_width max_y = tf.to_float(tf.reduce_max(new_coords[0, :])) / image_height max_x = tf.to_float(tf.reduce_max(new_coords[1, :])) / image_width # Clip the bboxes to be sure the fall between [0, 1]. min_y, min_x, max_y, max_x = _clip_bbox(min_y, min_x, max_y, max_x) min_y, min_x, max_y, max_x = _check_bbox_area(min_y, min_x, max_y, max_x) return tf.stack([min_y, min_x, max_y, max_x]) def shear_with_bboxes(image, bboxes, level, replace, shear_horizontal): """Applies Shear Transformation to the image and shifts the bboxes. Args: image: 3D uint8 Tensor. bboxes: 2D Tensor that is a list of the bboxes in the image. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float with values between [0, 1]. level: Float. How much to shear the image. This value will be between -0.3 to 0.3. replace: A one or three value 1D tensor to fill empty pixels. shear_horizontal: Boolean. If true then shear in X dimension else shear in the Y dimension. Returns: A tuple containing a 3D uint8 Tensor that will be the result of shearing image by level. The second element of the tuple is bboxes, where now the coordinates will be shifted to reflect the sheared image. """ if shear_horizontal: image = shear_x(image, level, replace) else: image = shear_y(image, level, replace) # Convert bbox coordinates to pixel values. image_height = tf.shape(image)[0] image_width = tf.shape(image)[1] # pylint:disable=g-long-lambda wrapped_shear_bbox = lambda bbox: _shear_bbox( bbox, image_height, image_width, level, shear_horizontal) # pylint:enable=g-long-lambda bboxes = tf.map_fn(wrapped_shear_bbox, bboxes) return image, bboxes def autocontrast(image): """Implements Autocontrast function from PIL using TF ops. Args: image: A 3D uint8 tensor. Returns: The image after it has had autocontrast applied to it and will be of type uint8. """ def scale_channel(image): """Scale the 2D image using the autocontrast rule.""" # A possibly cheaper version can be done using cumsum/unique_with_counts # over the histogram values, rather than iterating over the entire image. # to compute mins and maxes. lo = tf.to_float(tf.reduce_min(image)) hi = tf.to_float(tf.reduce_max(image)) # Scale the image, making the lowest value 0 and the highest value 255. def scale_values(im): scale = 255.0 / (hi - lo) offset = -lo * scale im = tf.to_float(im) * scale + offset im = tf.clip_by_value(im, 0.0, 255.0) return tf.cast(im, tf.uint8) result = tf.cond(hi > lo, lambda: scale_values(image), lambda: image) return result # Assumes RGB for now. Scales each channel independently # and then stacks the result. s1 = scale_channel(image[:, :, 0]) s2 = scale_channel(image[:, :, 1]) s3 = scale_channel(image[:, :, 2]) image = tf.stack([s1, s2, s3], 2) return image def sharpness(image, factor): """Implements Sharpness function from PIL using TF ops.""" orig_image = image image = tf.cast(image, tf.float32) # Make image 4D for conv operation. image = tf.expand_dims(image, 0) # SMOOTH PIL Kernel. kernel = tf.constant( [[1, 1, 1], [1, 5, 1], [1, 1, 1]], dtype=tf.float32, shape=[3, 3, 1, 1]) / 13. # Tile across channel dimension. kernel = tf.tile(kernel, [1, 1, 3, 1]) strides = [1, 1, 1, 1] with tf.device('/cpu:0'): degenerate = tf.nn.depthwise_conv2d( image, kernel, strides, padding='VALID', rate=[1, 1]) degenerate = tf.clip_by_value(degenerate, 0.0, 255.0) degenerate = tf.squeeze(tf.cast(degenerate, tf.uint8), [0]) # For the borders of the resulting image, fill in the values of the # original image. mask = tf.ones_like(degenerate) padded_mask = tf.pad(mask, [[1, 1], [1, 1], [0, 0]]) padded_degenerate = tf.pad(degenerate, [[1, 1], [1, 1], [0, 0]]) result = tf.where(tf.equal(padded_mask, 1), padded_degenerate, orig_image) # Blend the final result. return blend(result, orig_image, factor) def equalize(image): """Implements Equalize function from PIL using TF ops.""" def scale_channel(im, c): """Scale the data in the channel to implement equalize.""" im = tf.cast(im[:, :, c], tf.int32) # Compute the histogram of the image channel. histo = tf.histogram_fixed_width(im, [0, 255], nbins=256) # For the purposes of computing the step, filter out the nonzeros. nonzero = tf.where(tf.not_equal(histo, 0)) nonzero_histo = tf.reshape(tf.gather(histo, nonzero), [-1]) step = (tf.reduce_sum(nonzero_histo) - nonzero_histo[-1]) // 255 def build_lut(histo, step): # Compute the cumulative sum, shifting by step // 2 # and then normalization by step. lut = (tf.cumsum(histo) + (step // 2)) // step # Shift lut, prepending with 0. lut = tf.concat([[0], lut[:-1]], 0) # Clip the counts to be in range. This is done # in the C code for image.point. return tf.clip_by_value(lut, 0, 255) # If step is zero, return the original image. Otherwise, build # lut from the full histogram and step and then index from it. result = tf.cond(tf.equal(step, 0), lambda: im, lambda: tf.gather(build_lut(histo, step), im)) return tf.cast(result, tf.uint8) # Assumes RGB for now. Scales each channel independently # and then stacks the result. s1 = scale_channel(image, 0) s2 = scale_channel(image, 1) s3 = scale_channel(image, 2) image = tf.stack([s1, s2, s3], 2) return image def wrap(image): """Returns 'image' with an extra channel set to all 1s.""" shape = tf.shape(image) extended_channel = tf.ones([shape[0], shape[1], 1], image.dtype) extended = tf.concat([image, extended_channel], 2) return extended def unwrap(image, replace): """Unwraps an image produced by wrap. Where there is a 0 in the last channel for every spatial position, the rest of the three channels in that spatial dimension are grayed (set to 128). Operations like translate and shear on a wrapped Tensor will leave 0s in empty locations. Some transformations look at the intensity of values to do preprocessing, and we want these empty pixels to assume the 'average' value, rather than pure black. Args: image: A 3D Image Tensor with 4 channels. replace: A one or three value 1D tensor to fill empty pixels. Returns: image: A 3D image Tensor with 3 channels. """ image_shape = tf.shape(image) # Flatten the spatial dimensions. flattened_image = tf.reshape(image, [-1, image_shape[2]]) # Find all pixels where the last channel is zero. alpha_channel = flattened_image[:, 3] replace = tf.concat([replace, tf.ones([1], image.dtype)], 0) # Where they are zero, fill them in with 'replace'. flattened_image = tf.where( tf.equal(alpha_channel, 0), tf.ones_like(flattened_image, dtype=image.dtype) * replace, flattened_image) image = tf.reshape(flattened_image, image_shape) image = tf.slice(image, [0, 0, 0], [image_shape[0], image_shape[1], 3]) return image def _cutout_inside_bbox(image, bbox, pad_fraction): """Generates cutout mask and the mean pixel value of the bbox. First a location is randomly chosen within the image as the center where the cutout mask will be applied. Note this can be towards the boundaries of the image, so the full cutout mask may not be applied. Args: image: 3D uint8 Tensor. bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. pad_fraction: Float that specifies how large the cutout mask should be in in reference to the size of the original bbox. If pad_fraction is 0.25, then the cutout mask will be of shape (0.25 * bbox height, 0.25 * bbox width). Returns: A tuple. Fist element is a tensor of the same shape as image where each element is either a 1 or 0 that is used to determine where the image will have cutout applied. The second element is the mean of the pixels in the image where the bbox is located. """ image_height = tf.maximum(tf.shape(image)[0], 10) image_width = tf.maximum(tf.shape(image)[1], 10) # Transform from shape [1, 4] to [4]. bbox = tf.squeeze(bbox) min_y = tf.to_int32(tf.to_float(image_height) * bbox[0]) min_x = tf.to_int32(tf.to_float(image_width) * bbox[1]) max_y = tf.to_int32(tf.to_float(image_height) * bbox[2]) max_x = tf.to_int32(tf.to_float(image_width) * bbox[3]) # Calculate the mean pixel values in the bounding box, which will be used # to fill the cutout region. mean = tf.reduce_mean(image[min_y:max_y + 1, min_x:max_x + 1], reduction_indices=[0, 1]) # Cutout mask will be size pad_size_heigh * 2 by pad_size_width * 2 if the # region lies entirely within the bbox. box_height = max_y - min_y + 1 box_width = max_x - min_x + 1 pad_size_height = tf.to_int32(pad_fraction * (box_height / 2)) pad_size_width = tf.to_int32(pad_fraction * (box_width / 2)) # Sample the center location in the image where the zero mask will be applied. cutout_center_height = tf.random_uniform( shape=[], minval=min_y, maxval=max_y+1, dtype=tf.int32) cutout_center_width = tf.random_uniform( shape=[], minval=min_x, maxval=max_x+1, dtype=tf.int32) lower_pad = tf.maximum( 0, cutout_center_height - pad_size_height) upper_pad = tf.maximum( 0, image_height - cutout_center_height - pad_size_height) left_pad = tf.maximum( 0, cutout_center_width - pad_size_width) right_pad = tf.maximum( 0, image_width - cutout_center_width - pad_size_width) cutout_shape = [image_height - (lower_pad + upper_pad), image_width - (left_pad + right_pad)] padding_dims = [[lower_pad, upper_pad], [left_pad, right_pad]] mask = tf.pad( tf.zeros(cutout_shape, dtype=image.dtype), padding_dims, constant_values=1) mask = tf.expand_dims(mask, 2) mask = tf.tile(mask, [1, 1, 3]) return mask, mean def bbox_cutout(image, bboxes, pad_fraction, replace_with_mean): """Applies cutout to the image according to bbox information. This is a cutout variant that using bbox information to make more informed decisions on where to place the cutout mask. Args: image: 3D uint8 Tensor. bboxes: 2D Tensor that is a list of the bboxes in the image. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float with values between [0, 1]. pad_fraction: Float that specifies how large the cutout mask should be in in reference to the size of the original bbox. If pad_fraction is 0.25, then the cutout mask will be of shape (0.25 * bbox height, 0.25 * bbox width). replace_with_mean: Boolean that specified what value should be filled in where the cutout mask is applied. Since the incoming image will be of uint8 and will not have had any mean normalization applied, by default we set the value to be 128. If replace_with_mean is True then we find the mean pixel values across the channel dimension and use those to fill in where the cutout mask is applied. Returns: A tuple. First element is a tensor of the same shape as image that has cutout applied to it. Second element is the bboxes that were passed in that will be unchanged. """ def apply_bbox_cutout(image, bboxes, pad_fraction): """Applies cutout to a single bounding box within image.""" # Choose a single bounding box to apply cutout to. random_index = tf.random_uniform( shape=[], maxval=tf.shape(bboxes)[0], dtype=tf.int32) # Select the corresponding bbox and apply cutout. chosen_bbox = tf.gather(bboxes, random_index) mask, mean = _cutout_inside_bbox(image, chosen_bbox, pad_fraction) # When applying cutout we either set the pixel value to 128 or to the mean # value inside the bbox. replace = mean if replace_with_mean else 128 # Apply the cutout mask to the image. Where the mask is 0 we fill it with # `replace`. image = tf.where( tf.equal(mask, 0), tf.cast(tf.ones_like(image, dtype=image.dtype) * replace, dtype=image.dtype), image) return image # Check to see if there are boxes, if so then apply boxcutout. image = tf.cond(tf.equal(tf.shape(bboxes)[0], 0), lambda: image, lambda: apply_bbox_cutout(image, bboxes, pad_fraction)) return image, bboxes NAME_TO_FUNC = { 'AutoContrast': autocontrast, 'Equalize': equalize, 'Posterize': posterize, 'Solarize': solarize, 'SolarizeAdd': solarize_add, 'Color': color, 'Contrast': contrast, 'Brightness': brightness, 'Sharpness': sharpness, 'Cutout': cutout, 'BBox_Cutout': bbox_cutout, 'Rotate_BBox': rotate_with_bboxes, # pylint:disable=g-long-lambda 'TranslateX_BBox': lambda image, bboxes, pixels, replace: translate_bbox( image, bboxes, pixels, replace, shift_horizontal=True), 'TranslateY_BBox': lambda image, bboxes, pixels, replace: translate_bbox( image, bboxes, pixels, replace, shift_horizontal=False), 'ShearX_BBox': lambda image, bboxes, level, replace: shear_with_bboxes( image, bboxes, level, replace, shear_horizontal=True), 'ShearY_BBox': lambda image, bboxes, level, replace: shear_with_bboxes( image, bboxes, level, replace, shear_horizontal=False), # pylint:enable=g-long-lambda 'Rotate_Only_BBoxes': rotate_only_bboxes, 'ShearX_Only_BBoxes': shear_x_only_bboxes, 'ShearY_Only_BBoxes': shear_y_only_bboxes, 'TranslateX_Only_BBoxes': translate_x_only_bboxes, 'TranslateY_Only_BBoxes': translate_y_only_bboxes, 'Flip_Only_BBoxes': flip_only_bboxes, 'Solarize_Only_BBoxes': solarize_only_bboxes, 'Equalize_Only_BBoxes': equalize_only_bboxes, 'Cutout_Only_BBoxes': cutout_only_bboxes, } def _randomly_negate_tensor(tensor): """With 50% prob turn the tensor negative.""" should_flip = tf.cast(tf.floor(tf.random_uniform([]) + 0.5), tf.bool) final_tensor = tf.cond(should_flip, lambda: tensor, lambda: -tensor) return final_tensor def _rotate_level_to_arg(level): level = (level/_MAX_LEVEL) * 30. level = _randomly_negate_tensor(level) return (level,) def _shrink_level_to_arg(level): """Converts level to ratio by which we shrink the image content.""" if level == 0: return (1.0,) # if level is zero, do not shrink the image # Maximum shrinking ratio is 2.9. level = 2. / (_MAX_LEVEL / level) + 0.9 return (level,) def _enhance_level_to_arg(level): return ((level/_MAX_LEVEL) * 1.8 + 0.1,) def _shear_level_to_arg(level): level = (level/_MAX_LEVEL) * 0.3 # Flip level to negative with 50% chance. level = _randomly_negate_tensor(level) return (level,) def _translate_level_to_arg(level, translate_const): level = (level/_MAX_LEVEL) * float(translate_const) # Flip level to negative with 50% chance. level = _randomly_negate_tensor(level) return (level,) def _bbox_cutout_level_to_arg(level, hparams): cutout_pad_fraction = (level/_MAX_LEVEL) * hparams.cutout_max_pad_fraction return (cutout_pad_fraction, hparams.cutout_bbox_replace_with_mean) def level_to_arg(hparams): return { 'AutoContrast': lambda level: (), 'Equalize': lambda level: (), 'Posterize': lambda level: (int((level/_MAX_LEVEL) * 4),), 'Solarize': lambda level: (int((level/_MAX_LEVEL) * 256),), 'SolarizeAdd': lambda level: (int((level/_MAX_LEVEL) * 110),), 'Color': _enhance_level_to_arg, 'Contrast': _enhance_level_to_arg, 'Brightness': _enhance_level_to_arg, 'Sharpness': _enhance_level_to_arg, 'Cutout': lambda level: (int((level/_MAX_LEVEL) * hparams.cutout_const),), # pylint:disable=g-long-lambda 'BBox_Cutout': lambda level: _bbox_cutout_level_to_arg( level, hparams), 'TranslateX_BBox': lambda level: _translate_level_to_arg( level, hparams.translate_const), 'TranslateY_BBox': lambda level: _translate_level_to_arg( level, hparams.translate_const), # pylint:enable=g-long-lambda 'ShearX_BBox': _shear_level_to_arg, 'ShearY_BBox': _shear_level_to_arg, 'Rotate_BBox': _rotate_level_to_arg, 'Rotate_Only_BBoxes': _rotate_level_to_arg, 'ShearX_Only_BBoxes': _shear_level_to_arg, 'ShearY_Only_BBoxes': _shear_level_to_arg, # pylint:disable=g-long-lambda 'TranslateX_Only_BBoxes': lambda level: _translate_level_to_arg( level, hparams.translate_bbox_const), 'TranslateY_Only_BBoxes': lambda level: _translate_level_to_arg( level, hparams.translate_bbox_const), # pylint:enable=g-long-lambda 'Flip_Only_BBoxes': lambda level: (), 'Solarize_Only_BBoxes': lambda level: (int((level/_MAX_LEVEL) * 256),), 'Equalize_Only_BBoxes': lambda level: (), # pylint:disable=g-long-lambda 'Cutout_Only_BBoxes': lambda level: ( int((level/_MAX_LEVEL) * hparams.cutout_bbox_const),), # pylint:enable=g-long-lambda } def bbox_wrapper(func): """Adds a bboxes function argument to func and returns unchanged bboxes.""" def wrapper(images, bboxes, *args, **kwargs): return (func(images, *args, **kwargs), bboxes) return wrapper def _parse_policy_info(name, prob, level, replace_value, augmentation_hparams): """Return the function that corresponds to `name` and update `level` param.""" func = NAME_TO_FUNC[name] args = level_to_arg(augmentation_hparams)[name](level) # Check to see if prob is passed into function. This is used for operations # where we alter bboxes independently. # pytype:disable=wrong-arg-types if 'prob' in inspect.getfullargspec(func)[0]: args = tuple([prob] + list(args)) # pytype:enable=wrong-arg-types # Add in replace arg if it is required for the function that is being called. if 'replace' in inspect.getfullargspec(func)[0]: # Make sure replace is the final argument assert 'replace' == inspect.getfullargspec(func)[0][-1] args = tuple(list(args) + [replace_value]) # Add bboxes as the second positional argument for the function if it does # not already exist. if 'bboxes' not in inspect.getfullargspec(func)[0]: func = bbox_wrapper(func) return (func, prob, args) def _apply_func_with_prob(func, image, args, prob, bboxes): """Apply `func` to image w/ `args` as input with probability `prob`.""" assert isinstance(args, tuple) assert 'bboxes' == inspect.getfullargspec(func)[0][1] # If prob is a function argument, then this randomness is being handled # inside the function, so make sure it is always called. if 'prob' in inspect.getfullargspec(func)[0]: prob = 1.0 # Apply the function with probability `prob`. should_apply_op = tf.cast( tf.floor(tf.random_uniform([], dtype=tf.float32) + prob), tf.bool) augmented_image, augmented_bboxes = tf.cond( should_apply_op, lambda: func(image, bboxes, *args), lambda: (image, bboxes)) return augmented_image, augmented_bboxes def select_and_apply_random_policy(policies, image, bboxes): """Select a random policy from `policies` and apply it to `image`.""" policy_to_select = tf.random_uniform([], maxval=len(policies), dtype=tf.int32) # Note that using tf.case instead of tf.conds would result in significantly # larger graphs and would even break export for some larger policies. for (i, policy) in enumerate(policies): image, bboxes = tf.cond( tf.equal(i, policy_to_select), lambda selected_policy=policy: selected_policy(image, bboxes), lambda: (image, bboxes)) return (image, bboxes) def build_and_apply_nas_policy(policies, image, bboxes, augmentation_hparams): """Build a policy from the given policies passed in and apply to image. Args: policies: list of lists of tuples in the form `(func, prob, level)`, `func` is a string name of the augmentation function, `prob` is the probability of applying the `func` operation, `level` is the input argument for `func`. image: tf.Tensor that the resulting policy will be applied to. bboxes: tf.Tensor of shape [N, 4] representing ground truth boxes that are normalized between [0, 1]. augmentation_hparams: Hparams associated with the NAS learned policy. Returns: A version of image that now has data augmentation applied to it based on the `policies` pass into the function. Additionally, returns bboxes if a value for them is passed in that is not None """ replace_value = [128, 128, 128] # func is the string name of the augmentation function, prob is the # probability of applying the operation and level is the parameter associated # with the tf op. # tf_policies are functions that take in an image and return an augmented # image. tf_policies = [] for policy in policies: tf_policy = [] # Link string name to the correct python function and make sure the correct # argument is passed into that function. for policy_info in policy: policy_info = list(policy_info) + [replace_value, augmentation_hparams] tf_policy.append(_parse_policy_info(*policy_info)) # Now build the tf policy that will apply the augmentation procedue # on image. def make_final_policy(tf_policy_): def final_policy(image_, bboxes_): for func, prob, args in tf_policy_: image_, bboxes_ = _apply_func_with_prob( func, image_, args, prob, bboxes_) return image_, bboxes_ return final_policy tf_policies.append(make_final_policy(tf_policy)) augmented_images, augmented_bboxes = select_and_apply_random_policy( tf_policies, image, bboxes) # If no bounding boxes were specified, then just return the images. return (augmented_images, augmented_bboxes) @tf.autograph.experimental.do_not_convert def distort_image_with_autoaugment(image, bboxes, augmentation_name): """Applies the AutoAugment policy to `image` and `bboxes`. Args: image: `Tensor` of shape [height, width, 3] representing an image. bboxes: `Tensor` of shape [N, 4] representing ground truth boxes that are normalized between [0, 1]. augmentation_name: The name of the AutoAugment policy to use. The available options are `v0`, `v1`, `v2`, `v3` and `test`. `v0` is the policy used for all of the results in the paper and was found to achieve the best results on the COCO dataset. `v1`, `v2` and `v3` are additional good policies found on the COCO dataset that have slight variation in what operations were used during the search procedure along with how many operations are applied in parallel to a single image (2 vs 3). Returns: A tuple containing the augmented versions of `image` and `bboxes`. """ logging.info('Using autoaugmention policy: %s', augmentation_name) available_policies = {'v0': policy_v0, 'v1': policy_v1, 'v2': policy_v2, 'v3': policy_v3, 'test': policy_vtest} if augmentation_name not in available_policies: raise ValueError('Invalid augmentation_name: {}'.format(augmentation_name)) policy = available_policies[augmentation_name]() # Hparams that will be used for AutoAugment. augmentation_hparams = hparams_config.Config(dict( cutout_max_pad_fraction=0.75, cutout_bbox_replace_with_mean=False, cutout_const=100, translate_const=250, cutout_bbox_const=50, translate_bbox_const=120)) return build_and_apply_nas_policy(policy, image, bboxes, augmentation_hparams) def distort_image_with_randaugment(image, bboxes, num_layers, magnitude): """Applies the RandAugment to `image` and `bboxes`.""" replace_value = [128, 128, 128] tf.logging.info('Using RandAugment.') augmentation_hparams = hparams_config.Config( dict( cutout_max_pad_fraction=0.75, cutout_bbox_replace_with_mean=False, cutout_const=100, translate_const=250, cutout_bbox_const=50, translate_bbox_const=120)) available_ops = [ 'Equalize', 'Solarize', 'Color', 'Cutout', 'SolarizeAdd', 'TranslateX_BBox', 'TranslateY_BBox', 'ShearX_BBox', 'ShearY_BBox', 'Rotate_BBox'] if bboxes is None: bboxes = tf.constant(0.0) for layer_num in range(num_layers): op_to_select = tf.random_uniform( [], maxval=len(available_ops), dtype=tf.int32) random_magnitude = float(magnitude) with tf.name_scope('randaug_layer_{}'.format(layer_num)): for (i, op_name) in enumerate(available_ops): prob = tf.random_uniform([], minval=0.2, maxval=0.8, dtype=tf.float32) func, _, args = _parse_policy_info(op_name, prob, random_magnitude, replace_value, augmentation_hparams) image, bboxes = tf.cond( tf.equal(i, op_to_select), lambda fn=func, fn_args=args: fn(image, bboxes, *fn_args), lambda: (image, bboxes)) return (image, bboxes)
efficientdet/aug/autoaugment.py
66,666
Applies augmentation_func to the subsection of image indicated by bbox. Args: image: 3D uint8 Tensor. bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. augmentation_func: Augmentation function that will be applied to the subsection of image. *args: Additional parameters that will be passed into augmentation_func when it is called. Returns: A modified version of image, where the bbox location in the image will have `ugmentation_func applied to it. Applies _apply_bbox_augmentation with probability prob. Args: image: 3D uint8 Tensor. bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. new_bboxes: 2D Tensor that is a list of the bboxes in the image after they have been altered by aug_func. These will only be changed when func_changes_bbox is set to true. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float that are the normalized bbox coordinates between 0 and 1. prob: Float that is the probability of applying _apply_bbox_augmentation. augmentation_func: Augmentation function that will be applied to the subsection of image. func_changes_bbox: Boolean. Does augmentation_func return bbox in addition to image. *args: Additional parameters that will be passed into augmentation_func when it is called. Returns: A tuple. Fist element is a modified version of image, where the bbox location in the image will have augmentation_func applied to it if it is chosen to be called with probability `prob`. The second element is a Tensor of Tensors of length 4 that will contain the altered bbox after applying augmentation_func. Apply `func` to image w/ `args` as input with probability `prob`. Applies aug_func to the image for each bbox in bboxes. Args: image: 3D uint8 Tensor. bboxes: 2D Tensor that is a list of the bboxes in the image. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float. prob: Float that is the probability of applying aug_func to a specific bounding box within the image. aug_func: Augmentation function that will be applied to the subsections of image indicated by the bbox values in bboxes. func_changes_bbox: Boolean. Does augmentation_func return bbox in addition to image. *args: Additional parameters that will be passed into augmentation_func when it is called. Returns: A modified version of image, where each bbox location in the image will have augmentation_func applied to it if it is chosen to be called with probability prob independently across all bboxes. Also the final bboxes are returned that will be unchanged if func_changes_bbox is set to false and if true, the new altered ones will be returned. Checks to be sure num bboxes > 0 before calling inner function. Adjusts bbox coordinates to make sure the area is > 0. Args: min_y: Normalized bbox coordinate of type float between 0 and 1. min_x: Normalized bbox coordinate of type float between 0 and 1. max_y: Normalized bbox coordinate of type float between 0 and 1. max_x: Normalized bbox coordinate of type float between 0 and 1. delta: Float, this is used to create a gap of size 2 * delta between bbox min/max coordinates that are the same on the boundary. This prevents the bbox from having an area of zero. Returns: Tuple of new bbox coordinates between 0 and 1 that will now have a guaranteed area > 0. Clip bounding box coordinates between 0 and 1. Args: min_y: Normalized bbox coordinate of type float between 0 and 1. min_x: Normalized bbox coordinate of type float between 0 and 1. max_y: Normalized bbox coordinate of type float between 0 and 1. max_x: Normalized bbox coordinate of type float between 0 and 1. Returns: Clipped coordinate values between 0 and 1. Helper function that concats bbox to bboxes along the first dimension. Generates cutout mask and the mean pixel value of the bbox. First a location is randomly chosen within the image as the center where the cutout mask will be applied. Note this can be towards the boundaries of the image, so the full cutout mask may not be applied. Args: image: 3D uint8 Tensor. bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. pad_fraction: Float that specifies how large the cutout mask should be in in reference to the size of the original bbox. If pad_fraction is 0.25, then the cutout mask will be of shape (0.25 * bbox height, 0.25 * bbox width). Returns: A tuple. Fist element is a tensor of the same shape as image where each element is either a 1 or 0 that is used to determine where the image will have cutout applied. The second element is the mean of the pixels in the image where the bbox is located. Return the function that corresponds to `name` and update `level` param. With 50% prob turn the tensor negative. Rotates the bbox coordinated by degrees. Args: bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. image_height: Int, height of the image. image_width: Int, height of the image. degrees: Float, a scalar angle in degrees to rotate all images by. If degrees is positive the image will be rotated clockwise otherwise it will be rotated counterclockwise. Returns: A tensor of the same shape as bbox, but now with the rotated coordinates. Reduce the probability of the bbox-only operation. Probability is reduced so that we do not distort the content of too many bounding boxes that are close to each other. The value of 3.0 was a chosen hyper parameter when designing the autoaugment algorithm that we found empirically to work well. Args: prob: Float that is the probability of applying the bbox-only operation. Returns: Reduced probability. Shifts the bbox according to how the image was sheared. Args: bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. image_height: Int, height of the image. image_width: Int, height of the image. level: Float. How much to shear the image. shear_horizontal: If true then shear in X dimension else shear in the Y dimension. Returns: A tensor of the same shape as bbox, but now with the shifted coordinates. Shifts the bbox coordinates by pixels. Args: bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. image_height: Int, height of the image. image_width: Int, width of the image. pixels: An int. How many pixels to shift the bbox. shift_horizontal: Boolean. If true then shift in X dimension else shift in Y dimension. Returns: A tensor of the same shape as bbox, but now with the shifted coordinates. Converts level to ratio by which we shrink the image content. Applies cutout to a single bounding box within image. Implements Autocontrast function from PIL using TF ops. Args: image: A 3D uint8 tensor. Returns: The image after it has had autocontrast applied to it and will be of type uint8. Applies cutout to the image according to bbox information. This is a cutout variant that using bbox information to make more informed decisions on where to place the cutout mask. Args: image: 3D uint8 Tensor. bboxes: 2D Tensor that is a list of the bboxes in the image. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float with values between [0, 1]. pad_fraction: Float that specifies how large the cutout mask should be in in reference to the size of the original bbox. If pad_fraction is 0.25, then the cutout mask will be of shape (0.25 * bbox height, 0.25 * bbox width). replace_with_mean: Boolean that specified what value should be filled in where the cutout mask is applied. Since the incoming image will be of uint8 and will not have had any mean normalization applied, by default we set the value to be 128. If replace_with_mean is True then we find the mean pixel values across the channel dimension and use those to fill in where the cutout mask is applied. Returns: A tuple. First element is a tensor of the same shape as image that has cutout applied to it. Second element is the bboxes that were passed in that will be unchanged. Adds a bboxes function argument to func and returns unchanged bboxes. Blend image1 and image2 using 'factor'. Factor can be above 0.0. A value of 0.0 means only image1 is used. A value of 1.0 means only image2 is used. A value between 0.0 and 1.0 means we linearly interpolate the pixel values between the two images. A value greater than 1.0 "extrapolates" the difference between the two pixel values, and we clip the results to values between 0 and 255. Args: image1: An image Tensor of type uint8. image2: An image Tensor of type uint8. factor: A floating point value above 0.0. Returns: A blended image Tensor of type uint8. Equivalent of PIL Brightness. Build a policy from the given policies passed in and apply to image. Args: policies: list of lists of tuples in the form `(func, prob, level)`, `func` is a string name of the augmentation function, `prob` is the probability of applying the `func` operation, `level` is the input argument for `func`. image: tf.Tensor that the resulting policy will be applied to. bboxes: tf.Tensor of shape [N, 4] representing ground truth boxes that are normalized between [0, 1]. augmentation_hparams: Hparams associated with the NAS learned policy. Returns: A version of image that now has data augmentation applied to it based on the `policies` pass into the function. Additionally, returns bboxes if a value for them is passed in that is not None Equivalent of PIL Color. Equivalent of PIL Contrast. Apply cutout (https://arxiv.org/abs/1708.04552) to image. This operation applies a (2*pad_size x 2*pad_size) mask of zeros to a random location within `img`. The pixel values filled in will be of the value `replace`. The located where the mask will be applied is randomly chosen uniformly over the whole image. Args: image: An image Tensor of type uint8. pad_size: Specifies how big the zero mask that will be generated is that is applied to the image. The mask will be of size (2*pad_size x 2*pad_size). replace: What pixel value to fill in the image in the area that has the cutout mask applied to it. Returns: An image Tensor that is of type uint8. Apply cutout to each bbox in the image with probability prob. Applies the AutoAugment policy to `image` and `bboxes`. Args: image: `Tensor` of shape [height, width, 3] representing an image. bboxes: `Tensor` of shape [N, 4] representing ground truth boxes that are normalized between [0, 1]. augmentation_name: The name of the AutoAugment policy to use. The available options are `v0`, `v1`, `v2`, `v3` and `test`. `v0` is the policy used for all of the results in the paper and was found to achieve the best results on the COCO dataset. `v1`, `v2` and `v3` are additional good policies found on the COCO dataset that have slight variation in what operations were used during the search procedure along with how many operations are applied in parallel to a single image (2 vs 3). Returns: A tuple containing the augmented versions of `image` and `bboxes`. Applies the RandAugment to `image` and `bboxes`. Implements Equalize function from PIL using TF ops. Apply equalize to each bbox in the image with probability prob. Apply flip_lr to each bbox in the image with probability prob. Applies mask to bbox region in image then adds content_tensor to it. Autoaugment policy that was used in AutoAugment Detection Paper. Autoaugment policy that was used in AutoAugment Detection Paper. Additional policy that performs well on object detection. "Additional policy that performs well on object detection. Autoaugment test policy for debugging. Equivalent of PIL Posterize. Move the bbox and the image content to a slightly new random location. Args: image: 3D uint8 Tensor. bbox: 1D Tensor that has 4 elements (min_y, min_x, max_y, max_x) of type float that represents the normalized coordinates between 0 and 1. The potential values for the new min corner of the bbox will be between [old_min - pixel_scaling * bbox_height/2, old_min - pixel_scaling * bbox_height/2]. pixel_scaling: A float between 0 and 1 that specifies the pixel range that the new bbox location will be sampled from. replace: A one or three value 1D tensor to fill empty pixels. new_min_bbox_coords: If not None, then this is a tuple that specifies the (min_y, min_x) coordinates of the new bbox. Normally this is randomly specified, but this allows it to be manually set. The coordinates are the absolute coordinates between 0 and image height/width and are int32. Returns: The new image that will have the shifted bbox location in it along with the new bbox that contains the new coordinates. Rotates the image by degrees either clockwise or counterclockwise. Args: image: An image Tensor of type uint8. degrees: Float, a scalar angle in degrees to rotate all images by. If degrees is positive the image will be rotated clockwise otherwise it will be rotated counterclockwise. replace: A one or three value 1D tensor to fill empty pixels caused by the rotate operation. Returns: The rotated version of image. Apply rotate to each bbox in the image with probability prob. Equivalent of PIL Rotate that rotates the image and bbox. Args: image: 3D uint8 Tensor. bboxes: 2D Tensor that is a list of the bboxes in the image. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float. degrees: Float, a scalar angle in degrees to rotate all images by. If degrees is positive the image will be rotated clockwise otherwise it will be rotated counterclockwise. replace: A one or three value 1D tensor to fill empty pixels. Returns: A tuple containing a 3D uint8 Tensor that will be the result of rotating image by degrees. The second element of the tuple is bboxes, where now the coordinates will be shifted to reflect the rotated image. Scale the 2D image using the autocontrast rule. Scale the data in the channel to implement equalize. Select a random policy from `policies` and apply it to `image`. Implements Sharpness function from PIL using TF ops. Applies Shear Transformation to the image and shifts the bboxes. Args: image: 3D uint8 Tensor. bboxes: 2D Tensor that is a list of the bboxes in the image. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float with values between [0, 1]. level: Float. How much to shear the image. This value will be between -0.3 to 0.3. replace: A one or three value 1D tensor to fill empty pixels. shear_horizontal: Boolean. If true then shear in X dimension else shear in the Y dimension. Returns: A tuple containing a 3D uint8 Tensor that will be the result of shearing image by level. The second element of the tuple is bboxes, where now the coordinates will be shifted to reflect the sheared image. Equivalent of PIL Shearing in X dimension. Apply shear_x to each bbox in the image with probability prob. Equivalent of PIL Shearing in Y dimension. Apply shear_y to each bbox in the image with probability prob. Apply solarize to each bbox in the image with probability prob. Equivalent of PIL Translate in X/Y dimension that shifts image and bbox. Args: image: 3D uint8 Tensor. bboxes: 2D Tensor that is a list of the bboxes in the image. Each bbox has 4 elements (min_y, min_x, max_y, max_x) of type float with values between [0, 1]. pixels: An int. How many pixels to shift the image and bboxes replace: A one or three value 1D tensor to fill empty pixels. shift_horizontal: Boolean. If true then shift in X dimension else shift in Y dimension. Returns: A tuple containing a 3D uint8 Tensor that will be the result of translating image by pixels. The second element of the tuple is bboxes, where now the coordinates will be shifted to reflect the shifted image. Equivalent of PIL Translate in X dimension. Apply translate_x to each bbox in the image with probability prob. Equivalent of PIL Translate in Y dimension. Apply translate_y to each bbox in the image with probability prob. Unwraps an image produced by wrap. Where there is a 0 in the last channel for every spatial position, the rest of the three channels in that spatial dimension are grayed (set to 128). Operations like translate and shear on a wrapped Tensor will leave 0s in empty locations. Some transformations look at the intensity of values to do preprocessing, and we want these empty pixels to assume the 'average' value, rather than pure black. Args: image: A 3D Image Tensor with 4 channels. replace: A one or three value 1D tensor to fill empty pixels. Returns: image: A 3D image Tensor with 3 channels. Returns 'image' with an extra channel set to all 1s. AutoAugment. [1] Barret, et al. Learning Data Augmentation Strategies for Object Detection. Arxiv: https://arxiv.org/abs/1906.11172 Copyright 2020 Google Research. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ============================================================================== This signifies the max integer that the controller RNN could predict for the augmentation scheme. Represents an invalid bounding box that is used for checking for padding lists of bounding box coordinates for a few augmentation operations Each tuple is an augmentation operation of the form (operation, probability, magnitude). Each element in policy is a sub-policy that will be applied sequentially on the image. Each tuple is an augmentation operation of the form (operation, probability, magnitude). Each element in policy is a sub-policy that will be applied sequentially on the image. Each tuple is an augmentation operation of the form (operation, probability, magnitude). Each element in policy is a sub-policy that will be applied sequentially on the image. Each tuple is an augmentation operation of the form (operation, probability, magnitude). Each element in policy is a sub-policy that will be applied sequentially on the image. Each tuple is an augmentation operation of the form (operation, probability, magnitude). Each element in policy is a sub-policy that will be applied sequentially on the image. [('Equalize', 0.0, 4), ('Cutout', 0.8, 10)], [('Rotate_BBox', 1.0, 10), ('Cutout', 1.0, 10)], Do addition in float. Interpolate Interpolation means we always stay within 0 and 255. Extrapolate: We need to clip and then cast. Sample the center location in the image where the zero mask will be applied. For each pixel in the image, select the pixel if the value is less than the threshold. Otherwise, subtract 255 from the pixel. For each pixel in the image less than threshold we add 'addition' amount to it and then clip the pixel value to be between 0 and 255. The value of 'addition' is between -128 and 128. Cast before calling tf.histogram. Compute the grayscale histogram, then compute the mean pixel value, and create a constant image size of that value. Use that as the blending degenerate target of the original image. Convert from degrees to radians. In practice, we should randomize the rotation degrees by flipping it negatively half the time, but that's done on 'degrees' outside of the function. Obtains image height and width and create helper clip functions. Convert bbox to pixel coordinates. Select the new min/max bbox ranges that are used for sampling the new min x/y coordinates of the shifted bbox. Sample and calculate the new unclipped min/max coordinates of the new bbox. Determine if any of the new bbox was shifted outside the current image. This is used for determining if any of the original bbox content should be discarded. Create the new bbox tensor by converting pixel integer values to floats. Copy the contents in the bbox and fill the old bbox location with gray (128). Zero out original bbox location. Fill in bbox content to new bbox location. Make sure max is never 0 and min is never 1. Clip to be sure the max values do not fall out of range. Get the sub-tensor that is the image within the bounding box region. Apply the augmentation function to the bbox portion of the image. Pad the augmented_bbox_content and the mask to match the shape of original image. Create a mask that will be used to zero out a part of the original image. Replace the old bbox content with the new augmented content. Note if all elements in bboxes are -1 (_INVALID_BOX), then this means we discard bboxes and start the bboxes Tensor with the current bbox. This check will be true when it is an _INVALID_BOX Will keep track of the new altered bboxes after aug_func is repeatedly applied. The -1 values are a dummy value and this first Tensor will be removed upon appending the first real bbox. If the bboxes are empty, then just give it _INVALID_BOX. The result will be thrown away. pylint:disable=g-long-lambda pylint:disable=line-too-long pylint:enable=g-long-lambda pylint:enable=line-too-long Setup the while_loop. We loop until we go over all bboxes. Counter for the while loop. Conditional function when to end the loop once we go over all bboxes images_and_bboxes contain (_image, _new_bboxes) Shuffle the bboxes so that the augmentation order is not deterministic if we are not changing the bboxes with aug_func. Main function of while_loop where we repeatedly apply augmentation on the bboxes in the image. pylint:disable=g-long-lambda pylint:enable=g-long-lambda Either return the altered bboxes or the original ones depending on if we altered them in anyway. pylint:disable=g-long-lambda pylint:enable=g-long-lambda Convert from degrees to radians. Translate the bbox to the center of the image and turn the normalized 0-1 coordinates to absolute pixel locations. Y coordinates are made negative as the y axis of images goes down with increasing pixel values, so we negate to make sure x axis and y axis points are in the traditionally positive direction. Rotate the coordinates according to the rotation matrix clockwise if radians is positive, else negative Find min/max values and convert them back to normalized 0-1 floats. Clip the bboxes to be sure the fall between [0, 1]. Rotate the image. Convert bbox coordinates to pixel values. pylint:disable=g-long-lambda pylint:enable=g-long-lambda Convert bbox to integer pixel locations. Convert bbox back to floats. Clip the bboxes to be sure the fall between [0, 1]. Convert bbox coordinates to pixel values. pylint:disable=g-long-lambda pylint:enable=g-long-lambda Shear parallel to x axis is a projective transform with a matrix form of: [1 level 0 1]. Shear parallel to y axis is a projective transform with a matrix form of: [1 0 level 1]. Change bbox coordinates to be pixels. Shear the coordinates according to the translation matrix. Find min/max values and convert them back to floats. Clip the bboxes to be sure the fall between [0, 1]. Convert bbox coordinates to pixel values. pylint:disable=g-long-lambda pylint:enable=g-long-lambda A possibly cheaper version can be done using cumsum/unique_with_counts over the histogram values, rather than iterating over the entire image. to compute mins and maxes. Scale the image, making the lowest value 0 and the highest value 255. Assumes RGB for now. Scales each channel independently and then stacks the result. Make image 4D for conv operation. SMOOTH PIL Kernel. Tile across channel dimension. For the borders of the resulting image, fill in the values of the original image. Blend the final result. Compute the histogram of the image channel. For the purposes of computing the step, filter out the nonzeros. Compute the cumulative sum, shifting by step // 2 and then normalization by step. Shift lut, prepending with 0. Clip the counts to be in range. This is done in the C code for image.point. If step is zero, return the original image. Otherwise, build lut from the full histogram and step and then index from it. Assumes RGB for now. Scales each channel independently and then stacks the result. Flatten the spatial dimensions. Find all pixels where the last channel is zero. Where they are zero, fill them in with 'replace'. Transform from shape [1, 4] to [4]. Calculate the mean pixel values in the bounding box, which will be used to fill the cutout region. Cutout mask will be size pad_size_heigh * 2 by pad_size_width * 2 if the region lies entirely within the bbox. Sample the center location in the image where the zero mask will be applied. Choose a single bounding box to apply cutout to. Select the corresponding bbox and apply cutout. When applying cutout we either set the pixel value to 128 or to the mean value inside the bbox. Apply the cutout mask to the image. Where the mask is 0 we fill it with `replace`. Check to see if there are boxes, if so then apply boxcutout. pylint:disable=g-long-lambda pylint:enable=g-long-lambda if level is zero, do not shrink the image Maximum shrinking ratio is 2.9. Flip level to negative with 50% chance. Flip level to negative with 50% chance. pylint:disable=g-long-lambda pylint:enable=g-long-lambda pylint:disable=g-long-lambda pylint:enable=g-long-lambda pylint:disable=g-long-lambda pylint:enable=g-long-lambda Check to see if prob is passed into function. This is used for operations where we alter bboxes independently. pytype:disable=wrong-arg-types pytype:enable=wrong-arg-types Add in replace arg if it is required for the function that is being called. Make sure replace is the final argument Add bboxes as the second positional argument for the function if it does not already exist. If prob is a function argument, then this randomness is being handled inside the function, so make sure it is always called. Apply the function with probability `prob`. Note that using tf.case instead of tf.conds would result in significantly larger graphs and would even break export for some larger policies. func is the string name of the augmentation function, prob is the probability of applying the operation and level is the parameter associated with the tf op. tf_policies are functions that take in an image and return an augmented image. Link string name to the correct python function and make sure the correct argument is passed into that function. Now build the tf policy that will apply the augmentation procedue on image. If no bounding boxes were specified, then just return the images. Hparams that will be used for AutoAugment.
27,376
en
0.829549
""" mbed SDK Copyright (c) 2011-2015 ARM Limited Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. Author: Przemyslaw Wirkus <Przemyslaw.Wirkus@arm.com> """ from mbed_os_tools.test.host_tests_plugins.module_copy_silabs import ( HostTestPluginCopyMethod_Silabs, load_plugin, )
packages/mbed-host-tests/mbed_host_tests/host_tests_plugins/module_copy_silabs.py
761
mbed SDK Copyright (c) 2011-2015 ARM Limited Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. Author: Przemyslaw Wirkus <Przemyslaw.Wirkus@arm.com>
624
en
0.801932
""" Tests for the linecache module """ import linecache import unittest import os.path from test import support FILENAME = linecache.__file__ INVALID_NAME = '!@$)(!@#_1' EMPTY = '' TESTS = 'inspect_fodder inspect_fodder2 mapping_tests' TESTS = TESTS.split() TEST_PATH = os.path.dirname(__file__) MODULES = "linecache abc".split() MODULE_PATH = os.path.dirname(FILENAME) SOURCE_1 = ''' " Docstring " def function(): return result ''' SOURCE_2 = ''' def f(): return 1 + 1 a = f() ''' SOURCE_3 = ''' def f(): return 3''' # No ending newline class LineCacheTests(unittest.TestCase): def test_getline(self): getline = linecache.getline # Bad values for line number should return an empty string self.assertEqual(getline(FILENAME, 2**15), EMPTY) self.assertEqual(getline(FILENAME, -1), EMPTY) # Float values currently raise TypeError, should it? self.assertRaises(TypeError, getline, FILENAME, 1.1) # Bad filenames should return an empty string self.assertEqual(getline(EMPTY, 1), EMPTY) self.assertEqual(getline(INVALID_NAME, 1), EMPTY) # Check whether lines correspond to those from file iteration for entry in TESTS: filename = os.path.join(TEST_PATH, entry) + '.py' with open(filename) as file: for index, line in enumerate(file): self.assertEqual(line, getline(filename, index + 1)) # Check module loading for entry in MODULES: filename = os.path.join(MODULE_PATH, entry) + '.py' with open(filename) as file: for index, line in enumerate(file): self.assertEqual(line, getline(filename, index + 1)) # Check that bogus data isn't returned (issue #1309567) empty = linecache.getlines('a/b/c/__init__.py') self.assertEqual(empty, []) def test_no_ending_newline(self): self.addCleanup(support.unlink, support.TESTFN) with open(support.TESTFN, "w") as fp: fp.write(SOURCE_3) lines = linecache.getlines(support.TESTFN) self.assertEqual(lines, ["\n", "def f():\n", " return 3\n"]) def test_clearcache(self): cached = [] for entry in TESTS: filename = os.path.join(TEST_PATH, entry) + '.py' cached.append(filename) linecache.getline(filename, 1) # Are all files cached? cached_empty = [fn for fn in cached if fn not in linecache.cache] self.assertEqual(cached_empty, []) # Can we clear the cache? linecache.clearcache() cached_empty = [fn for fn in cached if fn in linecache.cache] self.assertEqual(cached_empty, []) def test_checkcache(self): getline = linecache.getline # Create a source file and cache its contents source_name = support.TESTFN + '.py' self.addCleanup(support.unlink, source_name) with open(source_name, 'w') as source: source.write(SOURCE_1) getline(source_name, 1) # Keep a copy of the old contents source_list = [] with open(source_name) as source: for index, line in enumerate(source): self.assertEqual(line, getline(source_name, index + 1)) source_list.append(line) with open(source_name, 'w') as source: source.write(SOURCE_2) # Try to update a bogus cache entry linecache.checkcache('dummy') # Check that the cache matches the old contents for index, line in enumerate(source_list): self.assertEqual(line, getline(source_name, index + 1)) # Update the cache and check whether it matches the new source file linecache.checkcache(source_name) with open(source_name) as source: for index, line in enumerate(source): self.assertEqual(line, getline(source_name, index + 1)) source_list.append(line) def test_memoryerror(self): lines = linecache.getlines(FILENAME) self.assertTrue(lines) def raise_memoryerror(*args, **kwargs): raise MemoryError with support.swap_attr(linecache, 'updatecache', raise_memoryerror): lines2 = linecache.getlines(FILENAME) self.assertEqual(lines2, lines) linecache.clearcache() with support.swap_attr(linecache, 'updatecache', raise_memoryerror): lines3 = linecache.getlines(FILENAME) self.assertEqual(lines3, []) self.assertEqual(linecache.getlines(FILENAME), lines) if __name__ == "__main__": unittest.main()
Src/StdLib/Lib/test/test_linecache.py
4,691
Tests for the linecache module No ending newline Bad values for line number should return an empty string Float values currently raise TypeError, should it? Bad filenames should return an empty string Check whether lines correspond to those from file iteration Check module loading Check that bogus data isn't returned (issue 1309567) Are all files cached? Can we clear the cache? Create a source file and cache its contents Keep a copy of the old contents Try to update a bogus cache entry Check that the cache matches the old contents Update the cache and check whether it matches the new source file
605
en
0.719135
import torch import os import os.path import shutil import numpy as np import soundfile as sf from pathlib import PurePath from torch import nn from torch.utils.data import DataLoader, random_split from asteroid.data import TimitDataset from asteroid.data.utils import CachedWavSet, RandomMixtureSet, FixedMixtureSet from tqdm import tqdm from torch import optim from pytorch_lightning import Trainer, seed_everything, loggers as pl_loggers from pytorch_lightning.callbacks.early_stopping import EarlyStopping from pytorch_lightning.callbacks import ModelCheckpoint from asteroid_filterbanks.transforms import mag from asteroid.engine import System from asteroid.losses import singlesrc_neg_sisdr from egs.whamr.TasNet.model import TasNet BATCH_SIZE = 8 # could be more on cluster, test if larger one work SAMPLE_RATE = 8000 # as agreed upon CROP_LEN = 24000 # average track len in TIMIT SEED = 42 # magic number :) def sisdr_loss_wrapper(est_target, target): return singlesrc_neg_sisdr(est_target.squeeze(1), target).mean() def train_val_split(ds, val_fraction=0.1, random_seed=SEED): assert val_fraction > 0 and val_fraction < 0.5 len_train = int(len(ds) * (1 - val_fraction)) len_val = len(ds) - len_train return random_split(ds, [len_train, len_val], generator=torch.Generator().manual_seed(random_seed)) DRONE_NOISE_DIR = '/jmain01/home/JAD007/txk02/aaa18-txk02/Datasets/noises-train-drones' # fixed SNRs for validation set TRAIN_SNRS = [-25, -20, -15, -10, -5] TIMIT_DIR = PurePath('/jmain01/home/JAD007/txk02/aaa18-txk02/Datasets/TIMIT') TIMIT_DIR_8kHZ = PurePath('/jmain01/home/JAD007/txk02/aaa18-txk02/Datasets/TIMIT_8kHZ') # Reproducibility - fix all random seeds seed_everything(SEED) # Load noises, resample and save into the memory noises = CachedWavSet(DRONE_NOISE_DIR, sample_rate=SAMPLE_RATE, precache=True) # Load clean data and split it into train and val timit = TimitDataset(TIMIT_DIR_8kHZ, subset='train', sample_rate=SAMPLE_RATE, with_path=False) timit_train, timit_val = train_val_split(timit, val_fraction=0.1, random_seed=SEED) # Training data mixes crops randomly on the fly with random SNR in range (effectively infinite training data) # `repeat_factor=20` means that the dataset contains 20 copies of itself - it is the easiest way to make the epoch longer timit_train = RandomMixtureSet(timit_train, noises, random_seed=SEED, snr_range=(-25, -5), crop_length=CROP_LEN, repeat_factor=30) # Validation data is fixed (for stability): mix every clean clip with all the noises in the folder # Argument `mixtures_per_clean` regulates with how many different noise files each clean file will be mixed timit_val = FixedMixtureSet(timit_val, noises, snrs=TRAIN_SNRS, random_seed=SEED, mixtures_per_clean=5, crop_length=CROP_LEN) NUM_WORKERS = 5 train_loader = DataLoader(timit_train, shuffle=True, batch_size=BATCH_SIZE, num_workers=NUM_WORKERS, drop_last=True) val_loader = DataLoader(timit_val, batch_size=BATCH_SIZE, num_workers=NUM_WORKERS, drop_last=True) # some random parameters, does it look sensible? LR = 1e-3 REDUCE_LR_PATIENCE = 5 EARLY_STOP_PATIENCE = 20 MAX_EPOCHS = 20 # the model here should be constructed in the script accordingly to the passed config (including the model type) # most of the models accept `sample_rate` parameter for encoders, which is important (default is 16000, override) model = TasNet(fb_conf={'n_filters': 512, 'kernel_size': 40, 'stride': 20}, mask_conf ={'n_layers': 4, 'n_units': 500, 'dropout': 0.3, "n_src": 1}) optimizer = optim.Adam(model.parameters(), lr=LR) scheduler = optim.lr_scheduler.ReduceLROnPlateau(optimizer, patience=REDUCE_LR_PATIENCE) early_stopping = EarlyStopping(monitor='val_loss', patience=EARLY_STOP_PATIENCE) checkpoint = ModelCheckpoint( filename='{epoch:02d}-{val_loss:.2f}', monitor="val_loss", mode="min", save_top_k=5, verbose=True ) # Probably we also need to subclass `System`, in order to log the target metrics on the validation set (PESQ/STOI) system = System(model, optimizer, sisdr_loss_wrapper, train_loader, val_loader, scheduler) # log dir and model name are also part of the config, of course LOG_DIR = 'logs' logger = pl_loggers.TensorBoardLogger(LOG_DIR, name='TIMIT-drones-TasNet-random_test', version=1) # choose the proper accelerator for JADE, probably `ddp` (also, `auto_select_gpus=True` might be useful) trainer = Trainer(max_epochs=MAX_EPOCHS, gpus=-1, logger=logger, callbacks=[early_stopping, checkpoint], deterministic=True, gradient_clip_val=5.0,) trainer.fit(system) #torch.save(model.serialize(), 'tasnet_model.pt')
notebooks/train_tasnet.py
4,814
could be more on cluster, test if larger one work as agreed upon average track len in TIMIT magic number :) fixed SNRs for validation set Reproducibility - fix all random seeds Load noises, resample and save into the memory Load clean data and split it into train and val Training data mixes crops randomly on the fly with random SNR in range (effectively infinite training data) `repeat_factor=20` means that the dataset contains 20 copies of itself - it is the easiest way to make the epoch longer Validation data is fixed (for stability): mix every clean clip with all the noises in the folder Argument `mixtures_per_clean` regulates with how many different noise files each clean file will be mixed some random parameters, does it look sensible? the model here should be constructed in the script accordingly to the passed config (including the model type) most of the models accept `sample_rate` parameter for encoders, which is important (default is 16000, override) Probably we also need to subclass `System`, in order to log the target metrics on the validation set (PESQ/STOI) log dir and model name are also part of the config, of course choose the proper accelerator for JADE, probably `ddp` (also, `auto_select_gpus=True` might be useful)torch.save(model.serialize(), 'tasnet_model.pt')
1,298
en
0.852977
# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. import logging import datetime import sqlalchemy as sa from random import randint from flask import Blueprint from flask import current_app from flask import redirect from flask import url_for from relengapi.blueprints.archiver import tables from relengapi.blueprints.archiver.tasks import TASK_TIME_OUT from relengapi.blueprints.archiver.tasks import create_and_upload_archive from relengapi.blueprints.archiver.types import MozharnessArchiveTask from relengapi.lib import api from relengapi.lib import badpenny from relengapi.lib.time import now bp = Blueprint('archiver', __name__) log = logging.getLogger(__name__) GET_EXPIRES_IN = 300 PENDING_EXPIRES_IN = 60 FINISHED_STATES = ['SUCCESS', 'FAILURE', 'REVOKED'] def delete_tracker(tracker): session = current_app.db.session('relengapi') log.info("deleting tracker with id: {}".format(tracker.task_id)) session.delete(tracker) session.commit() def update_tracker_state(tracker, state): session = current_app.db.session('relengapi') log.info("updating tracker with id: {} to state: {}".format(tracker.id, state)) try: tracker.state = state session.commit() except sa.exc.IntegrityError: session.rollback() @badpenny.periodic_task(seconds=TASK_TIME_OUT) def cleanup_old_tasks(job_status): """delete any tracker task if it is older than the time a task can live for.""" session = current_app.db.session('relengapi') expiry_cutoff = now() - datetime.timedelta(seconds=TASK_TIME_OUT) table = tables.ArchiverTask for tracker in session.query(table).order_by(table.created_at): if tracker.created_at < expiry_cutoff: delete_tracker(tracker) else: break def renew_tracker_pending_expiry(tracker): pending_expires_at = now() + datetime.timedelta(seconds=PENDING_EXPIRES_IN) session = current_app.db.session('relengapi') log.info("renewing tracker {} with pending expiry: {}".format(tracker.id, pending_expires_at)) tracker.pending_expires_at = pending_expires_at session.commit() @bp.route('/status/<task_id>') @api.apimethod(MozharnessArchiveTask, unicode) def task_status(task_id): """ Check and return the current state of the create_and_upload_archive celery task with task id of <task_id>. If the task is unknown, state will be PENDING. Once the task starts it will be updated to STARTED and finally, if it completes, it will be either SUCCESS (no exceptions), or FAILURE. See update_state() within create_and_upload_archive and http://celery.readthedocs.org/en/latest/reference/celery.states.html for more details. If state is SUCCESS, it is safe to check response['s3_urls'] for the archives submitted to s3 """ task = create_and_upload_archive.AsyncResult(task_id) task_tracker = tables.ArchiverTask.query.filter(tables.ArchiverTask.task_id == task_id).first() log.info("checking status of task id {}: current state {}".format(task_id, task.state)) task_info = task.info or {} response = { 'state': task.state, } if task.state != 'FAILURE': response['status'] = task_info.get('status', 'no status available at this point.') response['src_url'] = task_info.get('src_url', '') response['s3_urls'] = task_info.get('s3_urls', {}) else: # something went wrong response['status'] = str(task.info) # this is the exception raised response['src_url'] = '' response['s3_urls'] = {} # archiver does not create any custom states, so we can assume to have only the defaults: # http://docs.celeryproject.org/en/latest/userguide/tasks.html#task-states # therefore, delete our state_id tracker from the db if the celery state is in a final state: # e.g. not RETRY, STARTED, or PENDING if task_tracker: if task.state in FINISHED_STATES: delete_tracker(task_tracker) elif task.state == "PENDING" and task_tracker.pending_expires_at < now(): log.info("Task {} has expired from pending too long. Re-creating task".format(task.id)) renew_tracker_pending_expiry(task_tracker) # let exceptions bubble up before moving on create_and_upload_archive.apply_async(args=[task_tracker.src_url, task_tracker.s3_key], task_id=task.id) response['state'] = 'RETRY' response['status'] = 'Task has expired from pending for too long. Re-creating task.' elif task_tracker.state != task.state: update_tracker_state(task_tracker, task.state) return MozharnessArchiveTask(**response) @bp.route('/hgmo/<path:repo>/<rev>') @api.apimethod(None, unicode, unicode, unicode, unicode, unicode, status_code=302) def get_hgmo_archive(repo, rev, subdir=None, suffix='tar.gz', preferred_region=None): """ An archiver for hg.mozilla.org related requests. Uses relengapi.blueprints.archiver.get_archive :param repo: the repo location off of hg.mozilla.org/ :param rev: the rev associated with the repo :param subdir: optional subdir path to only archive a portion of the repo :param suffix: the archive extension type. defaulted to tar.gz :param preferred_region: the preferred s3 region to use """ # allow for the short hash and full hash to be passed rev = rev[0:12] src_url = current_app.config['ARCHIVER_HGMO_URL_TEMPLATE'].format( repo=repo, rev=rev, suffix=suffix, subdir=subdir or '' ) # though slightly odd to append the archive suffix extension with a subdir, this: # 1) allows us to have archives based on different subdir locations from the same repo and rev # 2) is aligned with the hg.mozilla.org format key = '{repo}-{rev}.{suffix}'.format(repo=repo, rev=rev, suffix=suffix) if subdir: key += '/{}'.format(subdir) return get_archive(src_url, key, preferred_region) def get_archive(src_url, key, preferred_region): """ A generic getter for retrieving an s3 location of an archive where the archive is based off a src_url. sub-dir: hg.mozilla.org supports archives of sub directories within a repository. This flexibility allows for creating archives of only a portion of what would normally be an entire repo archive. logic flow: If their is already a key within s3, a re-direct link is given for the s3 location. If the key does not exist, download the archive from src url, upload it to s3 for each region supported and return all uploaded s3 url locations. When the key does not exist, the remaining work will be assigned to a celery background task with a url location returned immediately for obtaining task state updates. """ buckets = current_app.config['ARCHIVER_S3_BUCKETS'] random_region = buckets.keys()[randint(0, len(buckets.keys()) - 1)] # use preferred region if available otherwise choose a valid one at random region = preferred_region if preferred_region and preferred_region in buckets else random_region bucket = buckets[region] s3 = current_app.aws.connect_to('s3', region) session = current_app.db.session('relengapi') # first, see if the key exists if not s3.get_bucket(bucket).get_key(key): task_id = key.replace('/', '_') # keep things simple and avoid slashes in task url # can't use unique support: # api.pub.build.mozilla.org/docs/development/databases/#unique-row-support-get-or-create # because we want to know when the row doesn't exist before creating it tracker = tables.ArchiverTask.query.filter(tables.ArchiverTask.task_id == task_id).first() if tracker and tracker.state in FINISHED_STATES: log.info('Task tracker: {} exists but finished with state: ' '{}'.format(task_id, tracker.state)) # remove tracker and try celery task again delete_tracker(tracker) tracker = None if not tracker: log.info("Creating new celery task and task tracker for: {}".format(task_id)) task = create_and_upload_archive.apply_async(args=[src_url, key], task_id=task_id) if task and task.id: pending_expires_at = now() + datetime.timedelta(seconds=PENDING_EXPIRES_IN) session.add(tables.ArchiverTask(task_id=task.id, s3_key=key, created_at=now(), pending_expires_at=pending_expires_at, src_url=src_url, state="PENDING")) session.commit() else: return {}, 500 return {}, 202, {'Location': url_for('archiver.task_status', task_id=task.id)} log.info("generating GET URL to {}, expires in {}s".format(key, GET_EXPIRES_IN)) # return 302 pointing to s3 url with archive signed_url = s3.generate_url( method='GET', expires_in=GET_EXPIRES_IN, bucket=bucket, key=key ) return redirect(signed_url)
relengapi/blueprints/archiver/__init__.py
9,256
delete any tracker task if it is older than the time a task can live for. A generic getter for retrieving an s3 location of an archive where the archive is based off a src_url. sub-dir: hg.mozilla.org supports archives of sub directories within a repository. This flexibility allows for creating archives of only a portion of what would normally be an entire repo archive. logic flow: If their is already a key within s3, a re-direct link is given for the s3 location. If the key does not exist, download the archive from src url, upload it to s3 for each region supported and return all uploaded s3 url locations. When the key does not exist, the remaining work will be assigned to a celery background task with a url location returned immediately for obtaining task state updates. An archiver for hg.mozilla.org related requests. Uses relengapi.blueprints.archiver.get_archive :param repo: the repo location off of hg.mozilla.org/ :param rev: the rev associated with the repo :param subdir: optional subdir path to only archive a portion of the repo :param suffix: the archive extension type. defaulted to tar.gz :param preferred_region: the preferred s3 region to use Check and return the current state of the create_and_upload_archive celery task with task id of <task_id>. If the task is unknown, state will be PENDING. Once the task starts it will be updated to STARTED and finally, if it completes, it will be either SUCCESS (no exceptions), or FAILURE. See update_state() within create_and_upload_archive and http://celery.readthedocs.org/en/latest/reference/celery.states.html for more details. If state is SUCCESS, it is safe to check response['s3_urls'] for the archives submitted to s3 This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. something went wrong this is the exception raised archiver does not create any custom states, so we can assume to have only the defaults: http://docs.celeryproject.org/en/latest/userguide/tasks.htmltask-states therefore, delete our state_id tracker from the db if the celery state is in a final state: e.g. not RETRY, STARTED, or PENDING let exceptions bubble up before moving on allow for the short hash and full hash to be passed though slightly odd to append the archive suffix extension with a subdir, this: 1) allows us to have archives based on different subdir locations from the same repo and rev 2) is aligned with the hg.mozilla.org format use preferred region if available otherwise choose a valid one at random first, see if the key exists keep things simple and avoid slashes in task url can't use unique support: api.pub.build.mozilla.org/docs/development/databases/unique-row-support-get-or-create because we want to know when the row doesn't exist before creating it remove tracker and try celery task again return 302 pointing to s3 url with archive
2,972
en
0.854273
# Copyright 2017 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Output command line arguments and json-encoded TF_CONFIGs. Usage: `make_tf_configs.py --workers="server1:1234" --ps="server3:2134,server4:2334"` Outputs 1 line per job to stdout, first the workers, then the parameter servers. Each line has the TF_CONFIG, then a tab, then the command line flags for that job. If there is a single worker, workers will have the `--sync` flag. """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import json # Dependency imports import six import tensorflow as tf flags = tf.flags FLAGS = flags.FLAGS flags.DEFINE_string("workers", "", "Comma-separated list of worker addresses") flags.DEFINE_string("ps", "", "Comma-separated list of ps addresses") def main(_): if not (FLAGS.workers and FLAGS.ps): raise ValueError("Must provide --workers and --ps") workers = FLAGS.workers.split(",") ps = FLAGS.ps.split(",") cluster = {"ps": ps, "worker": workers} for task_type, jobs in six.iteritems(cluster): for idx, job in enumerate(jobs): if task_type == "worker": cmd_line_flags = " ".join([ "--master=%s" % job, "--ps_replicas=%d" % len(ps), "--worker_replicas=%d" % len(workers), "--worker_gpu=1", "--worker_id=%d" % idx, "--ps_gpu=1", "--schedule=train", "--sync" if len(workers) == 1 else "", ]) else: cmd_line_flags = " ".join([ "--schedule=run_std_server", ]) tf_config = json.dumps({ "cluster": cluster, "task": { "type": task_type, "index": idx } }) print(tf_config + "\t" + cmd_line_flags) if __name__ == "__main__": tf.app.run()
tensor2tensor/bin/make_tf_configs.py
2,369
Output command line arguments and json-encoded TF_CONFIGs. Usage: `make_tf_configs.py --workers="server1:1234" --ps="server3:2134,server4:2334"` Outputs 1 line per job to stdout, first the workers, then the parameter servers. Each line has the TF_CONFIG, then a tab, then the command line flags for that job. If there is a single worker, workers will have the `--sync` flag. Copyright 2017 Google Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. Dependency imports
948
en
0.795617
# coding: utf-8 from __future__ import unicode_literals import re import base64 from .common import InfoExtractor from ..compat import ( compat_urlparse, compat_parse_qs, ) from ..utils import ( clean_html, ExtractorError, int_or_none, unsmuggle_url, smuggle_url, ) class KalturaIE(InfoExtractor): _VALID_URL = r'''(?x) (?: kaltura:(?P<partner_id>\d+):(?P<id>[0-9a-z_]+)| https?:// (:?(?:www|cdnapi(?:sec)?)\.)?kaltura\.com(?::\d+)?/ (?: (?: # flash player index\.php/(?:kwidget|extwidget/preview)| # html5 player html5/html5lib/[^/]+/mwEmbedFrame\.php ) )(?:/(?P<path>[^?]+))?(?:\?(?P<query>.*))? ) ''' _SERVICE_URL = 'http://cdnapi.kaltura.com' _SERVICE_BASE = '/api_v3/index.php' # See https://github.com/kaltura/server/blob/master/plugins/content/caption/base/lib/model/enums/CaptionType.php _CAPTION_TYPES = { 1: 'srt', 2: 'ttml', 3: 'vtt', } _TESTS = [ { 'url': 'kaltura:269692:1_1jc2y3e4', 'md5': '3adcbdb3dcc02d647539e53f284ba171', 'info_dict': { 'id': '1_1jc2y3e4', 'ext': 'mp4', 'title': 'Straight from the Heart', 'upload_date': '20131219', 'uploader_id': 'mlundberg@wolfgangsvault.com', 'description': 'The Allman Brothers Band, 12/16/1981', 'thumbnail': 're:^https?://.*/thumbnail/.*', 'timestamp': int, }, }, { 'url': 'http://www.kaltura.com/index.php/kwidget/cache_st/1300318621/wid/_269692/uiconf_id/3873291/entry_id/1_1jc2y3e4', 'only_matching': True, }, { 'url': 'https://cdnapisec.kaltura.com/index.php/kwidget/wid/_557781/uiconf_id/22845202/entry_id/1_plr1syf3', 'only_matching': True, }, { 'url': 'https://cdnapisec.kaltura.com/html5/html5lib/v2.30.2/mwEmbedFrame.php/p/1337/uiconf_id/20540612/entry_id/1_sf5ovm7u?wid=_243342', 'only_matching': True, }, { # video with subtitles 'url': 'kaltura:111032:1_cw786r8q', 'only_matching': True, }, { # video with ttml subtitles (no fileExt) 'url': 'kaltura:1926081:0_l5ye1133', 'info_dict': { 'id': '0_l5ye1133', 'ext': 'mp4', 'title': 'What Can You Do With Python?', 'upload_date': '20160221', 'uploader_id': 'stork', 'thumbnail': 're:^https?://.*/thumbnail/.*', 'timestamp': int, 'subtitles': { 'en': [{ 'ext': 'ttml', }], }, }, 'skip': 'Gone. Maybe https://www.safaribooksonline.com/library/tutorials/introduction-to-python-anon/3469/', 'params': { 'skip_download': True, }, }, { 'url': 'https://www.kaltura.com/index.php/extwidget/preview/partner_id/1770401/uiconf_id/37307382/entry_id/0_58u8kme7/embed/iframe?&flashvars[streamerType]=auto', 'only_matching': True, }, { 'url': 'https://www.kaltura.com:443/index.php/extwidget/preview/partner_id/1770401/uiconf_id/37307382/entry_id/0_58u8kme7/embed/iframe?&flashvars[streamerType]=auto', 'only_matching': True, } ] @staticmethod def _extract_url(webpage): # Embed codes: https://knowledge.kaltura.com/embedding-kaltura-media-players-your-site mobj = ( re.search( r"""(?xs) kWidget\.(?:thumb)?[Ee]mbed\( \{.*? (?P<q1>['"])wid(?P=q1)\s*:\s* (?P<q2>['"])_?(?P<partner_id>(?:(?!(?P=q2)).)+)(?P=q2),.*? (?P<q3>['"])entry_?[Ii]d(?P=q3)\s*:\s* (?P<q4>['"])(?P<id>(?:(?!(?P=q4)).)+)(?P=q4)(?:,|\s*\}) """, webpage) or re.search( r'''(?xs) (?P<q1>["']) (?:https?:)?//cdnapi(?:sec)?\.kaltura\.com(?::\d+)?/(?:(?!(?P=q1)).)*\b(?:p|partner_id)/(?P<partner_id>\d+)(?:(?!(?P=q1)).)* (?P=q1).*? (?: (?: entry_?[Ii]d| (?P<q2>["'])entry_?[Ii]d(?P=q2) )\s*:\s*| \[\s*(?P<q2_1>["'])entry_?[Ii]d(?P=q2_1)\s*\]\s*=\s* ) (?P<q3>["'])(?P<id>(?:(?!(?P=q3)).)+)(?P=q3) ''', webpage) or re.search( r'''(?xs) <(?:iframe[^>]+src|meta[^>]+\bcontent)=(?P<q1>["']) (?:https?:)?//(?:(?:www|cdnapi(?:sec)?)\.)?kaltura\.com/(?:(?!(?P=q1)).)*\b(?:p|partner_id)/(?P<partner_id>\d+) (?:(?!(?P=q1)).)* [?&;]entry_id=(?P<id>(?:(?!(?P=q1))[^&])+) (?:(?!(?P=q1)).)* (?P=q1) ''', webpage) ) if mobj: embed_info = mobj.groupdict() for k, v in embed_info.items(): embed_info[k] = v.strip() url = 'kaltura:%(partner_id)s:%(id)s' % embed_info escaped_pid = re.escape(embed_info['partner_id']) service_url = re.search( r'<script[^>]+src=["\']((?:https?:)?//.+?)/p/%s/sp/%s00/embedIframeJs' % (escaped_pid, escaped_pid), webpage) if service_url: url = smuggle_url(url, {'service_url': service_url.group(1)}) return url def _kaltura_api_call(self, video_id, actions, service_url=None, *args, **kwargs): params = actions[0] if len(actions) > 1: for i, a in enumerate(actions[1:], start=1): for k, v in a.items(): params['%d:%s' % (i, k)] = v data = self._download_json( (service_url or self._SERVICE_URL) + self._SERVICE_BASE, video_id, query=params, *args, **kwargs) status = data if len(actions) == 1 else data[0] if status.get('objectType') == 'KalturaAPIException': raise ExtractorError( '%s said: %s' % (self.IE_NAME, status['message'])) return data def _get_video_info(self, video_id, partner_id, service_url=None): actions = [ { 'action': 'null', 'apiVersion': '3.1.5', 'clientTag': 'kdp:v3.8.5', 'format': 1, # JSON, 2 = XML, 3 = PHP 'service': 'multirequest', }, { 'expiry': 86400, 'service': 'session', 'action': 'startWidgetSession', 'widgetId': '_%s' % partner_id, }, { 'action': 'get', 'entryId': video_id, 'service': 'baseentry', 'ks': '{1:result:ks}', 'responseProfile:fields': 'createdAt,dataUrl,duration,name,plays,thumbnailUrl,userId', 'responseProfile:type': 1, }, { 'action': 'getbyentryid', 'entryId': video_id, 'service': 'flavorAsset', 'ks': '{1:result:ks}', }, { 'action': 'list', 'filter:entryIdEqual': video_id, 'service': 'caption_captionasset', 'ks': '{1:result:ks}', }, ] return self._kaltura_api_call( video_id, actions, service_url, note='Downloading video info JSON') def _real_extract(self, url): url, smuggled_data = unsmuggle_url(url, {}) mobj = re.match(self._VALID_URL, url) partner_id, entry_id = mobj.group('partner_id', 'id') ks = None captions = None if partner_id and entry_id: _, info, flavor_assets, captions = self._get_video_info(entry_id, partner_id, smuggled_data.get('service_url')) else: path, query = mobj.group('path', 'query') if not path and not query: raise ExtractorError('Invalid URL', expected=True) params = {} if query: params = compat_parse_qs(query) if path: splitted_path = path.split('/') params.update(dict((zip(splitted_path[::2], [[v] for v in splitted_path[1::2]])))) if 'wid' in params: partner_id = params['wid'][0][1:] elif 'p' in params: partner_id = params['p'][0] elif 'partner_id' in params: partner_id = params['partner_id'][0] else: raise ExtractorError('Invalid URL', expected=True) if 'entry_id' in params: entry_id = params['entry_id'][0] _, info, flavor_assets, captions = self._get_video_info(entry_id, partner_id) elif 'uiconf_id' in params and 'flashvars[referenceId]' in params: reference_id = params['flashvars[referenceId]'][0] webpage = self._download_webpage(url, reference_id) entry_data = self._parse_json(self._search_regex( r'window\.kalturaIframePackageData\s*=\s*({.*});', webpage, 'kalturaIframePackageData'), reference_id)['entryResult'] info, flavor_assets = entry_data['meta'], entry_data['contextData']['flavorAssets'] entry_id = info['id'] # Unfortunately, data returned in kalturaIframePackageData lacks # captions so we will try requesting the complete data using # regular approach since we now know the entry_id try: _, info, flavor_assets, captions = self._get_video_info( entry_id, partner_id) except ExtractorError: # Regular scenario failed but we already have everything # extracted apart from captions and can process at least # with this pass else: raise ExtractorError('Invalid URL', expected=True) ks = params.get('flashvars[ks]', [None])[0] source_url = smuggled_data.get('source_url') if source_url: referrer = base64.b64encode( '://'.join(compat_urlparse.urlparse(source_url)[:2]) .encode('utf-8')).decode('utf-8') else: referrer = None def sign_url(unsigned_url): if ks: unsigned_url += '/ks/%s' % ks if referrer: unsigned_url += '?referrer=%s' % referrer return unsigned_url data_url = info['dataUrl'] if '/flvclipper/' in data_url: data_url = re.sub(r'/flvclipper/.*', '/serveFlavor', data_url) formats = [] for f in flavor_assets: # Continue if asset is not ready if f.get('status') != 2: continue # Original format that's not available (e.g. kaltura:1926081:0_c03e1b5g) # skip for now. if f.get('fileExt') == 'chun': continue # DRM-protected video, cannot be decrypted if f.get('fileExt') == 'wvm': continue if not f.get('fileExt'): # QT indicates QuickTime; some videos have broken fileExt if f.get('containerFormat') == 'qt': f['fileExt'] = 'mov' else: f['fileExt'] = 'mp4' video_url = sign_url( '%s/flavorId/%s' % (data_url, f['id'])) # audio-only has no videoCodecId (e.g. kaltura:1926081:0_c03e1b5g # -f mp4-56) vcodec = 'none' if 'videoCodecId' not in f and f.get( 'frameRate') == 0 else f.get('videoCodecId') formats.append({ 'format_id': '%(fileExt)s-%(bitrate)s' % f, 'ext': f.get('fileExt'), 'tbr': int_or_none(f['bitrate']), 'fps': int_or_none(f.get('frameRate')), 'filesize_approx': int_or_none(f.get('size'), invscale=1024), 'container': f.get('containerFormat'), 'vcodec': vcodec, 'height': int_or_none(f.get('height')), 'width': int_or_none(f.get('width')), 'url': video_url, }) if '/playManifest/' in data_url: m3u8_url = sign_url(data_url.replace( 'format/url', 'format/applehttp')) formats.extend(self._extract_m3u8_formats( m3u8_url, entry_id, 'mp4', 'm3u8_native', m3u8_id='hls', fatal=False)) self._sort_formats(formats) subtitles = {} if captions: for caption in captions.get('objects', []): # Continue if caption is not ready if caption.get('status') != 2: continue if not caption.get('id'): continue caption_format = int_or_none(caption.get('format')) subtitles.setdefault(caption.get('languageCode') or caption.get('language'), []).append({ 'url': '%s/api_v3/service/caption_captionasset/action/serve/captionAssetId/%s' % (self._SERVICE_URL, caption['id']), 'ext': caption.get('fileExt') or self._CAPTION_TYPES.get(caption_format) or 'ttml', }) return { 'id': entry_id, 'title': info['name'], 'formats': formats, 'subtitles': subtitles, 'description': clean_html(info.get('description')), 'thumbnail': info.get('thumbnailUrl'), 'duration': info.get('duration'), 'timestamp': info.get('createdAt'), 'uploader_id': info.get('userId') if info.get('userId') != 'None' else None, 'view_count': info.get('plays'), }
youtube_dl/extractor/kaltura.py
14,728
coding: utf-8 See https://github.com/kaltura/server/blob/master/plugins/content/caption/base/lib/model/enums/CaptionType.php video with subtitles video with ttml subtitles (no fileExt) Embed codes: https://knowledge.kaltura.com/embedding-kaltura-media-players-your-site JSON, 2 = XML, 3 = PHP Unfortunately, data returned in kalturaIframePackageData lacks captions so we will try requesting the complete data using regular approach since we now know the entry_id Regular scenario failed but we already have everything extracted apart from captions and can process at least with this Continue if asset is not ready Original format that's not available (e.g. kaltura:1926081:0_c03e1b5g) skip for now. DRM-protected video, cannot be decrypted QT indicates QuickTime; some videos have broken fileExt audio-only has no videoCodecId (e.g. kaltura:1926081:0_c03e1b5g -f mp4-56) Continue if caption is not ready
903
en
0.776964
from conans import ConanFile, CMake, tools import os class TestPackageConan(ConanFile): settings = "os", "arch", "compiler", "build_type" generators = "cmake", "cmake_find_package_multi" def build_requirements(self): if self.settings.os == "Macos" and self.settings.arch == "armv8": # Workaround for CMake bug with error message: # Attempting to use @rpath without CMAKE_SHARED_LIBRARY_RUNTIME_C_FLAG being # set. This could be because you are using a Mac OS X version less than 10.5 # or because CMake's platform configuration is corrupt. # FIXME: Remove once CMake on macOS/M1 CI runners is upgraded. self.build_requires("cmake/3.22.3") def build(self): cmake = CMake(self) cmake.configure() cmake.build() def test(self): if not tools.cross_building(self): bin_path = os.path.join("bin", "test_package") self.run(bin_path, run_environment=True)
recipes/librasterlite/all/test_package/conanfile.py
1,008
Workaround for CMake bug with error message: Attempting to use @rpath without CMAKE_SHARED_LIBRARY_RUNTIME_C_FLAG being set. This could be because you are using a Mac OS X version less than 10.5 or because CMake's platform configuration is corrupt. FIXME: Remove once CMake on macOS/M1 CI runners is upgraded.
309
en
0.898917
""" Tests for grph.py """ import os import platform import random import re import string from subprocess import getstatusoutput PRG = './grph.py' RUN = f'python {PRG}' if platform.system() == 'Windows' else PRG SAMPLE1 = './tests/inputs/1.fa' SAMPLE2 = './tests/inputs/2.fa' SAMPLE3 = './tests/inputs/3.fa' # -------------------------------------------------- def test_exists() -> None: """ Program exists """ assert os.path.isfile(PRG) # -------------------------------------------------- def test_usage() -> None: """ Usage """ rv, out = getstatusoutput(RUN) assert rv > 0 assert out.lower().startswith('usage:') # -------------------------------------------------- def test_bad_k() -> None: """ Dies on bad k """ k = random.choice(range(-10, 1)) rv, out = getstatusoutput(f'{RUN} -k {k} {SAMPLE1}') assert rv != 0 assert out.lower().startswith('usage:') assert re.search(f'-k "{k}" must be > 0', out) # -------------------------------------------------- def test_bad_file() -> None: """ Dies on bad file """ bad = random_string() rv, out = getstatusoutput('{} {}'.format(RUN, bad)) assert rv != 0 assert out.lower().startswith('usage:') assert re.search(f"No such file or directory: '{bad}'", out) # -------------------------------------------------- def run(in_file: str, k: int) -> None: """ Run with args """ out_file = '.'.join([in_file, str(k), 'out']) assert os.path.isfile(out_file) expected = open(out_file).read().rstrip() cmd = '{} -k {} {} | sort'.format(RUN, k, in_file) rv, out = getstatusoutput(cmd) assert rv == 0 assert out.rstrip() == expected # -------------------------------------------------- def test_01(): """ Runs OK """ run(SAMPLE1, 3) # -------------------------------------------------- def test_02() -> None: """ Runs OK """ run(SAMPLE1, 4) # -------------------------------------------------- def test_03() -> None: """ Runs OK """ run(SAMPLE1, 5) # -------------------------------------------------- def test_04() -> None: """ Runs OK """ run(SAMPLE2, 3) # -------------------------------------------------- def test_05() -> None: """ Runs OK """ run(SAMPLE2, 4) # -------------------------------------------------- def test_06() -> None: """ Runs OK """ run(SAMPLE2, 5) # -------------------------------------------------- def test_07() -> None: """ Runs OK """ run(SAMPLE3, 3) # -------------------------------------------------- def test_08() -> None: """ Runs OK """ run(SAMPLE3, 4) # -------------------------------------------------- def test_09() -> None: """ Runs OK """ run(SAMPLE3, 5) # -------------------------------------------------- def random_string() -> str: """Generate a random string""" return ''.join( random.sample(string.ascii_letters + string.digits, k=random.randint(5, 10)))
09_grph/tests/grph_test.py
2,997
Generate a random string Run with args Runs OK Runs OK Runs OK Runs OK Runs OK Runs OK Runs OK Runs OK Runs OK Dies on bad file Dies on bad k Program exists Usage Tests for grph.py -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- -------------------------------------------------- --------------------------------------------------
962
en
0.250365
import six import pytest from pymemcache.test.utils import MockMemcacheClient @pytest.mark.unit() def test_get_set(): client = MockMemcacheClient() assert client.get(b"hello") is None client.set(b"hello", 12) assert client.get(b"hello") == 12 @pytest.mark.unit() def test_get_set_unicide_key(): client = MockMemcacheClient() assert client.get(u"hello") is None client.set(b"hello", 12) assert client.get(u"hello") == 12 @pytest.mark.unit() def test_get_set_non_ascii_value(): client = MockMemcacheClient() assert client.get(b"hello") is None # This is the value of msgpack.packb('non_ascii') non_ascii_str = b'\xa9non_ascii' client.set(b"hello", non_ascii_str) assert client.get(b"hello") == non_ascii_str @pytest.mark.unit() def test_get_many_set_many(): client = MockMemcacheClient() client.set(b"h", 1) result = client.get_many([b"h", b"e", b"l", b"o"]) assert result == {b"h": 1} # Convert keys into bytes d = dict((k.encode('ascii'), v) for k, v in six.iteritems(dict(h=1, e=2, z=3))) client.set_many(d) assert client.get_many([b"h", b"e", b"z", b"o"]) == d @pytest.mark.unit() def test_get_many_set_many_non_ascii_values(): client = MockMemcacheClient() # These are the values of calling msgpack.packb() on '1', '2', and '3' non_ascii_1 = b'\xa11' non_ascii_2 = b'\xa12' non_ascii_3 = b'\xa13' client.set(b"h", non_ascii_1) result = client.get_many([b"h", b"e", b"l", b"o"]) assert result == {b"h": non_ascii_1} # Convert keys into bytes d = dict((k.encode('ascii'), v) for k, v in six.iteritems( dict(h=non_ascii_1, e=non_ascii_2, z=non_ascii_3) )) client.set_many(d) assert client.get_many([b"h", b"e", b"z", b"o"]) == d @pytest.mark.unit() def test_add(): client = MockMemcacheClient() client.add(b"k", 2) assert client.get(b"k") == 2 client.add(b"k", 25) assert client.get(b"k") == 2 @pytest.mark.unit() def test_delete(): client = MockMemcacheClient() client.add(b"k", 2) assert client.get(b"k") == 2 client.delete(b"k") assert client.get(b"k") is None @pytest.mark.unit() def test_incr_decr(): client = MockMemcacheClient() client.add(b"k", 2) client.incr(b"k", 4) assert client.get(b"k") == 6 client.decr(b"k", 2) assert client.get(b"k") == 4 @pytest.mark.unit() def test_prepand_append(): client = MockMemcacheClient() client.set(b"k", '1') client.append(b"k", 'a') client.prepend(b"k", 'p') assert client.get(b"k") == b'p1a'
pymemcache/test/test_utils.py
2,650
This is the value of msgpack.packb('non_ascii') Convert keys into bytes These are the values of calling msgpack.packb() on '1', '2', and '3' Convert keys into bytes
164
en
0.747367
# Copyright 2021 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Async pipeline for testing.""" from tfx.dsl.compiler import compiler from tfx.dsl.component.experimental.annotations import InputArtifact from tfx.dsl.component.experimental.annotations import OutputArtifact from tfx.dsl.component.experimental.annotations import Parameter from tfx.dsl.component.experimental.decorators import component from tfx.orchestration import pipeline as pipeline_lib from tfx.proto.orchestration import pipeline_pb2 from tfx.types import standard_artifacts @component def _example_gen(examples: OutputArtifact[standard_artifacts.Examples]): del examples # pytype: disable=wrong-arg-types @component def _transform( examples: InputArtifact[standard_artifacts.Examples], transform_graph: OutputArtifact[standard_artifacts.TransformGraph], a_param: Parameter[int]): del examples, transform_graph, a_param # pytype: enable=wrong-arg-types @component def _trainer(examples: InputArtifact[standard_artifacts.Examples], transform_graph: InputArtifact[standard_artifacts.TransformGraph], model: OutputArtifact[standard_artifacts.Model]): del examples, transform_graph, model def create_pipeline() -> pipeline_pb2.Pipeline: """Creates an async pipeline for testing.""" # pylint: disable=no-value-for-parameter example_gen = _example_gen().with_id('my_example_gen') transform = _transform( examples=example_gen.outputs['examples'], a_param=10).with_id('my_transform') trainer = _trainer( examples=example_gen.outputs['examples'], transform_graph=transform.outputs['transform_graph']).with_id( 'my_trainer') # pylint: enable=no-value-for-parameter pipeline = pipeline_lib.Pipeline( pipeline_name='my_pipeline', pipeline_root='/path/to/root', components=[ example_gen, transform, trainer, ], execution_mode=pipeline_lib.ExecutionMode.ASYNC) dsl_compiler = compiler.Compiler() compiled_pipeline: pipeline_pb2.Pipeline = dsl_compiler.compile(pipeline) # Compiler does not support setting min_count yet, so we mutate the proto # explicitly for testing. trainer = compiled_pipeline.nodes[2].pipeline_node assert trainer.node_info.id == 'my_trainer' for value in trainer.inputs.inputs.values(): value.min_count = 1 return compiled_pipeline
tfx/orchestration/experimental/core/testing/test_async_pipeline.py
2,933
Creates an async pipeline for testing. Async pipeline for testing. Copyright 2021 Google LLC. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. pytype: disable=wrong-arg-types pytype: enable=wrong-arg-types pylint: disable=no-value-for-parameter pylint: enable=no-value-for-parameter Compiler does not support setting min_count yet, so we mutate the proto explicitly for testing.
874
en
0.809199
'''Escreva um programa que receba um número inteiro na entrada e verifique se o número recebido possui ao menos um dígito com um dígito adjacente igual a ele. Caso exista, imprima "sim"; se não existir, imprima "não". ''' num = int(input('Digite um número inteiro: ')) alg1 = num % 10 while num > 0: num = num // 10 alg2 = num % 10 if alg1 == alg2: print('sim') break else: if num > 0: alg1 = alg2 else: break if num == 0: print('não')
Exercicios/digitos_adjacentes.py
524
Escreva um programa que receba um número inteiro na entrada e verifique se o número recebido possui ao menos um dígito com um dígito adjacente igual a ele. Caso exista, imprima "sim"; se não existir, imprima "não".
216
pt
0.979442
#from fake_useragent import UserAgent from selenium import webdriver from selenium.webdriver.common.by import By import time import os # 订阅,0关闭 dingyue = 1 # 短信,0关闭 duanxin = 1 # 设置翻页范围 start_page = 1 end_page = 1 def main(): # 当前版本仅支持订阅一个手机号码 _list = [ ['18888888888', '张三'] ] _run(_list) def _run(_list): delay = 0 option = webdriver.ChromeOptions() # option.add_argument('--disable-gpu') option.add_argument('--incognito') option.add_argument('blink-settings=imagesEnabled=false') option.add_argument('--headless') ua = 'Mozilla/5.0 (Windows; U; MSIE 9.0; Windows NT 9.0; en-US)' option.add_argument('--user-agent={}'.format(ua)) driver = webdriver.Chrome(options=option) driver.execute_cdp_cmd('Page.addScriptToEvaluateOnNewDocument', { 'source': 'Object.defineProperty(navigator, "webdriver", {get: () => undefined})'}) driver.maximize_window() driver.get("https://www.baidu.com/") citys = ["jn", "qingdao", "yantai", "dongying", "weihai", "zibo", "weifang", "liaocheng", "heze", "rizhao", "linyi", "dezhou", "jining", "zaozhuang", "binzhou", "taian", "laiwu", "laiyang", "xintai", "hf", "liuan", "anqing", "xuancheng", "ahsuzhou", "chaohu", "chizhou", "huainan", "chuzhou", "wuhu", "bengbu", "tongling", "fuyang", "maanshan", "huangshan", "huaibei", "bozhou", "dangtu", "huoqiu", "sh", "hz", "ningbo", "shaoxing", "taizhou", "jinhua", "wenzhou", "jiaxing", "lishui", "huzhou", "zhoushan", "quzhou", "nj", "suzhou", "wuxi", "xuzhou", "changzhou", "nantong", "huaian", "changshu", "kunshan", "yx", "yangzhou", "jiangyin", "tz", "zhenjiang", "suqian", "lyg", "taicang", "yancheng", "zhangjiagang", "liyang", "donghai", "baoying", "rugao", "yizheng", "haian", "haimen", "gaoyou", "qidong", "rudong", "xinghua", "jingjiang", "taixing", "nc", "ganzhou", "shangrao", "jxfc", "jiujiang", "jingan", "yichun", "xinyu", "ruijin", "jian", "pingxiang", "jingdezhen", "fuzhou", "yingtan", "ys", "quanzhou", "fz", "xiamen", "zhangzhou", "putian", "longyan", "nanping", "sanming", "ningde", "gz", "sz", "foshan", "zhongshan", "dongguan", "zhuhai", "shantou", "zhaoqing", "jiangmen", "huizhou", "jieyang", "meizhou", "heyuan", "qingyuan", "zhanjiang", "yangjiang", "shaoguan", "maoming", "yunfu", "chaozhou", "shanwei", "puning", "sanya", "hn", "hk", "guilin", "nn", "liuzhou", "beihai", "qinzhou", "laibin", "yulin", "fangchenggang", "guigang", "wuzhou", "chongzuo", "hechi", "baise", "hezhou", "hengxian", "by", "wh", "yichang", "huangshi", "jingzhou", "lhk", "yicheng", "zaoyang", "shiyan", "xianning", "xiaogan", "jingmen", "xiangyang", "huanggang", "ezhou", "suizhou", "xiantao", "tianmen", "qianjiang", "enshi", "shennongjia", "zz", "luoyang", "jiaozuo", "xinyang", "nanyang", "shangqiu", "anyang", "pingdingshan", "ruzhou", "xinxiang", "wugang", "zhumadian", "kaifeng", "xuchang", "luohe", "hebi", "sanmenxia", "zhoukou", "puyang", "yuzhou", "yongcheng", "changge", "yanling", "cs", "xiangtan", "zhuzhou", "changde", "chenzhou", "hengyang", "huaihua", "loudi", "yiyang", "yongzhou", "yueyang", "zjj", "shaoyang", "xiangxi", "changshada", "heb", "hailin", "daqing", "zhaoyuan", "zhaozhou", "mudanjiang", "qiqihaer", "suihua", "jixi", "jiamusi", "heihe", "hljyichun", "zhaodong", "anda", "shuangyashan", "hegang", "qitaihe", "sy", "dl", "anshan", "panjin", "yingkou", "huludao", "wafangdian", "zhuanghe", "jinzhou", "dandong", "chaoyang", "benxi", "liaoyang", "fushun", "fuxin", "tieling", "cc", "jl", "huadian", "yanbian", "siping", "baishan", "baicheng", "songyuan", "dehui", "tonghua", "nongan", "liaoyuan", "jlys", "gongzhuling", "gy", "zunyi", "bijie", "qianxinan", "qiandongnan", "qiannan", "tongren", "anshun", "lps", "cd", "mianyang", "leshan", "nanchong", "meishan", "yibin", "bazhong", "guangyuan", "luzhou", "guanghan", "renshou", "deyang", "neijiang", "anyue", "suining", "guangan", "dazhou", "ziyang", "panzhihua", "zigong", "yaan", "liangshan", "ganzi", "abazhou", "cq", "hechuan", "km", "dali", "qujing", "yuxi", "xishuangbanna", "lijiang", "baoshan", "dehong", "zhaotong", "wenshan", "chuxiong", "honghe", "puer", "lincang", "diqing", "lasa", "rkz", "changdu", "linzhi", "ty", "yuncheng", "huairen", "linfen", "datong", "jinzhong", "changzhi", "yangquan", "lvliang", "shuozhou", "jc", "xinzhou", "hhht", "eerduosi", "baotou", "wlcb", "byne", "chifeng", "xam", "wuhai", "tl", "hlbe", "xlglm", "alsm", "sjz", "tangshan", "baoding", "qinhuangdao", "handan", "zhangjiakou", "hengshui", "xingtai", "langfang", "chengde", "sanhe", "yuxian", "cangzhou", "wenan", "yanjiao", "dachang", "zhuozhou", "guan", "xianghe", "yongqing", "hbbz", "dingzhou", "gaobeidian", "changli", "xiongan", "bj", "tj", "wuqing", "wlmq", "bazhou", "yili", "kashi", "akesu", "changji", "kzls", "hami", "klmy", "bedl", "ht", "shz", "tlf", "wjq", "qt", "kel", "alt", "tc", "lz", "gannan", "tianshui", "qingyang", "dingxi", "pingliang", "zhangye", "jiuquan", "wuwei", "longnan", "yongdeng", "baiyin", "yuzhong", "jinchang", "jiayuguan", "xn", "haidong", "guoluo", "haibei", "haixi", "yushu", "huangnan", "yc", "guyuan", "wuzhong", "zhongwei", "shizuishan", "xa", "xianyang", "baoji", "weinan", "hanzhong", "ankang", "tongchuan", "shangluo", "yl", "yanan"] #citys = ['xm'] # 遍历城市列表 for city in citys: # 打开page页 for page in range(start_page, end_page+1): href = "http://"+city + \ ".jiwu.com/loupan/list-page"+str(page)+".html" driver.execute_script(f'window.open("{href}", "_blank");') driver.switch_to.window(driver.window_handles[0]) driver.close() driver.switch_to.window(driver.window_handles[0]) time.sleep(3) try: driver.switch_to.alert.dismiss() time.sleep(delay) except: pass urls = driver.find_elements( By.XPATH, "//div[@class='box new_house']/div[@class='img']/a") houses = [] # 从page页获取楼盘链接存入houses[] for url in urls: houses.append(url.get_attribute("href")) # 逐个打开楼盘页并处理 for house in houses: os.system('cls') print( '正在第{}/{}个城市{}订阅第{}/{}页,当页进度:{}/{}'.format(str(citys.index(city)+1), str(len(citys)), city, str(page), str(end_page), str(houses.index(house)+1), str(len(houses)))) try: # 打开楼盘页 driver.execute_script(f'window.open("{house}", "_blank");') driver.switch_to.window(driver.window_handles[0]) driver.close() driver.switch_to.window(driver.window_handles[0]) # 删除遮挡元素 try: js = "document.getElementsByClassName('dialing-box')[0].remove();" driver.execute_script(js) except: pass # 操作 try: for person in _list: if(duanxin != 0): ## 发送地址给手机 ## try: driver.find_element( By.LINK_TEXT, "[发地址到手机]").click() except: pass # 输入电话 try: driver.find_element( By.CSS_SELECTOR, "[type='tel']").send_keys(person[0]) except: pass # 点击领取 try: driver.find_element( By.CSS_SELECTOR, "[class='btn btn-green ok']").click() except: pass # 关闭窗口 try: driver.find_element( By.CSS_SELECTOR, "[class='close-pop']").click() except: pass time.sleep(delay) if(dingyue != 0): ## 降价通知 ## try: driver.find_element( By.LINK_TEXT, "降价通知我").click() except: pass # 输入电话 try: driver.find_element( By.CSS_SELECTOR, "[type='tel']").send_keys(person[0]) except: pass # 点击提交 try: driver.find_element( By.CSS_SELECTOR, "[class='btn btn-green ok']").click() except: pass # 关闭窗口 try: driver.find_element( By.CSS_SELECTOR, "[class='close-pop']").click() except: pass time.sleep(delay) ## 组团砍价 ## try: driver.find_element( By.LINK_TEXT, "立即报名").click() except: pass # 输入电话 try: driver.find_element( By.CSS_SELECTOR, "[type='tel']").send_keys(person[0]) except: pass # 点击提交 try: driver.find_element( By.CSS_SELECTOR, "[class='btn btn-green ok']").click() except: pass # 关闭窗口 try: driver.find_element( By.CSS_SELECTOR, "[class='close-pop']").click() except: pass time.sleep(delay) ## 新信息通知 ## try: driver.find_element( By.LINK_TEXT, "新信息通知").click() except: pass # 输入电话 try: driver.find_element( By.CSS_SELECTOR, "[type='tel']").send_keys(person[0]) except: pass # 点击提交 try: driver.find_element( By.CSS_SELECTOR, "[class='btn btn-green ok']").click() except: pass # 关闭窗口 try: driver.find_element( By.CSS_SELECTOR, "[class='close-pop']").click() except: pass time.sleep(delay) ## 开盘通知我 ## try: driver.find_element( By.LINK_TEXT, "[开盘通知我]").click() except: pass # 输入电话 try: driver.find_element( By.CSS_SELECTOR, "[type='tel']").send_keys(person[0]) except: pass # 点击领取 try: driver.find_element( By.CSS_SELECTOR, "[class='btn btn-green ok']").click() except: pass # 关闭窗口 try: driver.find_element( By.CSS_SELECTOR, "[class='close-pop']").click() except: pass time.sleep(delay) ## 新动态通知我 ## try: driver.find_element( By.LINK_TEXT, "新动态通知我").click() except: pass # 输入电话 try: driver.find_element( By.CSS_SELECTOR, "[type='tel']").send_keys(person[0]) except: pass # 点击提交 try: driver.find_element( By.CSS_SELECTOR, "[class='btn btn-green ok']").click() except: pass # 关闭窗口 try: driver.find_element( By.CSS_SELECTOR, "[class='close-pop']").click() except: pass except: pass except: pass if __name__ == "__main__": main()
main/JW/main.py
16,048
from fake_useragent import UserAgent 订阅,0关闭 短信,0关闭 设置翻页范围 当前版本仅支持订阅一个手机号码 option.add_argument('--disable-gpu')citys = ['xm'] 遍历城市列表 打开page页 从page页获取楼盘链接存入houses[] 逐个打开楼盘页并处理 打开楼盘页 删除遮挡元素 操作 发送地址给手机 输入电话 点击领取 关闭窗口 降价通知 输入电话 点击提交 关闭窗口 组团砍价 输入电话 点击提交 关闭窗口 新信息通知 输入电话 点击提交 关闭窗口 开盘通知我 输入电话 点击领取 关闭窗口 新动态通知我 输入电话 点击提交 关闭窗口
322
zh
0.961164
#!/usr/bin/env python from django.urls import reverse_lazy from django.shortcuts import Http404 from django.utils.translation import ugettext as _ from vanilla import ListView, CreateView, DetailView, UpdateView, DeleteView, TemplateView from .forms import ArticleForm, ArticleSearchForm from .models import Article, Folder from haystack.generic_views import SearchView from haystack.query import SearchQuerySet class ArticleList(ListView): model = Article paginate_by = 20 #class ArticleCreate(CreateView): # model = Article # form_class = ArticleForm # success_url = reverse_lazy('bibloi:list') class ArticleDetail(DetailView): model = Article def get_context_data(self, **kwargs): context = super(ArticleDetail, self).get_context_data(**kwargs) return context #class ArticleUpdate(UpdateView): # model = Article # form_class = ArticleForm # success_url = reverse_lazy('bibloi:list') #class ArticleDelete(DeleteView): # model = Article # success_url = reverse_lazy('bibloi:list') class ArticleSearch(SearchView): template_name = 'search/search.html' form_class = ArticleSearchForm queryset = SearchQuerySet().order_by('-date') paginate_by = 5 def get_context_data(self, **kwargs): context = super(ArticleSearch, self).get_context_data(**kwargs) return context class FolderView(ListView): model = Article template_name = 'bibloi/folder_browse.html' parent = None def get_queryset(self): path = self.kwargs.get('path', '') folders = path.split('/') for folder in folders: try: if not self.parent: if folder: self.parent = Folder.objects.get(name=folder) else: self.parent = self.parent.get_children().get(name=folder) except Folder.DoesNotExist: raise Http404(_('Folder does not exist')) return self.model.objects.filter(folder=self.parent) def get_context_data(self, **kwargs): context = super(FolderView, self).get_context_data(**kwargs) context['parent_folders'] = self.parent.parent_folders if self.parent else [] context['current_folder'] = self.parent if self.parent: context['folders'] = self.parent.get_children() else: context['folders'] = Folder.objects.filter(parent=self.parent) return context class TasksView(TemplateView): template_name = 'tasks.html'
pergamum/bibloi/views.py
2,543
!/usr/bin/env pythonclass ArticleCreate(CreateView): model = Article form_class = ArticleForm success_url = reverse_lazy('bibloi:list')class ArticleUpdate(UpdateView): model = Article form_class = ArticleForm success_url = reverse_lazy('bibloi:list')class ArticleDelete(DeleteView): model = Article success_url = reverse_lazy('bibloi:list')
364
en
0.477143
""" Get absolute URLs for static assets. """ from urllib.parse import ParseResult, urlparse, urlunparse from django import template from django.templatetags.static import StaticNode from django.urls import reverse register = template.Library() class AbstaticNode(StaticNode): """ {% abstatic %} is like Django's {% static %} tag, but always returns an absolute URI. """ def url(self, context) -> str: url_to_asset = super().url(context) parsed_url = urlparse(url_to_asset) assert parsed_url.path if is_absolute_uri(parsed_url): return url_to_asset # Delegate to Django to provide its own schema and authority: path_and_file = to_pf_url(parsed_url) return context["request"].build_absolute_uri(path_and_file) def is_absolute_uri(url: ParseResult) -> bool: """ Returns True if the parsed result is an "absolute URI". We define an "absolute URI" as containing at mimimum a **scheme** and an **host** (a.k.a., an authority). It must contain SH according to the nomenclature defined in this proposal: https://gist.github.com/andrewdotn/eebeaa60d48c3c0f6f9fc75f0ede8d03#proposal Examples of absolute URIs: [SH ] https://example.com [SHP ] https://example.com/ [SHPF] https://example.com/foo/cat.gif What are NOT absolute URIs: [ F] cat.gif [ P ] / [ PF] /foo/cat.gif [ HPF] //example.com/foo/cat.gif† [S F] https:cat.gif (uncommon) [S PF] https:/foo/cat.gif (uncommon) †: This is called a "network-path reference, and relies on inferring the scheme based on an existing base URI. For our purposes, this is not "absolute" enough! Source: https://tools.ietf.org/html/rfc3986#section-4.2 """ # netloc == authority, i.e., [username[:password]@]example.com[:443] if url.scheme and url.netloc: return True return False def to_pf_url(url: ParseResult): """ Returns *P*ath and *F*ile as defined here: https://gist.github.com/andrewdotn/eebeaa60d48c3c0f6f9fc75f0ede8d03#proposal """ return urlunparse(url._replace(scheme="", netloc="")) @register.tag def abstatic(parser, token): """ Given a relative path to a static asset, return the absolute path to the asset. Derived from: https://github.com/django/django/blob/635d53a86a36cde7866b9caefeb64d809e6bfcd9/django/templatetags/static.py#L143-L159 """ return AbstaticNode.handle_token(parser, token)
src/CreeDictionary/CreeDictionary/templatetags/url_extras.py
2,553
{% abstatic %} is like Django's {% static %} tag, but always returns an absolute URI. Given a relative path to a static asset, return the absolute path to the asset. Derived from: https://github.com/django/django/blob/635d53a86a36cde7866b9caefeb64d809e6bfcd9/django/templatetags/static.py#L143-L159 Returns True if the parsed result is an "absolute URI". We define an "absolute URI" as containing at mimimum a **scheme** and an **host** (a.k.a., an authority). It must contain SH according to the nomenclature defined in this proposal: https://gist.github.com/andrewdotn/eebeaa60d48c3c0f6f9fc75f0ede8d03#proposal Examples of absolute URIs: [SH ] https://example.com [SHP ] https://example.com/ [SHPF] https://example.com/foo/cat.gif What are NOT absolute URIs: [ F] cat.gif [ P ] / [ PF] /foo/cat.gif [ HPF] //example.com/foo/cat.gif† [S F] https:cat.gif (uncommon) [S PF] https:/foo/cat.gif (uncommon) †: This is called a "network-path reference, and relies on inferring the scheme based on an existing base URI. For our purposes, this is not "absolute" enough! Source: https://tools.ietf.org/html/rfc3986#section-4.2 Returns *P*ath and *F*ile as defined here: https://gist.github.com/andrewdotn/eebeaa60d48c3c0f6f9fc75f0ede8d03#proposal Get absolute URLs for static assets. Delegate to Django to provide its own schema and authority: netloc == authority, i.e., [username[:password]@]example.com[:443]
1,467
en
0.769448
""" hubspot engagements api """ from hubspot3.base import BaseClient from hubspot3.utils import get_log from typing import Dict, List ENGAGEMENTS_API_VERSION = "1" class EngagementsClient(BaseClient): """ The hubspot3 Engagements client uses the _make_request method to call the API for data. It returns a python object translated from the json returned """ def __init__(self, *args, **kwargs) -> None: super(EngagementsClient, self).__init__(*args, **kwargs) self.log = get_log("hubspot3.engagements") def _get_path(self, subpath: str) -> str: """get full subpath""" return f"engagements/v{self.options.get('version') or ENGAGEMENTS_API_VERSION}/{subpath}" def get(self, engagement_id, **options): """Get a HubSpot engagement.""" return self._call(f"engagements/{engagement_id}", method="GET", **options) def get_associated(self, object_type, object_id, **options) -> List[Dict]: """ get all engagements associated with the given object :param object_type: type of object to get associations on [CONTACT, COMPANY, DEAL] :param object_id: ID of the object to get associations on """ finished = False output = [] # type: List[Dict] query_limit = 100 # Max value according to docs offset = 0 while not finished: batch = self._call( f"engagements/associated/{object_type}/{object_id}/paged", method="GET", params={"limit": query_limit, "offset": offset}, **options, ) output.extend(batch["results"]) finished = not batch["hasMore"] offset = batch["offset"] return output def create(self, data=None, **options): data = data or {} return self._call("engagements", data=data, method="POST", **options) def update(self, key, data=None, **options): data = data or {} return self._call(f"engagements/{key}", data=data, method="PUT", **options) def patch(self, key, data=None, **options): data = data or {} return self._call(f"engagements/{key}", data=data, method="PATCH", **options) def get_all(self, **options) -> List[Dict]: """get all engagements""" finished = False output = [] # type: List[Dict] query_limit = 250 # Max value according to docs offset = 0 while not finished: batch = self._call( "engagements/paged", method="GET", params={"limit": query_limit, "offset": offset}, **options, ) output.extend(batch["results"]) finished = not batch["hasMore"] offset = batch["offset"] return output def get_recently_modified(self, since, **options) -> List[Dict]: """get recently modified engagements""" finished = False output = [] # type: List[Dict] query_limit = 100 # Max value according to docs offset = 0 while not finished: batch = self._call( "engagements/recent/modified", method="GET", params={"limit": query_limit, "offset": offset, "since": since}, **options, ) output.extend(batch["results"]) finished = not batch["hasMore"] offset = batch["offset"] return output
hubspot3/engagements.py
3,513
The hubspot3 Engagements client uses the _make_request method to call the API for data. It returns a python object translated from the json returned get full subpath Get a HubSpot engagement. get all engagements get all engagements associated with the given object :param object_type: type of object to get associations on [CONTACT, COMPANY, DEAL] :param object_id: ID of the object to get associations on get recently modified engagements hubspot engagements api type: List[Dict] Max value according to docs type: List[Dict] Max value according to docs type: List[Dict] Max value according to docs
601
en
0.74581
# coding: utf-8 """ Copyright 2017 Square, Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from pprint import pformat from six import iteritems import re class V1ListItemsRequest(object): """ NOTE: This class is auto generated by the swagger code generator program. Do not edit the class manually. """ def __init__(self, batch_token=None): """ V1ListItemsRequest - a model defined in Swagger :param dict swaggerTypes: The key is attribute name and the value is attribute type. :param dict attributeMap: The key is attribute name and the value is json key in definition. """ self.swagger_types = { 'batch_token': 'str' } self.attribute_map = { 'batch_token': 'batch_token' } self._batch_token = batch_token @property def batch_token(self): """ Gets the batch_token of this V1ListItemsRequest. A pagination cursor to retrieve the next set of results for your original query to the endpoint. :return: The batch_token of this V1ListItemsRequest. :rtype: str """ return self._batch_token @batch_token.setter def batch_token(self, batch_token): """ Sets the batch_token of this V1ListItemsRequest. A pagination cursor to retrieve the next set of results for your original query to the endpoint. :param batch_token: The batch_token of this V1ListItemsRequest. :type: str """ self._batch_token = batch_token def to_dict(self): """ Returns the model properties as a dict """ result = {} for attr, _ in iteritems(self.swagger_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], item[1].to_dict()) if hasattr(item[1], "to_dict") else item, value.items() )) else: result[attr] = value return result def to_str(self): """ Returns the string representation of the model """ return pformat(self.to_dict()) def __repr__(self): """ For `print` and `pprint` """ return self.to_str() def __eq__(self, other): """ Returns true if both objects are equal """ return self.__dict__ == other.__dict__ def __ne__(self, other): """ Returns true if both objects are not equal """ return not self == other
squareconnect/models/v1_list_items_request.py
3,533
NOTE: This class is auto generated by the swagger code generator program. Do not edit the class manually. Returns true if both objects are equal V1ListItemsRequest - a model defined in Swagger :param dict swaggerTypes: The key is attribute name and the value is attribute type. :param dict attributeMap: The key is attribute name and the value is json key in definition. Returns true if both objects are not equal For `print` and `pprint` Gets the batch_token of this V1ListItemsRequest. A pagination cursor to retrieve the next set of results for your original query to the endpoint. :return: The batch_token of this V1ListItemsRequest. :rtype: str Sets the batch_token of this V1ListItemsRequest. A pagination cursor to retrieve the next set of results for your original query to the endpoint. :param batch_token: The batch_token of this V1ListItemsRequest. :type: str Returns the model properties as a dict Returns the string representation of the model Copyright 2017 Square, Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. coding: utf-8
1,616
en
0.793423
# coding: utf-8 """ Swaggy Jenkins Jenkins API clients generated from Swagger / Open API specification # noqa: E501 The version of the OpenAPI document: 1.1.2-pre.0 Contact: blah@cliffano.com Generated by: https://openapi-generator.tech """ from __future__ import absolute_import import unittest import datetime import openapi_client from openapi_client.models.queue import Queue # noqa: E501 from openapi_client.rest import ApiException class TestQueue(unittest.TestCase): """Queue unit test stubs""" def setUp(self): pass def tearDown(self): pass def make_instance(self, include_optional): """Test Queue include_option is a boolean, when False only required params are included, when True both required and optional params are included """ # model = openapi_client.models.queue.Queue() # noqa: E501 if include_optional : return Queue( _class = '', items = [ openapi_client.models.queue_blocked_item.QueueBlockedItem( _class = '', actions = [ openapi_client.models.cause_action.CauseAction( _class = '', causes = [ openapi_client.models.cause_user_id_cause.CauseUserIdCause( _class = '', short_description = '', user_id = '', user_name = '', ) ], ) ], blocked = True, buildable = True, id = 56, in_queue_since = 56, params = '', stuck = True, task = openapi_client.models.free_style_project.FreeStyleProject( _class = '', name = '', url = '', color = '', description = '', display_name = '', display_name_or_null = '', full_display_name = '', full_name = '', buildable = True, builds = [ openapi_client.models.free_style_build.FreeStyleBuild( _class = '', number = 56, url = '', building = True, description = '', display_name = '', duration = 56, estimated_duration = 56, executor = '', full_display_name = '', id = '', keep_log = True, queue_id = 56, result = '', timestamp = 56, built_on = '', change_set = openapi_client.models.empty_change_log_set.EmptyChangeLogSet( _class = '', kind = '', ), ) ], first_build = openapi_client.models.free_style_build.FreeStyleBuild( _class = '', number = 56, url = '', building = True, description = '', display_name = '', duration = 56, estimated_duration = 56, executor = '', full_display_name = '', id = '', keep_log = True, queue_id = 56, result = '', timestamp = 56, built_on = '', ), health_report = [ openapi_client.models.free_style_projecthealth_report.FreeStyleProjecthealthReport( description = '', icon_class_name = '', icon_url = '', score = 56, _class = '', ) ], in_queue = True, keep_dependencies = True, last_build = openapi_client.models.free_style_build.FreeStyleBuild( _class = '', number = 56, url = '', building = True, description = '', display_name = '', duration = 56, estimated_duration = 56, executor = '', full_display_name = '', id = '', keep_log = True, queue_id = 56, result = '', timestamp = 56, built_on = '', ), last_completed_build = openapi_client.models.free_style_build.FreeStyleBuild( _class = '', number = 56, url = '', building = True, description = '', display_name = '', duration = 56, estimated_duration = 56, executor = '', full_display_name = '', id = '', keep_log = True, queue_id = 56, result = '', timestamp = 56, built_on = '', ), last_failed_build = '', last_stable_build = openapi_client.models.free_style_build.FreeStyleBuild( _class = '', number = 56, url = '', building = True, description = '', display_name = '', duration = 56, estimated_duration = 56, executor = '', full_display_name = '', id = '', keep_log = True, queue_id = 56, result = '', timestamp = 56, built_on = '', ), last_successful_build = openapi_client.models.free_style_build.FreeStyleBuild( _class = '', number = 56, url = '', building = True, description = '', display_name = '', duration = 56, estimated_duration = 56, executor = '', full_display_name = '', id = '', keep_log = True, queue_id = 56, result = '', timestamp = 56, built_on = '', ), last_unstable_build = '', last_unsuccessful_build = '', next_build_number = 56, queue_item = '', concurrent_build = True, scm = openapi_client.models.null_scm.NullSCM( _class = '', ), ), url = '', why = '', buildable_start_milliseconds = 56, ) ] ) else : return Queue( ) def testQueue(self): """Test Queue""" inst_req_only = self.make_instance(include_optional=False) inst_req_and_optional = self.make_instance(include_optional=True) if __name__ == '__main__': unittest.main()
clients/python-legacy/generated/test/test_queue.py
9,671
Queue unit test stubs Test Queue include_option is a boolean, when False only required params are included, when True both required and optional params are included Test Queue Swaggy Jenkins Jenkins API clients generated from Swagger / Open API specification # noqa: E501 The version of the OpenAPI document: 1.1.2-pre.0 Contact: blah@cliffano.com Generated by: https://openapi-generator.tech coding: utf-8 noqa: E501 model = openapi_client.models.queue.Queue() noqa: E501
480
en
0.709678
# Code generated by lark_sdk_gen. DO NOT EDIT. import typing from pylark.lark_request import Response from pylark.api_service_hire_job_get import ( GetHireJobReq, GetHireJobResp, _gen_get_hire_job_req, ) from pylark.api_service_hire_job_manager_get import ( GetHireJobManagerReq, GetHireJobManagerResp, _gen_get_hire_job_manager_req, ) from pylark.api_service_hire_talent_get import ( GetHireTalentReq, GetHireTalentResp, _gen_get_hire_talent_req, ) from pylark.api_service_hire_attachment_get import ( GetHireAttachmentReq, GetHireAttachmentResp, _gen_get_hire_attachment_req, ) from pylark.api_service_hire_attachment_preview_get import ( GetHireAttachmentPreviewReq, GetHireAttachmentPreviewResp, _gen_get_hire_attachment_preview_req, ) from pylark.api_service_hire_resume_sources_get import ( GetHireResumeSourceReq, GetHireResumeSourceResp, _gen_get_hire_resume_source_req, ) from pylark.api_service_hire_note_create import ( CreateHireNoteReq, CreateHireNoteResp, _gen_create_hire_note_req, ) from pylark.api_service_hire_note_update import ( UpdateHireNoteReq, UpdateHireNoteResp, _gen_update_hire_note_req, ) from pylark.api_service_hire_note_get import ( GetHireNoteReq, GetHireNoteResp, _gen_get_hire_note_req, ) from pylark.api_service_hire_note_list import ( GetHireNoteListReq, GetHireNoteListResp, _gen_get_hire_note_list_req, ) from pylark.api_service_hire_referral_get_by_application import ( GetHireReferralByApplicationReq, GetHireReferralByApplicationResp, _gen_get_hire_referral_by_application_req, ) from pylark.api_service_hire_job_process_list import ( GetHireJobProcessListReq, GetHireJobProcessListResp, _gen_get_hire_job_process_list_req, ) from pylark.api_service_hire_application_create import ( CreateHireApplicationReq, CreateHireApplicationResp, _gen_create_hire_application_req, ) from pylark.api_service_hire_application_terminate import ( TerminateHireApplicationReq, TerminateHireApplicationResp, _gen_terminate_hire_application_req, ) from pylark.api_service_hire_application_get import ( GetHireApplicationReq, GetHireApplicationResp, _gen_get_hire_application_req, ) from pylark.api_service_hire_application_list import ( GetHireApplicationListReq, GetHireApplicationListResp, _gen_get_hire_application_list_req, ) from pylark.api_service_hire_application_interview_list import ( GetHireApplicationInterviewListReq, GetHireApplicationInterviewListResp, _gen_get_hire_application_interview_list_req, ) from pylark.api_service_hire_offer_get_by_application import ( GetHireOfferByApplicationReq, GetHireOfferByApplicationResp, _gen_get_hire_offer_by_application_req, ) from pylark.api_service_hire_offer_schema_get import ( GetHireOfferSchemaReq, GetHireOfferSchemaResp, _gen_get_hire_offer_schema_req, ) from pylark.api_service_hire_transfer_onboard_by_application import ( MakeHireTransferOnboardByApplicationReq, MakeHireTransferOnboardByApplicationResp, _gen_make_hire_transfer_onboard_by_application_req, ) from pylark.api_service_hire_employee_update import ( UpdateHireEmployeeReq, UpdateHireEmployeeResp, _gen_update_hire_employee_req, ) from pylark.api_service_hire_employee_get_by_application import ( GetHireEmployeeByApplicationReq, GetHireEmployeeByApplicationResp, _gen_get_hire_employee_by_application_req, ) from pylark.api_service_hire_employee_get import ( GetHireEmployeeReq, GetHireEmployeeResp, _gen_get_hire_employee_req, ) if typing.TYPE_CHECKING: from lark import Lark class LarkHireService(object): cli: "Lark" def __init__(self, cli: "Lark"): self.cli = cli def get_hire_job( self, request: GetHireJobReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireJobResp, Response]: return self.cli.raw_request(_gen_get_hire_job_req(request, options)) def get_hire_job_manager( self, request: GetHireJobManagerReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireJobManagerResp, Response]: return self.cli.raw_request(_gen_get_hire_job_manager_req(request, options)) def get_hire_talent( self, request: GetHireTalentReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireTalentResp, Response]: return self.cli.raw_request(_gen_get_hire_talent_req(request, options)) def get_hire_attachment( self, request: GetHireAttachmentReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireAttachmentResp, Response]: return self.cli.raw_request(_gen_get_hire_attachment_req(request, options)) def get_hire_attachment_preview( self, request: GetHireAttachmentPreviewReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireAttachmentPreviewResp, Response]: return self.cli.raw_request( _gen_get_hire_attachment_preview_req(request, options) ) def get_hire_resume_source( self, request: GetHireResumeSourceReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireResumeSourceResp, Response]: return self.cli.raw_request(_gen_get_hire_resume_source_req(request, options)) def create_hire_note( self, request: CreateHireNoteReq, options: typing.List[str] = None ) -> typing.Tuple[CreateHireNoteResp, Response]: return self.cli.raw_request(_gen_create_hire_note_req(request, options)) def update_hire_note( self, request: UpdateHireNoteReq, options: typing.List[str] = None ) -> typing.Tuple[UpdateHireNoteResp, Response]: return self.cli.raw_request(_gen_update_hire_note_req(request, options)) def get_hire_note( self, request: GetHireNoteReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireNoteResp, Response]: return self.cli.raw_request(_gen_get_hire_note_req(request, options)) def get_hire_note_list( self, request: GetHireNoteListReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireNoteListResp, Response]: return self.cli.raw_request(_gen_get_hire_note_list_req(request, options)) def get_hire_referral_by_application( self, request: GetHireReferralByApplicationReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireReferralByApplicationResp, Response]: return self.cli.raw_request( _gen_get_hire_referral_by_application_req(request, options) ) def get_hire_job_process_list( self, request: GetHireJobProcessListReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireJobProcessListResp, Response]: return self.cli.raw_request( _gen_get_hire_job_process_list_req(request, options) ) def create_hire_application( self, request: CreateHireApplicationReq, options: typing.List[str] = None ) -> typing.Tuple[CreateHireApplicationResp, Response]: return self.cli.raw_request(_gen_create_hire_application_req(request, options)) def terminate_hire_application( self, request: TerminateHireApplicationReq, options: typing.List[str] = None ) -> typing.Tuple[TerminateHireApplicationResp, Response]: return self.cli.raw_request( _gen_terminate_hire_application_req(request, options) ) def get_hire_application( self, request: GetHireApplicationReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireApplicationResp, Response]: return self.cli.raw_request(_gen_get_hire_application_req(request, options)) def get_hire_application_list( self, request: GetHireApplicationListReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireApplicationListResp, Response]: return self.cli.raw_request( _gen_get_hire_application_list_req(request, options) ) def get_hire_application_interview_list( self, request: GetHireApplicationInterviewListReq, options: typing.List[str] = None, ) -> typing.Tuple[GetHireApplicationInterviewListResp, Response]: return self.cli.raw_request( _gen_get_hire_application_interview_list_req(request, options) ) def get_hire_offer_by_application( self, request: GetHireOfferByApplicationReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireOfferByApplicationResp, Response]: return self.cli.raw_request( _gen_get_hire_offer_by_application_req(request, options) ) def get_hire_offer_schema( self, request: GetHireOfferSchemaReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireOfferSchemaResp, Response]: return self.cli.raw_request(_gen_get_hire_offer_schema_req(request, options)) def make_hire_transfer_onboard_by_application( self, request: MakeHireTransferOnboardByApplicationReq, options: typing.List[str] = None, ) -> typing.Tuple[MakeHireTransferOnboardByApplicationResp, Response]: return self.cli.raw_request( _gen_make_hire_transfer_onboard_by_application_req(request, options) ) def update_hire_employee( self, request: UpdateHireEmployeeReq, options: typing.List[str] = None ) -> typing.Tuple[UpdateHireEmployeeResp, Response]: return self.cli.raw_request(_gen_update_hire_employee_req(request, options)) def get_hire_employee_by_application( self, request: GetHireEmployeeByApplicationReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireEmployeeByApplicationResp, Response]: return self.cli.raw_request( _gen_get_hire_employee_by_application_req(request, options) ) def get_hire_employee( self, request: GetHireEmployeeReq, options: typing.List[str] = None ) -> typing.Tuple[GetHireEmployeeResp, Response]: return self.cli.raw_request(_gen_get_hire_employee_req(request, options))
pylark/api_service_hire.py
10,069
Code generated by lark_sdk_gen. DO NOT EDIT.
44
en
0.515403
"""Run experiment. This module is intended to be run as a script: $ python src/experiment.py """ import os import pandas as pd from bert import BERT from constants import LANGUAGES, MASK, MISSING from filenames import CLOZE_DIR, EXPERIMENTS_DIR, FEATURES_DIR from utils import refresh ENGLISH_MODEL = "bert-base-cased" MULTILINGUAL_MODEL = "bert-base-multilingual-cased" def index_of_masked_word(sentence, bert): """Return index of the masked word in `sentence` using `bert`'s' tokenizer. We use this function to calculate the linear distance between the target and controller as BERT sees it. Parameters ---------- sentence : str Returns ------- int """ tokens = bert.tokenize(sentence) try: return tokens.index(MASK) except ValueError: # MASK not in sentence return -1 def run(language, force_multilingual=False, fold_case=True, gpu=True): """Run the experiment for `language`. Parameters ---------- language : str force_multilingual : bool Whether to use the multilingual model even on English fold_case : bool Whether to ignore caseing differences after making predictions gpu : bool Whether to run on GPU or not (useful for debugging) Returns ------- pd.DataFrame """ if (language == "English") and (not force_multilingual): bert = BERT(ENGLISH_MODEL, gpu=gpu) else: bert = BERT(MULTILINGUAL_MODEL, gpu=gpu) vocab = bert.vocab if fold_case: vocab = [word.lower() for word in vocab] code = LANGUAGES[language] cloze = pd.read_csv(os.path.join(CLOZE_DIR, f"{code}.csv")) num_examples = len(cloze) * 2 # because we mask out both words print(f"\n\nNumber of examples for {language}: {num_examples}") print_every = num_examples // 100 features = pd.read_csv( os.path.join(FEATURES_DIR, f"{code}.csv"), dtype={"person": str} ) # remove any words that aren't in the vocab features = features[features["word"].isin(vocab)] # if we are masking out the controller, we know that the masked word is # also a noun or a pronoun, so we can remove everything else from features # features = features[features['pos'].isin(['NOUN', 'PRON'])] cols = ["number", "gender", "case", "person"] result = [] count, total = 0, 0 for _, example in cloze.iterrows(): for mask in ["masked", "other_masked"]: try: predictions = bert.predict(example[mask], fold_case) except ValueError: # MASK not in sentence continue predictions = features.merge( predictions, how="left", left_on="word", right_index=True ) # only keep words of the same POS category as the masked word predictions = predictions[predictions["pos"] == example["pos"]] # A word is correct if all its features are identical with the features # of the masked word. predictions["correct"] = (predictions[cols] == example[cols]).all(axis=1) # If a word form has multiple feature bundles and at least one of them # is correct, then we count that word form as correct. The values of # 'p' for the differently valued but identical word forms will be # identical (because BERT predicts word forms). I want to include the # 'p' in the resulting dataframe so I just take the first value. predictions = predictions.groupby("word").agg( {"correct": any, "p": "first"} ) # we compute the average (unnormalized) probability of all the word # forms BERT got correct and all it got incorrect. mean = predictions.groupby("correct")["p"].mean() try: example["correct"] = mean[True] except KeyError: example["correct"] = 0.0 try: example["incorrect"] = mean[False] except KeyError: example["incorrect"] = 0.0 # add in the linear distance between masked and other word masked_index = index_of_masked_word(example["masked"], bert) other_index = index_of_masked_word(example["other_masked"], bert) example["distance"] = abs(masked_index - other_index) result.append(example) if example["correct"] > example["incorrect"]: count += 1 total += 1 if total % print_every == 0: percent_correct = round(100 * (count / total), 3) percent_done = round(100 * (total / num_examples), 3) print(f"{percent_correct}% correct with {percent_done}% done") result = pd.DataFrame(result) result["right"] = result["correct"] > result["incorrect"] file_name = os.path.join(EXPERIMENTS_DIR, f"{code}.csv") result.to_csv(file_name, index=False) return result if __name__ == "__main__": # # # refresh(EXPERIMENTS_DIR) # don't uncomment me! # # run experiments for languages with fewer cloze examples first # ORDER = { # language: len(pd.read_csv(os.path.join(CLOZE_DIR, f'{code}.csv'))) # for language, code in LANGUAGES.items() # } ORDER = {"Czech": 0, "German": 1} for language in sorted(ORDER, key=ORDER.get): try: result = run(language) proportion_correct = result["right"].value_counts(normalize=True)[True] print(language, round(proportion_correct, 2)) except: # noqa print(f"Error with {language}")
src/experiment.py
5,693
Return index of the masked word in `sentence` using `bert`'s' tokenizer. We use this function to calculate the linear distance between the target and controller as BERT sees it. Parameters ---------- sentence : str Returns ------- int Run the experiment for `language`. Parameters ---------- language : str force_multilingual : bool Whether to use the multilingual model even on English fold_case : bool Whether to ignore caseing differences after making predictions gpu : bool Whether to run on GPU or not (useful for debugging) Returns ------- pd.DataFrame Run experiment. This module is intended to be run as a script: $ python src/experiment.py MASK not in sentence because we mask out both words remove any words that aren't in the vocab if we are masking out the controller, we know that the masked word is also a noun or a pronoun, so we can remove everything else from features features = features[features['pos'].isin(['NOUN', 'PRON'])] MASK not in sentence only keep words of the same POS category as the masked word A word is correct if all its features are identical with the features of the masked word. If a word form has multiple feature bundles and at least one of them is correct, then we count that word form as correct. The values of 'p' for the differently valued but identical word forms will be identical (because BERT predicts word forms). I want to include the 'p' in the resulting dataframe so I just take the first value. we compute the average (unnormalized) probability of all the word forms BERT got correct and all it got incorrect. add in the linear distance between masked and other word refresh(EXPERIMENTS_DIR) don't uncomment me! run experiments for languages with fewer cloze examples first ORDER = { language: len(pd.read_csv(os.path.join(CLOZE_DIR, f'{code}.csv'))) for language, code in LANGUAGES.items() } noqa
1,910
en
0.84095
# Python Object Oriented Programming import datetime class Employee: num_of_emps = 0 # Class Variable raise_amount = 1.02 # Class Variable def __init__(self, FirstName, LastName, salary): self.FirstName = FirstName self.LastName = LastName self.salary = int(salary) self.email = FirstName + "." + LastName + "@email.com" Employee.num_of_emps += 1 # Class Variable def FullName(self): return "{} {}".format(self.FirstName, self.LastName) def apply_raise(self): self.salary = int(self.salary * self.raise_amount) @classmethod def set_raise_amount(cls, amount): cls.raise_amount = amount @classmethod def from_string(cls, emp_str): first, last, salary = emp_str.split("-") return cls(first, last, salary) @staticmethod def is_workday(day): if day.weekday() == 5 or day.weekday() == 6: return False else: return True def __repr__(self): return "Employee('{}', '{}', '{}')".format(self.FirstName, self.LastName, self.salary) def __str__(self): return "{} - {}".format(self.FullName(), self.email) def __add__(self, other): return self.salary + other.salary def __len__(self): return len(self.FullName()) class Developer(Employee): def __init__(self, FirstName, LastName, salary, prog_lang): super().__init__(FirstName, LastName, salary) self.prog_lang = prog_lang @classmethod def from_string(cls, dev_str): first, last, salary, prog_lang = dev_str.split("-") return cls(first, last, salary, prog_lang) class Manager(Employee): def __init__(self, FirstName, LastName, salary, employees = None): super().__init__(FirstName, LastName, salary) if employees is None: self.employees = [] else: self.employees = employees def add_emp(self, emp): if emp not in self.employees: self.employees.append(emp) def remove_emp(self, emp): if emp in self.employees: self.employees.remove(emp) def print_emps(self): for emp in self.employees: print("\n --> {}".format(emp.FullName())) emp_1 = Employee("Sudani", "Coder", 100500) emp_2 = Employee("Root", "Admin", 100500) print() print(emp_1, end = "\n\n") print(repr(emp_2), end = "\n\n") print(str(emp_2), end = "\n\n") print(emp_1 + emp_2, end = "\n\n") print(len(emp_2), end = "\n\n")
Object Oriented Programming/5 - Special Methods/index.py
2,539
Python Object Oriented Programming Class Variable Class Variable Class Variable
79
en
0.522726
# -*- coding: utf-8 -*- """ Key classification using multiclass Support Vector Machine (SVM) reference: Date: Jun 05, 2017 @author: Thuong Tran @Library: scikit-learn """ import os, glob, random import numpy as np from pandas import DataFrame from sklearn.feature_extraction.text import CountVectorizer, TfidfTransformer from sklearn.naive_bayes import MultinomialNB from sklearn.svm import LinearSVC from sklearn.metrics import confusion_matrix, precision_score from sklearn.pipeline import Pipeline from sklearn.externals import joblib from sklearn.cross_validation import KFold import time import codecs import matplotlib.pyplot as plt import itertools NEW_LINE = '\r\n' TRAIN_SIZE = 0.8 def build_data_frame(data_dir): # folders = [d for d in os.listdir(data_dir) if os.path.isdir(os.path.join(data_dir, d))] dirs = next(os.walk(data_dir))[1] # [0]: path, [1]: folders list, [2] files list class_names = [] total_amount = [] train_amount = [] test_amount = [] train_data = DataFrame({'value': [], 'class': []}) test_data = DataFrame({'value': [], 'class': []}) for d in dirs: tmp_dir = os.path.join(data_dir, d) rows = [] index = [] for f in glob.glob(os.path.join(tmp_dir, '*.txt')): with open(f, encoding="latin1") as fc: value = [line.replace('\n', '').replace('\r', '').replace('\t', '') for line in fc.readlines()] value = '. '.join(value) rows.append({'value': value, 'class': d}) index.append(f) tmp_df = DataFrame(rows, index=index) size = int(len(tmp_df) * TRAIN_SIZE) train_df, test_df = tmp_df.iloc[:size], tmp_df.iloc[size:] train_data = train_data.append(train_df) test_data = test_data.append(test_df) class_names.append(d) total_amount.append(len(os.listdir(tmp_dir))) train_amount.append(len(train_df)) test_amount.append(len(test_df)) tmp_arr = np.array([total_amount, train_amount, test_amount]) print (DataFrame(tmp_arr, ['Total', 'Train', 'Test'], class_names)) train_data = train_data.reindex(np.random.permutation(train_data.index)) test_data = test_data.reindex(np.random.permutation(test_data.index)) return train_data, test_data, class_names def plot_confusion_matrix(cm, classes, normalize=False, title='Confusion matrix', cmap=plt.cm.Blues): """ This function prints and plots the confusion matrix. Normalization can be applied by setting `normalize=True`. """ plt.imshow(cm, interpolation='nearest', cmap=cmap) plt.title(title) plt.colorbar() tick_marks = np.arange(len(classes)) plt.xticks(tick_marks, classes, rotation=45) plt.yticks(tick_marks, classes) if normalize: cm = cm.astype('float') / cm.sum(axis=1)[:, np.newaxis] print("Normalized confusion matrix") else: print('Confusion matrix, without normalization') print(cm) thresh = cm.max() / 2. for i, j in itertools.product(range(cm.shape[0]), range(cm.shape[1])): plt.text(j, i, cm[i, j], horizontalalignment="center", color="white" if cm[i, j] > thresh else "black") plt.tight_layout() plt.ylabel('True label') plt.xlabel('Predicted label') def main(): data_dir = '/Users/thuong/Documents/tmp_datasets/SI/TrainValue' train_data_df, test_data_df, class_names = build_data_frame(data_dir) pipeline = Pipeline([ ('vectorizer', CountVectorizer()), ('tfidf_transformer', TfidfTransformer()), ('classifier', LinearSVC())]) ######### One-KFolds ############################## train_data, test_data = train_data_df['value'].values, test_data_df['value'].values train_target, test_target = train_data_df['class'].values, test_data_df['class'].values pipeline.fit(train_data, train_target) predictions = pipeline.predict(test_data) cnf_matrix = confusion_matrix(test_target, predictions) print('Confusion matrix with one-fold: ') print(cnf_matrix) print("Score with one-fold: %s" % precision_score(test_target, predictions, average = 'weighted')) print("Score with one-fold: %s" % precision_score(test_target, predictions, average = None)) # ######### KFolds ############################## # k_fold = KFold(n=len(data_frame), n_folds=6) # scores = [] # confusion = np.array([[0, 0], [0, 0]]) # for train_indices, test_indices in k_fold: # train_text = data_frame.iloc[train_indices]['text'].values # train_label = data_frame.iloc[train_indices]['class'].values # test_text = data_frame.iloc[test_indices]['text'].values # test_label = data_frame.iloc[test_indices]['class'].values # pipeline.fit(train_text, train_label) # predictions = pipeline.predict(test_text) # confusion += confusion_matrix(test_label, predictions) # score = f1_score(test_label, predictions, pos_label = SPAM) # scores.append(score) # print('Confusion matrix with 6-fold: ') # print(confusion) # print('Score with 6-fold: %s' % (sum(scores)/len(scores))) # Plot non-normalized confusion matrix plt.figure() plot_confusion_matrix(cnf_matrix, classes=class_names, title='Confusion matrix, without normalization') # Plot normalized confusion matrix plt.figure() plot_confusion_matrix(cnf_matrix, classes=class_names, normalize=True, title='Normalized confusion matrix') plt.show() if __name__ == "__main__": main()
TrainValue/multiclass_svm.py
5,500
This function prints and plots the confusion matrix. Normalization can be applied by setting `normalize=True`. Key classification using multiclass Support Vector Machine (SVM) reference: Date: Jun 05, 2017 @author: Thuong Tran @Library: scikit-learn -*- coding: utf-8 -*- folders = [d for d in os.listdir(data_dir) if os.path.isdir(os.path.join(data_dir, d))] [0]: path, [1]: folders list, [2] files list One-KFolds KFolds k_fold = KFold(n=len(data_frame), n_folds=6) scores = [] confusion = np.array([[0, 0], [0, 0]]) for train_indices, test_indices in k_fold: train_text = data_frame.iloc[train_indices]['text'].values train_label = data_frame.iloc[train_indices]['class'].values test_text = data_frame.iloc[test_indices]['text'].values test_label = data_frame.iloc[test_indices]['class'].values pipeline.fit(train_text, train_label) predictions = pipeline.predict(test_text) confusion += confusion_matrix(test_label, predictions) score = f1_score(test_label, predictions, pos_label = SPAM) scores.append(score) print('Confusion matrix with 6-fold: ') print(confusion) print('Score with 6-fold: %s' % (sum(scores)/len(scores))) Plot non-normalized confusion matrix Plot normalized confusion matrix
1,227
en
0.491376
import math import collections from gym_jsbsim import utils class BoundedProperty(collections.namedtuple('BoundedProperty', ['name', 'description', 'min', 'max'])): def get_legal_name(self): return utils.AttributeFormatter.translate(self.name) class Property(collections.namedtuple('Property', ['name', 'description'])): def get_legal_name(self): return utils.AttributeFormatter.translate(self.name) # position and attitude altitude_sl_ft = BoundedProperty('position/h-sl-ft', 'altitude above mean sea level [ft]', -1400, 85000) pitch_rad = BoundedProperty('attitude/pitch-rad', 'pitch [rad]', -0.5 * math.pi, 0.5 * math.pi) roll_rad = BoundedProperty('attitude/roll-rad', 'roll [rad]', -math.pi, math.pi) heading_deg = BoundedProperty('attitude/psi-deg', 'heading [deg]', 0, 360) sideslip_deg = BoundedProperty('aero/beta-deg', 'sideslip [deg]', -180, +180) lat_geod_deg = BoundedProperty('position/lat-geod-deg', 'geocentric latitude [deg]', -90, 90) lng_geoc_deg = BoundedProperty('position/long-gc-deg', 'geodesic longitude [deg]', -180, 180) dist_travel_m = Property('position/distance-from-start-mag-mt', 'distance travelled from starting position [m]') # velocities u_fps = BoundedProperty('velocities/u-fps', 'body frame x-axis velocity [ft/s]', -2200, 2200) v_fps = BoundedProperty('velocities/v-fps', 'body frame y-axis velocity [ft/s]', -2200, 2200) w_fps = BoundedProperty('velocities/w-fps', 'body frame z-axis velocity [ft/s]', -2200, 2200) v_north_fps = BoundedProperty('velocities/v-north-fps', 'velocity true north [ft/s]', float('-inf'), float('+inf')) v_east_fps = BoundedProperty('velocities/v-east-fps', 'velocity east [ft/s]', float('-inf'), float('+inf')) v_down_fps = BoundedProperty('velocities/v-down-fps', 'velocity downwards [ft/s]', float('-inf'), float('+inf')) vc_fps = BoundedProperty("velocities/vc-fps", "airspeed in knots", 0, 4400) p_radps = BoundedProperty('velocities/p-rad_sec', 'roll rate [rad/s]', -2 * math.pi, 2 * math.pi) q_radps = BoundedProperty('velocities/q-rad_sec', 'pitch rate [rad/s]', -2 * math.pi, 2 * math.pi) r_radps = BoundedProperty('velocities/r-rad_sec', 'yaw rate [rad/s]', -2 * math.pi, 2 * math.pi) altitude_rate_fps = Property('velocities/h-dot-fps', 'Rate of altitude change [ft/s]') # accelerations accelerations_n_pilot_x_norm = Property('accelerations/n-pilot-x-norm', 'pilot body x-axis acceleration, normalised') accelerations_n_pilot_y_norm = Property('accelerations/n-pilot-y-norm', 'pilot body y-axis acceleration, normalised') accelerations_n_pilot_z_norm = Property('accelerations/n-pilot-z-norm', 'pilot body z-axis acceleration, normalised') # controls state aileron_left = BoundedProperty('fcs/left-aileron-pos-norm', 'left aileron position, normalised', -1, 1) aileron_right = BoundedProperty('fcs/right-aileron-pos-norm', 'right aileron position, normalised', -1, 1) elevator = BoundedProperty('fcs/elevator-pos-norm', 'elevator position, normalised', -1, 1) rudder = BoundedProperty('fcs/rudder-pos-norm', 'rudder position, normalised', -1, 1) throttle = BoundedProperty('fcs/throttle-pos-norm', 'throttle position, normalised', 0, 1) gear = BoundedProperty('gear/gear-pos-norm', 'landing gear position, normalised', 0, 1) # engines engine_running = Property('propulsion/engine/set-running', 'engine running (0/1 bool)') all_engine_running = Property('propulsion/set-running', 'set engine running (-1 for all engines)') engine_thrust_lbs = Property('propulsion/engine/thrust-lbs', 'engine thrust [lb]') # controls command aileron_cmd = BoundedProperty('fcs/aileron-cmd-norm', 'aileron commanded position, normalised', -1., 1.) elevator_cmd = BoundedProperty('fcs/elevator-cmd-norm', 'elevator commanded position, normalised', -1., 1.) rudder_cmd = BoundedProperty('fcs/rudder-cmd-norm', 'rudder commanded position, normalised', -1., 1.) throttle_cmd = BoundedProperty('fcs/throttle-cmd-norm', 'throttle commanded position, normalised', 0., 1.) mixture_cmd = BoundedProperty('fcs/mixture-cmd-norm', 'engine mixture setting, normalised', 0., 1.) gear_all_cmd = BoundedProperty('gear/gear-cmd-norm', 'all landing gear commanded position, normalised', 0, 1) # simulation sim_dt = Property('simulation/dt', 'JSBSim simulation timestep [s]') sim_time_s = Property('simulation/sim-time-sec', 'Simulation time [s]') # initial conditions initial_altitude_ft = Property('ic/h-sl-ft', 'initial altitude MSL [ft]') initial_terrain_altitude_ft = Property('ic/terrain-elevation-ft', 'initial terrain alt [ft]') initial_longitude_geoc_deg = Property('ic/long-gc-deg', 'initial geocentric longitude [deg]') initial_latitude_geod_deg = Property('ic/lat-geod-deg', 'initial geodesic latitude [deg]') initial_u_fps = Property('ic/u-fps', 'body frame x-axis velocity; positive forward [ft/s]') initial_v_fps = Property('ic/v-fps', 'body frame y-axis velocity; positive right [ft/s]') initial_w_fps = Property('ic/w-fps', 'body frame z-axis velocity; positive down [ft/s]') initial_p_radps = Property('ic/p-rad_sec', 'roll rate [rad/s]') initial_q_radps = Property('ic/q-rad_sec', 'pitch rate [rad/s]') initial_r_radps = Property('ic/r-rad_sec', 'yaw rate [rad/s]') initial_roc_fpm = Property('ic/roc-fpm', 'initial rate of climb [ft/min]') initial_heading_deg = Property('ic/psi-true-deg', 'initial (true) heading [deg]') class Vector2(object): def __init__(self, x: float, y: float): self.x = x self.y = y def heading_deg(self): """ Calculate heading in degrees of vector from origin """ heading_rad = math.atan2(self.x, self.y) heading_deg_normalised = (math.degrees(heading_rad) + 360) % 360 return heading_deg_normalised @staticmethod def from_sim(sim: 'simulation.Simulation') -> 'Vector2': return Vector2(sim[v_east_fps], sim[v_north_fps]) class GeodeticPosition(object): def __init__(self, latitude_deg: float, longitude_deg: float): self.lat = latitude_deg self.lon = longitude_deg def heading_deg_to(self, destination: 'GeodeticPosition') -> float: """ Determines heading in degrees of course between self and destination """ difference_vector = destination - self return difference_vector.heading_deg() @staticmethod def from_sim(sim: 'simulation.Simulation') -> 'GeodeticPosition': """ Return a GeodeticPosition object with lat and lon from simulation """ lat_deg = sim[lat_geod_deg] lon_deg = sim[lng_geoc_deg] return GeodeticPosition(lat_deg, lon_deg) def __sub__(self, other) -> Vector2: """ Returns difference between two coords as (delta_lat, delta_long) """ return Vector2(self.lon - other.lon, self.lat - other.lat)
gym_jsbsim/properties.py
6,721
Returns difference between two coords as (delta_lat, delta_long) Return a GeodeticPosition object with lat and lon from simulation Calculate heading in degrees of vector from origin Determines heading in degrees of course between self and destination position and attitude velocities accelerations controls state engines controls command simulation initial conditions
373
en
0.887022
import re import os import time import tweepy import neologdn import emoji TRAINFILE = "./data/train.tsv" DEVFILE = "./data/dev.tsv" TESTFILE = "./data/test.tsv" dup = [] if os.path.exists(TRAINFILE): with open(TRAINFILE) as f: dup += f.readlines() if os.path.exists(DEVFILE): with open(DEVFILE) as f: dup += f.readlines() if os.path.exists(TESTFILE): with open(TESTFILE) as f: dup += f.readlines() class Tweet: def __init__(self, status): self.in_reply_to_status_id = status.in_reply_to_status_id self.text = status.text self.created_at = status.created_at self.screen_name = status.user.screen_name self.username = status.user.name self.user_id = status.user.id def is_valid_tweet(status): # is bot if "bot" in status.user.screen_name: return False # include URL if re.search(r"https?://", status.text): return False # is hashtag if re.search(r"#(\w+)", status.text): return False # reply to multi user tweet = re.sub(r"@([A-Za-z0-9_]+)", "<unk>", status.text) if tweet.split().count("<unk>") > 1: return False # too long if len(tweet.replace("<unk>", "")) > 20: return False return True def normalize(text): text = text.replace(" ", "") text = text.replace("\n", "") text = neologdn.normalize(text) text = "".join(["" if c in emoji.UNICODE_EMOJI["en"].keys() else c for c in text]) tmp = re.sub(r"(\d)([,.])(\d+)", r"\1\3", text) # text = re.sub(r"\d+", "0", tmp) tmp = re.sub(r"[!-/:-@[-`{-~]", r" ", text) text = re.sub(u"[■-♯]", " ", tmp) text = text.strip() return text def main(): query = input("Search Query: ") max_tw = int(input("Tweet Count: ")) CK = os.getenv("TW_CK") CS = os.getenv("TW_CS") auth = tweepy.AppAuthHandler(CK, CS) api = tweepy.API(auth) got = 0 filtered = 0 saved = 0 lookup_ids = [] replies = {} max_id = None while saved <= max_tw: try: statuses = api.search_tweets( q=query, lang="ja", count=100, max_id=max_id ) max_id = statuses[-1].id for status in statuses: got += 1 # is not reply if not status.in_reply_to_status_id: continue # filter if not is_valid_tweet(status): continue # append lookup id lookup_ids.append(status.in_reply_to_status_id) replies[status.in_reply_to_status_id] = Tweet(status) filtered += 1 print(f"\r{got} => {filtered} => {saved}", end="") # collect 100 tweets if len(lookup_ids) >= 100: pstatuses = api.lookup_statuses(lookup_ids) for pstatus in pstatuses: if not is_valid_tweet(pstatus): continue reply = replies[pstatus.id] # is same user if pstatus.user.id == reply.user_id: continue intext = re.sub(r"@([A-Za-z0-9_]+)", "", pstatus.text) intext = normalize(intext) outtext = re.sub(r"@([A-Za-z0-9_]+)", "", reply.text) outtext = normalize(outtext) if not intext or not outtext: continue if f"{intext}\t{outtext}\n" in dup: continue if saved <= max_tw * .9: path = TRAINFILE elif saved <= max_tw * .95: path = DEVFILE else: path = TESTFILE with open(path, "a") as f: f.write(f"{intext}\t{outtext}\n") saved += 1 print(f"\r{got} => {filtered} => {saved}", end="") if saved > max_tw: exit() lookup_ids = [] replies = {} except Exception: print() limit_status = api.rate_limit_status( )["resources"]["search"]["/search/tweets"] while limit_status["reset"] >= int(time.time()): print("\rLimited: " + (" " + str(limit_status["reset"] - int(time.time())))[-3:] + "s", end="") time.sleep(.5) print() if __name__ == "__main__": main()
collect_tweet.py
4,746
is bot include URL is hashtag reply to multi user too long text = re.sub(r"\d+", "0", tmp) is not reply filter append lookup id collect 100 tweets is same user
159
en
0.917604
""" WSGI config for kubeoperator project. It exposes the WSGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/2.1/howto/deployment/wsgi/ """ import os from django.core.wsgi import get_wsgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'kubeoperator.settings') application = get_wsgi_application()
core/apps/kubeoperator/wsgi.py
401
WSGI config for kubeoperator project. It exposes the WSGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/2.1/howto/deployment/wsgi/
218
en
0.776478
"""The tests for the MQTT switch platform.""" import json from unittest.mock import ANY from asynctest import patch import pytest from homeassistant.components import mqtt, switch from homeassistant.components.mqtt.discovery import async_start from homeassistant.const import ( ATTR_ASSUMED_STATE, STATE_OFF, STATE_ON, STATE_UNAVAILABLE) import homeassistant.core as ha from homeassistant.setup import async_setup_component from tests.common import ( MockConfigEntry, async_fire_mqtt_message, async_mock_mqtt_component, mock_coro, mock_registry) from tests.components.switch import common @pytest.fixture def mock_publish(hass): """Initialize components.""" yield hass.loop.run_until_complete(async_mock_mqtt_component(hass)) async def test_controlling_state_via_topic(hass, mock_publish): """Test the controlling state via topic.""" assert await async_setup_component(hass, switch.DOMAIN, { switch.DOMAIN: { 'platform': 'mqtt', 'name': 'test', 'state_topic': 'state-topic', 'command_topic': 'command-topic', 'payload_on': 1, 'payload_off': 0 } }) state = hass.states.get('switch.test') assert STATE_OFF == state.state assert not state.attributes.get(ATTR_ASSUMED_STATE) async_fire_mqtt_message(hass, 'state-topic', '1') await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_ON == state.state async_fire_mqtt_message(hass, 'state-topic', '0') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_OFF == state.state async def test_sending_mqtt_commands_and_optimistic(hass, mock_publish): """Test the sending MQTT commands in optimistic mode.""" fake_state = ha.State('switch.test', 'on') with patch('homeassistant.helpers.restore_state.RestoreEntity' '.async_get_last_state', return_value=mock_coro(fake_state)): assert await async_setup_component(hass, switch.DOMAIN, { switch.DOMAIN: { 'platform': 'mqtt', 'name': 'test', 'command_topic': 'command-topic', 'payload_on': 'beer on', 'payload_off': 'beer off', 'qos': '2' } }) state = hass.states.get('switch.test') assert STATE_ON == state.state assert state.attributes.get(ATTR_ASSUMED_STATE) common.turn_on(hass, 'switch.test') await hass.async_block_till_done() mock_publish.async_publish.assert_called_once_with( 'command-topic', 'beer on', 2, False) mock_publish.async_publish.reset_mock() state = hass.states.get('switch.test') assert STATE_ON == state.state common.turn_off(hass, 'switch.test') await hass.async_block_till_done() await hass.async_block_till_done() mock_publish.async_publish.assert_called_once_with( 'command-topic', 'beer off', 2, False) state = hass.states.get('switch.test') assert STATE_OFF == state.state async def test_controlling_state_via_topic_and_json_message( hass, mock_publish): """Test the controlling state via topic and JSON message.""" assert await async_setup_component(hass, switch.DOMAIN, { switch.DOMAIN: { 'platform': 'mqtt', 'name': 'test', 'state_topic': 'state-topic', 'command_topic': 'command-topic', 'payload_on': 'beer on', 'payload_off': 'beer off', 'value_template': '{{ value_json.val }}' } }) state = hass.states.get('switch.test') assert STATE_OFF == state.state async_fire_mqtt_message(hass, 'state-topic', '{"val":"beer on"}') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_ON == state.state async_fire_mqtt_message(hass, 'state-topic', '{"val":"beer off"}') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_OFF == state.state async def test_default_availability_payload(hass, mock_publish): """Test the availability payload.""" assert await async_setup_component(hass, switch.DOMAIN, { switch.DOMAIN: { 'platform': 'mqtt', 'name': 'test', 'state_topic': 'state-topic', 'command_topic': 'command-topic', 'availability_topic': 'availability_topic', 'payload_on': 1, 'payload_off': 0 } }) state = hass.states.get('switch.test') assert STATE_UNAVAILABLE == state.state async_fire_mqtt_message(hass, 'availability_topic', 'online') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_OFF == state.state assert not state.attributes.get(ATTR_ASSUMED_STATE) async_fire_mqtt_message(hass, 'availability_topic', 'offline') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_UNAVAILABLE == state.state async_fire_mqtt_message(hass, 'state-topic', '1') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_UNAVAILABLE == state.state async_fire_mqtt_message(hass, 'availability_topic', 'online') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_ON == state.state async def test_custom_availability_payload(hass, mock_publish): """Test the availability payload.""" assert await async_setup_component(hass, switch.DOMAIN, { switch.DOMAIN: { 'platform': 'mqtt', 'name': 'test', 'state_topic': 'state-topic', 'command_topic': 'command-topic', 'availability_topic': 'availability_topic', 'payload_on': 1, 'payload_off': 0, 'payload_available': 'good', 'payload_not_available': 'nogood' } }) state = hass.states.get('switch.test') assert STATE_UNAVAILABLE == state.state async_fire_mqtt_message(hass, 'availability_topic', 'good') await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_OFF == state.state assert not state.attributes.get(ATTR_ASSUMED_STATE) async_fire_mqtt_message(hass, 'availability_topic', 'nogood') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_UNAVAILABLE == state.state async_fire_mqtt_message(hass, 'state-topic', '1') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_UNAVAILABLE == state.state async_fire_mqtt_message(hass, 'availability_topic', 'good') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_ON == state.state async def test_custom_state_payload(hass, mock_publish): """Test the state payload.""" assert await async_setup_component(hass, switch.DOMAIN, { switch.DOMAIN: { 'platform': 'mqtt', 'name': 'test', 'state_topic': 'state-topic', 'command_topic': 'command-topic', 'payload_on': 1, 'payload_off': 0, 'state_on': "HIGH", 'state_off': "LOW", } }) state = hass.states.get('switch.test') assert STATE_OFF == state.state assert not state.attributes.get(ATTR_ASSUMED_STATE) async_fire_mqtt_message(hass, 'state-topic', 'HIGH') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_ON == state.state async_fire_mqtt_message(hass, 'state-topic', 'LOW') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.test') assert STATE_OFF == state.state async def test_setting_attribute_via_mqtt_json_message(hass, mqtt_mock): """Test the setting of attribute via MQTT with JSON payload.""" assert await async_setup_component(hass, switch.DOMAIN, { switch.DOMAIN: { 'platform': 'mqtt', 'name': 'test', 'command_topic': 'test-topic', 'json_attributes_topic': 'attr-topic' } }) async_fire_mqtt_message(hass, 'attr-topic', '{ "val": "100" }') await hass.async_block_till_done() state = hass.states.get('switch.test') assert '100' == state.attributes.get('val') async def test_update_with_json_attrs_not_dict(hass, mqtt_mock, caplog): """Test attributes get extracted from a JSON result.""" assert await async_setup_component(hass, switch.DOMAIN, { switch.DOMAIN: { 'platform': 'mqtt', 'name': 'test', 'command_topic': 'test-topic', 'json_attributes_topic': 'attr-topic' } }) async_fire_mqtt_message(hass, 'attr-topic', '[ "list", "of", "things"]') await hass.async_block_till_done() state = hass.states.get('switch.test') assert state.attributes.get('val') is None assert 'JSON result was not a dictionary' in caplog.text async def test_update_with_json_attrs_bad_JSON(hass, mqtt_mock, caplog): """Test attributes get extracted from a JSON result.""" assert await async_setup_component(hass, switch.DOMAIN, { switch.DOMAIN: { 'platform': 'mqtt', 'name': 'test', 'command_topic': 'test-topic', 'json_attributes_topic': 'attr-topic' } }) async_fire_mqtt_message(hass, 'attr-topic', 'This is not JSON') await hass.async_block_till_done() state = hass.states.get('switch.test') assert state.attributes.get('val') is None assert 'Erroneous JSON: This is not JSON' in caplog.text async def test_discovery_update_attr(hass, mqtt_mock, caplog): """Test update of discovered MQTTAttributes.""" entry = MockConfigEntry(domain=mqtt.DOMAIN) await async_start(hass, 'homeassistant', {}, entry) data1 = ( '{ "name": "Beer",' ' "command_topic": "test_topic",' ' "json_attributes_topic": "attr-topic1" }' ) data2 = ( '{ "name": "Beer",' ' "command_topic": "test_topic",' ' "json_attributes_topic": "attr-topic2" }' ) async_fire_mqtt_message(hass, 'homeassistant/switch/bla/config', data1) await hass.async_block_till_done() async_fire_mqtt_message(hass, 'attr-topic1', '{ "val": "100" }') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.beer') assert '100' == state.attributes.get('val') # Change json_attributes_topic async_fire_mqtt_message(hass, 'homeassistant/switch/bla/config', data2) await hass.async_block_till_done() await hass.async_block_till_done() # Verify we are no longer subscribing to the old topic async_fire_mqtt_message(hass, 'attr-topic1', '{ "val": "50" }') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.beer') assert '100' == state.attributes.get('val') # Verify we are subscribing to the new topic async_fire_mqtt_message(hass, 'attr-topic2', '{ "val": "75" }') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.beer') assert '75' == state.attributes.get('val') async def test_unique_id(hass): """Test unique id option only creates one switch per unique_id.""" await async_mock_mqtt_component(hass) assert await async_setup_component(hass, switch.DOMAIN, { switch.DOMAIN: [{ 'platform': 'mqtt', 'name': 'Test 1', 'state_topic': 'test-topic', 'command_topic': 'command-topic', 'unique_id': 'TOTALLY_UNIQUE' }, { 'platform': 'mqtt', 'name': 'Test 2', 'state_topic': 'test-topic', 'command_topic': 'command-topic', 'unique_id': 'TOTALLY_UNIQUE' }] }) async_fire_mqtt_message(hass, 'test-topic', 'payload') await hass.async_block_till_done() await hass.async_block_till_done() assert len(hass.states.async_entity_ids()) == 2 # all switches group is 1, unique id created is 1 async def test_discovery_removal_switch(hass, mqtt_mock, caplog): """Test removal of discovered switch.""" entry = MockConfigEntry(domain=mqtt.DOMAIN) await async_start(hass, 'homeassistant', {}, entry) data = ( '{ "name": "Beer",' ' "state_topic": "test_topic",' ' "command_topic": "test_topic" }' ) async_fire_mqtt_message(hass, 'homeassistant/switch/bla/config', data) await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.beer') assert state is not None assert state.name == 'Beer' async_fire_mqtt_message(hass, 'homeassistant/switch/bla/config', '') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.beer') assert state is None async def test_discovery_update_switch(hass, mqtt_mock, caplog): """Test update of discovered switch.""" entry = MockConfigEntry(domain=mqtt.DOMAIN) await async_start(hass, 'homeassistant', {}, entry) data1 = ( '{ "name": "Beer",' ' "state_topic": "test_topic",' ' "command_topic": "test_topic" }' ) data2 = ( '{ "name": "Milk",' ' "state_topic": "test_topic",' ' "command_topic": "test_topic" }' ) async_fire_mqtt_message(hass, 'homeassistant/switch/bla/config', data1) await hass.async_block_till_done() state = hass.states.get('switch.beer') assert state is not None assert state.name == 'Beer' async_fire_mqtt_message(hass, 'homeassistant/switch/bla/config', data2) await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.beer') assert state is not None assert state.name == 'Milk' state = hass.states.get('switch.milk') assert state is None async def test_discovery_broken(hass, mqtt_mock, caplog): """Test handling of bad discovery message.""" entry = MockConfigEntry(domain=mqtt.DOMAIN) await async_start(hass, 'homeassistant', {}, entry) data1 = ( '{ "name": "Beer" }' ) data2 = ( '{ "name": "Milk",' ' "state_topic": "test_topic",' ' "command_topic": "test_topic" }' ) async_fire_mqtt_message(hass, 'homeassistant/switch/bla/config', data1) await hass.async_block_till_done() state = hass.states.get('switch.beer') assert state is None async_fire_mqtt_message(hass, 'homeassistant/switch/bla/config', data2) await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.milk') assert state is not None assert state.name == 'Milk' state = hass.states.get('switch.beer') assert state is None async def test_entity_device_info_with_identifier(hass, mqtt_mock): """Test MQTT switch device registry integration.""" entry = MockConfigEntry(domain=mqtt.DOMAIN) entry.add_to_hass(hass) await async_start(hass, 'homeassistant', {}, entry) registry = await hass.helpers.device_registry.async_get_registry() data = json.dumps({ 'platform': 'mqtt', 'name': 'Test 1', 'state_topic': 'test-topic', 'command_topic': 'test-command-topic', 'device': { 'identifiers': ['helloworld'], 'connections': [ ["mac", "02:5b:26:a8:dc:12"], ], 'manufacturer': 'Whatever', 'name': 'Beer', 'model': 'Glass', 'sw_version': '0.1-beta', }, 'unique_id': 'veryunique' }) async_fire_mqtt_message(hass, 'homeassistant/switch/bla/config', data) await hass.async_block_till_done() await hass.async_block_till_done() device = registry.async_get_device({('mqtt', 'helloworld')}, set()) assert device is not None assert device.identifiers == {('mqtt', 'helloworld')} assert device.connections == {('mac', "02:5b:26:a8:dc:12")} assert device.manufacturer == 'Whatever' assert device.name == 'Beer' assert device.model == 'Glass' assert device.sw_version == '0.1-beta' async def test_entity_device_info_update(hass, mqtt_mock): """Test device registry update.""" entry = MockConfigEntry(domain=mqtt.DOMAIN) entry.add_to_hass(hass) await async_start(hass, 'homeassistant', {}, entry) registry = await hass.helpers.device_registry.async_get_registry() config = { 'platform': 'mqtt', 'name': 'Test 1', 'state_topic': 'test-topic', 'command_topic': 'test-command-topic', 'device': { 'identifiers': ['helloworld'], 'connections': [ ["mac", "02:5b:26:a8:dc:12"], ], 'manufacturer': 'Whatever', 'name': 'Beer', 'model': 'Glass', 'sw_version': '0.1-beta', }, 'unique_id': 'veryunique' } data = json.dumps(config) async_fire_mqtt_message(hass, 'homeassistant/switch/bla/config', data) await hass.async_block_till_done() await hass.async_block_till_done() device = registry.async_get_device({('mqtt', 'helloworld')}, set()) assert device is not None assert device.name == 'Beer' config['device']['name'] = 'Milk' data = json.dumps(config) async_fire_mqtt_message(hass, 'homeassistant/switch/bla/config', data) await hass.async_block_till_done() await hass.async_block_till_done() device = registry.async_get_device({('mqtt', 'helloworld')}, set()) assert device is not None assert device.name == 'Milk' async def test_entity_id_update(hass, mqtt_mock): """Test MQTT subscriptions are managed when entity_id is updated.""" registry = mock_registry(hass, {}) mock_mqtt = await async_mock_mqtt_component(hass) assert await async_setup_component(hass, switch.DOMAIN, { switch.DOMAIN: [{ 'platform': 'mqtt', 'name': 'beer', 'state_topic': 'test-topic', 'command_topic': 'command-topic', 'availability_topic': 'avty-topic', 'unique_id': 'TOTALLY_UNIQUE' }] }) state = hass.states.get('switch.beer') assert state is not None assert mock_mqtt.async_subscribe.call_count == 2 mock_mqtt.async_subscribe.assert_any_call('test-topic', ANY, 0, 'utf-8') mock_mqtt.async_subscribe.assert_any_call('avty-topic', ANY, 0, 'utf-8') mock_mqtt.async_subscribe.reset_mock() registry.async_update_entity('switch.beer', new_entity_id='switch.milk') await hass.async_block_till_done() await hass.async_block_till_done() state = hass.states.get('switch.beer') assert state is None state = hass.states.get('switch.milk') assert state is not None assert mock_mqtt.async_subscribe.call_count == 2 mock_mqtt.async_subscribe.assert_any_call('test-topic', ANY, 0, 'utf-8') mock_mqtt.async_subscribe.assert_any_call('avty-topic', ANY, 0, 'utf-8')
tests/components/mqtt/test_switch.py
20,076
Initialize components. The tests for the MQTT switch platform. Change json_attributes_topic Verify we are no longer subscribing to the old topic Verify we are subscribing to the new topic all switches group is 1, unique id created is 1
237
en
0.816256
import torch from utils.distmat import compute_distmat def init_feedback_indices(q, g, device=None): return torch.zeros((q, g), dtype=torch.bool, device=device) def init_feedback_indices_qg(q, g, positive=False, device=None): indices = torch.zeros(q, q + g, dtype=torch.bool, device=device) if positive: indices[torch.arange(q), torch.arange(q)] = True return indices def greedy_feedback(distmat, q_pids, g_pids, positive_indices, negative_indices, inplace=True): """ Update positive_indices, negative_indices with one round of feedback. Provide feedback for top-ranked gallery. Note that distmat is corrupted if inplace=True. :param distmat: q x g Tensor (adjusted query to gallery) :param q_pids: q :param g_pids: g :param positive_indices: q x g :param negative_indices: q x g :return: (positive_indices, negative_indices, matches) """ q, g = tuple(distmat.shape) if not inplace: distmat = distmat.clone().detach() positive_indices = positive_indices.copy() negative_indices = negative_indices.copy() distmat[positive_indices] = float("inf") distmat[negative_indices] = float("inf") indices = distmat.argmin(dim=1) pmap = g_pids[indices] == q_pids positive_q = torch.arange(0, q)[pmap] negative_q = torch.arange(0, q)[pmap == False] positive_g = indices[pmap] negative_g = indices[pmap == False] existing = positive_indices[positive_q, positive_g] assert (not existing.any()) positive_indices[positive_q, positive_g] = True existing = negative_indices[negative_q, negative_g] assert (not existing.any()) negative_indices[negative_q, negative_g] = True return positive_indices, negative_indices, pmap def naive_round(qf, gf, q_pids, g_pids, positive_indices=None, negative_indices=None, inplace=True, previous_distmat=None, device=None): """ qf: q x m gf: g x m q_pids: q g_pids: g positive_indices: q x g negative_indices: q x g previous_distmat: adjusted distmat (== compute_distmat(qf, gf) only at init) """ q, g = qf.shape[0], gf.shape[0] assert (qf.shape[1] == gf.shape[1]) if positive_indices is None: positive_indices = init_feedback_indices(q, g, device=device) if negative_indices is None: negative_indices = init_feedback_indices(q, g, device=device) if previous_distmat is None: distmat = compute_distmat(qf, gf) else: distmat = previous_distmat res = greedy_feedback(distmat, q_pids, g_pids, positive_indices, negative_indices, inplace=inplace) positive_indices, negative_indices, matches = res distmat = compute_distmat(qf, gf) distmat[positive_indices] = 0 distmat[negative_indices] = float("inf") return distmat, positive_indices, negative_indices, matches
hitl/feedback.py
2,870
Update positive_indices, negative_indices with one round of feedback. Provide feedback for top-ranked gallery. Note that distmat is corrupted if inplace=True. :param distmat: q x g Tensor (adjusted query to gallery) :param q_pids: q :param g_pids: g :param positive_indices: q x g :param negative_indices: q x g :return: (positive_indices, negative_indices, matches) qf: q x m gf: g x m q_pids: q g_pids: g positive_indices: q x g negative_indices: q x g previous_distmat: adjusted distmat (== compute_distmat(qf, gf) only at init)
532
en
0.651046
#Licenced under MIT License #charset = "utf-8" #Language = "Python3" #Bot Framework = "python-telegram-bot" #The Code is without Proxy, Actual code contains Proxy #Proxy should be used is of the type SOCKS5 #Special thanks to cyberboySumanjay #The bot will work till you press ctrl+c in the terminal or command line., #import the required files import requests import logging from telegram import * from telegram.ext import * #enable logger (optional) logging.basicConfig( format='%(asctime)s - %(name)s - %(levelname)s - %(message)s', level=logging.INFO ) logger = logging.getLogger(__name__) TOKEN = "Bot Token Here" #CommandHandler for message "Start" def start(update: Update, context: CallbackContext) -> None: update.message.reply_text(f"""*Hi {update.effective_chat.first_name},* Welcome to the Torrent Searcher Bot. Here you will find all the torrents you search for :) Type /help to know how to use the bot Type /info to know about the developer""", parse_mode=ParseMode.MARKDOWN) #CommandHandler for message "Help" def help(update: Update, context: CallbackContext) -> None: update.message.reply_text("""Send me the query you want to search and i will do the rest! If any error occurs, feel free to pm me on https://t.me/SL_PUNSITH1""", parse_mode=ParseMode.MARKDOWN) #CommandHandler to get torrents for the query def find(update: Update, context: CallbackContext) -> None: try: update.message.reply_text("Searching results for 👉{}👈".format(update.message.text)) #1337x, torrent9 & eztv api url = "https://src.abirxo.com/index.php?name={}&x1337=true&x1337pages=1".format(update.message.text) results = requests.get(url).json() print(results) for item in results: link = item.get('link') name = item.get('name') pic = item.get('picture') update.message.reply_text(f"""*➲Name:* `{name}` *➲Link:* `{link}`""", parse_mode=ParseMode.MARKDOWN) update.message.reply_text("End of Results") except: update.message.reply_text("""Search Completed""") #CommandHnadler for message "info" def info(update: Update, context: CallbackContext) -> None: #Never Mind :-) update.message.reply_text("""Bot by @unkusr""", parse_mode=ParseMode.MARKDOWN) #Add all handlers to the main function. def main() -> None: updater = Updater(TOKEN, use_context=True) dispatcher = updater.dispatcher dispatcher.add_handler(CommandHandler("start", start)) dispatcher.add_handler(CommandHandler("help", help)) dispatcher.add_handler(CommandHandler("info", info)) dispatcher.add_handler(MessageHandler(Filters.text & (~Filters.command), find)) updater.start_polling() #set bot to polling, if you use webhooks, replace this statement with the url of webhook., updater.idle() #Call the main function if __name__ == '__main__': main()
mirror.py
2,906
Licenced under MIT Licensecharset = "utf-8"Language = "Python3"Bot Framework = "python-telegram-bot"The Code is without Proxy, Actual code contains ProxyProxy should be used is of the type SOCKS5Special thanks to cyberboySumanjayThe bot will work till you press ctrl+c in the terminal or command line.,import the required filesenable logger (optional)CommandHandler for message "Start"CommandHandler for message "Help"CommandHandler to get torrents for the query1337x, torrent9 & eztv apiCommandHnadler for message "info"Never Mind :-)Add all handlers to the main function.set bot to polling, if you use webhooks, replace this statement with the url of webhook.,Call the main function
684
en
0.529391
# -*- coding: utf-8 -*- # This file is part of the Ingram Micro Cloud Blue Connect connect-cli. # Copyright (c) 2021 Ingram Micro. All Rights Reserved. import os import sys def unimport(): for m in ('connect.cli.plugins.play.commands', 'connect.cli.ccli'): if m in sys.modules: del sys.modules[m] def test_play_commands(fs, mocker): os.environ['CCLI_SCRIPTS'] = os.path.join(os.path.dirname(__file__), 'scripts') unimport() from connect.cli.ccli import main mocker.patch('connect.cli.plugins.play.commands.PlayOptions.context_file', None) mocker.patch('sys.argv', ['cmd', 'play', 'script1']) main() def test_play_commands_rel(fs, mocker): os.environ['CCLI_SCRIPTS'] = 'tests/plugins/play/scripts' unimport() from connect.cli.ccli import main mocker.patch('connect.cli.plugins.play.commands.PlayOptions.context_file', None) mocker.patch('sys.argv', ['cmd', 'play', 'script1']) main()
tests/plugins/play/test_play_commands.py
967
-*- coding: utf-8 -*- This file is part of the Ingram Micro Cloud Blue Connect connect-cli. Copyright (c) 2021 Ingram Micro. All Rights Reserved.
145
en
0.824047
import json from urllib import urlencode from corehq.apps.registration.utils import create_30_day_trial from dimagi.utils.couch import CriticalSection from dimagi.utils.couch.resource_conflict import retry_resource from django.contrib.auth.decorators import login_required from django.core.urlresolvers import reverse from django.http import Http404, HttpResponseRedirect, HttpResponse from django.template.loader import render_to_string from django.shortcuts import render from django.contrib import messages from dimagi.utils.name_to_url import name_to_url from django.utils.translation import ugettext as _, ugettext_lazy from corehq.apps.app_manager.views.apps import clear_app_cache from corehq.apps.domain.decorators import require_superuser from corehq.apps.domain.exceptions import NameUnavailableException from corehq.elastic import es_query, parse_args_for_es, fill_mapping_with_facets from corehq.apps.domain.models import Domain from dimagi.utils.couch.database import apply_update from corehq.apps.fixtures.models import FixtureDataType SNAPSHOT_FACETS = ['project_type', 'license', 'author.exact', 'is_starter_app'] DEPLOYMENT_FACETS = ['deployment.region'] SNAPSHOT_MAPPING = [ ("", True, [ {"facet": "project_type", "name": ugettext_lazy("Category"), "expanded": True}, { "facet": "license", "name": ugettext_lazy("License"), "expanded": True, "mapping": { 'cc': 'CC BY', 'cc-sa': 'CC BY-SA', 'cc-nd': 'CC BY-ND', 'cc-nc': 'CC BY-NC', 'cc-nc-sa': 'CC BY-NC-SA', 'cc-nc-nd': 'CC BY-NC-ND', } }, {"facet": "author.exact", "name": ugettext_lazy("Author"), "expanded": True}, ]), ] DEPLOYMENT_MAPPING = [ ("", True, [ {"facet": "deployment.region", "name": "Region", "expanded": True}, ]), ] def rewrite_url(request, path): return HttpResponseRedirect('/exchange%s?%s' % (path, request.META['QUERY_STRING'])) def inverse_dict(d): return dict([(v, k) for k, v in d.iteritems()]) def can_view_app(req, dom): if not dom or not dom.is_snapshot or not dom.published: return False if not dom.is_approved and ( not getattr(req, "couch_user", "") or not req.couch_user.is_domain_admin(dom.copied_from.name) ): return False return True def project_info(request, domain, template="appstore/project_info.html"): dom = Domain.get(domain) if not can_view_app(request, dom): raise Http404() copies = dom.copies_of_parent() images = set() audio = set() return render(request, template, { "project": dom, "applications": dom.full_applications(include_builds=False), "fixtures": FixtureDataType.by_domain(dom.name), "copies": copies, "images": images, "audio": audio, "url_base": reverse('appstore'), 'display_import': True if getattr(request, "couch_user", "") and request.couch_user.get_domains() else False }) def deduplicate(hits): unique_names = set() unique_hits = [] for hit in hits: if not hit['_source']['name'] in unique_names: unique_hits.append(hit) unique_names.add(hit['_source']['name']) return unique_hits def appstore(request, template="appstore/appstore_base.html"): page_length = 10 include_unapproved = True if request.GET.get('is_approved', "") == "false" else False if include_unapproved and not request.user.is_superuser: raise Http404() params, _ = parse_args_for_es(request) page = params.pop('page', 1) page = int(page[0] if isinstance(page, list) else page) results = es_snapshot_query(params, SNAPSHOT_FACETS) hits = results.get('hits', {}).get('hits', []) hits = deduplicate(hits) d_results = [Domain.wrap(res['_source']) for res in hits] starter_apps = request.GET.get('is_starter_app', None) sort_by = request.GET.get('sort_by', None) if sort_by == 'newest': pass else: d_results = Domain.hit_sort(d_results) persistent_params = {} if sort_by: persistent_params["sort_by"] = sort_by if include_unapproved: persistent_params["is_approved"] = "false" persistent_params = urlencode(persistent_params) # json.dumps(persistent_params) more_pages = False if len(d_results) <= page * page_length else True facet_map = fill_mapping_with_facets(SNAPSHOT_MAPPING, results, params) vals = dict( apps=d_results[(page - 1) * page_length:page * page_length], page=page, prev_page=(page - 1), next_page=(page + 1), more_pages=more_pages, sort_by=sort_by, show_starter_apps=starter_apps, include_unapproved=include_unapproved, facet_map=facet_map, facets=results.get("facets", []), query_str=request.META['QUERY_STRING'], search_query=params.get('search', [""])[0], persistent_params=persistent_params, ) return render(request, template, vals) def appstore_api(request): params, facets = parse_args_for_es(request) results = es_snapshot_query(params, facets) return HttpResponse(json.dumps(results), content_type="application/json") def es_snapshot_query(params, facets=None, terms=None, sort_by="snapshot_time"): if terms is None: terms = ['is_approved', 'sort_by', 'search'] if facets is None: facets = [] q = {"sort": {sort_by: {"order": "desc"}}, "query": {"bool": {"must": [ {"match": {'doc_type': "Domain"}}, {"term": {"published": True}}, {"term": {"is_snapshot": True}} ]}}, "filter": {"and": [{"term": {"is_approved": params.get('is_approved', None) or True}}]}} search_query = params.get('search', "") if search_query: q['query']['bool']['must'].append({ "match": { "_all": { "query": search_query, "operator": "and" } } }) return es_query(params, facets, terms, q) @require_superuser def approve_app(request, domain): domain = Domain.get(domain) if request.GET.get('approve') == 'true': domain.is_approved = True domain.save() elif request.GET.get('approve') == 'false': domain.is_approved = False domain.save() return HttpResponseRedirect(request.META.get('HTTP_REFERER') or reverse('appstore')) @login_required @retry_resource(3) def import_app(request, domain): user = request.couch_user if not user.is_eula_signed(): messages.error(request, 'You must agree to our eula to download an app') return project_info(request, domain) from_project = Domain.get(domain) if request.method == 'POST' and from_project.is_snapshot: if not from_project.published: messages.error(request, "This project is not published and can't be downloaded") return project_info(request, domain) to_project_name = request.POST['project'] if not user.is_member_of(to_project_name): messages.error(request, _("You don't belong to that project")) return project_info(request, domain) full_apps = from_project.full_applications(include_builds=False) assert full_apps, 'Bad attempt to copy apps from a project without any!' for app in full_apps: new_doc = from_project.copy_component(app['doc_type'], app.get_id, to_project_name, user) clear_app_cache(request, to_project_name) from_project.downloads += 1 from_project.save() messages.success(request, render_to_string("appstore/partials/view_wiki.html", {"pre": _("Application successfully imported!")}), extra_tags="html") return HttpResponseRedirect(reverse('view_app', args=[to_project_name, new_doc.id])) else: return HttpResponseRedirect(reverse('project_info', args=[domain])) @login_required def copy_snapshot(request, domain): user = request.couch_user if not user.is_eula_signed(): messages.error(request, 'You must agree to our eula to download an app') return project_info(request, domain) dom = Domain.get(domain) if request.method == "POST" and dom.is_snapshot: assert dom.full_applications(include_builds=False), 'Bad attempt to copy project without any apps!' from corehq.apps.registration.forms import DomainRegistrationForm args = { 'domain_name': request.POST['new_project_name'], 'hr_name': request.POST['new_project_name'], 'eula_confirmed': True, } form = DomainRegistrationForm(args) if request.POST.get('new_project_name', ""): if not dom.published: messages.error(request, _("This project is not published and can't be downloaded")) return project_info(request, domain) if not form.is_valid(): messages.error(request, form.errors) return project_info(request, domain) new_domain_name = name_to_url(form.cleaned_data['hr_name'], "project") with CriticalSection(['copy_domain_snapshot_{}_to_{}'.format(dom.name, new_domain_name)]): try: new_domain = dom.save_copy(new_domain_name, new_hr_name=form.cleaned_data['hr_name'], user=user) except NameUnavailableException: messages.error(request, _("A project by that name already exists")) return project_info(request, domain) # sign new project up for trial create_30_day_trial(new_domain) def inc_downloads(d): d.downloads += 1 apply_update(dom, inc_downloads) messages.success(request, render_to_string("appstore/partials/view_wiki.html", {"pre": _("Project copied successfully!")}), extra_tags="html") return HttpResponseRedirect(reverse('view_app', args=[new_domain.name, new_domain.full_applications()[0].get_id])) else: messages.error(request, _("You must specify a name for the new project")) return project_info(request, domain) else: return HttpResponseRedirect(reverse('project_info', args=[domain])) def project_image(request, domain): project = Domain.get(domain) if project.image_path: image = project.fetch_attachment(project.image_path) return HttpResponse(image, content_type=project.image_type) else: raise Http404() def project_documentation_file(request, domain): project = Domain.get(domain) if project.documentation_file_path: documentation_file = project.fetch_attachment(project.documentation_file_path) return HttpResponse(documentation_file, content_type=project.documentation_file_type) else: raise Http404() @login_required def deployment_info(request, domain, template="appstore/deployment_info.html"): dom = Domain.get_by_name(domain) if not dom or not dom.deployment.public: raise Http404() # get facets results = es_deployments_query({}, DEPLOYMENT_FACETS) facet_map = fill_mapping_with_facets(DEPLOYMENT_MAPPING, results, {}) return render(request, template, { 'domain': dom, 'search_url': reverse('deployments'), 'url_base': reverse('deployments'), 'facet_map': facet_map, }) @login_required def deployments(request, template="appstore/deployments.html"): params, _ = parse_args_for_es(request) params = dict([(DEPLOYMENT_MAPPING.get(p, p), params[p]) for p in params]) page = int(params.pop('page', 1)) results = es_deployments_query(params, DEPLOYMENT_FACETS) d_results = [Domain.wrap(res['_source']) for res in results['hits']['hits']] more_pages = False if len(d_results) <= page * 10 else True facet_map = fill_mapping_with_facets(DEPLOYMENT_MAPPING, results, params) include_unapproved = True if request.GET.get('is_approved', "") == "false" else False vals = {'deployments': d_results[(page - 1) * 10:page * 10], 'page': page, 'prev_page': page - 1, 'next_page': (page + 1), 'more_pages': more_pages, 'include_unapproved': include_unapproved, 'facet_map': facet_map, 'query_str': request.META['QUERY_STRING'], 'search_url': reverse('deployments'), 'search_query': params.get('search', [""])[0]} return render(request, template, vals) def deployments_api(request): params, facets = parse_args_for_es(request) params = dict([(DEPLOYMENT_MAPPING.get(p, p), params[p]) for p in params]) results = es_deployments_query(params, facets) return HttpResponse(json.dumps(results), content_type="application/json") def es_deployments_query(params, facets=None, terms=None, sort_by="snapshot_time"): if terms is None: terms = ['is_approved', 'sort_by', 'search'] if facets is None: facets = [] q = {"query": {"bool": {"must": [{"match": {'doc_type': "Domain"}}, {"term": {"deployment.public": True}}]}}} search_query = params.get('search', "") if search_query: q['query']['bool']['must'].append({ "match": { "_all": { "query": search_query, "operator": "and" } } }) return es_query(params, facets, terms, q) def media_files(request, domain, template="appstore/media_files.html"): dom = Domain.get(domain) if not can_view_app(request, dom): raise Http404() return render(request, template, { "project": dom, "url_base": reverse('appstore') })
corehq/apps/appstore/views.py
14,299
json.dumps(persistent_params) sign new project up for trial get facets
70
en
0.595297
#! /usr/bin/env python # coding=utf-8 # Copyright (c) 2019 Uber Technologies, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== import logging import os import sys from functools import partial from multiprocessing import Pool from typing import Union import numpy as np import tensorflow as tf from ludwig.constants import * from ludwig.encoders.image_encoders import ENCODER_REGISTRY from ludwig.features.base_feature import InputFeature from ludwig.utils.data_utils import get_abs_path from ludwig.utils.fs_utils import upload_h5 from ludwig.utils.image_utils import greyscale from ludwig.utils.image_utils import num_channels_in_image from ludwig.utils.image_utils import resize_image from ludwig.utils.image_utils import get_image_from_path, read_image from ludwig.utils.misc_utils import set_default_value logger = logging.getLogger(__name__) image_scaling_registry = { 'pixel_normalization': lambda x: x * 1.0 / 255, 'pixel_standardization': lambda x: tf.map_fn( lambda f: tf.image.per_image_standardization(f), x) } class ImageFeatureMixin: type = IMAGE preprocessing_defaults = { 'missing_value_strategy': BACKFILL, 'in_memory': True, 'resize_method': 'interpolate', 'scaling': 'pixel_normalization', 'num_processes': 1, 'infer_image_dimensions': False, 'infer_image_max_height': 256, 'infer_image_max_width': 256, 'infer_image_sample_size': 100 } preprocessing_schema = { 'missing_value_strategy': {'type': 'string', 'enum': MISSING_VALUE_STRATEGY_OPTIONS}, 'in_memory': {'type': 'boolean'}, 'resize_method': {'type': 'string', 'enum': RESIZE_METHODS}, 'scaling': {'type': 'string', 'enum': list(image_scaling_registry.keys())}, 'num_processes': {'type': 'integer', 'minimum': 0}, 'height': {'type': 'integer', 'minimum': 0}, 'width': {'type': 'integer', 'minimum': 0}, 'num_channels': {'type': 'integer', 'minimum': 0}, 'infer_image_dimensions': {'type': 'boolean'}, 'infer_image_max_height': {'type': 'integer', 'minimum': 0}, 'infer_image_max_width': {'type': 'integer', 'minimum': 0}, 'infer_image_sample_size': {'type': 'integer', 'minimum': 0} } @staticmethod def cast_column(column, backend): return column @staticmethod def get_feature_meta(column, preprocessing_parameters, backend): return { PREPROCESSING: preprocessing_parameters } @staticmethod def _read_image_and_resize( img_entry: Union[str, 'numpy.array'], img_width: int, img_height: int, should_resize: bool, num_channels: int, resize_method: str, user_specified_num_channels: int ): """ :param img_entry Union[str, 'numpy.array']: if str file path to the image else numpy.array of the image itself :param img_width: expected width of the image :param img_height: expected height of the image :param should_resize: Should the image be resized? :param resize_method: type of resizing method :param num_channels: expected number of channels in the first image :param user_specified_num_channels: did the user specify num channels? :return: image object Helper method to read and resize an image according to model defn. If the user doesn't specify a number of channels, we use the first image in the dataset as the source of truth. If any image in the dataset doesn't have the same number of channels as the first image, raise an exception. If the user specifies a number of channels, we try to convert all the images to the specifications by dropping channels/padding 0 channels """ img = read_image(img_entry) img_num_channels = num_channels_in_image(img) if img_num_channels == 1: img = img.reshape((img.shape[0], img.shape[1], 1)) if should_resize: img = resize_image(img, (img_height, img_width), resize_method) if user_specified_num_channels is True: # convert to greyscale if needed if num_channels == 1 and ( img_num_channels == 3 or img_num_channels == 4): img = greyscale(img) img_num_channels = 1 # Number of channels is specified by the user img_padded = np.zeros((img_height, img_width, num_channels), dtype=np.uint8) min_num_channels = min(num_channels, img_num_channels) img_padded[:, :, :min_num_channels] = img[:, :, :min_num_channels] img = img_padded if img_num_channels != num_channels: logger.warning( "Image has {0} channels, where as {1} " "channels are expected. Dropping/adding channels " "with 0s as appropriate".format( img_num_channels, num_channels)) else: # If the image isn't like the first image, raise exception if img_num_channels != num_channels: raise ValueError( 'Image has {0} channels, unlike the first image, which ' 'has {1} channels. Make sure all the images have the same ' 'number of channels or use the num_channels property in ' 'image preprocessing'.format(img_num_channels, num_channels)) if img.shape[0] != img_height or img.shape[1] != img_width: raise ValueError( "Images are not of the same size. " "Expected size is {0}, " "current image size is {1}." "Images are expected to be all of the same size " "or explicit image width and height are expected " "to be provided. " "Additional information: " "https://ludwig-ai.github.io/ludwig-docs/user_guide/#image-features-preprocessing" .format([img_height, img_width, num_channels], img.shape) ) return img @staticmethod def _finalize_preprocessing_parameters( preprocessing_parameters: dict, first_img_entry: Union[str, 'numpy.array'], src_path: str, input_feature_col: np.array ): """ Helper method to determine the height, width and number of channels for preprocessing the image data. This is achieved by looking at the parameters provided by the user. When there are some missing parameters, we fall back on to the first image in the dataset. The assumption being that all the images in the data are expected be of the same size with the same number of channels """ first_image = read_image(first_img_entry) first_img_height = first_image.shape[0] first_img_width = first_image.shape[1] first_img_num_channels = num_channels_in_image(first_image) should_resize = False if (HEIGHT in preprocessing_parameters or WIDTH in preprocessing_parameters): should_resize = True try: height = int(preprocessing_parameters[HEIGHT]) width = int(preprocessing_parameters[WIDTH]) except ValueError as e: raise ValueError( 'Image height and width must be set and have ' 'positive integer values: ' + str(e) ) if height <= 0 or width <= 0: raise ValueError( 'Image height and width must be positive integers' ) else: # User hasn't specified height and width. # Default to first image, or infer from sample. height, width = first_img_height, first_img_width if preprocessing_parameters[INFER_IMAGE_DIMENSIONS]: should_resize = True sample_size = min(len(input_feature_col), preprocessing_parameters[INFER_IMAGE_SAMPLE_SIZE]) sample_images = [read_image(get_image_from_path(src_path, img)) for img in input_feature_col[:sample_size]] if sample_images: height_avg = min( sum(x.shape[0] for x in sample_images) / len(sample_images), preprocessing_parameters[INFER_IMAGE_MAX_HEIGHT]) width_avg = min( sum(x.shape[1] for x in sample_images) / len(sample_images), preprocessing_parameters[INFER_IMAGE_MAX_WIDTH]) height, width = round(height_avg), round(width_avg) logger.debug("Inferring height: {0} and width: {1}".format(height, width)) else: logger.warning("Sample set for inference is empty, default to height and width of first image") if NUM_CHANNELS in preprocessing_parameters: # User specified num_channels in the model/feature config user_specified_num_channels = True num_channels = preprocessing_parameters[NUM_CHANNELS] else: user_specified_num_channels = False num_channels = first_img_num_channels assert isinstance(num_channels, int), ValueError( 'Number of image channels needs to be an integer' ) return ( should_resize, width, height, num_channels, user_specified_num_channels, first_image ) @staticmethod def add_feature_data( feature, input_df, proc_df, metadata, preprocessing_parameters, backend, skip_save_processed_input ): in_memory = preprocessing_parameters['in_memory'] if PREPROCESSING in feature and 'in_memory' in feature[PREPROCESSING]: in_memory = feature[PREPROCESSING]['in_memory'] num_processes = preprocessing_parameters['num_processes'] if PREPROCESSING in feature and 'num_processes' in feature[ PREPROCESSING]: num_processes = feature[PREPROCESSING]['num_processes'] src_path = None if SRC in metadata: src_path = os.path.dirname(os.path.abspath(metadata.get(SRC))) num_images = len(input_df[feature[COLUMN]]) if num_images == 0: raise ValueError('There are no images in the dataset provided.') first_img_entry = next(iter(input_df[feature[COLUMN]])) logger.debug( 'Detected image feature type is {}'.format(type(first_img_entry)) ) if not isinstance(first_img_entry, str) \ and not isinstance(first_img_entry, np.ndarray): raise ValueError( 'Invalid image feature data type. Detected type is {}, ' 'expect either string for file path or numpy array.' .format(type(first_img_entry)) ) first_img_entry = get_image_from_path(src_path, first_img_entry) ( should_resize, width, height, num_channels, user_specified_num_channels, first_image ) = ImageFeatureMixin._finalize_preprocessing_parameters( preprocessing_parameters, first_img_entry, src_path, input_df[feature[COLUMN]] ) metadata[feature[NAME]][PREPROCESSING]['height'] = height metadata[feature[NAME]][PREPROCESSING]['width'] = width metadata[feature[NAME]][PREPROCESSING][ 'num_channels'] = num_channels read_image_and_resize = partial( ImageFeatureMixin._read_image_and_resize, img_width=width, img_height=height, should_resize=should_resize, num_channels=num_channels, resize_method=preprocessing_parameters['resize_method'], user_specified_num_channels=user_specified_num_channels ) # check to see if the active backend can support lazy loading of # image features from the hdf5 cache. backend.check_lazy_load_supported(feature) if in_memory or skip_save_processed_input: # Number of processes to run in parallel for preprocessing metadata[feature[NAME]][PREPROCESSING][ 'num_processes'] = num_processes metadata[feature[NAME]]['reshape'] = (height, width, num_channels) # Split the dataset into pools only if we have an explicit request to use # multiple processes. In case we have multiple input images use the # standard code anyway. if backend.supports_multiprocessing and ( num_processes > 1 or num_images > 1): all_img_entries = [get_abs_path(src_path, img_entry) if isinstance(img_entry, str) else img_entry for img_entry in input_df[feature[COLUMN]]] with Pool(num_processes) as pool: logger.debug( 'Using {} processes for preprocessing images'.format( num_processes ) ) proc_df[feature[PROC_COLUMN]] = pool.map( read_image_and_resize, all_img_entries ) else: # If we're not running multiple processes and we are only processing one # image just use this faster shortcut, bypassing multiprocessing.Pool.map logger.debug( 'No process pool initialized. Using internal process for preprocessing images' ) # helper function for handling single image def _get_processed_image(img_store): if isinstance(img_store, str): return read_image_and_resize( get_abs_path(src_path, img_store) ) else: return read_image_and_resize(img_store) proc_df[feature[PROC_COLUMN]] = backend.df_engine.map_objects( input_df[feature[COLUMN]], _get_processed_image ) else: all_img_entries = [get_abs_path(src_path, img_entry) if isinstance(img_entry, str) else img_entry for img_entry in input_df[feature[COLUMN]]] data_fp = backend.cache.get_cache_path( metadata.get(SRC), metadata.get(CHECKSUM), TRAINING ) with upload_h5(data_fp) as h5_file: # todo future add multiprocessing/multithreading image_dataset = h5_file.create_dataset( feature[PROC_COLUMN] + '_data', (num_images, height, width, num_channels), dtype=np.uint8 ) for i, img_entry in enumerate(all_img_entries): image_dataset[i, :height, :width, :] = ( read_image_and_resize(img_entry) ) h5_file.flush() proc_df[feature[PROC_COLUMN]] = np.arange(num_images) return proc_df class ImageInputFeature(ImageFeatureMixin, InputFeature): height = 0 width = 0 num_channels = 0 scaling = 'pixel_normalization' encoder = 'stacked_cnn' def __init__(self, feature, encoder_obj=None): super().__init__(feature) self.overwrite_defaults(feature) if encoder_obj: self.encoder_obj = encoder_obj else: self.encoder_obj = self.initialize_encoder(feature) def call(self, inputs, training=None, mask=None): assert isinstance(inputs, tf.Tensor) assert inputs.dtype in [tf.uint8, tf.int64] # casting and rescaling inputs = tf.cast(inputs, tf.float32) / 255 inputs_encoded = self.encoder_obj( inputs, training=training, mask=mask ) return inputs_encoded @classmethod def get_input_dtype(cls): return tf.uint8 def get_input_shape(self): return self.height, self.width, self.num_channels @staticmethod def update_config_with_metadata( input_feature, feature_metadata, *args, **kwargs ): for key in ['height', 'width', 'num_channels', 'scaling']: input_feature[key] = feature_metadata[PREPROCESSING][key] @staticmethod def populate_defaults(input_feature): set_default_value(input_feature, TIED, None) set_default_value(input_feature, PREPROCESSING, {}) encoder_registry = ENCODER_REGISTRY
ludwig/features/image_feature.py
17,799
Helper method to determine the height, width and number of channels for preprocessing the image data. This is achieved by looking at the parameters provided by the user. When there are some missing parameters, we fall back on to the first image in the dataset. The assumption being that all the images in the data are expected be of the same size with the same number of channels :param img_entry Union[str, 'numpy.array']: if str file path to the image else numpy.array of the image itself :param img_width: expected width of the image :param img_height: expected height of the image :param should_resize: Should the image be resized? :param resize_method: type of resizing method :param num_channels: expected number of channels in the first image :param user_specified_num_channels: did the user specify num channels? :return: image object Helper method to read and resize an image according to model defn. If the user doesn't specify a number of channels, we use the first image in the dataset as the source of truth. If any image in the dataset doesn't have the same number of channels as the first image, raise an exception. If the user specifies a number of channels, we try to convert all the images to the specifications by dropping channels/padding 0 channels ! /usr/bin/env python coding=utf-8 Copyright (c) 2019 Uber Technologies, Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ============================================================================== convert to greyscale if needed Number of channels is specified by the user If the image isn't like the first image, raise exception User hasn't specified height and width. Default to first image, or infer from sample. User specified num_channels in the model/feature config check to see if the active backend can support lazy loading of image features from the hdf5 cache. Number of processes to run in parallel for preprocessing Split the dataset into pools only if we have an explicit request to use multiple processes. In case we have multiple input images use the standard code anyway. If we're not running multiple processes and we are only processing one image just use this faster shortcut, bypassing multiprocessing.Pool.map helper function for handling single image todo future add multiprocessing/multithreading casting and rescaling
2,803
en
0.839745
#!/usr/bin/env python # -*- coding: utf-8 -*- # Author: David McCue import sqlite3, re, os from bottle import route, run, debug, template, request, static_file, error, response db_filename=os.path.dirname(os.path.realpath(__file__)) + '/db/kickstarter.db' # Validate currency values def valid_amount(amount): amount = re.sub(r"\$","",amount) amount_regex=re.compile('^[0-9]+(\.[0-9]{2})?$') if amount_regex.match(amount) and float(amount) > 0.0: return True # Validate names def valid_name(name): name_regex=re.compile('^[A-Za-z0-9-_]{4,20}$') if name_regex.match(name): return True # Calculate luhn checksum # Credit: wikipedia def luhn_checksum(card_number): def digits_of(n): return [int(d) for d in str(n)] digits = digits_of(card_number) odd_digits = digits[-1::-2] even_digits = digits[-2::-2] checksum = 0 checksum += sum(odd_digits) for d in even_digits: checksum += sum(digits_of(d*2)) return checksum % 10 # Validate credit card def valid_creditcard(creditcard): creditcard_regex=re.compile('^[0-9]{1,19}$') if creditcard_regex.match(creditcard): if luhn_checksum(creditcard) == 0: return True # retrieve project by name @route('/project/<project>') def project_list(project): conn = sqlite3.connect(db_filename) c = conn.cursor() c.execute("SELECT id, name, target FROM 'project' WHERE name = '" + project + "';") result = c.fetchall() c.close() if result: c = conn.cursor() c.execute("SELECT name, amount FROM 'transaction' WHERE projectid = " + str(result[0][0]) + ";") backers = c.fetchall() c.close() return {'project': result[0], 'backers': backers} else: response.status = 400 return {'msg':'ERROR: Project name not found'} # add project @route('/project', method='POST') def project_new(): project = request.json print "DEBUG: " + str(project) if not valid_name(project['name']): response.status = 400 return {'msg':'ERROR: Project name validation error'} if not valid_amount(project['target']): response.status = 400 return {'msg':'ERROR: Project target validation error'} conn = sqlite3.connect(db_filename) c = conn.cursor() c.execute("SELECT name FROM project WHERE name = '" + project['name'] + "';") result = c.fetchall() c.close() if result: response.status = 400 return {'msg':'ERROR: Project name already exists'} conn = sqlite3.connect(db_filename) conn.execute("INSERT INTO 'project' (name, target) VALUES ('" + project['name'] + "', '" + project['target'] + "');") conn.commit() conn.close() # add transaction @route('/back', method='POST') def back_new(): back = request.json # Validate inputs if not valid_name(back['name']): response.status = 412 return {'msg':'ERROR: Backer name validation error'} if not valid_amount(back['amount']): response.status = 412 return {'msg':'ERROR: Backer amount validation error'} if not valid_creditcard(back['cc']): response.status = 400 return {'msg':'ERROR: This card is invalid'} # Check credit card is unique to this user name conn = sqlite3.connect(db_filename) c = conn.cursor() c.execute("SELECT name FROM 'transaction' WHERE cc = '" + back['cc'] + "' AND name != '" + back['name'] + "';") result = c.fetchall() c.close() if result: response.status = 400 return {'msg':'ERROR: That card has already been added by another user!'} # Get project id from name c = conn.cursor() c.execute("SELECT id FROM 'project' WHERE name = '" + back['projectname'] + "';") result = c.fetchall() c.close() if not result: response.status = 400 return {'msg':'ERROR: Unable to find project name'} back['projectid'] = result[0][0] conn.execute("INSERT INTO 'transaction' (projectid, name, cc, amount) VALUES (" + str(back['projectid']) + ", '" + back['name'] + "', " + str(back['cc']) + ", " + str(back['amount']) + ");") conn.commit() # retrieve backer by name @route('/backer/<backer>') def backer_list(backer): conn = sqlite3.connect(db_filename) c = conn.cursor() c.execute("SELECT project.name, 'transaction'.name, 'transaction'.amount FROM 'transaction' JOIN project ON 'transaction'.projectid = project.id WHERE 'transaction'.name = '" + backer + "';") result = c.fetchall() c.close() if result: return {'backer': result} else: response.status = 400 return {'msg':'ERROR: Backer not found'} @error(403) def mistake403(code): response.status = 403 return {'msg':'ERROR: There is a mistake in your URL'} @error(404) def mistake404(code): response.status = 404 return {'msg':'ERROR: This page does not exist'} #debug(True) run(reloader=True, host='0.0.0.0')
server.py
4,742
!/usr/bin/env python -*- coding: utf-8 -*- Author: David McCue Validate currency values Validate names Calculate luhn checksum Credit: wikipedia Validate credit card retrieve project by name add project add transaction Validate inputs Check credit card is unique to this user name Get project id from name retrieve backer by namedebug(True)
340
en
0.630326
from logs import logDecorator as lD import jsonref, sqlite3 config = jsonref.load(open('../config/config.json')) logBase = config['logging']['logBase'] + '.databaseIO.sqLiteIO' @lD.log(logBase + '.getAllData') def getAllData(logger, query, values=None, dbName=None): '''query data from the database Query the data over here. If there is a problem with the data, it is going to return the value of None, and log the error. Your program needs to check whether there was an error with the query by checking for a None return value. Note that the location of the dataabses are assumed to be present within the file ``../config/db.json``. Parameters ---------- logger : {logging.logger} logging element query : {str} The query to be made to the databse values : {tuple or list-like}, optional Additional values to be passed to the query (the default is None) dbName : {str or None}, optional The name of the database to use. If this is None, the function will attempt to read the name from the ``defaultDB`` item within the file ``../config/db.json``. Returns ------- list or None A list of tuples containing the values is returned. In case there is an error, the error will be logged, and a None will be return ''' vals = None try: db = jsonref.load(open('../config/db.json')) # Check whether a dbName is available if (dbName is None) and ('defaultDB' in db): dbName = db['defaultDB'] # Check whether a dbName has been specified if dbName is None: logger.error('A database name has not been specified.') return None conn = sqlite3.connect(db[dbName]['connection']) cur = conn.cursor() except Exception as e: logger.error('Unable to connect to the database') logger.error(str(e)) return try: if values is None: cur.execute(query) else: cur.execute(query, values) # We assume that the data is small so we # can download the entire thing here ... # ------------------------------------------- vals = cur.fetchall() except Exception as e: logger.error('Unable to obtain data from the database for:\n query: {}\n{values}'.format(query, values)) logger.error(str(e)) try: cur.close() conn.close() except Exception as e: logger.error('Unable to disconnect to the database') logger.error(str(e)) return return vals @lD.log(logBase + '.getDataIterator') def getDataIterator(logger, query, values=None, chunks=100, dbName=None): '''Create an iterator from a largish query This is a generator that returns values in chunks of chunksize ``chunks``. Parameters ---------- logger : {logging.logger} logging element query : {str} The query to be made to the databse values : {tuple or list-like}, optional Additional values to be passed to the query (the default is None) chunks : {number}, optional This is the number of rows that the data is going to return at every call if __next__() to this function. (the default is 100) dbName : {str or None}, optional The name of the database to use. If this is None, the function will attempt to read the name from the ``defaultDB`` item within the file ``../config/db.json``. Yields ------ list of tuples A list of tuples from the query, with a maximum of ``chunks`` tuples returned at one time. ''' try: db = jsonref.load(open('../config/db.json')) # Check whether a dbName is available if (dbName is None) and ('defaultDB' in db): dbName = db['defaultDB'] # Check whether a dbName has been specified if dbName is None: logger.error('A database name has not been specified.') return None conn = sqlite3.connect(db[dbName]['connection']) cur = conn.cursor() except Exception as e: logger.error('Unable to connect to the database') logger.error(str(e)) return try: if values is None: cur.execute(query) else: cur.execute(query, values) while True: vals = cur.fetchmany(chunks) if len(vals) == 0: break yield vals except Exception as e: logger.error('Unable to obtain data from the database for:\n query: {}\nvalues'.format(query, values)) logger.error(str(e)) try: conn.close() except Exception as e: logger.error('Unable to disconnect to the database') logger.error(str(e)) return return @lD.log(logBase + '.getSingleDataIterator') def getSingleDataIterator(logger, query, values=None, dbName=None): '''Create an iterator from a largish query This is a generator that returns values in chunks of chunksize 1. Parameters ---------- logger : {logging.logger} logging element query : {str} The query to be made to the databse values : {tuple or list-like}, optional Additional values to be passed to the query (the default is None) dbName : {str or None}, optional The name of the database to use. If this is None, the function will attempt to read the name from the ``defaultDB`` item within the file ``../config/db.json``. Yields ------ list of tuples A list of tuples from the query, with a maximum of ``chunks`` tuples returned at one time. ''' try: db = jsonref.load(open('../config/db.json')) # Check whether a dbName is available if (dbName is None) and ('defaultDB' in db): dbName = db['defaultDB'] # Check whether a dbName has been specified if dbName is None: logger.error('A database name has not been specified.') return None conn = sqlite3.connect(db[dbName]['connection']) cur = conn.cursor() except Exception as e: logger.error('Unable to connect to the database') logger.error(str(e)) return try: if values is None: cur.execute(query) else: cur.execute(query, values) while True: vals = cur.fetchone() if vals is None: break yield vals except Exception as e: logger.error('Unable to obtain data from the database for:\n query: {}\nvalues'.format(query, values)) logger.error(str(e)) try: conn.close() except Exception as e: logger.error('Unable to disconnect to the database') logger.error(str(e)) return return @lD.log(logBase + '.commitData') def commitData(logger, query, values=None, dbName=None): '''query data from the database Query the data over here. If there is a problem with the data, it is going to return the value of ``None``, and log the error. Your program needs to check whether there was an error with the query by checking for a ``None`` return value Parameters ---------- logger : {logging.logger} logging element query : {str} The query to be made to the databse values : {tuple or list-like}, optional Additional values to be passed to the query (the default is None) dbName : {str or None}, optional The name of the database to use. If this is None, the function will attempt to read the name from the ``defaultDB`` item within the file ``../config/db.json``. Returns ------- True or None On successful completion, a ``True`` is returned. In case there is an error, the error will be logged, and a ``None`` will be returnd ''' vals = True try: db = jsonref.load(open('../config/db.json')) # Check whether a dbName is available if (dbName is None) and ('defaultDB' in db): dbName = db['defaultDB'] # Check whether a dbName has been specified if dbName is None: logger.error('A database name has not been specified.') return None conn = sqlite3.connect(db[dbName]['connection']) cur = conn.cursor() except Exception as e: logger.error('Unable to connect to the database') logger.error(str(e)) return None try: if values is None: cur.execute(query) else: cur.execute(query, values) except Exception as e: logger.error('Unable to obtain data from the database for:\n query: {}\nvalues'.format(query, values)) logger.error(str(e)) vals = None try: conn.commit() cur.close() conn.close() except Exception as e: logger.error('Unable to disconnect to the database') logger.error(str(e)) return return vals @lD.log(logBase + '.commitDataList') def commitDataList(logger, query, values, dbName=None): '''query data from the database Query the data over here. If there is a problem with the data, it is going to return the value of None, and log the error. Your program needs to check whether there was an error with the query by checking for a ``None`` return value Parameters ---------- logger : {logging.logger} logging element query : {str} The query to be made to the databse values : {tuple or list-like}, optional Additional values to be passed to the query (the default is None) dbName : {str or None}, optional The name of the database to use. If this is None, the function will attempt to read the name from the ``defaultDB`` item within the file ``../config/db.json``. Returns ------- True or None A successful completion of this function returns a ``True``. In case there is an error, the error will be logged, and a ``None`` will be returned ''' val = True try: db = jsonref.load(open('../config/db.json')) # Check whether a dbName is available if (dbName is None) and ('defaultDB' in db): dbName = db['defaultDB'] # Check whether a dbName has been specified if dbName is None: logger.error('A database name has not been specified.') return None conn = sqlite3.connect(db[dbName]['connection']) cur = conn.cursor() except Exception as e: logger.error('Unable to connect to the database') logger.error(str(e)) return None try: cur.executemany(query, values) except Exception as e: logger.error('Unable to execute query for:\n query: {}\nvalues'.format(query, values)) logger.error(str(e)) val = None try: conn.commit() cur.close() conn.close() except Exception as e: logger.error('Unable to disconnect to the database') logger.error(str(e)) return None return val
src/lib/databaseIO/sqLiteIO.py
11,830
query data from the database Query the data over here. If there is a problem with the data, it is going to return the value of ``None``, and log the error. Your program needs to check whether there was an error with the query by checking for a ``None`` return value Parameters ---------- logger : {logging.logger} logging element query : {str} The query to be made to the databse values : {tuple or list-like}, optional Additional values to be passed to the query (the default is None) dbName : {str or None}, optional The name of the database to use. If this is None, the function will attempt to read the name from the ``defaultDB`` item within the file ``../config/db.json``. Returns ------- True or None On successful completion, a ``True`` is returned. In case there is an error, the error will be logged, and a ``None`` will be returnd query data from the database Query the data over here. If there is a problem with the data, it is going to return the value of None, and log the error. Your program needs to check whether there was an error with the query by checking for a ``None`` return value Parameters ---------- logger : {logging.logger} logging element query : {str} The query to be made to the databse values : {tuple or list-like}, optional Additional values to be passed to the query (the default is None) dbName : {str or None}, optional The name of the database to use. If this is None, the function will attempt to read the name from the ``defaultDB`` item within the file ``../config/db.json``. Returns ------- True or None A successful completion of this function returns a ``True``. In case there is an error, the error will be logged, and a ``None`` will be returned query data from the database Query the data over here. If there is a problem with the data, it is going to return the value of None, and log the error. Your program needs to check whether there was an error with the query by checking for a None return value. Note that the location of the dataabses are assumed to be present within the file ``../config/db.json``. Parameters ---------- logger : {logging.logger} logging element query : {str} The query to be made to the databse values : {tuple or list-like}, optional Additional values to be passed to the query (the default is None) dbName : {str or None}, optional The name of the database to use. If this is None, the function will attempt to read the name from the ``defaultDB`` item within the file ``../config/db.json``. Returns ------- list or None A list of tuples containing the values is returned. In case there is an error, the error will be logged, and a None will be return Create an iterator from a largish query This is a generator that returns values in chunks of chunksize ``chunks``. Parameters ---------- logger : {logging.logger} logging element query : {str} The query to be made to the databse values : {tuple or list-like}, optional Additional values to be passed to the query (the default is None) chunks : {number}, optional This is the number of rows that the data is going to return at every call if __next__() to this function. (the default is 100) dbName : {str or None}, optional The name of the database to use. If this is None, the function will attempt to read the name from the ``defaultDB`` item within the file ``../config/db.json``. Yields ------ list of tuples A list of tuples from the query, with a maximum of ``chunks`` tuples returned at one time. Create an iterator from a largish query This is a generator that returns values in chunks of chunksize 1. Parameters ---------- logger : {logging.logger} logging element query : {str} The query to be made to the databse values : {tuple or list-like}, optional Additional values to be passed to the query (the default is None) dbName : {str or None}, optional The name of the database to use. If this is None, the function will attempt to read the name from the ``defaultDB`` item within the file ``../config/db.json``. Yields ------ list of tuples A list of tuples from the query, with a maximum of ``chunks`` tuples returned at one time. Check whether a dbName is available Check whether a dbName has been specified We assume that the data is small so we can download the entire thing here ... ------------------------------------------- Check whether a dbName is available Check whether a dbName has been specified Check whether a dbName is available Check whether a dbName has been specified Check whether a dbName is available Check whether a dbName has been specified Check whether a dbName is available Check whether a dbName has been specified
4,828
en
0.721411
#!\usr\bin\python from numpy import array from scipy.special import erf from scipy.optimize import minimize from math import pi, sin, cos, exp, sqrt #import dicom line_array = [] ## global def read_line (file_name ): with open( file_name ) as f: for line in f: line_array.append( [float( line.split()[0] ), float( line.split()[2] )] ) read_line("4mv_line.csv") line_len_2 = int(len(line_array)*0.5) ## global def pi(x, b): # 0 1 2 3 4 5 6 # b is np.array of these parameters: [sigma_1, sigma_2, w_1, x_sh, bkg, B, b] s_1 = 0.5*b[5]/(abs(b[2])*b[0]+abs(1-abs(b[2]))*b[1]) s_2 = abs(b[2])*b[0]*erf( (b[6]-x-b[3])/(sqrt(2)*b[0]) ) s_3 = abs(b[2])*b[0]*erf( (-b[6]-x-b[3])/(sqrt(2)*b[0]) ) s_4 = abs(1-abs(b[2]))*b[1]*erf( (b[6]-x-b[3])/(sqrt(2)*b[1]) ) s_5 = abs(1-abs(b[2]))*b[1]*erf( (-b[6]-x-b[3])/(sqrt(2)*b[1]) ) return s_1*(s_2 - s_3 + s_4 - s_5) + b[4] # x in mm def s(b): n_points_checked = 190 halv = int( n_points_checked*0.5 ) temp = 0.0 for i in range( n_points_checked ): x = (i-halv)*0.2481 a = pi(x, b) - line_array[ line_len_2 - halv +i ][1] temp += a*a return temp # [sigma_1, sigma_2, w_1, x_sh, bkg, B, b ] x0 = array([1.58, 0.58, 0.08, -0.03, 1047.0, 15031.0, 1.40]) # initial values for minimize print ( x0 ) res = minimize(s, x0, method='nelder-mead', options={'xtol': 1e-2, 'disp': True, 'maxfev':1e5, 'maxiter':1e5} ) print (res.x) print (res.fun * 1e-6) # print out the whole line for i in range(190): x = (i-95)*0.2481 # x in milimiters print(x,", ", line_array[line_len_2 - 95 + i][1],", ",pi(x,res.x) )
minimize/retic_xmm_2gauss/2minimize_4mv.py
1,736
!\usr\bin\pythonimport dicom global global 0 1 2 3 4 5 6 b is np.array of these parameters: [sigma_1, sigma_2, w_1, x_sh, bkg, B, b] x in mm [sigma_1, sigma_2, w_1, x_sh, bkg, B, b ] initial values for minimize print out the whole line x in milimiters
300
en
0.488345
import pandas as pd import numpy as np import scipy as sp from scipy.optimize import linprog, minimize, NonlinearConstraint from .pdf_quantile_functions import pdf_quantile_builder from .support import diffMatMetalog, pdfMetalog, quantileMetalog, newtons_method_metalog import time import warnings def a_vector_OLS_and_LP(m_dict, bounds, boundedness, term_limit, term_lower_bound, fit_method, alpha, diff_error = .001, diff_step = 0.001): """ Main workhorse function of pymetalog package. Called during metalog.__init__ method call. Args: m_dict (:obj:`dict` with keys ['params', 'dataValues', 'Y']): Initialized output_dict variable from metalog class. - m_dict['params']: (:obj:`dict` with keys ['bounds', 'boundedness', 'term_limit', 'term_lower_bound', 'step_len', 'fit_method']): * 'bounds': metalog.bounds * 'boundedness': metalog.boundedness * 'term_limit': metalog.term_limit * 'term_lower_bound': metalog.term_lower_bound * 'step_len': metalog.step_len * 'fit_method': metalog.fit_method - m_dict['dataValues']: (:obj:`pandas.DataFrame` with columns ['x','probs','z'] of type numeric): * 'x': metalog.x * 'probs': metalog.probs * 'z': column calculated in metalog.append_zvector method - depends on metalog.boundedness attribute - metalog.boundedness = 'u': * 'z' = metalog.x - metalog.boundedness = 'sl': * 'z' = log( (metalog.x-lower_bound) ) - metalog.boundedness = 'su': * 'z' = = log( (upper_bound-metalog.x) ) - metalog.boundedness = 'b': * 'z' = log( (metalog.x-lower_bound) / (upper_bound-metalog.x) ) - m_dict['Y']: (:obj:`pandas.DataFrame` with columns ['y1','y2','y3','y4', ... ,'yn'] of type numeric): * 'y1': numpy.array of ones with length equal to len(x) * 'y2': numpy.array of numeric values equal to the term attached to s in the logistic quantile function np.log(m_dict['dataValues']['probs'] / (1 - m_dict['dataValues']['probs'])) * 'y3': numpy.array of numeric values (m_dict['dataValues']['probs'] - 0.5) * m_dict['Y']['y2'] * 'y4': numpy.array of numeric values m_dict['Y']['y4'] = m_dict['dataValues']['probs'] - 0.5 * 'yn': numpy.array of numeric values: - if n in 'yn' is odd, m_dict['Y']['yn'] = m_dict['Y']['y4']**(int(i//2)) - if n in 'yn' is even, zn = 'y' + str(n-1) m_dict['Y'][yn] = m_dict['Y']['y2'] * m_dict['Y'][zn] bounds (:obj:`list`): Upper and lower limits to filter the data with before calculating metalog quantiles/pdfs. - should be set in conjunction with the `boundedness` parameter boundedness (:obj:`str`): String that is used to specify the type of metalog to fit. - must be in set ('u','sl','su','b') - Default: 'u' * Fits an unbounded metalog - 'sl' fits a strictly lower bounded metalog * len(bounds) must == 1 - 'su' fits a strictly upper bounded metalog * len(bounds) must == 1 - 'b' fits a upper/lower bounded metalog * len(bounds) must == 2 * bounds[1] must be > bounds[0] term_limit (:obj:`int`): The upper limit of the range of metalog terms to use to fit the data. - strictly > term_lower_bound - in range [3,30] term_lower_bound (:obj:`int`): The lower limit of the range of metalog terms to use to fit the data. - strictly < term_limit - in range [2,29] fit_method (:obj:`str`): Fit method to use to fit metalog distribution. - must be in set ('any','OLS','LP','MLE') - Default: 'any' * first tries 'OLS' method than 'LP' - 'OLS' only tries to fit by solving directly for a coefficients using ordinary least squares method - 'LP' only tries to estimate fit using simplex linear program optimization routine - 'MLE' first tries 'OLS' method than falls back to a maximum likelihood estimation routine alpha (:obj:`float`, optional): Regularization term to add to OLS fit - strictly >= 0. - should be set in conjunction with `penalty` parameter - Default: 0. (no regularization, OLS) diff_error (:obj:`float`, optional): Value used to in scipy.optimize.linprog method call to init the array of values representing the upper-bound of each inequality constraint (row) in A_ub. - #TODO: Insert maths diff_step (:obj:`float`, optional): Value passed to `step_len` parameter in support.py diffMatMetalog method call defines the bin width for the Reimann sum of the differences differentiation method - diffMatMetalog differentiates the metalog pdf * Differentiation reference: https://math.stackexchange.com/a/313135 Returns: m_dict: (:obj:`dict` with keys ['params', 'dataValues', 'Y', 'A', 'M', 'Validation']) - m_dict['A']: (:obj:`pandas.DataFrame` with columns ['a2','a3', ... ,'an'] of type numeric): * a2, a3, ... , an are our a coefficients returned by the method specified in `fit_method` - m_dict['M']: (:obj:`pandas.DataFrame` with columns 0:'pdf_1',1:'cdf_1',2:'pdf_2',3:'cdf_2', ...,((2*(term_limit-term_lower_bound))+1)-1:'pdf_n', ((2*(term_limit-term_lower_bound))+1):'cdf_n' where n is the total number of metalog fits determined by (term_limit-term_lower_bound)+1 ) * pdf_1, pdf_2, ... , pdf_n are the metalog pdfs returned by pdf_quantile_builder.pdfMetalog method * cdf_1, cdf_2, ... , cdf_n are the metalog quantiles returned by pdf_quantile_builder.quantileMetalog method - m_dict['y']: (:obj: `numpy.ndarray` of type float): * Array of bin widths for both the pdf_n and cdf_n - m_dict['Validation']: (:obj:`pandas.DataFrame` with columns ['term', 'valid', 'method'] of type str): * 'term': each metalog estimation given a number of terms * 'valid': boolean flag indicating if the metalog estimation was valid or not * 'method': a string indicating which method was used for the metalog estimation """ A = pd.DataFrame() c_a_names = [] c_m_names = [] Mh = pd.DataFrame() Validation = pd.DataFrame() df_MH_temp_list = list() df_A_temp_list = list() df_Validation_temp_list = list() # TODO: Large for-loop can probably be factored into smaller functions for i in range(term_lower_bound,term_limit+1): Y = m_dict['Y'].iloc[:,0:i] eye = np.eye(Y.shape[1]) z = m_dict['dataValues']['z'] y = m_dict['dataValues']['probs'] step_len = m_dict['params']['step_len'] methodFit = 'OLS' a_name = 'a'+str(i) m_name = 'm'+str(i) M_name = 'M'+str(i) c_m_names = np.append(c_m_names, [m_name, M_name]) c_a_names = np.append(c_a_names, a_name) if fit_method == 'any' or fit_method == 'MLE': try: temp = np.dot(np.dot(np.linalg.inv(np.dot(Y.T, Y) + alpha*eye), Y.T), z) except: # use LP solver if OLS breaks temp = a_vector_LP(m_dict, term_limit=i, term_lower_bound=i, diff_error=diff_error, diff_step=diff_step) methodFit = 'Linear Program' if fit_method == 'OLS': try: temp = np.dot(np.dot(np.linalg.inv(np.dot(Y.T, Y) + alpha*eye), Y.T), z) except: raise RuntimeError("OLS was unable to solve infeasible or poorly formulated problem") if fit_method == "LP": temp = a_vector_LP(m_dict, term_limit=i, term_lower_bound=i, diff_error=diff_error, diff_step=diff_step) methodFit = 'Linear Program' if fit_method == 'MLE': temp = a_vector_MLE(temp, y, i, m_dict, bounds, boundedness) temp = np.append(temp, np.zeros(term_limit-i)) # build a y vector for smaller data sets if len(z) < 100: y2 = np.linspace(step_len, 1 - step_len, int((1 - step_len) / step_len)) tailstep = step_len / 10 y1 = np.linspace(tailstep, (min(y2) - tailstep), int((min(y2) - tailstep) / tailstep)) y3 = np.linspace((max(y2) + tailstep), (max(y2) + tailstep * 9), int((tailstep * 9) / tailstep)) y = np.hstack((y1, y2, y3)) # Get the dict and quantile values back for validation temp_dict = pdf_quantile_builder(temp, y=y, term_limit=i, bounds=bounds, boundedness=boundedness) # If it not a valid pdf run and the OLS version was used the LP version if (temp_dict['valid'] == 'no') and (fit_method != 'OLS'): temp = a_vector_LP(m_dict, term_limit=i, term_lower_bound=i, diff_error=diff_error, diff_step=diff_step) temp = np.append(temp, np.zeros(term_limit-i)) methodFit = 'Linear Program' # Get the dict and quantile values back for validation temp_dict = pdf_quantile_builder(temp, y=y, term_limit=i, bounds=bounds, boundedness=boundedness) df_MH_temp_list.append(pd.DataFrame(temp_dict['m'])) df_MH_temp_list.append(pd.DataFrame(temp_dict['M'])) df_A_temp_list.append(pd.DataFrame(temp)) tempValidation = pd.DataFrame(data={'term': [i], 'valid': [temp_dict['valid']], 'method': [methodFit]}) df_Validation_temp_list.append(tempValidation) Validation = pd.concat(df_Validation_temp_list, axis=0) Mh = pd.concat(df_MH_temp_list, axis=1) A = pd.concat(df_A_temp_list, axis=1) A.columns = c_a_names Mh.columns = c_m_names m_dict['A'] = A m_dict['M'] = Mh m_dict['M']['y'] = temp_dict['y'] m_dict['Validation'] = Validation A = np.column_stack((np.repeat(1.,len(A)), A)) Est = np.dot(m_dict['Y'], A) ncols = A.shape[1] Z = np.column_stack((np.array(m_dict['dataValues']['z']),np.repeat(m_dict['dataValues']['z'].values,ncols-1).reshape(len(m_dict['dataValues']['z']),ncols-1))) m_dict['square_residual_error'] = ((Z-Est)**2).sum(axis=1) return m_dict def a_vector_LP(m_dict, term_limit, term_lower_bound, diff_error = .001, diff_step = 0.001): """TODO: write docstring """ cnames = np.array([]) for i in range(term_lower_bound, term_limit + 1): Y = m_dict['Y'].iloc[:, 0:i] z = m_dict['dataValues']['z'] # Bulding the objective function using abs value LP formulation Y_neg = -Y new_Y = pd.DataFrame({'y1': Y.iloc[:, 0], 'y1_neg': Y_neg.iloc[:, 0]}) for c in range(1,len(Y.iloc[0,:])): new_Y['y'+str(c+1)] = Y.iloc[:,c] new_Y['y' + str(c+1)+'_neg'] = Y_neg.iloc[:, c] a = np.array([''.join(['a', str(i)])]) cnames = np.append(cnames, a, axis=0) # Building the constraint matrix error_mat = np.array([]) for j in range(1,len(Y.iloc[:,0])+1): front_zeros = np.zeros(2 * (j - 1)) ones = [1, -1] trail_zeroes = np.zeros(2 * (len(Y.iloc[:, 1]) - j)) if j == 1: error_vars = np.append(ones, trail_zeroes) elif j != 1: error_vars = np.append(front_zeros, ones) error_vars = np.append(error_vars, trail_zeroes) if error_mat.size == 0: error_mat = np.append(error_mat, error_vars, axis=0) else: error_mat = np.vstack((error_mat, error_vars)) new = pd.concat((pd.DataFrame(data=error_mat), new_Y), axis=1) diff_mat = diffMatMetalog(i, diff_step) diff_zeros = [] for t in range(0,len(diff_mat.iloc[:, 0])): zeros_temp = np.zeros(2 * len(Y.iloc[:, 0])) if np.size(diff_zeros) == 0: diff_zeros = zeros_temp else: diff_zeros = np.vstack((zeros_temp, diff_zeros)) diff_mat = np.concatenate((diff_zeros, diff_mat), axis=1) # Combine the total constraint matrix lp_mat = np.concatenate((new, diff_mat), axis=0) # Objective function coeficients c = np.append(np.ones(2 * len(Y.iloc[:, 1])), np.zeros(2*i)) # Constraint matrices A_eq = lp_mat[:len(Y.iloc[:, 1]),:] A_ub = -1*lp_mat[len(Y.iloc[:, 1]):,:] b_eq = z b_ub = -1*np.repeat(diff_error, len(diff_mat[:,0])) # Solving the linear program w/ scipy (for now) lp_sol = linprog(c, A_ub=A_ub, b_ub=b_ub, A_eq=A_eq, b_eq=b_eq, method='simplex', options={"maxiter":5000, "tol":1.0e-5,"disp": False}) # Consolidating solution back into the a vector tempLP = lp_sol.x[(2 * len(Y.iloc[:, 1])):(len(lp_sol.x)+1)] temp = [] for r in range(0,((len(tempLP) // 2))): temp.append(tempLP[(r * 2)] - tempLP[(2 * r)+1]) return temp def a_vector_MLE(a, y, term, m_dict, bounds, boundedness): """TODO: write docstring """ ym = [newtons_method_metalog(a, xi, term, bounds, boundedness) for xi in m_dict['dataValues']['x']] def MLE_quantile_constraints(x): M = [quantileMetalog(x[:term], yi, term, bounds=bounds, boundedness=boundedness) for yi in x[term:]] return m_dict['dataValues']['x'] - M def MLE_objective_function(x, y, term, m_dict): return -np.sum([np.log10(pdfMetalog(x[:term], yi, term, bounds, boundedness)) for yi in np.absolute(x[term:])]) m_dict[str('MLE' + str(term))] = {} x0 = np.hstack((a[:term],ym)) m_dict[str('MLE' + str(term))]['oldobj'] = -MLE_objective_function(x0, y, term, m_dict) bnd = ((None, None),)*len(a)+((0, 1),)*(len(x0)-len(a)) con = NonlinearConstraint(MLE_quantile_constraints, 0, 0) mle = minimize(MLE_objective_function, x0, args=(y, term, m_dict), bounds=bnd, constraints=con) m_dict[str('MLE' + str(term))]['newobj'] = -MLE_objective_function(mle.x, y, term, m_dict) m_dict[str('MLE'+str(term))]['A'] = mle.x[:term] m_dict[str('MLE'+str(term))]['Y'] = mle.x[term:] m_dict[str('MLE' + str(term))]['oldA'] = a m_dict[str('MLE' + str(term))]['oldY'] = y out_temp = np.zeros_like(a) for i in range(term): out_temp[i] = mle.x[i] return out_temp
pymetalog/a_vector.py
15,097
TODO: write docstring TODO: write docstring Main workhorse function of pymetalog package. Called during metalog.__init__ method call. Args: m_dict (:obj:`dict` with keys ['params', 'dataValues', 'Y']): Initialized output_dict variable from metalog class. - m_dict['params']: (:obj:`dict` with keys ['bounds', 'boundedness', 'term_limit', 'term_lower_bound', 'step_len', 'fit_method']): * 'bounds': metalog.bounds * 'boundedness': metalog.boundedness * 'term_limit': metalog.term_limit * 'term_lower_bound': metalog.term_lower_bound * 'step_len': metalog.step_len * 'fit_method': metalog.fit_method - m_dict['dataValues']: (:obj:`pandas.DataFrame` with columns ['x','probs','z'] of type numeric): * 'x': metalog.x * 'probs': metalog.probs * 'z': column calculated in metalog.append_zvector method - depends on metalog.boundedness attribute - metalog.boundedness = 'u': * 'z' = metalog.x - metalog.boundedness = 'sl': * 'z' = log( (metalog.x-lower_bound) ) - metalog.boundedness = 'su': * 'z' = = log( (upper_bound-metalog.x) ) - metalog.boundedness = 'b': * 'z' = log( (metalog.x-lower_bound) / (upper_bound-metalog.x) ) - m_dict['Y']: (:obj:`pandas.DataFrame` with columns ['y1','y2','y3','y4', ... ,'yn'] of type numeric): * 'y1': numpy.array of ones with length equal to len(x) * 'y2': numpy.array of numeric values equal to the term attached to s in the logistic quantile function np.log(m_dict['dataValues']['probs'] / (1 - m_dict['dataValues']['probs'])) * 'y3': numpy.array of numeric values (m_dict['dataValues']['probs'] - 0.5) * m_dict['Y']['y2'] * 'y4': numpy.array of numeric values m_dict['Y']['y4'] = m_dict['dataValues']['probs'] - 0.5 * 'yn': numpy.array of numeric values: - if n in 'yn' is odd, m_dict['Y']['yn'] = m_dict['Y']['y4']**(int(i//2)) - if n in 'yn' is even, zn = 'y' + str(n-1) m_dict['Y'][yn] = m_dict['Y']['y2'] * m_dict['Y'][zn] bounds (:obj:`list`): Upper and lower limits to filter the data with before calculating metalog quantiles/pdfs. - should be set in conjunction with the `boundedness` parameter boundedness (:obj:`str`): String that is used to specify the type of metalog to fit. - must be in set ('u','sl','su','b') - Default: 'u' * Fits an unbounded metalog - 'sl' fits a strictly lower bounded metalog * len(bounds) must == 1 - 'su' fits a strictly upper bounded metalog * len(bounds) must == 1 - 'b' fits a upper/lower bounded metalog * len(bounds) must == 2 * bounds[1] must be > bounds[0] term_limit (:obj:`int`): The upper limit of the range of metalog terms to use to fit the data. - strictly > term_lower_bound - in range [3,30] term_lower_bound (:obj:`int`): The lower limit of the range of metalog terms to use to fit the data. - strictly < term_limit - in range [2,29] fit_method (:obj:`str`): Fit method to use to fit metalog distribution. - must be in set ('any','OLS','LP','MLE') - Default: 'any' * first tries 'OLS' method than 'LP' - 'OLS' only tries to fit by solving directly for a coefficients using ordinary least squares method - 'LP' only tries to estimate fit using simplex linear program optimization routine - 'MLE' first tries 'OLS' method than falls back to a maximum likelihood estimation routine alpha (:obj:`float`, optional): Regularization term to add to OLS fit - strictly >= 0. - should be set in conjunction with `penalty` parameter - Default: 0. (no regularization, OLS) diff_error (:obj:`float`, optional): Value used to in scipy.optimize.linprog method call to init the array of values representing the upper-bound of each inequality constraint (row) in A_ub. - #TODO: Insert maths diff_step (:obj:`float`, optional): Value passed to `step_len` parameter in support.py diffMatMetalog method call defines the bin width for the Reimann sum of the differences differentiation method - diffMatMetalog differentiates the metalog pdf * Differentiation reference: https://math.stackexchange.com/a/313135 Returns: m_dict: (:obj:`dict` with keys ['params', 'dataValues', 'Y', 'A', 'M', 'Validation']) - m_dict['A']: (:obj:`pandas.DataFrame` with columns ['a2','a3', ... ,'an'] of type numeric): * a2, a3, ... , an are our a coefficients returned by the method specified in `fit_method` - m_dict['M']: (:obj:`pandas.DataFrame` with columns 0:'pdf_1',1:'cdf_1',2:'pdf_2',3:'cdf_2', ...,((2*(term_limit-term_lower_bound))+1)-1:'pdf_n', ((2*(term_limit-term_lower_bound))+1):'cdf_n' where n is the total number of metalog fits determined by (term_limit-term_lower_bound)+1 ) * pdf_1, pdf_2, ... , pdf_n are the metalog pdfs returned by pdf_quantile_builder.pdfMetalog method * cdf_1, cdf_2, ... , cdf_n are the metalog quantiles returned by pdf_quantile_builder.quantileMetalog method - m_dict['y']: (:obj: `numpy.ndarray` of type float): * Array of bin widths for both the pdf_n and cdf_n - m_dict['Validation']: (:obj:`pandas.DataFrame` with columns ['term', 'valid', 'method'] of type str): * 'term': each metalog estimation given a number of terms * 'valid': boolean flag indicating if the metalog estimation was valid or not * 'method': a string indicating which method was used for the metalog estimation TODO: Large for-loop can probably be factored into smaller functions use LP solver if OLS breaks build a y vector for smaller data sets Get the dict and quantile values back for validation If it not a valid pdf run and the OLS version was used the LP version Get the dict and quantile values back for validation Bulding the objective function using abs value LP formulation Building the constraint matrix Combine the total constraint matrix Objective function coeficients Constraint matrices Solving the linear program w/ scipy (for now) Consolidating solution back into the a vector
6,771
en
0.562292
# -*- coding: utf-8 -*- # This code is part of Qiskit. # # (C) Copyright IBM 2019. # # This code is licensed under the Apache License, Version 2.0. You may # obtain a copy of this license in the LICENSE.txt file in the root directory # of this source tree or at http://www.apache.org/licenses/LICENSE-2.0. # # Any modifications or derivative works of this code must retain this # copyright notice, and modified files need to carry a notice indicating # that they have been altered from the originals. """Test cases for the pulse scheduler passes.""" from qiskit import QuantumRegister, ClassicalRegister, QuantumCircuit, schedule from qiskit.pulse import (Schedule, DriveChannel, AcquireChannel, Acquire, MeasureChannel, MemorySlot) from qiskit.test.mock import FakeOpenPulse2Q, FakeOpenPulse3Q from qiskit.test import QiskitTestCase class TestBasicSchedule(QiskitTestCase): """Scheduling tests.""" def setUp(self): self.backend = FakeOpenPulse2Q() self.cmd_def = self.backend.defaults().build_cmd_def() def test_alap_pass(self): """Test ALAP scheduling.""" q = QuantumRegister(2) c = ClassicalRegister(2) qc = QuantumCircuit(q, c) qc.u2(3.14, 1.57, q[0]) qc.u2(0.5, 0.25, q[1]) qc.barrier(q[1]) qc.u2(0.5, 0.25, q[1]) qc.barrier(q[0], q[1]) qc.cx(q[0], q[1]) qc.measure(q, c) sched = schedule(qc, self.backend) # X pulse on q0 should end at the start of the CNOT expected = Schedule( (28, self.cmd_def.get('u2', [0], 3.14, 1.57)), self.cmd_def.get('u2', [1], 0.5, 0.25), (28, self.cmd_def.get('u2', [1], 0.5, 0.25)), (56, self.cmd_def.get('cx', [0, 1])), (78, self.cmd_def.get('measure', [0, 1]))) for actual, expected in zip(sched.instructions, expected.instructions): self.assertEqual(actual[0], expected[0]) self.assertEqual(actual[1].command, expected[1].command) self.assertEqual(actual[1].channels, expected[1].channels) def test_alap_with_barriers(self): """Test that ALAP respects barriers on new qubits.""" q = QuantumRegister(2) c = ClassicalRegister(2) qc = QuantumCircuit(q, c) qc.u2(0, 0, q[0]) qc.barrier(q[0], q[1]) qc.u2(0, 0, q[1]) sched = schedule(qc, self.backend, method='alap') expected = Schedule( self.cmd_def.get('u2', [0], 0, 0), (28, self.cmd_def.get('u2', [1], 0, 0))) for actual, expected in zip(sched.instructions, expected.instructions): self.assertEqual(actual[0], expected[0]) self.assertEqual(actual[1].command, expected[1].command) self.assertEqual(actual[1].channels, expected[1].channels) def test_alap_aligns_end(self): """Test that ALAP always acts as though there is a final global barrier.""" q = QuantumRegister(2) c = ClassicalRegister(2) qc = QuantumCircuit(q, c) qc.u3(0, 0, 0, q[0]) qc.u2(0, 0, q[1]) sched = schedule(qc, self.backend, method='alap') expected_sched = Schedule( self.cmd_def.get('u2', [1], 0, 0), (26, self.cmd_def.get('u3', [0], 0, 0, 0))) for actual, expected in zip(sched.instructions, expected_sched.instructions): self.assertEqual(actual[0], expected[0]) self.assertEqual(actual[1].command, expected[1].command) self.assertEqual(actual[1].channels, expected[1].channels) self.assertEqual(sched.ch_duration(DriveChannel(0)), expected_sched.ch_duration(DriveChannel(1))) def test_asap_pass(self): """Test ASAP scheduling.""" q = QuantumRegister(2) c = ClassicalRegister(2) qc = QuantumCircuit(q, c) qc.u2(3.14, 1.57, q[0]) qc.u2(0.5, 0.25, q[1]) qc.barrier(q[1]) qc.u2(0.5, 0.25, q[1]) qc.barrier(q[0], q[1]) qc.cx(q[0], q[1]) qc.measure(q, c) sched = schedule(qc, self.backend, method="as_soon_as_possible") # X pulse on q0 should start at t=0 expected = Schedule( self.cmd_def.get('u2', [0], 3.14, 1.57), self.cmd_def.get('u2', [1], 0.5, 0.25), (28, self.cmd_def.get('u2', [1], 0.5, 0.25)), (56, self.cmd_def.get('cx', [0, 1])), (78, self.cmd_def.get('measure', [0, 1]))) self.assertEqual(sched.instructions, expected.instructions) def test_alap_resource_respecting(self): """Test that the ALAP pass properly respects busy resources when backwards scheduling. For instance, a CX on 0 and 1 followed by an X on only 1 must respect both qubits' timeline.""" q = QuantumRegister(2) c = ClassicalRegister(2) qc = QuantumCircuit(q, c) qc.cx(q[0], q[1]) qc.u2(0.5, 0.25, q[1]) sched = schedule(qc, self.backend, method="as_late_as_possible") insts = sched.instructions self.assertEqual(insts[0][0], 0) self.assertEqual(insts[4][0], 22) qc = QuantumCircuit(q, c) qc.cx(q[0], q[1]) qc.u2(0.5, 0.25, q[1]) qc.measure(q, c) sched = schedule(qc, self.backend, method="as_late_as_possible") self.assertEqual(sched.instructions[-1][0], 50) def test_cmd_def_schedules_unaltered(self): """Test that forward scheduling doesn't change relative timing with a command.""" q = QuantumRegister(2) c = ClassicalRegister(2) qc = QuantumCircuit(q, c) qc.cx(q[0], q[1]) sched1 = schedule(qc, self.backend, method="as_soon_as_possible") sched2 = schedule(qc, self.backend, method="as_late_as_possible") self.assertEqual(sched1.instructions, sched2.instructions) insts = sched1.instructions self.assertEqual(insts[0][0], 0) self.assertEqual(insts[1][0], 10) self.assertEqual(insts[2][0], 20) self.assertEqual(insts[3][0], 20) def test_measure_combined(self): """ Test to check for measure on the same qubit which generated another measure schedule. The measures on different qubits are combined, but measures on the same qubit adds another measure to the schedule. """ q = QuantumRegister(2) c = ClassicalRegister(2) qc = QuantumCircuit(q, c) qc.u2(3.14, 1.57, q[0]) qc.cx(q[0], q[1]) qc.measure(q[0], c[0]) qc.measure(q[1], c[1]) qc.measure(q[1], c[1]) sched = schedule(qc, self.backend, method="as_soon_as_possible") expected = Schedule( self.cmd_def.get('u2', [0], 3.14, 1.57), (28, self.cmd_def.get('cx', [0, 1])), (50, self.cmd_def.get('measure', [0, 1])), (60, self.cmd_def.get('measure', [0, 1]).filter(channels=[MeasureChannel(1)])), (60, Acquire(duration=10)([AcquireChannel(0), AcquireChannel(1)], [MemorySlot(0), MemorySlot(1)]))) self.assertEqual(sched.instructions, expected.instructions) def test_3q_schedule(self): """Test a schedule that was recommended by David McKay :D """ backend = FakeOpenPulse3Q() cmd_def = backend.defaults().build_cmd_def() q = QuantumRegister(3) c = ClassicalRegister(3) qc = QuantumCircuit(q, c) qc.cx(q[0], q[1]) qc.u2(0.778, 0.122, q[2]) qc.u3(3.14, 1.57, 0., q[0]) qc.u2(3.14, 1.57, q[1]) qc.cx(q[1], q[2]) qc.u2(0.778, 0.122, q[2]) sched = schedule(qc, backend) expected = Schedule( cmd_def.get('cx', [0, 1]), (22, cmd_def.get('u2', [1], 3.14, 1.57)), (46, cmd_def.get('u2', [2], 0.778, 0.122)), (50, cmd_def.get('cx', [1, 2])), (72, cmd_def.get('u2', [2], 0.778, 0.122)), (74, cmd_def.get('u3', [0], 3.14, 1.57))) self.assertEqual(sched.instructions, expected.instructions) def test_schedule_multi(self): """Test scheduling multiple circuits at once.""" q = QuantumRegister(2) c = ClassicalRegister(2) qc0 = QuantumCircuit(q, c) qc0.cx(q[0], q[1]) qc1 = QuantumCircuit(q, c) qc1.cx(q[0], q[1]) schedules = schedule([qc0, qc1], self.backend) expected_insts = schedule(qc0, self.backend).instructions self.assertEqual(schedules[0].instructions, expected_insts) def test_circuit_name_kept(self): """Test that the new schedule gets its name from the circuit.""" q = QuantumRegister(2) c = ClassicalRegister(2) qc = QuantumCircuit(q, c, name='CIRCNAME') qc.cx(q[0], q[1]) sched = schedule(qc, self.backend, method="asap") self.assertEqual(sched.name, qc.name) sched = schedule(qc, self.backend, method="alap") self.assertEqual(sched.name, qc.name) def test_can_add_gates_into_free_space(self): """The scheduler does some time bookkeeping to know when qubits are free to be scheduled. Make sure this works for qubits that are used in the future. This was a bug, uncovered by this example: q0 = - - - - |X| q1 = |X| |u2| |X| In ALAP scheduling, the next operation on qubit 0 would be added at t=0 rather than immediately before the X gate. """ qr = QuantumRegister(2) qc = QuantumCircuit(qr) for i in range(2): qc.u2(0, 0, [qr[i]]) qc.u1(3.14, [qr[i]]) qc.u2(0, 0, [qr[i]]) sched = schedule(qc, self.backend, method="alap") expected = Schedule( self.cmd_def.get('u2', [0], 0, 0), self.cmd_def.get('u2', [1], 0, 0), (28, self.cmd_def.get('u1', [0], 3.14)), (28, self.cmd_def.get('u1', [1], 3.14)), (28, self.cmd_def.get('u2', [0], 0, 0)), (28, self.cmd_def.get('u2', [1], 0, 0))) self.assertEqual(sched.instructions, expected.instructions) def test_barriers_in_middle(self): """As a follow on to `test_can_add_gates_into_free_space`, similar issues arose for barriers, specifically. """ qr = QuantumRegister(2) qc = QuantumCircuit(qr) for i in range(2): qc.u2(0, 0, [qr[i]]) qc.barrier(qr[i]) qc.u1(3.14, [qr[i]]) qc.barrier(qr[i]) qc.u2(0, 0, [qr[i]]) sched = schedule(qc, self.backend, method="alap") expected = Schedule( self.cmd_def.get('u2', [0], 0, 0), self.cmd_def.get('u2', [1], 0, 0), (28, self.cmd_def.get('u1', [0], 3.14)), (28, self.cmd_def.get('u1', [1], 3.14)), (28, self.cmd_def.get('u2', [0], 0, 0)), (28, self.cmd_def.get('u2', [1], 0, 0))) self.assertEqual(sched.instructions, expected.instructions) def test_only_needed_measures(self): """Test that `MeasureChannel`s are only added for measured qubits.""" q = QuantumRegister(2) c = ClassicalRegister(2) qc = QuantumCircuit(q, c) qc.measure(q[1], c[1]) sched_all_channels = schedule(qc, self.backend, method="as_soon_as_possible").channels deleted_channels = [MeasureChannel(0)] self.assertNotIn(sched_all_channels, deleted_channels) def test_user_mapping_for_memslots(self): """ Test that the new schedule only has required `MeasureChannel`s and that the `MemorySlot`s are mapped according to the input circuit. """ q = QuantumRegister(2) c = ClassicalRegister(2) qc = QuantumCircuit(q, c) qc.measure(q[0], c[1]) sched = schedule(qc, self.backend) expected = Schedule( self.cmd_def.get('measure', [0, 1]).filter(channels=[MeasureChannel(0)]), Acquire(duration=10)([AcquireChannel(0), AcquireChannel(1)], [MemorySlot(1), MemorySlot(0)])) self.assertEqual(sched.instructions, expected.instructions) def test_user_mapping_for_memslots_3Q(self): """Test measuring two of three qubits.""" backend = FakeOpenPulse3Q() cmd_def = backend.defaults().build_cmd_def() q = QuantumRegister(3) c = ClassicalRegister(3) qc = QuantumCircuit(q, c) qc.measure(q[1], c[2]) qc.measure(q[2], c[0]) sched = schedule(qc, backend) expected = Schedule( cmd_def.get('measure', [0, 1, 2]).filter( channels=[MeasureChannel(1), MeasureChannel(2)]), Acquire(duration=10)([AcquireChannel(0), AcquireChannel(1), AcquireChannel(2)], [MemorySlot(1), MemorySlot(2), MemorySlot(0)])) self.assertEqual(sched.instructions, expected.instructions) def test_multiple_measure_in_3Q(self): """Test multiple measure, user memslot mapping, 3Q.""" backend = FakeOpenPulse3Q() cmd_def = backend.defaults().build_cmd_def() q = QuantumRegister(3) c = ClassicalRegister(5) qc = QuantumCircuit(q, c) qc.measure(q[0], c[2]) qc.measure(q[0], c[4]) sched = schedule(qc, backend) expected = Schedule( cmd_def.get('measure', [0, 1, 2]).filter(channels=[MeasureChannel(0)]), Acquire(duration=10)([AcquireChannel(0), AcquireChannel(1), AcquireChannel(2)], [MemorySlot(2), MemorySlot(0), MemorySlot(1)]), (10, cmd_def.get('measure', [0, 1, 2]).filter(channels=[MeasureChannel(0)])), (10, Acquire(duration=10)([AcquireChannel(0), AcquireChannel(1), AcquireChannel(2)], [MemorySlot(4), MemorySlot(0), MemorySlot(1)]))) self.assertEqual(sched.instructions, expected.instructions)
artifacts/old_dataset_versions/minimal_commits/qiskit-terra/qiskit-terra#2704/after/test_basic_scheduler.py
14,036
Scheduling tests. Test a schedule that was recommended by David McKay :D Test that ALAP always acts as though there is a final global barrier. Test ALAP scheduling. Test that the ALAP pass properly respects busy resources when backwards scheduling. For instance, a CX on 0 and 1 followed by an X on only 1 must respect both qubits' timeline. Test that ALAP respects barriers on new qubits. Test ASAP scheduling. As a follow on to `test_can_add_gates_into_free_space`, similar issues arose for barriers, specifically. The scheduler does some time bookkeeping to know when qubits are free to be scheduled. Make sure this works for qubits that are used in the future. This was a bug, uncovered by this example: q0 = - - - - |X| q1 = |X| |u2| |X| In ALAP scheduling, the next operation on qubit 0 would be added at t=0 rather than immediately before the X gate. Test that the new schedule gets its name from the circuit. Test that forward scheduling doesn't change relative timing with a command. Test to check for measure on the same qubit which generated another measure schedule. The measures on different qubits are combined, but measures on the same qubit adds another measure to the schedule. Test multiple measure, user memslot mapping, 3Q. Test that `MeasureChannel`s are only added for measured qubits. Test scheduling multiple circuits at once. Test that the new schedule only has required `MeasureChannel`s and that the `MemorySlot`s are mapped according to the input circuit. Test measuring two of three qubits. Test cases for the pulse scheduler passes. -*- coding: utf-8 -*- This code is part of Qiskit. (C) Copyright IBM 2019. This code is licensed under the Apache License, Version 2.0. You may obtain a copy of this license in the LICENSE.txt file in the root directory of this source tree or at http://www.apache.org/licenses/LICENSE-2.0. Any modifications or derivative works of this code must retain this copyright notice, and modified files need to carry a notice indicating that they have been altered from the originals. X pulse on q0 should end at the start of the CNOT X pulse on q0 should start at t=0
2,136
en
0.933343
""" >>> sd = SliceDump() >>> sd[1] 1 >>> sd[2:5] slice(2, 5, None) >>> sd[:2] slice(None, 2, None) >>> sd[7:] slice(7, None, None) >>> sd[:] slice(None, None, None) >>> sd[1:9:3] slice(1, 9, 3) >>> sd[1:9:3, 2:3] (slice(1, 9, 3), slice(2, 3, None)) >>> s = sd[1:9:3] >>> s.indices(20) (1, 9, 3) >>> s.indices(5) (1, 5, 3) >>> s.indices(1) (1, 1, 3) >>> s.indices(0) (0, 0, 3) """ class SliceDump: def __getitem__(self, pos): return pos
attic/sequences/slice_dump.py
548
>>> sd = SliceDump() >>> sd[1] 1 >>> sd[2:5] slice(2, 5, None) >>> sd[:2] slice(None, 2, None) >>> sd[7:] slice(7, None, None) >>> sd[:] slice(None, None, None) >>> sd[1:9:3] slice(1, 9, 3) >>> sd[1:9:3, 2:3] (slice(1, 9, 3), slice(2, 3, None)) >>> s = sd[1:9:3] >>> s.indices(20) (1, 9, 3) >>> s.indices(5) (1, 5, 3) >>> s.indices(1) (1, 1, 3) >>> s.indices(0) (0, 0, 3)
371
en
0.475071
""" utility functions for breaking down a given block of text into it's component syntactic parts. """ import nltk from nltk.tokenize import RegexpTokenizer from . import syllables_en TOKENIZER = RegexpTokenizer('(?u)\W+|\$[\d\.]+|\S+') SPECIAL_CHARS = ['.', ',', '!', '?'] def get_char_count(words): characters = 0 for word in words: characters += len(word) return characters def get_words(text=''): words = [] words = TOKENIZER.tokenize(text) filtered_words = [] for word in words: if word in SPECIAL_CHARS or word == " ": pass else: new_word = word.replace(",","").replace(".","") new_word = new_word.replace("!","").replace("?","") filtered_words.append(new_word) return filtered_words def get_sentences(text=''): tokenizer = nltk.data.load('tokenizers/punkt/english.pickle') sentences = tokenizer.tokenize(text) return sentences def count_syllables(words): syllableCount = 0 for word in words: syllableCount += syllables_en.count(word) return syllableCount #This method must be enhanced. At the moment it only #considers the number of syllables in a word. #This often results in that too many complex words are detected. def count_complex_words(text=''): words = get_words(text) sentences = get_sentences(text) complex_words = 0 found = False cur_word = [] for word in words: cur_word.append(word) if count_syllables(cur_word)>= 3: #Checking proper nouns. If a word starts with a capital letter #and is NOT at the beginning of a sentence we don't add it #as a complex word. if not(word[0].isupper()): complex_words += 1 else: for sentence in sentences: if str(sentence).startswith(word): found = True break if found: complex_words += 1 found = False cur_word.remove(word) return complex_words
opm/readability/utils.py
2,162
utility functions for breaking down a given block of text into it's component syntactic parts. This method must be enhanced. At the moment it onlyconsiders the number of syllables in a word.This often results in that too many complex words are detected.Checking proper nouns. If a word starts with a capital letterand is NOT at the beginning of a sentence we don't add itas a complex word.
390
en
0.911614
import requests import json from simplegist.mygist import Mygist from simplegist.do import Do from comments import Comments try: from simplegist.config import USERNAME, API_TOKEN, BASE_URL, GIST_URL except: pass class Simplegist: """ Gist Base Class This class is to used to instantiate the wrapper and authenticate. Authenticate with providing Github Username and API-Token to use it for all future API requests """ def __init__(self, **args): # Save our username and api_token (If given) for later use. if 'username' in args: self.username = args['username'] else: if not USERNAME: raise Exception('Please provide your Github username.') else: self.username = USERNAME if 'api_token' in args: self.api_token = args['api_token'] else: if not API_TOKEN: raise Exception('Please provide your Github API Token.') else: self.api_token = API_TOKEN # Set header information in every request. self.header = { 'X-Github-Username': self.username, 'Content-Type': 'application/json', 'Authorization': 'token %s' %self.api_token } def profile(self): return Mygist(self) def search(self, user): return Mygist(self,user=user) def do(self): return Do(self) def comments(self): return Comments(self) def create(self, **args): if 'description' in args: self.description = args['description'] else: self.description = '' if 'name' in args: self.gist_name = args['name'] else: self.gist_name = '' if 'public' in args: self.public = args['public'] else: self.public = 1 if 'content' in args: self.content = args['content'] else: raise Exception('Gist content can\'t be empty') url = '/gists' data = {"description": self.description, "public": self.public, "files": { self.gist_name: { "content": self.content } } } r = requests.post( '%s%s' % (BASE_URL, url), data=json.dumps(data), headers=self.header ) if (r.status_code == 201): response = { 'Gist-Link': '%s/%s/%s' %(GIST_URL,self.username,r.json()['id']), 'Clone-Link': '%s/%s.git' %(GIST_URL,r.json()['id']), 'Embed-Script': '<script src="%s/%s/%s.js"</script>' %(GIST_URL,self.username,r.json()['id']), 'id': r.json()['id'], 'created_at': r.json()['created_at'], } return response raise Exception('Gist not created: server response was [%s] %s' % (r.status_code, r.text))
simplegist/simplegist.py
2,459
Gist Base Class This class is to used to instantiate the wrapper and authenticate. Authenticate with providing Github Username and API-Token to use it for all future API requests Save our username and api_token (If given) for later use. Set header information in every request.
281
en
0.776873
from pathlib import Path from urllib.parse import urljoin, urlparse, unquote def _get_url_file_name(url): path = urlparse(url).path try: path = unquote(path) except (TypeError, ValueError): pass return Path(path).name def dav_index(context, data): """List files in a WebDAV directory.""" # This is made to work with ownCloud/nextCloud, but some rumor has # it they are "standards compliant" and it should thus work for # other DAV servers. url = data.get("url") context.log.info("Fetching WebDAV path: %s" % url) result = context.http.request("PROPFIND", url) for resp in result.xml.findall("./{DAV:}response"): href = resp.findtext("./{DAV:}href") if href is None: continue child_url = urljoin(url, href) if child_url == url: continue child = dict(data) child["url"] = child_url child["foreign_id"] = child_url child["file_name"] = _get_url_file_name(href) rule = "file" if resp.find(".//{DAV:}collection") is not None: rule = "folder" context.emit(data=child, rule=rule)
memorious/operations/webdav.py
1,167
List files in a WebDAV directory. This is made to work with ownCloud/nextCloud, but some rumor has it they are "standards compliant" and it should thus work for other DAV servers.
181
en
0.967719
""" Name : c9_01_optimize.py Book : Python for Finance (2nd ed.) Publisher: Packt Publishing Ltd. Author : Yuxing Yan Date : 6/6/2017 email : yany@canisius.edu paulyxy@hotmail.com """ from scipy.optimize import minimize def myFunction(x): return (3.2+5*x**2) x0=100 res = minimize(myFunction,x0,method='nelder-mead',options={'xtol':1e-8,'disp': True})
Chapter09/c9_01_optimize.py
402
Name : c9_01_optimize.py Book : Python for Finance (2nd ed.) Publisher: Packt Publishing Ltd. Author : Yuxing Yan Date : 6/6/2017 email : yany@canisius.edu paulyxy@hotmail.com
204
en
0.519011
# coding=utf-8 # Copyright 2020 The Real-World RL Suite Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for evaluators.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import os from absl.testing import absltest from absl.testing import parameterized import numpy as np import realworldrl_suite.environments as rwrl from realworldrl_suite.utils import evaluators class RandomAgent(object): def __init__(self, action_spec): self.action_spec = action_spec def action(self): return np.random.uniform( self.action_spec.minimum, self.action_spec.maximum, size=self.action_spec.shape) class EvaluatorsTest(parameterized.TestCase): def _gen_stats(self, domain_name, task_name): temp_dir = self.create_tempdir() env = rwrl.load( domain_name=domain_name, task_name=task_name, safety_spec={'enable': True}, log_output=os.path.join(temp_dir.full_path, 'test.pickle'), environment_kwargs=dict(log_safety_vars=True, flat_observation=True)) random_policy = RandomAgent(env.action_spec()).action for _ in range(3): timestep = env.step(random_policy()) while not timestep.last(): timestep = env.step(random_policy()) env.write_logs() return env.logs_path @parameterized.named_parameters(*rwrl.ALL_TASKS) def test_loading(self, domain_name, task_name): temp_path = self._gen_stats(domain_name, task_name) data_in = np.load(temp_path, allow_pickle=True) evaluators.Evaluators(data_in) def test_safety_evaluator(self): # TODO(dulacarnold): Make this test general to all envs. temp_path = self._gen_stats( domain_name='cartpole', task_name='realworld_balance') data_in = np.load(temp_path, allow_pickle=True) ev = evaluators.Evaluators(data_in) self.assertLen(ev.get_safety_evaluator(), 3) def test_standard_evaluators(self): # TODO(dulacarnold): Make this test general to all envs. temp_path = self._gen_stats( domain_name='cartpole', task_name='realworld_balance') data_in = np.load(temp_path, allow_pickle=True) ev = evaluators.Evaluators(data_in) self.assertLen(ev.get_standard_evaluators(), 5) @parameterized.named_parameters(*rwrl.ALL_TASKS) def test_safety_plot(self, domain_name, task_name): temp_path = self._gen_stats(domain_name, task_name) data_in = np.load(temp_path, allow_pickle=True) ev = evaluators.Evaluators(data_in) ev.get_safety_plot() if __name__ == '__main__': absltest.main()
realworldrl_suite/utils/evaluators_test.py
3,096
Tests for evaluators. coding=utf-8 Copyright 2020 The Real-World RL Suite Authors. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. TODO(dulacarnold): Make this test general to all envs. TODO(dulacarnold): Make this test general to all envs.
716
en
0.844294
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import random import warnings from dataclasses import dataclass from typing import Any, Callable, Dict, List, NewType, Optional, Tuple, Union import torch from torch.nn.utils.rnn import pad_sequence from ..file_utils import PaddingStrategy from ..modeling_utils import PreTrainedModel from ..models.bert import BertTokenizer, BertTokenizerFast from ..tokenization_utils_base import BatchEncoding, PreTrainedTokenizerBase InputDataClass = NewType("InputDataClass", Any) """ A DataCollator is a function that takes a list of samples from a Dataset and collate them into a batch, as a dictionary of Tensors. """ DataCollator = NewType("DataCollator", Callable[[List[InputDataClass]], Dict[str, torch.Tensor]]) def default_data_collator(features: List[InputDataClass]) -> Dict[str, torch.Tensor]: """ Very simple data collator that simply collates batches of dict-like objects and performs special handling for potential keys named: - ``label``: handles a single value (int or float) per object - ``label_ids``: handles a list of values per object Does not do any additional preprocessing: property names of the input object will be used as corresponding inputs to the model. See glue and ner for example of how it's useful. """ # In this function we'll make the assumption that all `features` in the batch # have the same attributes. # So we will look at the first element as a proxy for what attributes exist # on the whole batch. if not isinstance(features[0], (dict, BatchEncoding)): features = [vars(f) for f in features] first = features[0] batch = {} # Special handling for labels. # Ensure that tensor is created with the correct type # (it should be automatically the case, but let's make sure of it.) if "label" in first and first["label"] is not None: label = first["label"].item() if isinstance(first["label"], torch.Tensor) else first["label"] dtype = torch.long if isinstance(label, int) else torch.float batch["labels"] = torch.tensor([f["label"] for f in features], dtype=dtype) elif "label_ids" in first and first["label_ids"] is not None: if isinstance(first["label_ids"], torch.Tensor): batch["labels"] = torch.stack([f["label_ids"] for f in features]) else: dtype = torch.long if type(first["label_ids"][0]) is int else torch.float batch["labels"] = torch.tensor([f["label_ids"] for f in features], dtype=dtype) # Handling of all other possible keys. # Again, we will use the first element to figure out which key/values are not None for this model. for k, v in first.items(): if k not in ("label", "label_ids") and v is not None and not isinstance(v, str): if isinstance(v, torch.Tensor): batch[k] = torch.stack([f[k] for f in features]) else: batch[k] = torch.tensor([f[k] for f in features]) return batch @dataclass class DataCollatorWithPadding: """ Data collator that will dynamically pad the inputs received. Args: tokenizer (:class:`~transformers.PreTrainedTokenizer` or :class:`~transformers.PreTrainedTokenizerFast`): The tokenizer used for encoding the data. padding (:obj:`bool`, :obj:`str` or :class:`~transformers.file_utils.PaddingStrategy`, `optional`, defaults to :obj:`True`): Select a strategy to pad the returned sequences (according to the model's padding side and padding index) among: * :obj:`True` or :obj:`'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence if provided). * :obj:`'max_length'`: Pad to a maximum length specified with the argument :obj:`max_length` or to the maximum acceptable input length for the model if that argument is not provided. * :obj:`False` or :obj:`'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different lengths). max_length (:obj:`int`, `optional`): Maximum length of the returned list and optionally padding length (see above). pad_to_multiple_of (:obj:`int`, `optional`): If set will pad the sequence to a multiple of the provided value. This is especially useful to enable the use of Tensor Cores on NVIDIA hardware with compute capability >= 7.5 (Volta). """ tokenizer: PreTrainedTokenizerBase padding: Union[bool, str, PaddingStrategy] = True max_length: Optional[int] = None pad_to_multiple_of: Optional[int] = None def __call__(self, features: List[Dict[str, Union[List[int], torch.Tensor]]]) -> Dict[str, torch.Tensor]: batch = self.tokenizer.pad( features, padding=self.padding, max_length=self.max_length, pad_to_multiple_of=self.pad_to_multiple_of, return_tensors="pt", ) if "label" in batch: batch["labels"] = batch["label"] del batch["label"] if "label_ids" in batch: batch["labels"] = batch["label_ids"] del batch["label_ids"] return batch @dataclass class DataCollatorForTokenClassification: """ Data collator that will dynamically pad the inputs received, as well as the labels. Args: tokenizer (:class:`~transformers.PreTrainedTokenizer` or :class:`~transformers.PreTrainedTokenizerFast`): The tokenizer used for encoding the data. padding (:obj:`bool`, :obj:`str` or :class:`~transformers.file_utils.PaddingStrategy`, `optional`, defaults to :obj:`True`): Select a strategy to pad the returned sequences (according to the model's padding side and padding index) among: * :obj:`True` or :obj:`'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence if provided). * :obj:`'max_length'`: Pad to a maximum length specified with the argument :obj:`max_length` or to the maximum acceptable input length for the model if that argument is not provided. * :obj:`False` or :obj:`'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different lengths). max_length (:obj:`int`, `optional`): Maximum length of the returned list and optionally padding length (see above). pad_to_multiple_of (:obj:`int`, `optional`): If set will pad the sequence to a multiple of the provided value. This is especially useful to enable the use of Tensor Cores on NVIDIA hardware with compute capability >= 7.5 (Volta). label_pad_token_id (:obj:`int`, `optional`, defaults to -100): The id to use when padding the labels (-100 will be automatically ignore by PyTorch loss functions). """ tokenizer: PreTrainedTokenizerBase padding: Union[bool, str, PaddingStrategy] = True max_length: Optional[int] = None pad_to_multiple_of: Optional[int] = None label_pad_token_id: int = -100 def __call__(self, features): label_name = "label" if "label" in features[0].keys() else "labels" labels = [feature[label_name] for feature in features] if label_name in features[0].keys() else None batch = self.tokenizer.pad( features, padding=self.padding, max_length=self.max_length, pad_to_multiple_of=self.pad_to_multiple_of, # Conversion to tensors will fail if we have labels as they are not of the same length yet. return_tensors="pt" if labels is None else None, ) if labels is None: return batch sequence_length = torch.tensor(batch["input_ids"]).shape[1] padding_side = self.tokenizer.padding_side if padding_side == "right": batch["labels"] = [label + [self.label_pad_token_id] * (sequence_length - len(label)) for label in labels] else: batch["labels"] = [[self.label_pad_token_id] * (sequence_length - len(label)) + label for label in labels] batch = {k: torch.tensor(v, dtype=torch.int64) for k, v in batch.items()} return batch def _collate_batch(examples, tokenizer, pad_to_multiple_of: Optional[int] = None): """Collate `examples` into a batch, using the information in `tokenizer` for padding if necessary.""" # Tensorize if necessary. if isinstance(examples[0], (list, tuple)): examples = [torch.tensor(e, dtype=torch.long) for e in examples] # Check if padding is necessary. length_of_first = examples[0].size(0) are_tensors_same_length = all(x.size(0) == length_of_first for x in examples) if are_tensors_same_length and (pad_to_multiple_of is None or length_of_first % pad_to_multiple_of == 0): return torch.stack(examples, dim=0) # If yes, check if we have a `pad_token`. if tokenizer._pad_token is None: raise ValueError( "You are attempting to pad samples but the tokenizer you are using" f" ({tokenizer.__class__.__name__}) does not have a pad token." ) # Creating the full tensor and filling it with our data. max_length = max(x.size(0) for x in examples) if pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): max_length = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of result = examples[0].new_full([len(examples), max_length], tokenizer.pad_token_id) for i, example in enumerate(examples): if tokenizer.padding_side == "right": result[i, : example.shape[0]] = example else: result[i, -example.shape[0] :] = example return result def tolist(x: Union[List[Any], torch.Tensor]): return x.tolist() if isinstance(x, torch.Tensor) else x @dataclass class DataCollatorForSeq2Seq: """ Data collator that will dynamically pad the inputs received, as well as the labels. Args: tokenizer (:class:`~transformers.PreTrainedTokenizer` or :class:`~transformers.PreTrainedTokenizerFast`): The tokenizer used for encoding the data. model (:class:`~transformers.PreTrainedModel`): The model that is being trained. If set and has the `prepare_decoder_input_ids_from_labels`, use it to prepare the `decoder_input_ids` This is useful when using `label_smoothing` to avoid calculating loss twice. padding (:obj:`bool`, :obj:`str` or :class:`~transformers.file_utils.PaddingStrategy`, `optional`, defaults to :obj:`True`): Select a strategy to pad the returned sequences (according to the model's padding side and padding index) among: * :obj:`True` or :obj:`'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence is provided). * :obj:`'max_length'`: Pad to a maximum length specified with the argument :obj:`max_length` or to the maximum acceptable input length for the model if that argument is not provided. * :obj:`False` or :obj:`'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different lengths). max_length (:obj:`int`, `optional`): Maximum length of the returned list and optionally padding length (see above). pad_to_multiple_of (:obj:`int`, `optional`): If set will pad the sequence to a multiple of the provided value. This is especially useful to enable the use of Tensor Cores on NVIDIA hardware with compute capability >= 7.5 (Volta). label_pad_token_id (:obj:`int`, `optional`, defaults to -100): The id to use when padding the labels (-100 will be automatically ignored by PyTorch loss functions). """ tokenizer: PreTrainedTokenizerBase model: Optional[PreTrainedModel] = None padding: Union[bool, str, PaddingStrategy] = True max_length: Optional[int] = None pad_to_multiple_of: Optional[int] = None label_pad_token_id: int = -100 def __call__(self, features): labels = [feature["labels"] for feature in features] if "labels" in features[0].keys() else None # We have to pad the labels before calling `tokenizer.pad` as this method won't pad them and needs them of the # same length to return tensors. if labels is not None: max_label_length = max(len(l) for l in labels) padding_side = self.tokenizer.padding_side for feature in features: remainder = [self.label_pad_token_id] * (max_label_length - len(feature["labels"])) feature["labels"] = ( feature["labels"] + remainder if padding_side == "right" else remainder + feature["labels"] ) features = self.tokenizer.pad( features, padding=self.padding, max_length=self.max_length, pad_to_multiple_of=self.pad_to_multiple_of, return_tensors="pt", ) # prepare decoder_input_ids if self.model is not None and hasattr(self.model, "prepare_decoder_input_ids_from_labels"): decoder_input_ids = self.model.prepare_decoder_input_ids_from_labels(labels=features["labels"]) features["decoder_input_ids"] = decoder_input_ids return features @dataclass class DataCollatorForLanguageModeling: """ Data collator used for language modeling. Inputs are dynamically padded to the maximum length of a batch if they are not all of the same length. Args: tokenizer (:class:`~transformers.PreTrainedTokenizer` or :class:`~transformers.PreTrainedTokenizerFast`): The tokenizer used for encoding the data. mlm (:obj:`bool`, `optional`, defaults to :obj:`True`): Whether or not to use masked language modeling. If set to :obj:`False`, the labels are the same as the inputs with the padding tokens ignored (by setting them to -100). Otherwise, the labels are -100 for non-masked tokens and the value to predict for the masked token. mlm_probability (:obj:`float`, `optional`, defaults to 0.15): The probability with which to (randomly) mask tokens in the input, when :obj:`mlm` is set to :obj:`True`. pad_to_multiple_of (:obj:`int`, `optional`): If set will pad the sequence to a multiple of the provided value. .. note:: For best performance, this data collator should be used with a dataset having items that are dictionaries or BatchEncoding, with the :obj:`"special_tokens_mask"` key, as returned by a :class:`~transformers.PreTrainedTokenizer` or a :class:`~transformers.PreTrainedTokenizerFast` with the argument :obj:`return_special_tokens_mask=True`. """ tokenizer: PreTrainedTokenizerBase mlm: bool = True mlm_probability: float = 0.15 pad_to_multiple_of: Optional[int] = None def __post_init__(self): if self.mlm and self.tokenizer.mask_token is None: raise ValueError( "This tokenizer does not have a mask token which is necessary for masked language modeling. " "You should pass `mlm=False` to train on causal language modeling instead." ) def __call__( self, examples: List[Union[List[int], torch.Tensor, Dict[str, torch.Tensor]]] ) -> Dict[str, torch.Tensor]: # Handle dict or lists with proper padding and conversion to tensor. if isinstance(examples[0], (dict, BatchEncoding)): batch = self.tokenizer.pad(examples, return_tensors="pt", pad_to_multiple_of=self.pad_to_multiple_of) else: batch = {"input_ids": _collate_batch(examples, self.tokenizer, pad_to_multiple_of=self.pad_to_multiple_of)} # If special token mask has been preprocessed, pop it from the dict. special_tokens_mask = batch.pop("special_tokens_mask", None) if self.mlm: batch["input_ids"], batch["labels"] = self.mask_tokens( batch["input_ids"], special_tokens_mask=special_tokens_mask ) else: labels = batch["input_ids"].clone() if self.tokenizer.pad_token_id is not None: labels[labels == self.tokenizer.pad_token_id] = -100 batch["labels"] = labels return batch def mask_tokens( self, inputs: torch.Tensor, special_tokens_mask: Optional[torch.Tensor] = None ) -> Tuple[torch.Tensor, torch.Tensor]: """ Prepare masked tokens inputs/labels for masked language modeling: 80% MASK, 10% random, 10% original. """ labels = inputs.clone() # We sample a few tokens in each sequence for MLM training (with probability `self.mlm_probability`) probability_matrix = torch.full(labels.shape, self.mlm_probability) if special_tokens_mask is None: special_tokens_mask = [ self.tokenizer.get_special_tokens_mask(val, already_has_special_tokens=True) for val in labels.tolist() ] special_tokens_mask = torch.tensor(special_tokens_mask, dtype=torch.bool) else: special_tokens_mask = special_tokens_mask.bool() probability_matrix.masked_fill_(special_tokens_mask, value=0.0) masked_indices = torch.bernoulli(probability_matrix).bool() labels[~masked_indices] = -100 # We only compute loss on masked tokens # 80% of the time, we replace masked input tokens with tokenizer.mask_token ([MASK]) indices_replaced = torch.bernoulli(torch.full(labels.shape, 0.8)).bool() & masked_indices inputs[indices_replaced] = self.tokenizer.convert_tokens_to_ids(self.tokenizer.mask_token) # 10% of the time, we replace masked input tokens with random word indices_random = torch.bernoulli(torch.full(labels.shape, 0.5)).bool() & masked_indices & ~indices_replaced random_words = torch.randint(len(self.tokenizer), labels.shape, dtype=torch.long) inputs[indices_random] = random_words[indices_random] # The rest of the time (10% of the time) we keep the masked input tokens unchanged return inputs, labels @dataclass class DataCollatorForWholeWordMask(DataCollatorForLanguageModeling): """ Data collator used for language modeling that masks entire words. - collates batches of tensors, honoring their tokenizer's pad_token - preprocesses batches for masked language modeling .. note:: This collator relies on details of the implementation of subword tokenization by :class:`~transformers.BertTokenizer`, specifically that subword tokens are prefixed with `##`. For tokenizers that do not adhere to this scheme, this collator will produce an output that is roughly equivalent to :class:`.DataCollatorForLanguageModeling`. """ def __call__( self, examples: List[Union[List[int], torch.Tensor, Dict[str, torch.Tensor]]] ) -> Dict[str, torch.Tensor]: if isinstance(examples[0], (dict, BatchEncoding)): input_ids = [e["input_ids"] for e in examples] else: input_ids = examples examples = [{"input_ids": e} for e in examples] batch_input = _collate_batch(input_ids, self.tokenizer) mask_labels = [] for e in examples: ref_tokens = [] for id in tolist(e["input_ids"]): token = self.tokenizer._convert_id_to_token(id) ref_tokens.append(token) # For Chinese tokens, we need extra inf to mark sub-word, e.g [喜,欢]-> [喜,##欢] if "chinese_ref" in e: ref_pos = tolist(e["chinese_ref"]) len_seq = len(e["input_ids"]) for i in range(len_seq): if i in ref_pos: ref_tokens[i] = "##" + ref_tokens[i] mask_labels.append(self._whole_word_mask(ref_tokens)) batch_mask = _collate_batch(mask_labels, self.tokenizer) inputs, labels = self.mask_tokens(batch_input, batch_mask) return {"input_ids": inputs, "labels": labels} def _whole_word_mask(self, input_tokens: List[str], max_predictions=512): """ Get 0/1 labels for masked tokens with whole word mask proxy """ if not isinstance(self.tokenizer, (BertTokenizer, BertTokenizerFast)): warnings.warn( "DataCollatorForWholeWordMask is only suitable for BertTokenizer-like tokenizers." "Please refer to the documentation for more information." ) cand_indexes = [] for (i, token) in enumerate(input_tokens): if token == "[CLS]" or token == "[SEP]": continue if len(cand_indexes) >= 1 and token.startswith("##"): cand_indexes[-1].append(i) else: cand_indexes.append([i]) random.shuffle(cand_indexes) num_to_predict = min(max_predictions, max(1, int(round(len(input_tokens) * self.mlm_probability)))) masked_lms = [] covered_indexes = set() for index_set in cand_indexes: if len(masked_lms) >= num_to_predict: break # If adding a whole-word mask would exceed the maximum number of # predictions, then just skip this candidate. if len(masked_lms) + len(index_set) > num_to_predict: continue is_any_index_covered = False for index in index_set: if index in covered_indexes: is_any_index_covered = True break if is_any_index_covered: continue for index in index_set: covered_indexes.add(index) masked_lms.append(index) assert len(covered_indexes) == len(masked_lms) mask_labels = [1 if i in covered_indexes else 0 for i in range(len(input_tokens))] return mask_labels def mask_tokens(self, inputs: torch.Tensor, mask_labels: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: """ Prepare masked tokens inputs/labels for masked language modeling: 80% MASK, 10% random, 10% original. Set 'mask_labels' means we use whole word mask (wwm), we directly mask idxs according to it's ref. """ if self.tokenizer.mask_token is None: raise ValueError( "This tokenizer does not have a mask token which is necessary for masked language modeling. Remove the --mlm flag if you want to use this tokenizer." ) labels = inputs.clone() # We sample a few tokens in each sequence for masked-LM training (with probability args.mlm_probability defaults to 0.15 in Bert/RoBERTa) probability_matrix = mask_labels special_tokens_mask = [ self.tokenizer.get_special_tokens_mask(val, already_has_special_tokens=True) for val in labels.tolist() ] probability_matrix.masked_fill_(torch.tensor(special_tokens_mask, dtype=torch.bool), value=0.0) if self.tokenizer._pad_token is not None: padding_mask = labels.eq(self.tokenizer.pad_token_id) probability_matrix.masked_fill_(padding_mask, value=0.0) masked_indices = probability_matrix.bool() labels[~masked_indices] = -100 # We only compute loss on masked tokens # 80% of the time, we replace masked input tokens with tokenizer.mask_token ([MASK]) indices_replaced = torch.bernoulli(torch.full(labels.shape, 0.8)).bool() & masked_indices inputs[indices_replaced] = self.tokenizer.convert_tokens_to_ids(self.tokenizer.mask_token) # 10% of the time, we replace masked input tokens with random word indices_random = torch.bernoulli(torch.full(labels.shape, 0.5)).bool() & masked_indices & ~indices_replaced random_words = torch.randint(len(self.tokenizer), labels.shape, dtype=torch.long) inputs[indices_random] = random_words[indices_random] # The rest of the time (10% of the time) we keep the masked input tokens unchanged return inputs, labels @dataclass class DataCollatorForSOP(DataCollatorForLanguageModeling): """ Data collator used for sentence order prediction task. - collates batches of tensors, honoring their tokenizer's pad_token - preprocesses batches for both masked language modeling and sentence order prediction """ def __init__(self, *args, **kwargs): warnings.warn( "DataCollatorForSOP is deprecated and will be removed in a future version, you can now use " "DataCollatorForLanguageModeling instead.", FutureWarning, ) def __call__(self, examples: List[Dict[str, torch.Tensor]]) -> Dict[str, torch.Tensor]: input_ids = [example["input_ids"] for example in examples] input_ids = _collate_batch(input_ids, self.tokenizer) input_ids, labels, attention_mask = self.mask_tokens(input_ids) token_type_ids = [example["token_type_ids"] for example in examples] # size of segment_ids varied because randomness, padding zero to the end as the original implementation token_type_ids = pad_sequence(token_type_ids, batch_first=True, padding_value=self.tokenizer.pad_token_id) sop_label_list = [example["sentence_order_label"] for example in examples] sentence_order_label = torch.stack(sop_label_list) return { "input_ids": input_ids, "labels": labels, "attention_mask": attention_mask, "token_type_ids": token_type_ids, "sentence_order_label": sentence_order_label, } def mask_tokens(self, inputs: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ Prepare masked tokens inputs/labels/attention_mask for masked language modeling: 80% MASK, 10% random, 10% original. N-gram not applied yet. """ if self.tokenizer.mask_token is None: raise ValueError( "This tokenizer does not have a mask token which is necessary for masked language modeling. Remove the --mlm flag if you want to use this tokenizer." ) labels = inputs.clone() # We sample a few tokens in each sequence for masked-LM training (with probability args.mlm_probability defaults to 0.15 in Bert/RoBERTa) probability_matrix = torch.full(labels.shape, self.mlm_probability) special_tokens_mask = [ self.tokenizer.get_special_tokens_mask(val, already_has_special_tokens=True) for val in labels.tolist() ] probability_matrix.masked_fill_(torch.tensor(special_tokens_mask, dtype=torch.bool), value=0.0) if self.tokenizer._pad_token is not None: padding_mask = labels.eq(self.tokenizer.pad_token_id) probability_matrix.masked_fill_(padding_mask, value=0.0) masked_indices = torch.bernoulli(probability_matrix).bool() # probability be `1` (masked), however in albert model attention mask `0` means masked, revert the value attention_mask = (~masked_indices).float() if self.tokenizer._pad_token is not None: attention_padding_mask = labels.eq(self.tokenizer.pad_token_id) attention_mask.masked_fill_(attention_padding_mask, value=1.0) labels[~masked_indices] = -100 # We only compute loss on masked tokens, -100 is default for CE compute # 80% of the time, we replace masked input tokens with tokenizer.mask_token ([MASK]) indices_replaced = torch.bernoulli(torch.full(labels.shape, 0.8)).bool() & masked_indices inputs[indices_replaced] = self.tokenizer.convert_tokens_to_ids(self.tokenizer.mask_token) # 10% of the time, we replace masked input tokens with random word indices_random = torch.bernoulli(torch.full(labels.shape, 0.5)).bool() & masked_indices & ~indices_replaced random_words = torch.randint(len(self.tokenizer), labels.shape, dtype=torch.long) inputs[indices_random] = random_words[indices_random] # The rest of the time (10% of the time) we keep the masked input tokens unchanged return inputs, labels, attention_mask @dataclass class DataCollatorForPermutationLanguageModeling: """ Data collator used for permutation language modeling. - collates batches of tensors, honoring their tokenizer's pad_token - preprocesses batches for permutation language modeling with procedures specific to XLNet """ tokenizer: PreTrainedTokenizerBase plm_probability: float = 1 / 6 max_span_length: int = 5 # maximum length of a span of masked tokens def __call__( self, examples: List[Union[List[int], torch.Tensor, Dict[str, torch.Tensor]]] ) -> Dict[str, torch.Tensor]: if isinstance(examples[0], (dict, BatchEncoding)): examples = [e["input_ids"] for e in examples] batch = _collate_batch(examples, self.tokenizer) inputs, perm_mask, target_mapping, labels = self.mask_tokens(batch) return {"input_ids": inputs, "perm_mask": perm_mask, "target_mapping": target_mapping, "labels": labels} def mask_tokens(self, inputs: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]: """ The masked tokens to be predicted for a particular sequence are determined by the following algorithm: 0. Start from the beginning of the sequence by setting ``cur_len = 0`` (number of tokens processed so far). 1. Sample a ``span_length`` from the interval ``[1, max_span_length]`` (length of span of tokens to be masked) 2. Reserve a context of length ``context_length = span_length / plm_probability`` to surround span to be masked 3. Sample a starting point ``start_index`` from the interval ``[cur_len, cur_len + context_length - span_length]`` and mask tokens ``start_index:start_index + span_length`` 4. Set ``cur_len = cur_len + context_length``. If ``cur_len < max_len`` (i.e. there are tokens remaining in the sequence to be processed), repeat from Step 1. """ if self.tokenizer.mask_token is None: raise ValueError( "This tokenizer does not have a mask token which is necessary for permutation language modeling. Please add a mask token if you want to use this tokenizer." ) if inputs.size(1) % 2 != 0: raise ValueError( "This collator requires that sequence lengths be even to create a leakage-free perm_mask. Please see relevant comments in source code for details." ) labels = inputs.clone() # Creating the mask and target_mapping tensors masked_indices = torch.full(labels.shape, 0, dtype=torch.bool) target_mapping = torch.zeros((labels.size(0), labels.size(1), labels.size(1)), dtype=torch.float32) for i in range(labels.size(0)): # Start from the beginning of the sequence by setting `cur_len = 0` (number of tokens processed so far). cur_len = 0 max_len = labels.size(1) while cur_len < max_len: # Sample a `span_length` from the interval `[1, max_span_length]` (length of span of tokens to be masked) span_length = torch.randint(1, self.max_span_length + 1, (1,)).item() # Reserve a context of length `context_length = span_length / plm_probability` to surround the span to be masked context_length = int(span_length / self.plm_probability) # Sample a starting point `start_index` from the interval `[cur_len, cur_len + context_length - span_length]` and mask tokens `start_index:start_index + span_length` start_index = cur_len + torch.randint(context_length - span_length + 1, (1,)).item() masked_indices[i, start_index : start_index + span_length] = 1 # Set `cur_len = cur_len + context_length` cur_len += context_length # Since we're replacing non-masked tokens with -100 in the labels tensor instead of skipping them altogether, # the i-th predict corresponds to the i-th token. target_mapping[i] = torch.eye(labels.size(1)) special_tokens_mask = torch.tensor( [self.tokenizer.get_special_tokens_mask(val, already_has_special_tokens=True) for val in labels.tolist()], dtype=torch.bool, ) masked_indices.masked_fill_(special_tokens_mask, value=0.0) if self.tokenizer._pad_token is not None: padding_mask = labels.eq(self.tokenizer.pad_token_id) masked_indices.masked_fill_(padding_mask, value=0.0) # Mask indicating non-functional tokens, where functional tokens are [SEP], [CLS], padding, etc. non_func_mask = ~(padding_mask | special_tokens_mask) inputs[masked_indices] = self.tokenizer.mask_token_id labels[~masked_indices] = -100 # We only compute loss on masked tokens perm_mask = torch.zeros((labels.size(0), labels.size(1), labels.size(1)), dtype=torch.float32) for i in range(labels.size(0)): # Generate permutation indices i.e. sample a random factorisation order for the sequence. This will # determine which tokens a given token can attend to (encoded in `perm_mask`). # Note: Length of token sequence being permuted has to be less than or equal to reused sequence length # (see documentation for `mems`), otherwise information may leak through due to reuse. In this implementation, # we assume that reused length is half of sequence length and permutation length is equal to reused length. # This requires that the sequence length be even. # Create a linear factorisation order perm_index = torch.arange(labels.size(1)) # Split this into two halves, assuming that half the sequence is reused each time perm_index = perm_index.reshape((-1, labels.size(1) // 2)).transpose(0, 1) # Permute the two halves such that they do not cross over perm_index = perm_index[torch.randperm(labels.size(1) // 2)] # Flatten this out into the desired permuted factorisation order perm_index = torch.flatten(perm_index.transpose(0, 1)) # Set the permutation indices of non-masked (non-functional) tokens to the # smallest index (-1) so that: # (1) They can be seen by all other positions # (2) They cannot see masked positions, so there won't be information leak perm_index.masked_fill_(~masked_indices[i] & non_func_mask[i], -1) # The logic for whether the i-th token can attend on the j-th token based on the factorisation order: # 0 (can attend): If perm_index[i] > perm_index[j] or j is neither masked nor a functional token # 1 (cannot attend): If perm_index[i] <= perm_index[j] and j is either masked or a functional token perm_mask[i] = ( perm_index.reshape((labels.size(1), 1)) <= perm_index.reshape((1, labels.size(1))) ) & masked_indices[i] return inputs.long(), perm_mask, target_mapping, labels.long()
src/transformers/data/data_collator.py
36,353
Data collator used for language modeling. Inputs are dynamically padded to the maximum length of a batch if they are not all of the same length. Args: tokenizer (:class:`~transformers.PreTrainedTokenizer` or :class:`~transformers.PreTrainedTokenizerFast`): The tokenizer used for encoding the data. mlm (:obj:`bool`, `optional`, defaults to :obj:`True`): Whether or not to use masked language modeling. If set to :obj:`False`, the labels are the same as the inputs with the padding tokens ignored (by setting them to -100). Otherwise, the labels are -100 for non-masked tokens and the value to predict for the masked token. mlm_probability (:obj:`float`, `optional`, defaults to 0.15): The probability with which to (randomly) mask tokens in the input, when :obj:`mlm` is set to :obj:`True`. pad_to_multiple_of (:obj:`int`, `optional`): If set will pad the sequence to a multiple of the provided value. .. note:: For best performance, this data collator should be used with a dataset having items that are dictionaries or BatchEncoding, with the :obj:`"special_tokens_mask"` key, as returned by a :class:`~transformers.PreTrainedTokenizer` or a :class:`~transformers.PreTrainedTokenizerFast` with the argument :obj:`return_special_tokens_mask=True`. Data collator used for permutation language modeling. - collates batches of tensors, honoring their tokenizer's pad_token - preprocesses batches for permutation language modeling with procedures specific to XLNet Data collator used for sentence order prediction task. - collates batches of tensors, honoring their tokenizer's pad_token - preprocesses batches for both masked language modeling and sentence order prediction Data collator that will dynamically pad the inputs received, as well as the labels. Args: tokenizer (:class:`~transformers.PreTrainedTokenizer` or :class:`~transformers.PreTrainedTokenizerFast`): The tokenizer used for encoding the data. model (:class:`~transformers.PreTrainedModel`): The model that is being trained. If set and has the `prepare_decoder_input_ids_from_labels`, use it to prepare the `decoder_input_ids` This is useful when using `label_smoothing` to avoid calculating loss twice. padding (:obj:`bool`, :obj:`str` or :class:`~transformers.file_utils.PaddingStrategy`, `optional`, defaults to :obj:`True`): Select a strategy to pad the returned sequences (according to the model's padding side and padding index) among: * :obj:`True` or :obj:`'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence is provided). * :obj:`'max_length'`: Pad to a maximum length specified with the argument :obj:`max_length` or to the maximum acceptable input length for the model if that argument is not provided. * :obj:`False` or :obj:`'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different lengths). max_length (:obj:`int`, `optional`): Maximum length of the returned list and optionally padding length (see above). pad_to_multiple_of (:obj:`int`, `optional`): If set will pad the sequence to a multiple of the provided value. This is especially useful to enable the use of Tensor Cores on NVIDIA hardware with compute capability >= 7.5 (Volta). label_pad_token_id (:obj:`int`, `optional`, defaults to -100): The id to use when padding the labels (-100 will be automatically ignored by PyTorch loss functions). Data collator that will dynamically pad the inputs received, as well as the labels. Args: tokenizer (:class:`~transformers.PreTrainedTokenizer` or :class:`~transformers.PreTrainedTokenizerFast`): The tokenizer used for encoding the data. padding (:obj:`bool`, :obj:`str` or :class:`~transformers.file_utils.PaddingStrategy`, `optional`, defaults to :obj:`True`): Select a strategy to pad the returned sequences (according to the model's padding side and padding index) among: * :obj:`True` or :obj:`'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence if provided). * :obj:`'max_length'`: Pad to a maximum length specified with the argument :obj:`max_length` or to the maximum acceptable input length for the model if that argument is not provided. * :obj:`False` or :obj:`'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different lengths). max_length (:obj:`int`, `optional`): Maximum length of the returned list and optionally padding length (see above). pad_to_multiple_of (:obj:`int`, `optional`): If set will pad the sequence to a multiple of the provided value. This is especially useful to enable the use of Tensor Cores on NVIDIA hardware with compute capability >= 7.5 (Volta). label_pad_token_id (:obj:`int`, `optional`, defaults to -100): The id to use when padding the labels (-100 will be automatically ignore by PyTorch loss functions). Data collator used for language modeling that masks entire words. - collates batches of tensors, honoring their tokenizer's pad_token - preprocesses batches for masked language modeling .. note:: This collator relies on details of the implementation of subword tokenization by :class:`~transformers.BertTokenizer`, specifically that subword tokens are prefixed with `##`. For tokenizers that do not adhere to this scheme, this collator will produce an output that is roughly equivalent to :class:`.DataCollatorForLanguageModeling`. Data collator that will dynamically pad the inputs received. Args: tokenizer (:class:`~transformers.PreTrainedTokenizer` or :class:`~transformers.PreTrainedTokenizerFast`): The tokenizer used for encoding the data. padding (:obj:`bool`, :obj:`str` or :class:`~transformers.file_utils.PaddingStrategy`, `optional`, defaults to :obj:`True`): Select a strategy to pad the returned sequences (according to the model's padding side and padding index) among: * :obj:`True` or :obj:`'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence if provided). * :obj:`'max_length'`: Pad to a maximum length specified with the argument :obj:`max_length` or to the maximum acceptable input length for the model if that argument is not provided. * :obj:`False` or :obj:`'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different lengths). max_length (:obj:`int`, `optional`): Maximum length of the returned list and optionally padding length (see above). pad_to_multiple_of (:obj:`int`, `optional`): If set will pad the sequence to a multiple of the provided value. This is especially useful to enable the use of Tensor Cores on NVIDIA hardware with compute capability >= 7.5 (Volta). Collate `examples` into a batch, using the information in `tokenizer` for padding if necessary. Get 0/1 labels for masked tokens with whole word mask proxy Very simple data collator that simply collates batches of dict-like objects and performs special handling for potential keys named: - ``label``: handles a single value (int or float) per object - ``label_ids``: handles a list of values per object Does not do any additional preprocessing: property names of the input object will be used as corresponding inputs to the model. See glue and ner for example of how it's useful. Prepare masked tokens inputs/labels for masked language modeling: 80% MASK, 10% random, 10% original. Prepare masked tokens inputs/labels for masked language modeling: 80% MASK, 10% random, 10% original. Set 'mask_labels' means we use whole word mask (wwm), we directly mask idxs according to it's ref. Prepare masked tokens inputs/labels/attention_mask for masked language modeling: 80% MASK, 10% random, 10% original. N-gram not applied yet. The masked tokens to be predicted for a particular sequence are determined by the following algorithm: 0. Start from the beginning of the sequence by setting ``cur_len = 0`` (number of tokens processed so far). 1. Sample a ``span_length`` from the interval ``[1, max_span_length]`` (length of span of tokens to be masked) 2. Reserve a context of length ``context_length = span_length / plm_probability`` to surround span to be masked 3. Sample a starting point ``start_index`` from the interval ``[cur_len, cur_len + context_length - span_length]`` and mask tokens ``start_index:start_index + span_length`` 4. Set ``cur_len = cur_len + context_length``. If ``cur_len < max_len`` (i.e. there are tokens remaining in the sequence to be processed), repeat from Step 1. Copyright 2020 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. In this function we'll make the assumption that all `features` in the batch have the same attributes. So we will look at the first element as a proxy for what attributes exist on the whole batch. Special handling for labels. Ensure that tensor is created with the correct type (it should be automatically the case, but let's make sure of it.) Handling of all other possible keys. Again, we will use the first element to figure out which key/values are not None for this model. Conversion to tensors will fail if we have labels as they are not of the same length yet. Tensorize if necessary. Check if padding is necessary. If yes, check if we have a `pad_token`. Creating the full tensor and filling it with our data. We have to pad the labels before calling `tokenizer.pad` as this method won't pad them and needs them of the same length to return tensors. prepare decoder_input_ids Handle dict or lists with proper padding and conversion to tensor. If special token mask has been preprocessed, pop it from the dict. We sample a few tokens in each sequence for MLM training (with probability `self.mlm_probability`) We only compute loss on masked tokens 80% of the time, we replace masked input tokens with tokenizer.mask_token ([MASK]) 10% of the time, we replace masked input tokens with random word The rest of the time (10% of the time) we keep the masked input tokens unchanged For Chinese tokens, we need extra inf to mark sub-word, e.g [喜,欢]-> [喜,欢] If adding a whole-word mask would exceed the maximum number of predictions, then just skip this candidate. We sample a few tokens in each sequence for masked-LM training (with probability args.mlm_probability defaults to 0.15 in Bert/RoBERTa) We only compute loss on masked tokens 80% of the time, we replace masked input tokens with tokenizer.mask_token ([MASK]) 10% of the time, we replace masked input tokens with random word The rest of the time (10% of the time) we keep the masked input tokens unchanged size of segment_ids varied because randomness, padding zero to the end as the original implementation We sample a few tokens in each sequence for masked-LM training (with probability args.mlm_probability defaults to 0.15 in Bert/RoBERTa) probability be `1` (masked), however in albert model attention mask `0` means masked, revert the value We only compute loss on masked tokens, -100 is default for CE compute 80% of the time, we replace masked input tokens with tokenizer.mask_token ([MASK]) 10% of the time, we replace masked input tokens with random word The rest of the time (10% of the time) we keep the masked input tokens unchanged maximum length of a span of masked tokens Creating the mask and target_mapping tensors Start from the beginning of the sequence by setting `cur_len = 0` (number of tokens processed so far). Sample a `span_length` from the interval `[1, max_span_length]` (length of span of tokens to be masked) Reserve a context of length `context_length = span_length / plm_probability` to surround the span to be masked Sample a starting point `start_index` from the interval `[cur_len, cur_len + context_length - span_length]` and mask tokens `start_index:start_index + span_length` Set `cur_len = cur_len + context_length` Since we're replacing non-masked tokens with -100 in the labels tensor instead of skipping them altogether, the i-th predict corresponds to the i-th token. Mask indicating non-functional tokens, where functional tokens are [SEP], [CLS], padding, etc. We only compute loss on masked tokens Generate permutation indices i.e. sample a random factorisation order for the sequence. This will determine which tokens a given token can attend to (encoded in `perm_mask`). Note: Length of token sequence being permuted has to be less than or equal to reused sequence length (see documentation for `mems`), otherwise information may leak through due to reuse. In this implementation, we assume that reused length is half of sequence length and permutation length is equal to reused length. This requires that the sequence length be even. Create a linear factorisation order Split this into two halves, assuming that half the sequence is reused each time Permute the two halves such that they do not cross over Flatten this out into the desired permuted factorisation order Set the permutation indices of non-masked (non-functional) tokens to the smallest index (-1) so that: (1) They can be seen by all other positions (2) They cannot see masked positions, so there won't be information leak The logic for whether the i-th token can attend on the j-th token based on the factorisation order: 0 (can attend): If perm_index[i] > perm_index[j] or j is neither masked nor a functional token 1 (cannot attend): If perm_index[i] <= perm_index[j] and j is either masked or a functional token
14,317
en
0.767689
# -*- coding: utf-8 -*- ########################################################################### # Copyright (c), The AiiDA team. All rights reserved. # # This file is part of the AiiDA code. # # # # The code is hosted on GitHub at https://github.com/aiidateam/aiida-core # # For further information on the license, see the LICENSE.txt file # # For further information please visit http://www.aiida.net # ########################################################################### """Tests for the `verdi group` command.""" from aiida import orm from aiida.backends.testbase import AiidaTestCase from aiida.common import exceptions from aiida.cmdline.commands.cmd_group import ( group_list, group_create, group_delete, group_relabel, group_description, group_add_nodes, group_remove_nodes, group_show, group_copy ) class TestVerdiGroup(AiidaTestCase): """Tests for the `verdi group` command.""" @classmethod def setUpClass(cls, *args, **kwargs): super().setUpClass(*args, **kwargs) for group in ['dummygroup1', 'dummygroup2', 'dummygroup3', 'dummygroup4']: orm.Group(label=group).store() def setUp(self): """Create runner object to run tests.""" from click.testing import CliRunner self.cli_runner = CliRunner() def test_help(self): """Tests help text for all group sub commands.""" options = ['--help'] # verdi group list result = self.cli_runner.invoke(group_list, options) self.assertIsNone(result.exception, result.output) self.assertIn('Usage', result.output) # verdi group create result = self.cli_runner.invoke(group_create, options) self.assertIsNone(result.exception, result.output) self.assertIn('Usage', result.output) # verdi group delete result = self.cli_runner.invoke(group_delete, options) self.assertIsNone(result.exception, result.output) self.assertIn('Usage', result.output) # verdi group relabel result = self.cli_runner.invoke(group_relabel, options) self.assertIsNone(result.exception, result.output) self.assertIn('Usage', result.output) # verdi group description result = self.cli_runner.invoke(group_description, options) self.assertIsNone(result.exception, result.output) self.assertIn('Usage', result.output) # verdi group addnodes result = self.cli_runner.invoke(group_add_nodes, options) self.assertIsNone(result.exception, result.output) self.assertIn('Usage', result.output) # verdi group removenodes result = self.cli_runner.invoke(group_remove_nodes, options) self.assertIsNone(result.exception, result.output) self.assertIn('Usage', result.output) # verdi group show result = self.cli_runner.invoke(group_show, options) self.assertIsNone(result.exception, result.output) self.assertIn('Usage', result.output) # verdi group copy result = self.cli_runner.invoke(group_copy, options) self.assertIsNone(result.exception, result.output) self.assertIn('Usage', result.output) def test_create(self): """Test `verdi group create` command.""" result = self.cli_runner.invoke(group_create, ['dummygroup5']) self.assertClickResultNoException(result) # check if newly added group in present in list result = self.cli_runner.invoke(group_list) self.assertClickResultNoException(result) self.assertIn('dummygroup5', result.output) def test_list(self): """Test `verdi group list` command.""" result = self.cli_runner.invoke(group_list) self.assertClickResultNoException(result) for grp in ['dummygroup1', 'dummygroup2']: self.assertIn(grp, result.output) def test_copy(self): """Test `verdi group copy` command.""" result = self.cli_runner.invoke(group_copy, ['dummygroup1', 'dummygroup2']) self.assertClickResultNoException(result) self.assertIn('Success', result.output) def test_delete(self): """Test `verdi group delete` command.""" orm.Group(label='group_test_delete_01').store() orm.Group(label='group_test_delete_02').store() result = self.cli_runner.invoke(group_delete, ['--force', 'group_test_delete_01']) self.assertClickResultNoException(result) # Verify that removed group is not present in list result = self.cli_runner.invoke(group_list) self.assertClickResultNoException(result) self.assertNotIn('group_test_delete_01', result.output) node_01 = orm.CalculationNode().store() node_02 = orm.CalculationNode().store() # Add some nodes and then use `verdi group delete --clear` to delete a node even when it contains nodes group = orm.load_group(label='group_test_delete_02') group.add_nodes([node_01, node_02]) self.assertEqual(group.count(), 2) # Calling delete on a group without the `--clear` option should raise result = self.cli_runner.invoke(group_delete, ['--force', 'group_test_delete_02']) self.assertIsNotNone(result.exception, result.output) # With `--clear` option should delete group and nodes result = self.cli_runner.invoke(group_delete, ['--force', '--clear', 'group_test_delete_02']) self.assertClickResultNoException(result) with self.assertRaises(exceptions.NotExistent): group = orm.load_group(label='group_test_delete_02') def test_show(self): """Test `verdi group show` command.""" result = self.cli_runner.invoke(group_show, ['dummygroup1']) self.assertClickResultNoException(result) for grpline in [ 'Group label', 'dummygroup1', 'Group type_string', 'user', 'Group description', '<no description>' ]: self.assertIn(grpline, result.output) def test_description(self): """Test `verdi group description` command.""" description = 'It is a new description' group = orm.load_group(label='dummygroup2') self.assertNotEqual(group.description, description) # Change the description of the group result = self.cli_runner.invoke(group_description, [group.label, description]) self.assertClickResultNoException(result) self.assertEqual(group.description, description) # When no description argument is passed the command should just echo the current description result = self.cli_runner.invoke(group_description, [group.label]) self.assertClickResultNoException(result) self.assertIn(description, result.output) def test_relabel(self): """Test `verdi group relabel` command.""" result = self.cli_runner.invoke(group_relabel, ['dummygroup4', 'relabeled_group']) self.assertIsNone(result.exception, result.output) # check if group list command shows changed group name result = self.cli_runner.invoke(group_list) self.assertClickResultNoException(result) self.assertNotIn('dummygroup4', result.output) self.assertIn('relabeled_group', result.output) def test_add_remove_nodes(self): """Test `verdi group remove-nodes` command.""" node_01 = orm.CalculationNode().store() node_02 = orm.CalculationNode().store() node_03 = orm.CalculationNode().store() result = self.cli_runner.invoke(group_add_nodes, ['--force', '--group=dummygroup1', node_01.uuid]) self.assertClickResultNoException(result) # Check if node is added in group using group show command result = self.cli_runner.invoke(group_show, ['dummygroup1']) self.assertClickResultNoException(result) self.assertIn('CalculationNode', result.output) self.assertIn(str(node_01.pk), result.output) # Remove same node result = self.cli_runner.invoke(group_remove_nodes, ['--force', '--group=dummygroup1', node_01.uuid]) self.assertIsNone(result.exception, result.output) # Check if node is added in group using group show command result = self.cli_runner.invoke(group_show, ['-r', 'dummygroup1']) self.assertClickResultNoException(result) self.assertNotIn('CalculationNode', result.output) self.assertNotIn(str(node_01.pk), result.output) # Add all three nodes and then use `verdi group remove-nodes --clear` to remove them all group = orm.load_group(label='dummygroup1') group.add_nodes([node_01, node_02, node_03]) self.assertEqual(group.count(), 3) result = self.cli_runner.invoke(group_remove_nodes, ['--force', '--clear', '--group=dummygroup1']) self.assertClickResultNoException(result) self.assertEqual(group.count(), 0) def test_copy_existing_group(self): """Test user is prompted to continue if destination group exists and is not empty""" source_label = 'source_copy_existing_group' dest_label = 'dest_copy_existing_group' # Create source group with nodes calc_s1 = orm.CalculationNode().store() calc_s2 = orm.CalculationNode().store() nodes_source_group = {str(node.uuid) for node in [calc_s1, calc_s2]} source_group = orm.Group(label=source_label).store() source_group.add_nodes([calc_s1, calc_s2]) # Copy using `verdi group copy` - making sure all is successful options = [source_label, dest_label] result = self.cli_runner.invoke(group_copy, options) self.assertClickResultNoException(result) self.assertIn( 'Success: Nodes copied from group<{}> to group<{}>'.format(source_label, dest_label), result.output, result.exception ) # Check destination group exists with source group's nodes dest_group = orm.load_group(label=dest_label) self.assertEqual(dest_group.count(), 2) nodes_dest_group = {str(node.uuid) for node in dest_group.nodes} self.assertSetEqual(nodes_source_group, nodes_dest_group) # Copy again, making sure an abort error is raised, since no user input can be made and default is abort result = self.cli_runner.invoke(group_copy, options) self.assertIsNotNone(result.exception, result.output) self.assertIn( 'Warning: Destination group<{}> already exists and is not empty.'.format(dest_label), result.output, result.exception ) # Check destination group is unchanged dest_group = orm.load_group(label=dest_label) self.assertEqual(dest_group.count(), 2) nodes_dest_group = {str(node.uuid) for node in dest_group.nodes} self.assertSetEqual(nodes_source_group, nodes_dest_group)
tests/cmdline/commands/test_group.py
11,059
Tests for the `verdi group` command. Create runner object to run tests. Test `verdi group remove-nodes` command. Test `verdi group copy` command. Test user is prompted to continue if destination group exists and is not empty Test `verdi group create` command. Test `verdi group delete` command. Test `verdi group description` command. Tests help text for all group sub commands. Test `verdi group list` command. Test `verdi group relabel` command. Test `verdi group show` command. Tests for the `verdi group` command. -*- coding: utf-8 -*- Copyright (c), The AiiDA team. All rights reserved. This file is part of the AiiDA code. The code is hosted on GitHub at https://github.com/aiidateam/aiida-core For further information on the license, see the LICENSE.txt file For further information please visit http://www.aiida.net verdi group list verdi group create verdi group delete verdi group relabel verdi group description verdi group addnodes verdi group removenodes verdi group show verdi group copy check if newly added group in present in list Verify that removed group is not present in list Add some nodes and then use `verdi group delete --clear` to delete a node even when it contains nodes Calling delete on a group without the `--clear` option should raise With `--clear` option should delete group and nodes Change the description of the group When no description argument is passed the command should just echo the current description check if group list command shows changed group name Check if node is added in group using group show command Remove same node Check if node is added in group using group show command Add all three nodes and then use `verdi group remove-nodes --clear` to remove them all Create source group with nodes Copy using `verdi group copy` - making sure all is successful Check destination group exists with source group's nodes Copy again, making sure an abort error is raised, since no user input can be made and default is abort Check destination group is unchanged
2,163
en
0.833617
''' This module provides little helper functions that interpret the Handle Server's response codes and the HTTP status codes. All helpers functions test one possible outcome and return True or False. Author: Merret Buurman (DKRZ), 2015-2016 ''' import json from b2handle.compatibility_helper import decoded_response def is_redirect_from_http_to_https(response): if response.status_code == 302: oldurl = response.url newurl = response.headers['location'] if oldurl.startswith('http://') and oldurl.replace('http', 'https') == newurl: return True return False def is_temporary_redirect(response): if response.status_code in [301, 302, 303, 307]: return True return False def handle_success(response): response_content = decoded_response(response) if (response.status_code == 200 or response.status_code == 201) and json.loads(response_content)["responseCode"] == 1: return True return False def does_handle_exist(response): if handle_success(response): return True return False def is_handle_empty(response): response_content = decoded_response(response) if response.status_code == 200 and json.loads(response_content)["responseCode"] == 200: return True return False def was_handle_created(response): response_content = decoded_response(response) if response.status_code == 201 and json.loads(response_content)["responseCode"] == 1: return True return False def handle_not_found(response): response_content = decoded_response(response) if response.status_code == 404 and json.loads(response_content)["responseCode"] == 100: return True return False def not_authenticated(response): response_content = decoded_response(response) try: if response.status_code == 401 or json.loads(response_content)["responseCode"] == 402: # need to put 'OR' because the HS responseCode is not always received! return True except ValueError as e: # If there is no JSON response. pass return False def values_not_found(response): response_content = decoded_response(response) if response.status_code == 400 and json.loads(response_content)["responseCode"] == 200: return True return False def handle_already_exists(response): response_content = decoded_response(response) if response.status_code == 409 & json.loads(response_content)["responseCode"] == 101: return True return False
b2handle/hsresponses.py
2,540
This module provides little helper functions that interpret the Handle Server's response codes and the HTTP status codes. All helpers functions test one possible outcome and return True or False. Author: Merret Buurman (DKRZ), 2015-2016 need to put 'OR' because the HS responseCode is not always received! If there is no JSON response.
350
en
0.719631
""" auth.blueprints ~~~~~~~~~~~~~~~ """ from .simpleidp import blueprint as idpblueprint
auth/blueprints/__init__.py
98
auth.blueprints ~~~~~~~~~~~~~~~
31
en
0.262702
# -*- coding: utf-8 -*- # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. from pyarrow.compat import unittest, u # noqa import pyarrow import datetime class TestConvertList(unittest.TestCase): def test_boolean(self): expected = [True, None, False, None] arr = pyarrow.from_pylist(expected) assert len(arr) == 4 assert arr.null_count == 2 assert arr.type == pyarrow.bool_() assert arr.to_pylist() == expected def test_empty_list(self): arr = pyarrow.from_pylist([]) assert len(arr) == 0 assert arr.null_count == 0 assert arr.type == pyarrow.null() assert arr.to_pylist() == [] def test_all_none(self): arr = pyarrow.from_pylist([None, None]) assert len(arr) == 2 assert arr.null_count == 2 assert arr.type == pyarrow.null() assert arr.to_pylist() == [None, None] def test_integer(self): expected = [1, None, 3, None] arr = pyarrow.from_pylist(expected) assert len(arr) == 4 assert arr.null_count == 2 assert arr.type == pyarrow.int64() assert arr.to_pylist() == expected def test_garbage_collection(self): import gc # Force the cyclic garbage collector to run gc.collect() bytes_before = pyarrow.total_allocated_bytes() pyarrow.from_pylist([1, None, 3, None]) gc.collect() assert pyarrow.total_allocated_bytes() == bytes_before def test_double(self): data = [1.5, 1, None, 2.5, None, None] arr = pyarrow.from_pylist(data) assert len(arr) == 6 assert arr.null_count == 3 assert arr.type == pyarrow.double() assert arr.to_pylist() == data def test_unicode(self): data = [u'foo', u'bar', None, u'mañana'] arr = pyarrow.from_pylist(data) assert len(arr) == 4 assert arr.null_count == 1 assert arr.type == pyarrow.string() assert arr.to_pylist() == data def test_bytes(self): u1 = b'ma\xc3\xb1ana' data = [b'foo', u1.decode('utf-8'), # unicode gets encoded, None] arr = pyarrow.from_pylist(data) assert len(arr) == 3 assert arr.null_count == 1 assert arr.type == pyarrow.binary() assert arr.to_pylist() == [b'foo', u1, None] def test_date(self): data = [datetime.date(2000, 1, 1), None, datetime.date(1970, 1, 1), datetime.date(2040, 2, 26)] arr = pyarrow.from_pylist(data) assert len(arr) == 4 assert arr.type == pyarrow.date64() assert arr.null_count == 1 assert arr[0].as_py() == datetime.date(2000, 1, 1) assert arr[1].as_py() is None assert arr[2].as_py() == datetime.date(1970, 1, 1) assert arr[3].as_py() == datetime.date(2040, 2, 26) def test_timestamp(self): data = [ datetime.datetime(2007, 7, 13, 1, 23, 34, 123456), None, datetime.datetime(2006, 1, 13, 12, 34, 56, 432539), datetime.datetime(2010, 8, 13, 5, 46, 57, 437699) ] arr = pyarrow.from_pylist(data) assert len(arr) == 4 assert arr.type == pyarrow.timestamp('us') assert arr.null_count == 1 assert arr[0].as_py() == datetime.datetime(2007, 7, 13, 1, 23, 34, 123456) assert arr[1].as_py() is None assert arr[2].as_py() == datetime.datetime(2006, 1, 13, 12, 34, 56, 432539) assert arr[3].as_py() == datetime.datetime(2010, 8, 13, 5, 46, 57, 437699) def test_mixed_nesting_levels(self): pyarrow.from_pylist([1, 2, None]) pyarrow.from_pylist([[1], [2], None]) pyarrow.from_pylist([[1], [2], [None]]) with self.assertRaises(pyarrow.ArrowException): pyarrow.from_pylist([1, 2, [1]]) with self.assertRaises(pyarrow.ArrowException): pyarrow.from_pylist([1, 2, []]) with self.assertRaises(pyarrow.ArrowException): pyarrow.from_pylist([[1], [2], [None, [1]]]) def test_list_of_int(self): data = [[1, 2, 3], [], None, [1, 2]] arr = pyarrow.from_pylist(data) assert len(arr) == 4 assert arr.null_count == 1 assert arr.type == pyarrow.list_(pyarrow.int64()) assert arr.to_pylist() == data
python/pyarrow/tests/test_convert_builtin.py
5,264
-*- coding: utf-8 -*- Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to you under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. noqa Force the cyclic garbage collector to run unicode gets encoded,
843
en
0.871013
# Most of this code is stolen from myself, but it's an app about pirating so whatev import os import sys import time import datetime import subprocess import pycurl import spotipy import spotipy.util from StringIO import StringIO from collections import deque import sys # If this seems like overkill, that's because it is. I wrote this get_args code # for general use, this function is not specific to this def get_args(): def is_flag(check): if check[:2] == '--': return True elif check[0] == '-': return True else: return False argdict = {} for i, arg in enumerate(sys.argv): if i == 0: # this is just the name of the program continue check = sys.argv[i] if not is_flag(check): continue elif check[:2] == '--': # single multi-character flag check = check[2:] # it makes my life easier if they do this if '=' in check: check = check.split('=') argdict[check[0]] = check[1] continue j = i+1 params = [] while not is_flag(sys.argv[j]): params.append(sys.argv[j]) j += 1 if j >= len(sys.argv): break if not params: # if params is empty, just make the value True params = True argdict[check] = params elif check[0] == '-': # multiple single character flags check = check[1:] # it makes my life easier if they do this if '=' in check: check = check.split('=') argdict[check[0]] = check[1] continue if len(check) == 1: # if there is just one character, the arg can have params j = i+1 params = [] while not is_flag(sys.argv[j]): params.append(sys.argv[j]) j += 1 if not params: # if params is empty, just make the value True params = True argdict[check] = params else: for c in check: argdict[c] = True return argdict # Hacky AF but it works class Foobar: def wait(self): return class Song: def __init__(self, title, artist, pl): self.title = title.replace("/", "&").replace(" -", ",").replace("-", "~") self.artist = artist.replace("/", "&") self.playlist = pl self.id = None class Pirate: ### CONSTRUCTOR ### def __init__(self, settings): self.downloadthreads = [] self.downloadqueue = deque() self.alive = True self.verbose = int(settings["verbose"]) self.downloadthreads = settings["download_threads"] self.dl_folder = os.path.realpath(settings["dl_folder"]) self.log_folder = os.path.realpath(settings["log_folder"]) self.err_folder = os.path.realpath(settings["err_folder"]) self.username = settings["username"] scope = "user-library-read" client_id = settings["client_id"] client_secret = settings["client_secret"] redirect_url = "http://localhost:8888/callback" # Request a token token = spotipy.util.prompt_for_user_token(self.username, scope, client_id, client_secret, redirect_url) self.sp = spotipy.Spotify(auth=token) args = get_args() if "plid" in args: self.plid = args["plid"] elif "p" in args: self.plid = args["p"] else: self.plid = None self.mode = "all playlists" if "mode" in args: self.mode = args["mode"] elif "m" in args: self.mode = args["m"] self.mode = self.mode.lower() if self.mode not in ["all playlists", "apl", "saved", "s", "everything", "e", "playlist", "pl"]: self.log_error("Invalid mode: %s" % self.mode, 0) self.terminate() if self.mode == "playlist" and self.plid is None: self.log_error("You need to supply a playlist to fetch", 0) self.terminate() ### THESE FUNCTIONS HANDLE STARTUP ### def init(self): if not os.path.exists(self.dl_folder): os.makedirs(self.dl_folder) #initiate logs if not os.path.exists(self.log_folder): os.makedirs(self.log_folder) if not os.path.exists(self.err_folder): os.makedirs(self.err_folder) self.now = lambda: str(datetime.datetime.now()) # Error log fname = self.now().replace(':', '-') + ".log" self.log_file = os.path.join(self.log_folder, fname) self.err_file = os.path.join(self.err_folder, fname) with open(self.log_file, 'w') as f: f.write("Activity log for yt-streamer:\n---------------------\n") with open(self.err_file, 'w') as f: f.write("Error log for yt-streamer:\n---------------------\n") def go(self): try: if self.mode in ["all playlists", "apl"]: self.fetchPls() elif self.mode in ["saved", "s"]: self.fetchSaved() elif self.mode in ["everything", "e"]: self.fetchAll() elif self.mode in ["playlist", "pl"]: playlists = self.sp.user_playlists(self.username) for playlist in playlists['items']: if playlist['name'].lower() == self.plid.lower(): self.fetch(playlist) else: self.log("Not the one: %s" % playlist['name'], 20) self.log("Looking for: %s" % self.plid, 20) else: self.log_error("Invalid mode: %s" % self.mode, 0) self.terminate() self.handle_downloading() except Exception as e: self.log_error(str(e), 10) except: self.log("Program terminated by user (Keyboard Interrupt)", 10) self.terminate() def addtrack(self, track, plname): name = track['name'] artists = "" if len(track['artists']) > 0: for artist in track['artists']: artists += artist['name'] + ", " artists = artists[:-2] else: artists = "Unknown Artist" # Add to download queue as a Song object s = Song(name, artists, plname) s.id = self.ytsearch(s) self.downloadqueue.append(s) self.log("Found song: %s by %s" % (s.title, s.artist), 10) def fetchAll(self): self.fetchSaved() self.fetchPls() def fetchSaved(self): # Fetch saved music results = self.sp.current_user_saved_tracks() for item in results['items']: track = item['track'] self.addtrack(track, "My Music") def fetchPls(self): self.log("Fetching all playlists of %s" % self.username, 10) # Fetch all spotify playlists of a user playlists = self.sp.user_playlists(self.username) for playlist in playlists['items']: if playlist['owner']['id'] == self.username: self.fetch(playlist) # pl is a spodipy playlist object def fetch(self, playlist): # Fetch all songs in a playlist self.log("Fetching playlist: %s" % playlist['name']) results = self.sp.user_playlist(self.username, playlist['id'], fields="tracks,next") tracks = results['tracks'] while True: for item in tracks['items']: # For each item, get the artist, playlist, and song name track = item['track'] self.addtrack(track, playlist["name"]) if tracks['next']: tracks = self.sp.next(tracks) else: break def handle_downloading(self): # Fix so queue never mutates when this is happening while self.alive: if (len(self.downloadqueue)): to_push = self.downloadqueue.popleft() # Get the downloading song self.log("Starting download: %s" % to_push.title, 10) # Start download, wait for it to finish self.ytdl(to_push).wait() self.log("Finished download: %s" % to_push.title, 10) # Edit ogg header info # comment[0]="ARTIST=me"; # comment[1]="TITLE=the sound of Vorbis"; else: self.terminate() def ytdl(self, song): pl = song.playlist name = song.title + " - " + song.artist vid = song.id # Need better error handling (exceptions) # Fix it later (or never) if vid is None: return Foobar() # TODO: replace dl with self.dl_folder ndir = "./%s/%s" % (self.dl_folder, pl) fname = "%s/%s.ogg" % (ndir, name) if not os.path.exists(ndir): os.makedirs(ndir) # check if file exists if not os.path.isfile(fname): # download the audio of a video based on the youtube id try: p = subprocess.Popen(["youtube-dl",\ "--extract-audio",\ "--audio-format", "vorbis",\ "-o", ndir + "/" + name + ".%(ext)s",\ "https://www.youtube.com/watch?v=" + vid],\ stdout=None) return p except Exception as e: return Foobar() return Foobar() def getbetween(self, search, left, right): for line in search.splitlines(): if not left in line: continue out = line.split(left)[1].split(right)[0] return out def curl(self, url): c = pycurl.Curl() c.setopt(c.URL, url) buf = StringIO() c.setopt(c.WRITEDATA, buf) c.perform() c.close() return buf.getvalue() def ytsearch(self, song): # Get query from song query = song.title + " " + song.artist # Get the first result from the query try: query = query.replace(" ", "+") url = "https://www.youtube.com/results?search_query=" url += query #url += '+-"music+video"+dirty' #url += '+lyrics+dirty+radio+edit' url += '+lyrics+dirty+' # Search for video, try to avoid music videos since they suck search = self.curl(url) left = "watch?v=" right = "\"" video_id = self.getbetween(search, left, right) return video_id except Exception as e: # This means no results (or no internet connection) self.log_error("Failed to ytdl for some reason", 10) return None def terminate(self): self.alive = False ### THESE FUNCTIONS HANDLE LOGGING ### def report(self, msg): sys.stdout.flush() print msg # Grab the lock with open(self.log_file, 'a') as f: f.write(self.now() + ':\t') f.write(msg + '\n') return msg def log(self, msg, log_level=0): # This function is too complicated to properly comment if log_level <= self.verbose: return self.report(msg) else: return msg # This is pythonic def log_error(self, e, log_level=0): # Grab the lock if log_level <= self.verbose: with open(self.err_file, 'a') as f: f.write(self.now() + ':\t') f.write(str(e) + ('\n')) self.report("An exception has been raised: %s" % (e,)) return e def read(filen): config = dict() try: with open(filen) as cfg: for line in cfg: line = line.split('#')[0] # Comments, yay! line = line.split('//')[0] # //Comments, yay! parts = line.split(':') if len(parts) == 2: config[parts[0].strip()] = parts[1].strip() else: pass # This is pythonic return config except Exception as e: print "Error opening settings file %s" % filen return None settings = read("./dat/pirate.conf") app = Pirate(settings) app.init() app.go()
pirate.py
12,633
Most of this code is stolen from myself, but it's an app about pirating so whatev If this seems like overkill, that's because it is. I wrote this get_args code for general use, this function is not specific to this this is just the name of the program single multi-character flag it makes my life easier if they do this if params is empty, just make the value True multiple single character flags it makes my life easier if they do this if there is just one character, the arg can have params if params is empty, just make the value True Hacky AF but it works CONSTRUCTOR Request a token THESE FUNCTIONS HANDLE STARTUP initiate logs Error log Add to download queue as a Song object Fetch saved music Fetch all spotify playlists of a user pl is a spodipy playlist object Fetch all songs in a playlist For each item, get the artist, playlist, and song name Fix so queue never mutates when this is happening Get the downloading song Start download, wait for it to finish Edit ogg header info comment[0]="ARTIST=me"; comment[1]="TITLE=the sound of Vorbis"; Need better error handling (exceptions) Fix it later (or never) TODO: replace dl with self.dl_folder check if file exists download the audio of a video based on the youtube id Get query from song Get the first result from the queryurl += '+-"music+video"+dirty'url += '+lyrics+dirty+radio+edit' Search for video, try to avoid music videos since they suck This means no results (or no internet connection) THESE FUNCTIONS HANDLE LOGGING Grab the lock This function is too complicated to properly comment This is pythonic Grab the lock Comments, yay! //Comments, yay! This is pythonic
1,637
en
0.851483
import time import torch import torch.nn as nn import torch.nn.parallel import torch.utils.data from torch import optim from torch.utils import data from tqdm import tqdm from networks import LATENT_CODE_SIZE, device # from train_autoencoder import SDFSampleDataset, save_checkpoint, SIGMA class SDFNet(nn.Module): def __init__(self, latent_code_size=LATENT_CODE_SIZE, dropout_prob=0.2, point_dim=3): super(SDFNet, self).__init__() SDF_NET_BREADTH = latent_code_size * 2 # the decoder should only have xyz information, without sdf values self.layers1 = nn.Sequential( nn.utils.weight_norm(nn.Linear(point_dim + latent_code_size, SDF_NET_BREADTH)), nn.ReLU(), nn.Dropout(dropout_prob), nn.utils.weight_norm(nn.Linear(SDF_NET_BREADTH, SDF_NET_BREADTH)), nn.ReLU(), nn.Dropout(dropout_prob), nn.utils.weight_norm(nn.Linear(SDF_NET_BREADTH, SDF_NET_BREADTH)), nn.ReLU(), nn.Dropout(dropout_prob), nn.utils.weight_norm(nn.Linear(SDF_NET_BREADTH, SDF_NET_BREADTH - latent_code_size - point_dim)), nn.ReLU(), nn.Dropout(dropout_prob), ) self.layers2 = nn.Sequential( nn.utils.weight_norm(nn.Linear(SDF_NET_BREADTH, SDF_NET_BREADTH)), nn.ReLU(), nn.Dropout(dropout_prob), nn.utils.weight_norm(nn.Linear(SDF_NET_BREADTH, SDF_NET_BREADTH)), nn.ReLU(), nn.Dropout(dropout_prob), nn.utils.weight_norm(nn.Linear(SDF_NET_BREADTH, SDF_NET_BREADTH)), nn.ReLU(), nn.Dropout(dropout_prob), nn.utils.weight_norm(nn.Linear(SDF_NET_BREADTH, SDF_NET_BREADTH)), nn.ReLU(), nn.Dropout(dropout_prob), nn.Linear(SDF_NET_BREADTH, 1), nn.Tanh() ) def forward(self, input): """ input: [B, N, latent_size + point_dim] :param latent_codes: [B, N, LATENT_CODE_DIM] :param points: [B, N, 3] :return: sdf_pred: [B, N] """ x = self.layers1(input) x = torch.cat((x, input), dim=-1) x = self.layers2(x) return x # if __name__ == '__main__': # experiment = '5_samples_latent_128_no_reg' # num_epochs = 500 # # decoder = SDFNet() # # optimiser = optim.Adam(decoder.parameters(), lr=1e-5) # # model, optimiser, start_epoch, training_loss = load_or_init_model(experiment) # dataset = SDFSampleDataset('data/SdfSamples/ShapeNetV2/03001627/', '5_sample.json') # batch_size = 5 # normal_distribution = torch.distributions.normal.Normal(0, 0.0001) # latent_codes = normal_distribution.sample((MODEL_COUNT, LATENT_CODE_SIZE)).to(device) # latent_codes.requires_grad = True # train_data = data.DataLoader(dataset, batch_size=batch_size, shuffle=True, num_workers=4) # # # training loop starts # training_loss = [] # for epoch in range(1, num_epochs + 1): # start_time = time.time() # running_loss = [] # # for i_batch, batch in tqdm(enumerate(train_data)): # optimiser.zero_grad() # batch = batch.to(device) # [B, point_dim, N] # sdf_pred, input_trans, latent_code = model(batch) # sdf_gt = batch[:, -1, :].squeeze() # # loss = l1_loss(sdf_gt, sdf_pred) # TODO: experiment with only the l1 loss # loss += SIGMA**2 * min(1, epoch / 100) * torch.mean(torch.norm(latent_code, dim=1)) # loss.backward() # optimiser.step() # running_loss.append(loss.item()) # # epoch_duration = time.time() - start_time # epoch_loss = np.mean(running_loss) # training_loss.append(epoch_loss) # # print("Epoch {:d}, {:.1f}s. Loss: {:.8f}".format(epoch, epoch_duration, epoch_loss)) # # if epoch_loss < 0.02: # save_checkpoint(epoch, model, optimiser, training_loss, experiment, filename='sub002') # # # always save the latest snapshot # save_checkpoint(epoch, model, optimiser, training_loss, experiment) # if epoch % 100 == 0: # save_checkpoint(epoch, model, optimiser, training_loss, experiment, filename=str(epoch))
networks/sdf_net_decoder.py
4,295
input: [B, N, latent_size + point_dim] :param latent_codes: [B, N, LATENT_CODE_DIM] :param points: [B, N, 3] :return: sdf_pred: [B, N] from train_autoencoder import SDFSampleDataset, save_checkpoint, SIGMA the decoder should only have xyz information, without sdf values if __name__ == '__main__': experiment = '5_samples_latent_128_no_reg' num_epochs = 500 decoder = SDFNet() optimiser = optim.Adam(decoder.parameters(), lr=1e-5) model, optimiser, start_epoch, training_loss = load_or_init_model(experiment) dataset = SDFSampleDataset('data/SdfSamples/ShapeNetV2/03001627/', '5_sample.json') batch_size = 5 normal_distribution = torch.distributions.normal.Normal(0, 0.0001) latent_codes = normal_distribution.sample((MODEL_COUNT, LATENT_CODE_SIZE)).to(device) latent_codes.requires_grad = True train_data = data.DataLoader(dataset, batch_size=batch_size, shuffle=True, num_workers=4) training loop starts training_loss = [] for epoch in range(1, num_epochs + 1): start_time = time.time() running_loss = [] for i_batch, batch in tqdm(enumerate(train_data)): optimiser.zero_grad() batch = batch.to(device) [B, point_dim, N] sdf_pred, input_trans, latent_code = model(batch) sdf_gt = batch[:, -1, :].squeeze() loss = l1_loss(sdf_gt, sdf_pred) TODO: experiment with only the l1 loss loss += SIGMA**2 * min(1, epoch / 100) * torch.mean(torch.norm(latent_code, dim=1)) loss.backward() optimiser.step() running_loss.append(loss.item()) epoch_duration = time.time() - start_time epoch_loss = np.mean(running_loss) training_loss.append(epoch_loss) print("Epoch {:d}, {:.1f}s. Loss: {:.8f}".format(epoch, epoch_duration, epoch_loss)) if epoch_loss < 0.02: save_checkpoint(epoch, model, optimiser, training_loss, experiment, filename='sub002') always save the latest snapshot save_checkpoint(epoch, model, optimiser, training_loss, experiment) if epoch % 100 == 0: save_checkpoint(epoch, model, optimiser, training_loss, experiment, filename=str(epoch))
2,227
en
0.453643
''' Copyright 2022 Airbus SAS Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ''' import unittest import numpy as np import pandas as pd from os.path import join, dirname from pandas import DataFrame, read_csv from scipy.interpolate import interp1d from sos_trades_core.execution_engine.execution_engine import ExecutionEngine from sos_trades_core.tests.core.abstract_jacobian_unit_test import AbstractJacobianUnittest class ServicesJacobianDiscTest(AbstractJacobianUnittest): #AbstractJacobianUnittest.DUMP_JACOBIAN = True def setUp(self): self.name = 'Test' self.ee = ExecutionEngine(self.name) self.year_start = 2020 self.year_end = 2050 self.time_step = 1 self.years = np.arange(self.year_start, self.year_end + 1, self.time_step) self.nb_per = round((self.year_end - self.year_start) / self.time_step + 1) # ------------------------- # input data_dir = join(dirname(__file__), 'data') global_data_dir = join(dirname(dirname(__file__)), 'data') total_workforce_df = read_csv(join(data_dir, 'workingage_population_df.csv')) total_workforce_df = total_workforce_df[total_workforce_df['years']<=self.year_end] #multiply ageworking pop by employment rate and by % in services workforce = total_workforce_df['population_1570']* 0.659 * 0.509 self.workforce_df = pd.DataFrame({'years': self.years, 'workforce': workforce}) #Energy_supply brut_net = 1/1.45 share_indus = 0.37 #prepare energy df energy_outlook = pd.DataFrame({ 'year': [2010, 2017, 2018, 2025, 2030, 2035, 2040, 2050, 2060, 2100], 'energy': [149.483879, 162.7848774, 166.4685636, 180.7072889, 189.6932084, 197.8418842, 206.1201182, 220.000, 250.0, 300.0]}) f2 = interp1d(energy_outlook['year'], energy_outlook['energy']) #Find values for 2020, 2050 and concat dfs energy_supply = f2(np.arange(self.year_start, self.year_end+1)) energy_supply_values = energy_supply * brut_net * share_indus energy_supply_df = pd.DataFrame({'years': self.years, 'Total production': energy_supply_values}) energy_supply_df.index = self.years self.energy_supply_df = energy_supply_df #energy_supply_df.loc[2020, 'Total production'] = 91.936 #Investment growth at 2% init_value = 25 invest_serie = [] invest_serie.append(init_value) for year in np.arange(1, self.nb_per): invest_serie.append(invest_serie[year - 1] * 1.02) self.total_invest = pd.DataFrame({'years': self.years, 'investment': invest_serie}) #damage self.damage_df = pd.DataFrame({'years': self.years, 'damages': np.zeros(self.nb_per), 'damage_frac_output': np.zeros(self.nb_per), 'base_carbon_price': np.zeros(self.nb_per)}) self.damage_df.index = self.years self.damage_df['damage_frac_output'] = 1e-2 def analytic_grad_entry(self): return [ self.test_services_analytic_grad, self.test_services_withotudamagetoproductivity ] def test_services_analytic_grad(self): self.model_name = 'Services' ns_dict = {'ns_witness': f'{self.name}', 'ns_energy_mix': f'{self.name}', 'ns_public': f'{self.name}', 'ns_functions': f'{self.name}', 'ns_ref':f'{self.name}' } self.ee.ns_manager.add_ns_def(ns_dict) mod_path = 'climateeconomics.sos_wrapping.sos_wrapping_sectors.services.services_discipline.ServicesDiscipline' builder = self.ee.factory.get_builder_from_module( self.model_name, mod_path) self.ee.factory.set_builders_to_coupling_builder(builder) self.ee.configure() self.ee.display_treeview_nodes() inputs_dict = {f'{self.name}.year_start': self.year_start, f'{self.name}.year_end': self.year_end, f'{self.name}.time_step': self.time_step, f'{self.name}.damage_to_productivity': True, f'{self.name}.frac_damage_prod': 0.3, f'{self.name}.energy_production': self.energy_supply_df, f'{self.name}.damage_df': self.damage_df, f'{self.name}.workforce_df': self.workforce_df, f'{self.name}.sector_investment': self.total_invest, f'{self.name}.alpha': 0.5 } self.ee.load_study_from_input_dict(inputs_dict) disc_techno = self.ee.root_process.sos_disciplines[0] self.check_jacobian(location=dirname(__file__), filename=f'jacobian_services_discipline.pkl', discipline=disc_techno, step=1e-15, derr_approx='complex_step', inputs=[f'{self.name}.energy_production', f'{self.name}.damage_df', f'{self.name}.workforce_df', f'{self.name}.sector_investment'], outputs=[f'{self.name}.production_df', f'{self.name}.capital_df', f'{self.name}.emax_enet_constraint']) def test_services_withotudamagetoproductivity(self): self.model_name = 'Services' ns_dict = {'ns_witness': f'{self.name}', 'ns_energy_mix': f'{self.name}', 'ns_public': f'{self.name}', 'ns_functions': f'{self.name}', 'ns_ref':f'{self.name}' } self.ee.ns_manager.add_ns_def(ns_dict) mod_path = 'climateeconomics.sos_wrapping.sos_wrapping_sectors.services.services_discipline.ServicesDiscipline' builder = self.ee.factory.get_builder_from_module( self.model_name, mod_path) self.ee.factory.set_builders_to_coupling_builder(builder) self.ee.configure() self.ee.display_treeview_nodes() inputs_dict = {f'{self.name}.year_start': self.year_start, f'{self.name}.year_end': self.year_end, f'{self.name}.time_step': self.time_step, f'{self.name}.damage_to_productivity': False, f'{self.name}.frac_damage_prod': 0.3, f'{self.name}.energy_production': self.energy_supply_df, f'{self.name}.damage_df': self.damage_df, f'{self.name}.workforce_df': self.workforce_df, f'{self.name}.sector_investment': self.total_invest, f'{self.name}.alpha': 0.5 } self.ee.load_study_from_input_dict(inputs_dict) disc_techno = self.ee.root_process.sos_disciplines[0] self.check_jacobian(location=dirname(__file__), filename=f'jacobian_services_discipline_withoutdamage.pkl', discipline=disc_techno, step=1e-15, derr_approx='complex_step', inputs=[f'{self.name}.energy_production', f'{self.name}.damage_df', f'{self.name}.workforce_df', f'{self.name}.sector_investment'], outputs=[f'{self.name}.production_df', f'{self.name}.capital_df', f'{self.name}.emax_enet_constraint'])
climateeconomics/tests/_l1_test_gradient_services_discipline.py
8,163
Copyright 2022 Airbus SAS Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. AbstractJacobianUnittest.DUMP_JACOBIAN = True ------------------------- inputmultiply ageworking pop by employment rate and by % in servicesEnergy_supplyprepare energy df Find values for 2020, 2050 and concat dfs energy_supply_df.loc[2020, 'Total production'] = 91.936Investment growth at 2% damage
851
en
0.811302
import logging import json from http import HTTPStatus from typing import Any, Dict # noqa: F401 from flask import Response, jsonify, make_response, request from flask import current_app as app from flask.blueprints import Blueprint from amundsen_application.log.action_log import action_logging from amundsen_application.api.utils.request_utils import get_query_param, request_search from amundsen_application.api.utils.search_utils import generate_query_json, has_filters, \ map_table_result, transform_filters from amundsen_application.models.user import load_user, dump_user LOGGER = logging.getLogger(__name__) REQUEST_SESSION_TIMEOUT_SEC = 3 search_blueprint = Blueprint('search', __name__, url_prefix='/explore/api/search/v0') SEARCH_TABLE_ENDPOINT = '/search' SEARCH_TABLE_FILTER_ENDPOINT = '/search_table' SEARCH_USER_ENDPOINT = '/search_user' @search_blueprint.route('/table', methods=['POST']) def search_table() -> Response: """ Parse the request arguments and call the helper method to execute a table search :return: a Response created with the results from the helper method """ try: request_json = request.get_json() search_term = get_query_param(request_json, 'term', '"term" parameter expected in request data') page_index = get_query_param(request_json, 'pageIndex', '"pageIndex" parameter expected in request data') search_type = request_json.get('searchType') transformed_filters = transform_filters(filters=request_json.get('filters', {})) results_dict = _search_table(filters=transformed_filters, search_term=search_term, page_index=page_index, search_type=search_type) return make_response(jsonify(results_dict), results_dict.get('status_code', HTTPStatus.INTERNAL_SERVER_ERROR)) except Exception as e: message = 'Encountered exception: ' + str(e) logging.exception(message) return make_response(jsonify(results_dict), HTTPStatus.INTERNAL_SERVER_ERROR) @action_logging def _search_table(*, search_term: str, page_index: int, filters: Dict, search_type: str) -> Dict[str, Any]: """ Call the search service endpoint and return matching results Search service logic defined here: https://github.com/lyft/amundsensearchlibrary/blob/master/search_service/api/table.py :return: a json output containing search results array as 'results' """ # Default results tables = { 'page_index': int(page_index), 'results': [], 'total_results': 0, } results_dict = { 'search_term': search_term, 'msg': '', 'tables': tables, } try: if has_filters(filters=filters): query_json = generate_query_json(filters=filters, page_index=page_index, search_term=search_term) url_base = app.config['SEARCHSERVICE_BASE'] + SEARCH_TABLE_FILTER_ENDPOINT response = request_search(url=url_base, headers={'Content-Type': 'application/json'}, method='POST', data=json.dumps(query_json)) else: url_base = app.config['SEARCHSERVICE_BASE'] + SEARCH_TABLE_ENDPOINT url = f'{url_base}?query_term={search_term}&page_index={page_index}' response = request_search(url=url) status_code = response.status_code if status_code == HTTPStatus.OK: results_dict['msg'] = 'Success' results = response.json().get('results') tables['results'] = [map_table_result(result) for result in results] tables['total_results'] = response.json().get('total_results') else: message = 'Encountered error: Search request failed' results_dict['msg'] = message logging.error(message) results_dict['status_code'] = status_code return results_dict except Exception as e: message = 'Encountered exception: ' + str(e) results_dict['msg'] = message logging.exception(message) return results_dict @search_blueprint.route('/user', methods=['GET']) def search_user() -> Response: """ Parse the request arguments and call the helper method to execute a user search :return: a Response created with the results from the helper method """ try: search_term = get_query_param(request.args, 'query', 'Endpoint takes a "query" parameter') page_index = get_query_param(request.args, 'page_index', 'Endpoint takes a "page_index" parameter') search_type = request.args.get('search_type') results_dict = _search_user(search_term=search_term, page_index=page_index, search_type=search_type) return make_response(jsonify(results_dict), results_dict.get('status_code', HTTPStatus.INTERNAL_SERVER_ERROR)) except Exception as e: message = 'Encountered exception: ' + str(e) logging.exception(message) return make_response(jsonify(results_dict), HTTPStatus.INTERNAL_SERVER_ERROR) @action_logging def _search_user(*, search_term: str, page_index: int, search_type: str) -> Dict[str, Any]: """ Call the search service endpoint and return matching results Search service logic defined here: https://github.com/lyft/amundsensearchlibrary/blob/master/search_service/api/user.py :return: a json output containing search results array as 'results' """ def _map_user_result(result: Dict) -> Dict: user_result = dump_user(load_user(result)) user_result['type'] = 'user' return user_result users = { 'page_index': int(page_index), 'results': [], 'total_results': 0, } results_dict = { 'search_term': search_term, 'msg': 'Success', 'status_code': HTTPStatus.OK, 'users': users, } try: url = '{0}?query_term={1}&page_index={2}'.format(app.config['SEARCHSERVICE_BASE'] + SEARCH_USER_ENDPOINT, search_term, page_index) response = request_search(url=url) status_code = response.status_code if status_code == HTTPStatus.OK: results_dict['msg'] = 'Success' results = response.json().get('results') users['results'] = [_map_user_result(result) for result in results] users['total_results'] = response.json().get('total_results') else: message = 'Encountered error: Search request failed' results_dict['msg'] = message logging.error(message) results_dict['status_code'] = status_code return results_dict except Exception as e: message = 'Encountered exception: ' + str(e) results_dict['msg'] = message results_dict['status_code'] = HTTPStatus.INTERNAL_SERVER_ERROR logging.exception(message) return results_dict # TODO - Implement def _search_dashboard(*, search_term: str, page_index: int, filters: Dict, search_type: str) -> Dict[str, Any]: return {}
amundsen_application/api/search/v0.py
7,294
Call the search service endpoint and return matching results Search service logic defined here: https://github.com/lyft/amundsensearchlibrary/blob/master/search_service/api/table.py :return: a json output containing search results array as 'results' Call the search service endpoint and return matching results Search service logic defined here: https://github.com/lyft/amundsensearchlibrary/blob/master/search_service/api/user.py :return: a json output containing search results array as 'results' Parse the request arguments and call the helper method to execute a table search :return: a Response created with the results from the helper method Parse the request arguments and call the helper method to execute a user search :return: a Response created with the results from the helper method noqa: F401 Default results TODO - Implement
843
en
0.691593
"""Python wrappers around TensorFlow ops. This file is MACHINE GENERATED! Do not edit. Original C++ source file: gen_dataset_ops.cc """ import collections as _collections import six as _six from tensorflow.python import pywrap_tensorflow as _pywrap_tensorflow from tensorflow.python.eager import context as _context from tensorflow.python.eager import core as _core from tensorflow.python.eager import execute as _execute from tensorflow.python.framework import dtypes as _dtypes from tensorflow.python.framework import errors as _errors from tensorflow.python.framework import tensor_shape as _tensor_shape from tensorflow.core.framework import op_def_pb2 as _op_def_pb2 # Needed to trigger the call to _set_call_cpp_shape_fn. from tensorflow.python.framework import common_shapes as _common_shapes from tensorflow.python.framework import op_def_registry as _op_def_registry from tensorflow.python.framework import ops as _ops from tensorflow.python.framework import op_def_library as _op_def_library from tensorflow.python.util.deprecation import deprecated_endpoints from tensorflow.python.util import dispatch as _dispatch from tensorflow.python.util.tf_export import tf_export from tensorflow.python.util.tf_export import kwarg_only as _kwarg_only from tensorflow.tools.docs import doc_controls as _doc_controls @_dispatch.add_dispatch_list @tf_export('ignite_dataset') def ignite_dataset(cache_name, host, port, local, part, page_size, schema, permutation, name=None): r"""IgniteDataset that allows to get data from Apache Ignite. Apache Ignite is a memory-centric distributed database, caching, and processing platform for transactional, analytical, and streaming workloads, delivering in-memory speeds at petabyte scale. This contrib package contains an integration between Apache Ignite and TensorFlow. The integration is based on tf.data from TensorFlow side and Binary Client Protocol from Apache Ignite side. It allows to use Apache Ignite as a datasource for neural network training, inference and all other computations supported by TensorFlow. Ignite Dataset is based on Apache Ignite Binary Client Protocol. Args: cache_name: A `Tensor` of type `string`. Ignite Cache Name. host: A `Tensor` of type `string`. Ignite Thin Client Host. port: A `Tensor` of type `int32`. Ignite Thin Client Port. local: A `Tensor` of type `bool`. Local flag that defines that data should be fetched from local host only. part: A `Tensor` of type `int32`. Partition data should be fetched from. page_size: A `Tensor` of type `int32`. Page size for Ignite Thin Client. schema: A `Tensor` of type `int32`. Internal structure that defines schema of cache objects. permutation: A `Tensor` of type `int32`. Internal structure that defines permutation of cache objects. name: A name for the operation (optional). Returns: A `Tensor` of type `variant`. """ _ctx = _context._context or _context.context() if _ctx is not None and _ctx._thread_local_data.is_eager: try: _result = _pywrap_tensorflow.TFE_Py_FastPathExecute( _ctx._context_handle, _ctx._thread_local_data.device_name, "IgniteDataset", name, _ctx.post_execution_callbacks, cache_name, host, port, local, part, page_size, schema, permutation) return _result except _core._FallbackException: try: return ignite_dataset_eager_fallback( cache_name, host, port, local, part, page_size, schema, permutation, name=name, ctx=_ctx) except _core._SymbolicException: pass # Add nodes to the TensorFlow graph. except (TypeError, ValueError): result = _dispatch.dispatch( ignite_dataset, cache_name=cache_name, host=host, port=port, local=local, part=part, page_size=page_size, schema=schema, permutation=permutation, name=name) if result is not _dispatch.OpDispatcher.NOT_SUPPORTED: return result raise except _core._NotOkStatusException as e: if name is not None: message = e.message + " name: " + name else: message = e.message _six.raise_from(_core._status_to_exception(e.code, message), None) # Add nodes to the TensorFlow graph. try: _, _, _op = _op_def_lib._apply_op_helper( "IgniteDataset", cache_name=cache_name, host=host, port=port, local=local, part=part, page_size=page_size, schema=schema, permutation=permutation, name=name) except (TypeError, ValueError): result = _dispatch.dispatch( ignite_dataset, cache_name=cache_name, host=host, port=port, local=local, part=part, page_size=page_size, schema=schema, permutation=permutation, name=name) if result is not _dispatch.OpDispatcher.NOT_SUPPORTED: return result raise _result = _op.outputs[:] _inputs_flat = _op.inputs _attrs = None _execute.record_gradient( "IgniteDataset", _inputs_flat, _attrs, _result, name) _result, = _result return _result def IgniteDataset(cache_name, host, port, local, part, page_size, schema, permutation, name=None): return ignite_dataset(cache_name=cache_name, host=host, port=port, local=local, part=part, page_size=page_size, schema=schema, permutation=permutation, name=name) IgniteDataset.__doc__ = ignite_dataset.__doc__ IgniteDataset = _doc_controls.do_not_generate_docs(_kwarg_only(IgniteDataset)) tf_export("raw_ops.IgniteDataset")(IgniteDataset) def ignite_dataset_eager_fallback(cache_name, host, port, local, part, page_size, schema, permutation, name=None, ctx=None): r"""This is the slowpath function for Eager mode. This is for function ignite_dataset """ _ctx = ctx if ctx else _context.context() cache_name = _ops.convert_to_tensor(cache_name, _dtypes.string) host = _ops.convert_to_tensor(host, _dtypes.string) port = _ops.convert_to_tensor(port, _dtypes.int32) local = _ops.convert_to_tensor(local, _dtypes.bool) part = _ops.convert_to_tensor(part, _dtypes.int32) page_size = _ops.convert_to_tensor(page_size, _dtypes.int32) schema = _ops.convert_to_tensor(schema, _dtypes.int32) permutation = _ops.convert_to_tensor(permutation, _dtypes.int32) _inputs_flat = [cache_name, host, port, local, part, page_size, schema, permutation] _attrs = None _result = _execute.execute(b"IgniteDataset", 1, inputs=_inputs_flat, attrs=_attrs, ctx=_ctx, name=name) _execute.record_gradient( "IgniteDataset", _inputs_flat, _attrs, _result, name) _result, = _result return _result _ops.RegisterShape("IgniteDataset")(None) def _InitOpDefLibrary(op_list_proto_bytes): op_list = _op_def_pb2.OpList() op_list.ParseFromString(op_list_proto_bytes) _op_def_registry.register_op_list(op_list) op_def_lib = _op_def_library.OpDefLibrary() op_def_lib.add_op_list(op_list) return op_def_lib # op { # name: "IgniteDataset" # input_arg { # name: "cache_name" # type: DT_STRING # } # input_arg { # name: "host" # type: DT_STRING # } # input_arg { # name: "port" # type: DT_INT32 # } # input_arg { # name: "local" # type: DT_BOOL # } # input_arg { # name: "part" # type: DT_INT32 # } # input_arg { # name: "page_size" # type: DT_INT32 # } # input_arg { # name: "schema" # type: DT_INT32 # } # input_arg { # name: "permutation" # type: DT_INT32 # } # output_arg { # name: "handle" # type: DT_VARIANT # } # is_stateful: true # } _op_def_lib = _InitOpDefLibrary(b"\n\203\001\n\rIgniteDataset\022\016\n\ncache_name\030\007\022\010\n\004host\030\007\022\010\n\004port\030\003\022\t\n\005local\030\n\022\010\n\004part\030\003\022\r\n\tpage_size\030\003\022\n\n\006schema\030\003\022\017\n\013permutation\030\003\032\n\n\006handle\030\025\210\001\001")
venv/lib/python3.7/site-packages/tensorflow_core/contrib/ignite/python/ops/gen_dataset_ops.py
8,002
IgniteDataset that allows to get data from Apache Ignite. Apache Ignite is a memory-centric distributed database, caching, and processing platform for transactional, analytical, and streaming workloads, delivering in-memory speeds at petabyte scale. This contrib package contains an integration between Apache Ignite and TensorFlow. The integration is based on tf.data from TensorFlow side and Binary Client Protocol from Apache Ignite side. It allows to use Apache Ignite as a datasource for neural network training, inference and all other computations supported by TensorFlow. Ignite Dataset is based on Apache Ignite Binary Client Protocol. Args: cache_name: A `Tensor` of type `string`. Ignite Cache Name. host: A `Tensor` of type `string`. Ignite Thin Client Host. port: A `Tensor` of type `int32`. Ignite Thin Client Port. local: A `Tensor` of type `bool`. Local flag that defines that data should be fetched from local host only. part: A `Tensor` of type `int32`. Partition data should be fetched from. page_size: A `Tensor` of type `int32`. Page size for Ignite Thin Client. schema: A `Tensor` of type `int32`. Internal structure that defines schema of cache objects. permutation: A `Tensor` of type `int32`. Internal structure that defines permutation of cache objects. name: A name for the operation (optional). Returns: A `Tensor` of type `variant`. This is the slowpath function for Eager mode. This is for function ignite_dataset Python wrappers around TensorFlow ops. This file is MACHINE GENERATED! Do not edit. Original C++ source file: gen_dataset_ops.cc Needed to trigger the call to _set_call_cpp_shape_fn. Add nodes to the TensorFlow graph. Add nodes to the TensorFlow graph. op { name: "IgniteDataset" input_arg { name: "cache_name" type: DT_STRING } input_arg { name: "host" type: DT_STRING } input_arg { name: "port" type: DT_INT32 } input_arg { name: "local" type: DT_BOOL } input_arg { name: "part" type: DT_INT32 } input_arg { name: "page_size" type: DT_INT32 } input_arg { name: "schema" type: DT_INT32 } input_arg { name: "permutation" type: DT_INT32 } output_arg { name: "handle" type: DT_VARIANT } is_stateful: true }
2,298
en
0.655406