file_name
large_stringlengths
4
140
prefix
large_stringlengths
0
12.1k
suffix
large_stringlengths
0
12k
middle
large_stringlengths
0
7.51k
fim_type
large_stringclasses
4 values
plat_win.py
# Copyright 2013 Hardcoded Software (http://www.hardcoded.net) # This software is licensed under the "BSD" License as described in the "LICENSE" file, # which should be included with this package. The terms are also available at
from ctypes.wintypes import HWND, UINT, LPCWSTR, BOOL import os.path as op from .compat import text_type shell32 = windll.shell32 SHFileOperationW = shell32.SHFileOperationW class SHFILEOPSTRUCTW(Structure): _fields_ = [ ("hwnd", HWND), ("wFunc", UINT), ("pFrom", LPCWSTR), ("pTo", LPCWSTR), ("fFlags", c_uint), ("fAnyOperationsAborted", BOOL), ("hNameMappings", c_uint), ("lpszProgressTitle", LPCWSTR), ] FO_MOVE = 1 FO_COPY = 2 FO_DELETE = 3 FO_RENAME = 4 FOF_MULTIDESTFILES = 1 FOF_SILENT = 4 FOF_NOCONFIRMATION = 16 FOF_ALLOWUNDO = 64 FOF_NOERRORUI = 1024 def send2trash(path): if not isinstance(path, text_type): path = text_type(path, 'mbcs') if not op.isabs(path): path = op.abspath(path) fileop = SHFILEOPSTRUCTW() fileop.hwnd = 0 fileop.wFunc = FO_DELETE fileop.pFrom = LPCWSTR(path + '\0') fileop.pTo = None fileop.fFlags = FOF_ALLOWUNDO | FOF_NOCONFIRMATION | FOF_NOERRORUI | FOF_SILENT fileop.fAnyOperationsAborted = 0 fileop.hNameMappings = 0 fileop.lpszProgressTitle = None result = SHFileOperationW(byref(fileop)) if result: msg = "Couldn't perform operation. Error code: %d" % result raise OSError(msg)
# http://www.hardcoded.net/licenses/bsd_license from __future__ import unicode_literals from ctypes import windll, Structure, byref, c_uint
random_line_split
plat_win.py
# Copyright 2013 Hardcoded Software (http://www.hardcoded.net) # This software is licensed under the "BSD" License as described in the "LICENSE" file, # which should be included with this package. The terms are also available at # http://www.hardcoded.net/licenses/bsd_license from __future__ import unicode_literals from ctypes import windll, Structure, byref, c_uint from ctypes.wintypes import HWND, UINT, LPCWSTR, BOOL import os.path as op from .compat import text_type shell32 = windll.shell32 SHFileOperationW = shell32.SHFileOperationW class SHFILEOPSTRUCTW(Structure): _fields_ = [ ("hwnd", HWND), ("wFunc", UINT), ("pFrom", LPCWSTR), ("pTo", LPCWSTR), ("fFlags", c_uint), ("fAnyOperationsAborted", BOOL), ("hNameMappings", c_uint), ("lpszProgressTitle", LPCWSTR), ] FO_MOVE = 1 FO_COPY = 2 FO_DELETE = 3 FO_RENAME = 4 FOF_MULTIDESTFILES = 1 FOF_SILENT = 4 FOF_NOCONFIRMATION = 16 FOF_ALLOWUNDO = 64 FOF_NOERRORUI = 1024 def send2trash(path):
if not isinstance(path, text_type): path = text_type(path, 'mbcs') if not op.isabs(path): path = op.abspath(path) fileop = SHFILEOPSTRUCTW() fileop.hwnd = 0 fileop.wFunc = FO_DELETE fileop.pFrom = LPCWSTR(path + '\0') fileop.pTo = None fileop.fFlags = FOF_ALLOWUNDO | FOF_NOCONFIRMATION | FOF_NOERRORUI | FOF_SILENT fileop.fAnyOperationsAborted = 0 fileop.hNameMappings = 0 fileop.lpszProgressTitle = None result = SHFileOperationW(byref(fileop)) if result: msg = "Couldn't perform operation. Error code: %d" % result raise OSError(msg)
identifier_body
plat_win.py
# Copyright 2013 Hardcoded Software (http://www.hardcoded.net) # This software is licensed under the "BSD" License as described in the "LICENSE" file, # which should be included with this package. The terms are also available at # http://www.hardcoded.net/licenses/bsd_license from __future__ import unicode_literals from ctypes import windll, Structure, byref, c_uint from ctypes.wintypes import HWND, UINT, LPCWSTR, BOOL import os.path as op from .compat import text_type shell32 = windll.shell32 SHFileOperationW = shell32.SHFileOperationW class SHFILEOPSTRUCTW(Structure): _fields_ = [ ("hwnd", HWND), ("wFunc", UINT), ("pFrom", LPCWSTR), ("pTo", LPCWSTR), ("fFlags", c_uint), ("fAnyOperationsAborted", BOOL), ("hNameMappings", c_uint), ("lpszProgressTitle", LPCWSTR), ] FO_MOVE = 1 FO_COPY = 2 FO_DELETE = 3 FO_RENAME = 4 FOF_MULTIDESTFILES = 1 FOF_SILENT = 4 FOF_NOCONFIRMATION = 16 FOF_ALLOWUNDO = 64 FOF_NOERRORUI = 1024 def send2trash(path): if not isinstance(path, text_type):
if not op.isabs(path): path = op.abspath(path) fileop = SHFILEOPSTRUCTW() fileop.hwnd = 0 fileop.wFunc = FO_DELETE fileop.pFrom = LPCWSTR(path + '\0') fileop.pTo = None fileop.fFlags = FOF_ALLOWUNDO | FOF_NOCONFIRMATION | FOF_NOERRORUI | FOF_SILENT fileop.fAnyOperationsAborted = 0 fileop.hNameMappings = 0 fileop.lpszProgressTitle = None result = SHFileOperationW(byref(fileop)) if result: msg = "Couldn't perform operation. Error code: %d" % result raise OSError(msg)
path = text_type(path, 'mbcs')
conditional_block
plat_win.py
# Copyright 2013 Hardcoded Software (http://www.hardcoded.net) # This software is licensed under the "BSD" License as described in the "LICENSE" file, # which should be included with this package. The terms are also available at # http://www.hardcoded.net/licenses/bsd_license from __future__ import unicode_literals from ctypes import windll, Structure, byref, c_uint from ctypes.wintypes import HWND, UINT, LPCWSTR, BOOL import os.path as op from .compat import text_type shell32 = windll.shell32 SHFileOperationW = shell32.SHFileOperationW class
(Structure): _fields_ = [ ("hwnd", HWND), ("wFunc", UINT), ("pFrom", LPCWSTR), ("pTo", LPCWSTR), ("fFlags", c_uint), ("fAnyOperationsAborted", BOOL), ("hNameMappings", c_uint), ("lpszProgressTitle", LPCWSTR), ] FO_MOVE = 1 FO_COPY = 2 FO_DELETE = 3 FO_RENAME = 4 FOF_MULTIDESTFILES = 1 FOF_SILENT = 4 FOF_NOCONFIRMATION = 16 FOF_ALLOWUNDO = 64 FOF_NOERRORUI = 1024 def send2trash(path): if not isinstance(path, text_type): path = text_type(path, 'mbcs') if not op.isabs(path): path = op.abspath(path) fileop = SHFILEOPSTRUCTW() fileop.hwnd = 0 fileop.wFunc = FO_DELETE fileop.pFrom = LPCWSTR(path + '\0') fileop.pTo = None fileop.fFlags = FOF_ALLOWUNDO | FOF_NOCONFIRMATION | FOF_NOERRORUI | FOF_SILENT fileop.fAnyOperationsAborted = 0 fileop.hNameMappings = 0 fileop.lpszProgressTitle = None result = SHFileOperationW(byref(fileop)) if result: msg = "Couldn't perform operation. Error code: %d" % result raise OSError(msg)
SHFILEOPSTRUCTW
identifier_name
revealer.rs
// Copyright 2013-2015, The Rust-GNOME Project Developers. // See the COPYRIGHT file at the top-level directory of this distribution. // Licensed under the MIT license, see the LICENSE file or <http://opensource.org/licenses/MIT> //! Hide and show with animation use ffi; use cast::{GTK_REVEALER}; use glib::{to_bool, to_gboolean}; /// GtkRevealer — Hide and show with animation struct_Widget!(Revealer); impl Revealer { pub fn new() -> Option<Revealer> { let tmp_pointer = unsafe { ffi::gtk_revealer_new() }; check_pointer!(tmp_pointer, Revealer) } pub fn get_reveal_child(&self) -> bool { unsafe { to_bool(ffi::gtk_revealer_get_reveal_child(GTK_REVEALER(self.pointer))) } } pub fn set_reveal_child(&self, reveal_child: bool) { unsafe { ffi::gtk_revealer_set_reveal_child(GTK_REVEALER(self.pointer), to_gboolean(reveal_child)) } } pub fn is_child_revealed(&self) -> bool {
pub fn get_transition_duration(&self) -> u32 { unsafe { ffi::gtk_revealer_get_transition_duration(GTK_REVEALER(self.pointer)) } } pub fn set_transition_duration(&self, duration: u32) { unsafe { ffi::gtk_revealer_set_transition_duration(GTK_REVEALER(self.pointer), duration) } } pub fn set_transition_type(&self, transition: ::RevealerTransitionType) { unsafe { ffi::gtk_revealer_set_transition_type(GTK_REVEALER(self.pointer), transition) } } pub fn get_transition_type(&self) -> ::RevealerTransitionType { unsafe { ffi::gtk_revealer_get_transition_type(GTK_REVEALER(self.pointer)) } } } impl_drop!(Revealer); impl_TraitWidget!(Revealer); impl ::ContainerTrait for Revealer {} impl ::BinTrait for Revealer {}
unsafe { to_bool(ffi::gtk_revealer_get_child_revealed(GTK_REVEALER(self.pointer))) } }
identifier_body
revealer.rs
// Copyright 2013-2015, The Rust-GNOME Project Developers. // See the COPYRIGHT file at the top-level directory of this distribution. // Licensed under the MIT license, see the LICENSE file or <http://opensource.org/licenses/MIT> //! Hide and show with animation use ffi; use cast::{GTK_REVEALER}; use glib::{to_bool, to_gboolean}; /// GtkRevealer — Hide and show with animation struct_Widget!(Revealer); impl Revealer { pub fn new() -> Option<Revealer> { let tmp_pointer = unsafe { ffi::gtk_revealer_new() }; check_pointer!(tmp_pointer, Revealer) } pub fn get_reveal_child(&self) -> bool { unsafe { to_bool(ffi::gtk_revealer_get_reveal_child(GTK_REVEALER(self.pointer))) } } pub fn set_reveal_child(&self, reveal_child: bool) { unsafe { ffi::gtk_revealer_set_reveal_child(GTK_REVEALER(self.pointer), to_gboolean(reveal_child)) } } pub fn is_child_revealed(&self) -> bool { unsafe { to_bool(ffi::gtk_revealer_get_child_revealed(GTK_REVEALER(self.pointer))) } } pub fn get_transition_duration(&self) -> u32 { unsafe { ffi::gtk_revealer_get_transition_duration(GTK_REVEALER(self.pointer)) } } pub fn set_transition_duration(&self, duration: u32) { unsafe { ffi::gtk_revealer_set_transition_duration(GTK_REVEALER(self.pointer), duration) } } pub fn set_transition_type(&self, transition: ::RevealerTransitionType) { unsafe { ffi::gtk_revealer_set_transition_type(GTK_REVEALER(self.pointer), transition) } } pub fn get_transition_type(&self) -> ::RevealerTransitionType { unsafe { ffi::gtk_revealer_get_transition_type(GTK_REVEALER(self.pointer)) } } }
impl ::ContainerTrait for Revealer {} impl ::BinTrait for Revealer {}
impl_drop!(Revealer); impl_TraitWidget!(Revealer);
random_line_split
revealer.rs
// Copyright 2013-2015, The Rust-GNOME Project Developers. // See the COPYRIGHT file at the top-level directory of this distribution. // Licensed under the MIT license, see the LICENSE file or <http://opensource.org/licenses/MIT> //! Hide and show with animation use ffi; use cast::{GTK_REVEALER}; use glib::{to_bool, to_gboolean}; /// GtkRevealer — Hide and show with animation struct_Widget!(Revealer); impl Revealer { pub fn new() -> Option<Revealer> { let tmp_pointer = unsafe { ffi::gtk_revealer_new() }; check_pointer!(tmp_pointer, Revealer) } pub fn get_reveal_child(&self) -> bool { unsafe { to_bool(ffi::gtk_revealer_get_reveal_child(GTK_REVEALER(self.pointer))) } } pub fn set_reveal_child(&self, reveal_child: bool) { unsafe { ffi::gtk_revealer_set_reveal_child(GTK_REVEALER(self.pointer), to_gboolean(reveal_child)) } } pub fn is_child_revealed(&self) -> bool { unsafe { to_bool(ffi::gtk_revealer_get_child_revealed(GTK_REVEALER(self.pointer))) } } pub fn get_transition_duration(&self) -> u32 { unsafe { ffi::gtk_revealer_get_transition_duration(GTK_REVEALER(self.pointer)) } } pub fn se
self, duration: u32) { unsafe { ffi::gtk_revealer_set_transition_duration(GTK_REVEALER(self.pointer), duration) } } pub fn set_transition_type(&self, transition: ::RevealerTransitionType) { unsafe { ffi::gtk_revealer_set_transition_type(GTK_REVEALER(self.pointer), transition) } } pub fn get_transition_type(&self) -> ::RevealerTransitionType { unsafe { ffi::gtk_revealer_get_transition_type(GTK_REVEALER(self.pointer)) } } } impl_drop!(Revealer); impl_TraitWidget!(Revealer); impl ::ContainerTrait for Revealer {} impl ::BinTrait for Revealer {}
t_transition_duration(&
identifier_name
recent_commits_test.py
# coding: utf-8 # # Copyright 2018 The Oppia Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS-IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for recent commit controllers.""" from __future__ import annotations from core import feconf from core.platform import models from core.tests import test_utils (exp_models,) = models.Registry.import_models([models.NAMES.exploration]) class RecentCommitsHandlerUnitTests(test_utils.GenericTestBase): """Test the RecentCommitsHandler class.""" def setUp(self): super(RecentCommitsHandlerUnitTests, self).setUp() self.signup(self.MODERATOR_EMAIL, self.MODERATOR_USERNAME) self.set_moderators([self.MODERATOR_USERNAME]) self.signup(self.VIEWER_EMAIL, self.VIEWER_USERNAME) self.committer_1_id = self.get_user_id_from_email(self.VIEWER_EMAIL) self.signup(self.NEW_USER_EMAIL, self.NEW_USER_USERNAME) self.committer_2_id = self.get_user_id_from_email(self.NEW_USER_EMAIL) commit1 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_1', 0, self.committer_1_id, 'create', 'created first commit', [], 'public', True) commit2 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_1', 1, self.committer_2_id, 'edit', 'edited commit', [],
'public', True) commit3 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_2', 0, self.committer_1_id, 'create', 'created second commit', [], 'private', False) commit1.exploration_id = 'exp_1' commit2.exploration_id = 'exp_1' commit3.exploration_id = 'exp_2' commit1.update_timestamps() commit1.put() commit2.update_timestamps() commit2.put() commit3.update_timestamps() commit3.put() def test_get_recent_commits(self): """Test that this method should return all nonprivate commits.""" self.login(self.MODERATOR_EMAIL) response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual(len(response_dict['results']), 2) self.assertDictContainsSubset( {'username': self.VIEWER_USERNAME, 'exploration_id': 'exp_1', 'post_commit_status': 'public', 'version': 0, 'commit_message': 'created first commit', 'commit_type': 'create'}, response_dict['results'][1]) self.assertDictContainsSubset( {'username': self.NEW_USER_USERNAME, 'exploration_id': 'exp_1', 'post_commit_status': 'public', 'version': 1, 'commit_message': 'edited commit', 'commit_type': 'edit'}, response_dict['results'][0]) self.logout() def test_get_recent_commits_explorations(self): """Test that the response dict contains the correct exploration.""" self.login(self.MODERATOR_EMAIL) self.save_new_default_exploration( 'exp_1', 'owner0', title='MyExploration') response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual(len(response_dict['exp_ids_to_exp_data']), 1) self.assertEqual( response_dict['exp_ids_to_exp_data']['exp_1']['title'], 'MyExploration') self.logout() def test_get_recent_commits_three_pages_with_cursor(self): self.login(self.MODERATOR_EMAIL) response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertFalse(response_dict['more']) for i in range(feconf.COMMIT_LIST_PAGE_SIZE * 2): entity_id = 'my_entity_%s' % i exp_id = 'exp_%s' % i commit_i = exp_models.ExplorationCommitLogEntryModel.create( entity_id, 0, self.committer_2_id, 'create', 'created commit', [], 'public', True) commit_i.exploration_id = exp_id commit_i.update_timestamps() commit_i.put() response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual( len(response_dict['results']), feconf.COMMIT_LIST_PAGE_SIZE) self.assertTrue(response_dict['more']) cursor = response_dict['cursor'] response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={ 'query_type': 'all_non_private_commits', 'cursor': cursor }) self.assertEqual( len(response_dict['results']), feconf.COMMIT_LIST_PAGE_SIZE) self.assertTrue(response_dict['more']) cursor = response_dict['cursor'] response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={ 'query_type': 'all_non_private_commits', 'cursor': cursor }) self.assertFalse(response_dict['more']) self.assertEqual(len(response_dict['results']), 2) self.logout() def test_get_recent_commits_with_invalid_query_type_returns_404_status( self): self.login(self.MODERATOR_EMAIL) self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'invalid_query_type'}, expected_status_int=404) self.logout()
random_line_split
recent_commits_test.py
# coding: utf-8 # # Copyright 2018 The Oppia Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS-IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for recent commit controllers.""" from __future__ import annotations from core import feconf from core.platform import models from core.tests import test_utils (exp_models,) = models.Registry.import_models([models.NAMES.exploration]) class RecentCommitsHandlerUnitTests(test_utils.GenericTestBase):
'created second commit', [], 'private', False) commit1.exploration_id = 'exp_1' commit2.exploration_id = 'exp_1' commit3.exploration_id = 'exp_2' commit1.update_timestamps() commit1.put() commit2.update_timestamps() commit2.put() commit3.update_timestamps() commit3.put() def test_get_recent_commits(self): """Test that this method should return all nonprivate commits.""" self.login(self.MODERATOR_EMAIL) response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual(len(response_dict['results']), 2) self.assertDictContainsSubset( {'username': self.VIEWER_USERNAME, 'exploration_id': 'exp_1', 'post_commit_status': 'public', 'version': 0, 'commit_message': 'created first commit', 'commit_type': 'create'}, response_dict['results'][1]) self.assertDictContainsSubset( {'username': self.NEW_USER_USERNAME, 'exploration_id': 'exp_1', 'post_commit_status': 'public', 'version': 1, 'commit_message': 'edited commit', 'commit_type': 'edit'}, response_dict['results'][0]) self.logout() def test_get_recent_commits_explorations(self): """Test that the response dict contains the correct exploration.""" self.login(self.MODERATOR_EMAIL) self.save_new_default_exploration( 'exp_1', 'owner0', title='MyExploration') response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual(len(response_dict['exp_ids_to_exp_data']), 1) self.assertEqual( response_dict['exp_ids_to_exp_data']['exp_1']['title'], 'MyExploration') self.logout() def test_get_recent_commits_three_pages_with_cursor(self): self.login(self.MODERATOR_EMAIL) response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertFalse(response_dict['more']) for i in range(feconf.COMMIT_LIST_PAGE_SIZE * 2): entity_id = 'my_entity_%s' % i exp_id = 'exp_%s' % i commit_i = exp_models.ExplorationCommitLogEntryModel.create( entity_id, 0, self.committer_2_id, 'create', 'created commit', [], 'public', True) commit_i.exploration_id = exp_id commit_i.update_timestamps() commit_i.put() response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual( len(response_dict['results']), feconf.COMMIT_LIST_PAGE_SIZE) self.assertTrue(response_dict['more']) cursor = response_dict['cursor'] response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={ 'query_type': 'all_non_private_commits', 'cursor': cursor }) self.assertEqual( len(response_dict['results']), feconf.COMMIT_LIST_PAGE_SIZE) self.assertTrue(response_dict['more']) cursor = response_dict['cursor'] response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={ 'query_type': 'all_non_private_commits', 'cursor': cursor }) self.assertFalse(response_dict['more']) self.assertEqual(len(response_dict['results']), 2) self.logout() def test_get_recent_commits_with_invalid_query_type_returns_404_status( self): self.login(self.MODERATOR_EMAIL) self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'invalid_query_type'}, expected_status_int=404) self.logout()
"""Test the RecentCommitsHandler class.""" def setUp(self): super(RecentCommitsHandlerUnitTests, self).setUp() self.signup(self.MODERATOR_EMAIL, self.MODERATOR_USERNAME) self.set_moderators([self.MODERATOR_USERNAME]) self.signup(self.VIEWER_EMAIL, self.VIEWER_USERNAME) self.committer_1_id = self.get_user_id_from_email(self.VIEWER_EMAIL) self.signup(self.NEW_USER_EMAIL, self.NEW_USER_USERNAME) self.committer_2_id = self.get_user_id_from_email(self.NEW_USER_EMAIL) commit1 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_1', 0, self.committer_1_id, 'create', 'created first commit', [], 'public', True) commit2 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_1', 1, self.committer_2_id, 'edit', 'edited commit', [], 'public', True) commit3 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_2', 0, self.committer_1_id, 'create',
identifier_body
recent_commits_test.py
# coding: utf-8 # # Copyright 2018 The Oppia Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS-IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for recent commit controllers.""" from __future__ import annotations from core import feconf from core.platform import models from core.tests import test_utils (exp_models,) = models.Registry.import_models([models.NAMES.exploration]) class
(test_utils.GenericTestBase): """Test the RecentCommitsHandler class.""" def setUp(self): super(RecentCommitsHandlerUnitTests, self).setUp() self.signup(self.MODERATOR_EMAIL, self.MODERATOR_USERNAME) self.set_moderators([self.MODERATOR_USERNAME]) self.signup(self.VIEWER_EMAIL, self.VIEWER_USERNAME) self.committer_1_id = self.get_user_id_from_email(self.VIEWER_EMAIL) self.signup(self.NEW_USER_EMAIL, self.NEW_USER_USERNAME) self.committer_2_id = self.get_user_id_from_email(self.NEW_USER_EMAIL) commit1 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_1', 0, self.committer_1_id, 'create', 'created first commit', [], 'public', True) commit2 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_1', 1, self.committer_2_id, 'edit', 'edited commit', [], 'public', True) commit3 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_2', 0, self.committer_1_id, 'create', 'created second commit', [], 'private', False) commit1.exploration_id = 'exp_1' commit2.exploration_id = 'exp_1' commit3.exploration_id = 'exp_2' commit1.update_timestamps() commit1.put() commit2.update_timestamps() commit2.put() commit3.update_timestamps() commit3.put() def test_get_recent_commits(self): """Test that this method should return all nonprivate commits.""" self.login(self.MODERATOR_EMAIL) response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual(len(response_dict['results']), 2) self.assertDictContainsSubset( {'username': self.VIEWER_USERNAME, 'exploration_id': 'exp_1', 'post_commit_status': 'public', 'version': 0, 'commit_message': 'created first commit', 'commit_type': 'create'}, response_dict['results'][1]) self.assertDictContainsSubset( {'username': self.NEW_USER_USERNAME, 'exploration_id': 'exp_1', 'post_commit_status': 'public', 'version': 1, 'commit_message': 'edited commit', 'commit_type': 'edit'}, response_dict['results'][0]) self.logout() def test_get_recent_commits_explorations(self): """Test that the response dict contains the correct exploration.""" self.login(self.MODERATOR_EMAIL) self.save_new_default_exploration( 'exp_1', 'owner0', title='MyExploration') response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual(len(response_dict['exp_ids_to_exp_data']), 1) self.assertEqual( response_dict['exp_ids_to_exp_data']['exp_1']['title'], 'MyExploration') self.logout() def test_get_recent_commits_three_pages_with_cursor(self): self.login(self.MODERATOR_EMAIL) response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertFalse(response_dict['more']) for i in range(feconf.COMMIT_LIST_PAGE_SIZE * 2): entity_id = 'my_entity_%s' % i exp_id = 'exp_%s' % i commit_i = exp_models.ExplorationCommitLogEntryModel.create( entity_id, 0, self.committer_2_id, 'create', 'created commit', [], 'public', True) commit_i.exploration_id = exp_id commit_i.update_timestamps() commit_i.put() response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual( len(response_dict['results']), feconf.COMMIT_LIST_PAGE_SIZE) self.assertTrue(response_dict['more']) cursor = response_dict['cursor'] response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={ 'query_type': 'all_non_private_commits', 'cursor': cursor }) self.assertEqual( len(response_dict['results']), feconf.COMMIT_LIST_PAGE_SIZE) self.assertTrue(response_dict['more']) cursor = response_dict['cursor'] response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={ 'query_type': 'all_non_private_commits', 'cursor': cursor }) self.assertFalse(response_dict['more']) self.assertEqual(len(response_dict['results']), 2) self.logout() def test_get_recent_commits_with_invalid_query_type_returns_404_status( self): self.login(self.MODERATOR_EMAIL) self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'invalid_query_type'}, expected_status_int=404) self.logout()
RecentCommitsHandlerUnitTests
identifier_name
recent_commits_test.py
# coding: utf-8 # # Copyright 2018 The Oppia Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS-IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for recent commit controllers.""" from __future__ import annotations from core import feconf from core.platform import models from core.tests import test_utils (exp_models,) = models.Registry.import_models([models.NAMES.exploration]) class RecentCommitsHandlerUnitTests(test_utils.GenericTestBase): """Test the RecentCommitsHandler class.""" def setUp(self): super(RecentCommitsHandlerUnitTests, self).setUp() self.signup(self.MODERATOR_EMAIL, self.MODERATOR_USERNAME) self.set_moderators([self.MODERATOR_USERNAME]) self.signup(self.VIEWER_EMAIL, self.VIEWER_USERNAME) self.committer_1_id = self.get_user_id_from_email(self.VIEWER_EMAIL) self.signup(self.NEW_USER_EMAIL, self.NEW_USER_USERNAME) self.committer_2_id = self.get_user_id_from_email(self.NEW_USER_EMAIL) commit1 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_1', 0, self.committer_1_id, 'create', 'created first commit', [], 'public', True) commit2 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_1', 1, self.committer_2_id, 'edit', 'edited commit', [], 'public', True) commit3 = exp_models.ExplorationCommitLogEntryModel.create( 'entity_2', 0, self.committer_1_id, 'create', 'created second commit', [], 'private', False) commit1.exploration_id = 'exp_1' commit2.exploration_id = 'exp_1' commit3.exploration_id = 'exp_2' commit1.update_timestamps() commit1.put() commit2.update_timestamps() commit2.put() commit3.update_timestamps() commit3.put() def test_get_recent_commits(self): """Test that this method should return all nonprivate commits.""" self.login(self.MODERATOR_EMAIL) response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual(len(response_dict['results']), 2) self.assertDictContainsSubset( {'username': self.VIEWER_USERNAME, 'exploration_id': 'exp_1', 'post_commit_status': 'public', 'version': 0, 'commit_message': 'created first commit', 'commit_type': 'create'}, response_dict['results'][1]) self.assertDictContainsSubset( {'username': self.NEW_USER_USERNAME, 'exploration_id': 'exp_1', 'post_commit_status': 'public', 'version': 1, 'commit_message': 'edited commit', 'commit_type': 'edit'}, response_dict['results'][0]) self.logout() def test_get_recent_commits_explorations(self): """Test that the response dict contains the correct exploration.""" self.login(self.MODERATOR_EMAIL) self.save_new_default_exploration( 'exp_1', 'owner0', title='MyExploration') response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual(len(response_dict['exp_ids_to_exp_data']), 1) self.assertEqual( response_dict['exp_ids_to_exp_data']['exp_1']['title'], 'MyExploration') self.logout() def test_get_recent_commits_three_pages_with_cursor(self): self.login(self.MODERATOR_EMAIL) response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertFalse(response_dict['more']) for i in range(feconf.COMMIT_LIST_PAGE_SIZE * 2):
response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'all_non_private_commits'}) self.assertEqual( len(response_dict['results']), feconf.COMMIT_LIST_PAGE_SIZE) self.assertTrue(response_dict['more']) cursor = response_dict['cursor'] response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={ 'query_type': 'all_non_private_commits', 'cursor': cursor }) self.assertEqual( len(response_dict['results']), feconf.COMMIT_LIST_PAGE_SIZE) self.assertTrue(response_dict['more']) cursor = response_dict['cursor'] response_dict = self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={ 'query_type': 'all_non_private_commits', 'cursor': cursor }) self.assertFalse(response_dict['more']) self.assertEqual(len(response_dict['results']), 2) self.logout() def test_get_recent_commits_with_invalid_query_type_returns_404_status( self): self.login(self.MODERATOR_EMAIL) self.get_json( feconf.RECENT_COMMITS_DATA_URL, params={'query_type': 'invalid_query_type'}, expected_status_int=404) self.logout()
entity_id = 'my_entity_%s' % i exp_id = 'exp_%s' % i commit_i = exp_models.ExplorationCommitLogEntryModel.create( entity_id, 0, self.committer_2_id, 'create', 'created commit', [], 'public', True) commit_i.exploration_id = exp_id commit_i.update_timestamps() commit_i.put()
conditional_block
test_uptime.py
import unittest try: from unittest import mock except ImportError: import mock from pi3bar.plugins.uptime import get_uptime_seconds, uptime_format, Uptime class GetUptimeSecondsTestCase(unittest.TestCase): def test(self): m = mock.mock_open(read_data='5') m.return_value.readline.return_value = '5' # py33 with mock.patch('pi3bar.plugins.uptime.open', m, create=True): seconds = get_uptime_seconds() self.assertEqual(5, seconds) class UptimeFormatTestCase(unittest.TestCase): def test_seconds(self): s = uptime_format(5) self.assertEqual('0:00:00:05', s) def test_minutes(self): s = uptime_format(3540) self.assertEqual('0:00:59:00', s) def test_hours(self): s = uptime_format(49020) self.assertEqual('0:13:37:00', s) def test_days(self):
def test_format_days_applied_to_hours(self): s = uptime_format(135420, '%H:%M:%S') self.assertEqual('37:37:00', s) def test_format_hours_applied_to_minutes(self): s = uptime_format(49020, '%M:%S') self.assertEqual('817:00', s) class UptimeTestCase(unittest.TestCase): def test(self): plugin = Uptime() self.assertEqual('%d days %H:%M:%S up', plugin.full_format) self.assertEqual('%dd %H:%M up', plugin.short_format) @mock.patch('pi3bar.plugins.uptime.get_uptime_seconds') def test_cycle(self, mock_get_uptime_seconds): plugin = Uptime() mock_get_uptime_seconds.return_value = 49020 plugin.cycle() self.assertEqual('0 days 13:37:00 up', plugin.full_text) self.assertEqual('0d 13:37 up', plugin.short_text)
s = uptime_format(135420) self.assertEqual('1:13:37:00', s)
identifier_body
test_uptime.py
import unittest try: from unittest import mock except ImportError: import mock from pi3bar.plugins.uptime import get_uptime_seconds, uptime_format, Uptime class GetUptimeSecondsTestCase(unittest.TestCase): def test(self): m = mock.mock_open(read_data='5') m.return_value.readline.return_value = '5' # py33 with mock.patch('pi3bar.plugins.uptime.open', m, create=True): seconds = get_uptime_seconds() self.assertEqual(5, seconds) class
(unittest.TestCase): def test_seconds(self): s = uptime_format(5) self.assertEqual('0:00:00:05', s) def test_minutes(self): s = uptime_format(3540) self.assertEqual('0:00:59:00', s) def test_hours(self): s = uptime_format(49020) self.assertEqual('0:13:37:00', s) def test_days(self): s = uptime_format(135420) self.assertEqual('1:13:37:00', s) def test_format_days_applied_to_hours(self): s = uptime_format(135420, '%H:%M:%S') self.assertEqual('37:37:00', s) def test_format_hours_applied_to_minutes(self): s = uptime_format(49020, '%M:%S') self.assertEqual('817:00', s) class UptimeTestCase(unittest.TestCase): def test(self): plugin = Uptime() self.assertEqual('%d days %H:%M:%S up', plugin.full_format) self.assertEqual('%dd %H:%M up', plugin.short_format) @mock.patch('pi3bar.plugins.uptime.get_uptime_seconds') def test_cycle(self, mock_get_uptime_seconds): plugin = Uptime() mock_get_uptime_seconds.return_value = 49020 plugin.cycle() self.assertEqual('0 days 13:37:00 up', plugin.full_text) self.assertEqual('0d 13:37 up', plugin.short_text)
UptimeFormatTestCase
identifier_name
test_uptime.py
import unittest try: from unittest import mock except ImportError: import mock from pi3bar.plugins.uptime import get_uptime_seconds, uptime_format, Uptime class GetUptimeSecondsTestCase(unittest.TestCase): def test(self): m = mock.mock_open(read_data='5')
class UptimeFormatTestCase(unittest.TestCase): def test_seconds(self): s = uptime_format(5) self.assertEqual('0:00:00:05', s) def test_minutes(self): s = uptime_format(3540) self.assertEqual('0:00:59:00', s) def test_hours(self): s = uptime_format(49020) self.assertEqual('0:13:37:00', s) def test_days(self): s = uptime_format(135420) self.assertEqual('1:13:37:00', s) def test_format_days_applied_to_hours(self): s = uptime_format(135420, '%H:%M:%S') self.assertEqual('37:37:00', s) def test_format_hours_applied_to_minutes(self): s = uptime_format(49020, '%M:%S') self.assertEqual('817:00', s) class UptimeTestCase(unittest.TestCase): def test(self): plugin = Uptime() self.assertEqual('%d days %H:%M:%S up', plugin.full_format) self.assertEqual('%dd %H:%M up', plugin.short_format) @mock.patch('pi3bar.plugins.uptime.get_uptime_seconds') def test_cycle(self, mock_get_uptime_seconds): plugin = Uptime() mock_get_uptime_seconds.return_value = 49020 plugin.cycle() self.assertEqual('0 days 13:37:00 up', plugin.full_text) self.assertEqual('0d 13:37 up', plugin.short_text)
m.return_value.readline.return_value = '5' # py33 with mock.patch('pi3bar.plugins.uptime.open', m, create=True): seconds = get_uptime_seconds() self.assertEqual(5, seconds)
random_line_split
minesweeperNN.js
// ==UserScript== // @name mineAI // @namespace minesAI // @include http://minesweeperonline.com/#beginner-night // @version 1 // @required http://localhost:8000/convnetjs.js // @grant none // ==/UserScript== // Load the library. var D = document; var appTarg = D.getElementsByTagName ('head')[0] || D.body || D.documentElement; var jsNode = D.createElement ('script'); jsNode.src = 'http://localhost:8000/convnetjs.js'; jsNode.addEventListener ("load", initConvNetJsOnDelay, false); appTarg.appendChild (jsNode); // Allow some time for the library to initialize after loading. function initConvNetJsOnDelay ()
// Call the library's start-up function, if any. Note needed use of unsafeWindow. function initConvNetJs () { // species a 2-layer neural network with one hidden layer of 20 neurons var layer_defs = []; // ConvNetJS works on 3-Dimensional volumes (sx, sy, depth), but if you're not dealing with images // then the first two dimensions (sx, sy) will always be kept at size 1 layer_defs.push({type:'input', out_sx:1, out_sy:1, out_depth:2}); // declare 4 neurons, followed by ReLU (rectified linear unit non-linearity) layer_defs.push({type:'fc', num_neurons:4, activation:'relu'}); // 3 more for good measure layer_defs.push({type:'fc', num_neurons:3, activation:'relu'}); // declare the linear classifier on top of the previous hidden layer layer_defs.push({type:'softmax', num_classes:2}); // defined our net with unsafeWindow for use in GreaseMonkey var net = new unsafeWindow.convnetjs.Net(); // create our net with layers as defined above net.makeLayers(layer_defs); // define trainer var trainer = new convnetjs.SGDTrainer(net, {learning_rate:0.01, l2_decay:0.001}); // define inputs (XOR) var t1 = new convnetjs.Vol([0, 0]); // class 0 var t2 = new convnetjs.Vol([0, 1]); // class 1 var t3 = new convnetjs.Vol([1, 0]); // class 1 var t4 = new convnetjs.Vol([1, 1]); // class 0 // train for 1000 iterations with corresponding classes for (var i = 0; i < 1000; i++) { trainer.train(t1, 0); trainer.train(t2, 1); trainer.train(t3, 1); trainer.train(t4, 0); } // learned probability var prob00 = net.forward(t1); var prob01 = net.forward(t2); var prob10 = net.forward(t3); var prob11 = net.forward(t4); // log probability console.log('p(0 | 00): ' + prob00.w[0] + ", p(1 | 00): " + prob00.w[1]); console.log('p(0 | 01): ' + prob01.w[0] + ", p(1 | 01): " + prob01.w[1]); console.log('p(0 | 10): ' + prob10.w[0] + ", p(1 | 10): " + prob10.w[1]); console.log('p(0 | 11): ' + prob11.w[0] + ", p(1 | 11): " + prob11.w[1]); } alert("Done");
{ setTimeout (initConvNetJs, 666); }
identifier_body
minesweeperNN.js
// ==UserScript== // @name mineAI // @namespace minesAI // @include http://minesweeperonline.com/#beginner-night // @version 1 // @required http://localhost:8000/convnetjs.js // @grant none // ==/UserScript== // Load the library. var D = document; var appTarg = D.getElementsByTagName ('head')[0] || D.body || D.documentElement; var jsNode = D.createElement ('script'); jsNode.src = 'http://localhost:8000/convnetjs.js'; jsNode.addEventListener ("load", initConvNetJsOnDelay, false); appTarg.appendChild (jsNode); // Allow some time for the library to initialize after loading. function initConvNetJsOnDelay () { setTimeout (initConvNetJs, 666); } // Call the library's start-up function, if any. Note needed use of unsafeWindow. function initConvNetJs () { // species a 2-layer neural network with one hidden layer of 20 neurons var layer_defs = []; // ConvNetJS works on 3-Dimensional volumes (sx, sy, depth), but if you're not dealing with images // then the first two dimensions (sx, sy) will always be kept at size 1 layer_defs.push({type:'input', out_sx:1, out_sy:1, out_depth:2}); // declare 4 neurons, followed by ReLU (rectified linear unit non-linearity) layer_defs.push({type:'fc', num_neurons:4, activation:'relu'}); // 3 more for good measure layer_defs.push({type:'fc', num_neurons:3, activation:'relu'}); // declare the linear classifier on top of the previous hidden layer layer_defs.push({type:'softmax', num_classes:2}); // defined our net with unsafeWindow for use in GreaseMonkey var net = new unsafeWindow.convnetjs.Net(); // create our net with layers as defined above net.makeLayers(layer_defs); // define trainer var trainer = new convnetjs.SGDTrainer(net, {learning_rate:0.01, l2_decay:0.001}); // define inputs (XOR) var t1 = new convnetjs.Vol([0, 0]); // class 0 var t2 = new convnetjs.Vol([0, 1]); // class 1 var t3 = new convnetjs.Vol([1, 0]); // class 1 var t4 = new convnetjs.Vol([1, 1]); // class 0 // train for 1000 iterations with corresponding classes for (var i = 0; i < 1000; i++)
// learned probability var prob00 = net.forward(t1); var prob01 = net.forward(t2); var prob10 = net.forward(t3); var prob11 = net.forward(t4); // log probability console.log('p(0 | 00): ' + prob00.w[0] + ", p(1 | 00): " + prob00.w[1]); console.log('p(0 | 01): ' + prob01.w[0] + ", p(1 | 01): " + prob01.w[1]); console.log('p(0 | 10): ' + prob10.w[0] + ", p(1 | 10): " + prob10.w[1]); console.log('p(0 | 11): ' + prob11.w[0] + ", p(1 | 11): " + prob11.w[1]); } alert("Done");
{ trainer.train(t1, 0); trainer.train(t2, 1); trainer.train(t3, 1); trainer.train(t4, 0); }
conditional_block
minesweeperNN.js
// ==UserScript== // @name mineAI // @namespace minesAI // @include http://minesweeperonline.com/#beginner-night // @version 1 // @required http://localhost:8000/convnetjs.js // @grant none // ==/UserScript== // Load the library. var D = document; var appTarg = D.getElementsByTagName ('head')[0] || D.body || D.documentElement; var jsNode = D.createElement ('script'); jsNode.src = 'http://localhost:8000/convnetjs.js'; jsNode.addEventListener ("load", initConvNetJsOnDelay, false); appTarg.appendChild (jsNode); // Allow some time for the library to initialize after loading. function initConvNetJsOnDelay () { setTimeout (initConvNetJs, 666); } // Call the library's start-up function, if any. Note needed use of unsafeWindow. function
() { // species a 2-layer neural network with one hidden layer of 20 neurons var layer_defs = []; // ConvNetJS works on 3-Dimensional volumes (sx, sy, depth), but if you're not dealing with images // then the first two dimensions (sx, sy) will always be kept at size 1 layer_defs.push({type:'input', out_sx:1, out_sy:1, out_depth:2}); // declare 4 neurons, followed by ReLU (rectified linear unit non-linearity) layer_defs.push({type:'fc', num_neurons:4, activation:'relu'}); // 3 more for good measure layer_defs.push({type:'fc', num_neurons:3, activation:'relu'}); // declare the linear classifier on top of the previous hidden layer layer_defs.push({type:'softmax', num_classes:2}); // defined our net with unsafeWindow for use in GreaseMonkey var net = new unsafeWindow.convnetjs.Net(); // create our net with layers as defined above net.makeLayers(layer_defs); // define trainer var trainer = new convnetjs.SGDTrainer(net, {learning_rate:0.01, l2_decay:0.001}); // define inputs (XOR) var t1 = new convnetjs.Vol([0, 0]); // class 0 var t2 = new convnetjs.Vol([0, 1]); // class 1 var t3 = new convnetjs.Vol([1, 0]); // class 1 var t4 = new convnetjs.Vol([1, 1]); // class 0 // train for 1000 iterations with corresponding classes for (var i = 0; i < 1000; i++) { trainer.train(t1, 0); trainer.train(t2, 1); trainer.train(t3, 1); trainer.train(t4, 0); } // learned probability var prob00 = net.forward(t1); var prob01 = net.forward(t2); var prob10 = net.forward(t3); var prob11 = net.forward(t4); // log probability console.log('p(0 | 00): ' + prob00.w[0] + ", p(1 | 00): " + prob00.w[1]); console.log('p(0 | 01): ' + prob01.w[0] + ", p(1 | 01): " + prob01.w[1]); console.log('p(0 | 10): ' + prob10.w[0] + ", p(1 | 10): " + prob10.w[1]); console.log('p(0 | 11): ' + prob11.w[0] + ", p(1 | 11): " + prob11.w[1]); } alert("Done");
initConvNetJs
identifier_name
minesweeperNN.js
// ==UserScript== // @name mineAI // @namespace minesAI // @include http://minesweeperonline.com/#beginner-night // @version 1 // @required http://localhost:8000/convnetjs.js // @grant none // ==/UserScript== // Load the library. var D = document; var appTarg = D.getElementsByTagName ('head')[0] || D.body || D.documentElement; var jsNode = D.createElement ('script'); jsNode.src = 'http://localhost:8000/convnetjs.js'; jsNode.addEventListener ("load", initConvNetJsOnDelay, false); appTarg.appendChild (jsNode); // Allow some time for the library to initialize after loading. function initConvNetJsOnDelay () { setTimeout (initConvNetJs, 666);
// species a 2-layer neural network with one hidden layer of 20 neurons var layer_defs = []; // ConvNetJS works on 3-Dimensional volumes (sx, sy, depth), but if you're not dealing with images // then the first two dimensions (sx, sy) will always be kept at size 1 layer_defs.push({type:'input', out_sx:1, out_sy:1, out_depth:2}); // declare 4 neurons, followed by ReLU (rectified linear unit non-linearity) layer_defs.push({type:'fc', num_neurons:4, activation:'relu'}); // 3 more for good measure layer_defs.push({type:'fc', num_neurons:3, activation:'relu'}); // declare the linear classifier on top of the previous hidden layer layer_defs.push({type:'softmax', num_classes:2}); // defined our net with unsafeWindow for use in GreaseMonkey var net = new unsafeWindow.convnetjs.Net(); // create our net with layers as defined above net.makeLayers(layer_defs); // define trainer var trainer = new convnetjs.SGDTrainer(net, {learning_rate:0.01, l2_decay:0.001}); // define inputs (XOR) var t1 = new convnetjs.Vol([0, 0]); // class 0 var t2 = new convnetjs.Vol([0, 1]); // class 1 var t3 = new convnetjs.Vol([1, 0]); // class 1 var t4 = new convnetjs.Vol([1, 1]); // class 0 // train for 1000 iterations with corresponding classes for (var i = 0; i < 1000; i++) { trainer.train(t1, 0); trainer.train(t2, 1); trainer.train(t3, 1); trainer.train(t4, 0); } // learned probability var prob00 = net.forward(t1); var prob01 = net.forward(t2); var prob10 = net.forward(t3); var prob11 = net.forward(t4); // log probability console.log('p(0 | 00): ' + prob00.w[0] + ", p(1 | 00): " + prob00.w[1]); console.log('p(0 | 01): ' + prob01.w[0] + ", p(1 | 01): " + prob01.w[1]); console.log('p(0 | 10): ' + prob10.w[0] + ", p(1 | 10): " + prob10.w[1]); console.log('p(0 | 11): ' + prob11.w[0] + ", p(1 | 11): " + prob11.w[1]); } alert("Done");
} // Call the library's start-up function, if any. Note needed use of unsafeWindow. function initConvNetJs () {
random_line_split
state.directive.ts
import {Directive, ElementRef, Renderer, Input, OnChanges} from '@angular/core'; @Directive({ selector: '[state]' }) export class StateDirective implements OnChanges {
constructor(renderer: Renderer, el: ElementRef) { this.element = el; this.renderer = renderer; } ngOnChanges() { let cssClass = `state-${this.itemState}`; let text = 'A livrer'; let elementNode = this.element.nativeElement; switch (this.itemState) { case 1 : text = "aaaaa"; break; case 2 : text = "xxxxx"; break; case 3 : text = "wwwww"; break; default: text = "zied !!!"; } this.renderer.setElementClass(elementNode, cssClass, true); this.renderer.setText(elementNode, text); } }
@Input('state') itemState: any; private element: ElementRef; private renderer: Renderer;
random_line_split
state.directive.ts
import {Directive, ElementRef, Renderer, Input, OnChanges} from '@angular/core'; @Directive({ selector: '[state]' }) export class StateDirective implements OnChanges { @Input('state') itemState: any; private element: ElementRef; private renderer: Renderer; constructor(renderer: Renderer, el: ElementRef)
ngOnChanges() { let cssClass = `state-${this.itemState}`; let text = 'A livrer'; let elementNode = this.element.nativeElement; switch (this.itemState) { case 1 : text = "aaaaa"; break; case 2 : text = "xxxxx"; break; case 3 : text = "wwwww"; break; default: text = "zied !!!"; } this.renderer.setElementClass(elementNode, cssClass, true); this.renderer.setText(elementNode, text); } }
{ this.element = el; this.renderer = renderer; }
identifier_body
state.directive.ts
import {Directive, ElementRef, Renderer, Input, OnChanges} from '@angular/core'; @Directive({ selector: '[state]' }) export class
implements OnChanges { @Input('state') itemState: any; private element: ElementRef; private renderer: Renderer; constructor(renderer: Renderer, el: ElementRef) { this.element = el; this.renderer = renderer; } ngOnChanges() { let cssClass = `state-${this.itemState}`; let text = 'A livrer'; let elementNode = this.element.nativeElement; switch (this.itemState) { case 1 : text = "aaaaa"; break; case 2 : text = "xxxxx"; break; case 3 : text = "wwwww"; break; default: text = "zied !!!"; } this.renderer.setElementClass(elementNode, cssClass, true); this.renderer.setText(elementNode, text); } }
StateDirective
identifier_name
vsoFormatterTests.ts
/* * Copyright 2013 Palantir Technologies, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ import { assert } from "chai"; import * as ts from "typescript"; import { IFormatter, TestUtils } from "../lint"; import { createFailure } from "./utils"; describe("VSO Formatter", () => { const TEST_FILE = "formatters/vsoFormatter.test.ts"; let sourceFile: ts.SourceFile; let formatter: IFormatter; before(() => { const Formatter = TestUtils.getFormatter("vso"); sourceFile = TestUtils.getSourceFile(TEST_FILE); formatter = new Formatter(); }); it("formats failures", () => { const maxPosition = sourceFile.getFullWidth(); const failures = [ createFailure(sourceFile, 0, 1, "first failure", "first-name", undefined, "error"), createFailure(sourceFile, 32, 36, "mid failure", "mid-name", undefined, "error"), createFailure(sourceFile, maxPosition - 1, maxPosition, "last failure", "last-name", undefined, "error"), ]; const expectedResult = getFailureString(TEST_FILE, 1, 1, "first failure", "first-name") + getFailureString(TEST_FILE, 2, 12, "mid failure", "mid-name") + getFailureString(TEST_FILE, 9, 2, "last failure", "last-name"); const actualResult = formatter.format(failures); assert.equal(actualResult, expectedResult); }); it("does not duplicate output for fixed failures", () => { const maxPosition = sourceFile.getFullWidth(); const failures = [ createFailure(sourceFile, 0, 1, "first failure", "first-name", undefined, "error"), createFailure(sourceFile, 32, 36, "mid failure", "mid-name", undefined, "error"), createFailure(sourceFile, maxPosition - 1, maxPosition, "last failure", "last-name", undefined, "error"), ]; const expectedResult = getFailureString(TEST_FILE, 1, 1, "first failure", "first-name") + getFailureString(TEST_FILE, 2, 12, "mid failure", "mid-name") + getFailureString(TEST_FILE, 9, 2, "last failure", "last-name"); const fixed = failures.slice(); const actualResult = formatter.format(failures, fixed); assert.equal(actualResult, expectedResult); }); it("handles no failures", () => { const result = formatter.format([]); assert.equal(result, "\n"); }); function getFailureString(file: string, line: number, character: number, reason: string, code: string)
});
{ return `##vso[task.logissue type=warning;sourcepath=${file};linenumber=${line};columnnumber=${character};code=${code};]${reason}\n`; }
identifier_body
vsoFormatterTests.ts
/* * Copyright 2013 Palantir Technologies, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ import { assert } from "chai"; import * as ts from "typescript"; import { IFormatter, TestUtils } from "../lint"; import { createFailure } from "./utils"; describe("VSO Formatter", () => { const TEST_FILE = "formatters/vsoFormatter.test.ts"; let sourceFile: ts.SourceFile; let formatter: IFormatter; before(() => { const Formatter = TestUtils.getFormatter("vso"); sourceFile = TestUtils.getSourceFile(TEST_FILE); formatter = new Formatter(); }); it("formats failures", () => { const maxPosition = sourceFile.getFullWidth(); const failures = [ createFailure(sourceFile, 0, 1, "first failure", "first-name", undefined, "error"), createFailure(sourceFile, 32, 36, "mid failure", "mid-name", undefined, "error"), createFailure(sourceFile, maxPosition - 1, maxPosition, "last failure", "last-name", undefined, "error"), ]; const expectedResult = getFailureString(TEST_FILE, 1, 1, "first failure", "first-name") + getFailureString(TEST_FILE, 2, 12, "mid failure", "mid-name") + getFailureString(TEST_FILE, 9, 2, "last failure", "last-name"); const actualResult = formatter.format(failures); assert.equal(actualResult, expectedResult); }); it("does not duplicate output for fixed failures", () => { const maxPosition = sourceFile.getFullWidth(); const failures = [ createFailure(sourceFile, 0, 1, "first failure", "first-name", undefined, "error"), createFailure(sourceFile, 32, 36, "mid failure", "mid-name", undefined, "error"), createFailure(sourceFile, maxPosition - 1, maxPosition, "last failure", "last-name", undefined, "error"), ]; const expectedResult = getFailureString(TEST_FILE, 1, 1, "first failure", "first-name") + getFailureString(TEST_FILE, 2, 12, "mid failure", "mid-name") + getFailureString(TEST_FILE, 9, 2, "last failure", "last-name"); const fixed = failures.slice(); const actualResult = formatter.format(failures, fixed); assert.equal(actualResult, expectedResult); }); it("handles no failures", () => { const result = formatter.format([]); assert.equal(result, "\n"); }); function
(file: string, line: number, character: number, reason: string, code: string) { return `##vso[task.logissue type=warning;sourcepath=${file};linenumber=${line};columnnumber=${character};code=${code};]${reason}\n`; } });
getFailureString
identifier_name
vsoFormatterTests.ts
/* * Copyright 2013 Palantir Technologies, Inc. *
* Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ import { assert } from "chai"; import * as ts from "typescript"; import { IFormatter, TestUtils } from "../lint"; import { createFailure } from "./utils"; describe("VSO Formatter", () => { const TEST_FILE = "formatters/vsoFormatter.test.ts"; let sourceFile: ts.SourceFile; let formatter: IFormatter; before(() => { const Formatter = TestUtils.getFormatter("vso"); sourceFile = TestUtils.getSourceFile(TEST_FILE); formatter = new Formatter(); }); it("formats failures", () => { const maxPosition = sourceFile.getFullWidth(); const failures = [ createFailure(sourceFile, 0, 1, "first failure", "first-name", undefined, "error"), createFailure(sourceFile, 32, 36, "mid failure", "mid-name", undefined, "error"), createFailure(sourceFile, maxPosition - 1, maxPosition, "last failure", "last-name", undefined, "error"), ]; const expectedResult = getFailureString(TEST_FILE, 1, 1, "first failure", "first-name") + getFailureString(TEST_FILE, 2, 12, "mid failure", "mid-name") + getFailureString(TEST_FILE, 9, 2, "last failure", "last-name"); const actualResult = formatter.format(failures); assert.equal(actualResult, expectedResult); }); it("does not duplicate output for fixed failures", () => { const maxPosition = sourceFile.getFullWidth(); const failures = [ createFailure(sourceFile, 0, 1, "first failure", "first-name", undefined, "error"), createFailure(sourceFile, 32, 36, "mid failure", "mid-name", undefined, "error"), createFailure(sourceFile, maxPosition - 1, maxPosition, "last failure", "last-name", undefined, "error"), ]; const expectedResult = getFailureString(TEST_FILE, 1, 1, "first failure", "first-name") + getFailureString(TEST_FILE, 2, 12, "mid failure", "mid-name") + getFailureString(TEST_FILE, 9, 2, "last failure", "last-name"); const fixed = failures.slice(); const actualResult = formatter.format(failures, fixed); assert.equal(actualResult, expectedResult); }); it("handles no failures", () => { const result = formatter.format([]); assert.equal(result, "\n"); }); function getFailureString(file: string, line: number, character: number, reason: string, code: string) { return `##vso[task.logissue type=warning;sourcepath=${file};linenumber=${line};columnnumber=${character};code=${code};]${reason}\n`; } });
random_line_split
mainStaticPersonTask.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- #------------------------------------------------------------------------------- # Name: # Purpose: This .py file is the main Framework file # It ranks images of a specific person of interest in a static manner # # Required libs: python-dateutil, numpy,matplotlib,pyparsing # Author: konkonst # # Created: 30/03/2014 # Copyright: (c) ITI (CERTH) 2014 # Licence: <apache licence 2.0> #------------------------------------------------------------------------------- import time,os,pickle,glob,shutil, personPopularity from staticCommPersonTask import communitystatic print('staticCommPersonCentered') print(time.asctime( time.localtime(time.time()) )) '''PARAMETERS''' #Construct the data class from scratch: 1-yes / 2- from the community detection/ else-perform only the ranking dataextract = 1 #Provide a time Limit (unix timestamp) about when the dataset begins in case you only want part of the dataset. If it is set to 0 the whole dataset is considered. timeLimit = 0 #1071561600 #Community detection method. 'Ahn','Demon' and 'Copra' for overlapping and 'Louvain' for non. Ahn carries a threshold. commDetectMethod = ['Demon', 0.66] #User sets desired number of displayed top images topImages = 8 #User sets desired number of most frequent people to retrieve topPeople = 200 #Provide people set or leave empty to retrieve images for the number of topPeople as set above peopleSet = ['justin_timberlake','oprah_winfrey','lady_gaga','justin_bieber','michael_schumacher','miley_cyrus','jk_rowling','zinedine_zidane','barack_obama','prince_william','brad_pitt_actor','leonardo_dicaprio','natalie_portman'] peopleSet.sort() ##peopleSet = [] #Uncomment this to activate the use of the rankedPeople.txt pool of users #Delete all previous folders containing results? (Does not apply to the html files) delFolders = 0 #If there are any nodes that should not be considered, please place them in './data/txt/stopNodes.txt' '''Functions''' t = time.time() filename = [f for f in os.listdir("./data/txt/")] for idx,files in enumerate(filename): print(str(idx+1) + '.' + files)
selection = int(input('Select a dataset from the above: '))-1 dataset_path_results = "./data/"+filename[selection][:-4]+"/staticPersonCentered_"+commDetectMethod[0]+"/results/" dataset_path_tmp = "./data/"+filename[selection][:-4]+"/staticPersonCentered_"+commDetectMethod[0]+"/tmp/" datasetFilename = './data/txt/'+filename[selection] if not os.path.exists(dataset_path_results): os.makedirs(dataset_path_results) os.makedirs(dataset_path_tmp) if not os.path.exists(dataset_path_results+"rankedPeople.txt"): personPopularity.popPerson(datasetFilename, dataset_path_results, dataset_path_tmp, commDetectMethod,timeLimit=timeLimit) if dataextract==1:#Start from scratch data = communitystatic.from_txt(datasetFilename,dataset_path_results,dataset_path_tmp,timeLimit=timeLimit) dataPck = open(dataset_path_tmp + "allPersondata.pck", "wb") pickle.dump(data, dataPck , protocol = 2) dataPck.close() del(data) elapsed = time.time() - t print('Stage 1: %.2f seconds' % elapsed) if dataextract==1 or dataextract==2:#If the basic data (authors, mentions, time) has been created data = pickle.load(open(dataset_path_tmp + "allPersondata.pck", "rb")) captiondict = data.captiondict print('static Community detection method selected is :'+commDetectMethod[0]) dataStatic=data.extraction(commDetectMethod) del(data) elapsed = time.time() - t print('\nStage 2: %.2f seconds' % elapsed) decisionforAll = input('\nRetrieve the topImages by screening them one by one???(y or n) ') if dataextract ==1 or dataextract ==2 or dataextract ==3:#Only ranking beyond this point data = pickle.load(open(dataset_path_tmp + "allPersondata.pck", "rb")) captiondict = data.captiondict del(data) dataStatic = pickle.load(open(dataset_path_tmp + 'comm_'+commDetectMethod[0]+'.pck','rb')) #delete folders if you're starting from scratch if delFolders == 1: result_files = glob.glob(dataset_path_results+'/analysis/*.txt') if result_files: for file in result_files: os.remove(file) if not peopleSet: with open(dataset_path_results+'rankedPeople.txt','r') as f: for lineId,line in enumerate(f): if lineId>topPeople-1: break line = line.split('\t') peopleSet.append(line[0]) for person in peopleSet: if decisionforAll != str('n') and not os.path.exists(dataset_path_results+'html/'+person): os.makedirs(dataset_path_results+'html/'+person) if decisionforAll != str('n'): personDecision = input('\nRetrieve images for '+person+'???(y or n) ') if decisionforAll == str('n'): print("\nRetrieval Commences for "+person) if decisionforAll == str('n') or personDecision == str('y'): dataStatic.photoRetrieval(topImages, person, captiondict,decisionforAll) dataStatic.popularity_coappearence(topImages, person, captiondict) elapsed = time.time() - t print('\nStage 3: %.2f seconds' % elapsed)
random_line_split
mainStaticPersonTask.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- #------------------------------------------------------------------------------- # Name: # Purpose: This .py file is the main Framework file # It ranks images of a specific person of interest in a static manner # # Required libs: python-dateutil, numpy,matplotlib,pyparsing # Author: konkonst # # Created: 30/03/2014 # Copyright: (c) ITI (CERTH) 2014 # Licence: <apache licence 2.0> #------------------------------------------------------------------------------- import time,os,pickle,glob,shutil, personPopularity from staticCommPersonTask import communitystatic print('staticCommPersonCentered') print(time.asctime( time.localtime(time.time()) )) '''PARAMETERS''' #Construct the data class from scratch: 1-yes / 2- from the community detection/ else-perform only the ranking dataextract = 1 #Provide a time Limit (unix timestamp) about when the dataset begins in case you only want part of the dataset. If it is set to 0 the whole dataset is considered. timeLimit = 0 #1071561600 #Community detection method. 'Ahn','Demon' and 'Copra' for overlapping and 'Louvain' for non. Ahn carries a threshold. commDetectMethod = ['Demon', 0.66] #User sets desired number of displayed top images topImages = 8 #User sets desired number of most frequent people to retrieve topPeople = 200 #Provide people set or leave empty to retrieve images for the number of topPeople as set above peopleSet = ['justin_timberlake','oprah_winfrey','lady_gaga','justin_bieber','michael_schumacher','miley_cyrus','jk_rowling','zinedine_zidane','barack_obama','prince_william','brad_pitt_actor','leonardo_dicaprio','natalie_portman'] peopleSet.sort() ##peopleSet = [] #Uncomment this to activate the use of the rankedPeople.txt pool of users #Delete all previous folders containing results? (Does not apply to the html files) delFolders = 0 #If there are any nodes that should not be considered, please place them in './data/txt/stopNodes.txt' '''Functions''' t = time.time() filename = [f for f in os.listdir("./data/txt/")] for idx,files in enumerate(filename): print(str(idx+1) + '.' + files) selection = int(input('Select a dataset from the above: '))-1 dataset_path_results = "./data/"+filename[selection][:-4]+"/staticPersonCentered_"+commDetectMethod[0]+"/results/" dataset_path_tmp = "./data/"+filename[selection][:-4]+"/staticPersonCentered_"+commDetectMethod[0]+"/tmp/" datasetFilename = './data/txt/'+filename[selection] if not os.path.exists(dataset_path_results): os.makedirs(dataset_path_results) os.makedirs(dataset_path_tmp) if not os.path.exists(dataset_path_results+"rankedPeople.txt"): personPopularity.popPerson(datasetFilename, dataset_path_results, dataset_path_tmp, commDetectMethod,timeLimit=timeLimit) if dataextract==1:#Start from scratch
if dataextract==1 or dataextract==2:#If the basic data (authors, mentions, time) has been created data = pickle.load(open(dataset_path_tmp + "allPersondata.pck", "rb")) captiondict = data.captiondict print('static Community detection method selected is :'+commDetectMethod[0]) dataStatic=data.extraction(commDetectMethod) del(data) elapsed = time.time() - t print('\nStage 2: %.2f seconds' % elapsed) decisionforAll = input('\nRetrieve the topImages by screening them one by one???(y or n) ') if dataextract ==1 or dataextract ==2 or dataextract ==3:#Only ranking beyond this point data = pickle.load(open(dataset_path_tmp + "allPersondata.pck", "rb")) captiondict = data.captiondict del(data) dataStatic = pickle.load(open(dataset_path_tmp + 'comm_'+commDetectMethod[0]+'.pck','rb')) #delete folders if you're starting from scratch if delFolders == 1: result_files = glob.glob(dataset_path_results+'/analysis/*.txt') if result_files: for file in result_files: os.remove(file) if not peopleSet: with open(dataset_path_results+'rankedPeople.txt','r') as f: for lineId,line in enumerate(f): if lineId>topPeople-1: break line = line.split('\t') peopleSet.append(line[0]) for person in peopleSet: if decisionforAll != str('n') and not os.path.exists(dataset_path_results+'html/'+person): os.makedirs(dataset_path_results+'html/'+person) if decisionforAll != str('n'): personDecision = input('\nRetrieve images for '+person+'???(y or n) ') if decisionforAll == str('n'): print("\nRetrieval Commences for "+person) if decisionforAll == str('n') or personDecision == str('y'): dataStatic.photoRetrieval(topImages, person, captiondict,decisionforAll) dataStatic.popularity_coappearence(topImages, person, captiondict) elapsed = time.time() - t print('\nStage 3: %.2f seconds' % elapsed)
data = communitystatic.from_txt(datasetFilename,dataset_path_results,dataset_path_tmp,timeLimit=timeLimit) dataPck = open(dataset_path_tmp + "allPersondata.pck", "wb") pickle.dump(data, dataPck , protocol = 2) dataPck.close() del(data) elapsed = time.time() - t print('Stage 1: %.2f seconds' % elapsed)
conditional_block
test_instructor_dashboard.py
""" End to end tests for Instructor Dashboard. """ from bok_choy.web_app_test import WebAppTest from regression.pages.lms.course_page_lms import CourseHomePageExtended from regression.pages.lms.dashboard_lms import DashboardPageExtended from regression.pages.lms.instructor_dashboard import InstructorDashboardPageExtended from regression.pages.lms.utils import get_course_key from regression.tests.helpers.api_clients import LmsLoginApi from regression.tests.helpers.utils import get_course_display_name, get_course_info
""" Regression tests on Analytics on Instructor Dashboard """ def setUp(self): super().setUp() login_api = LmsLoginApi() login_api.authenticate(self.browser) course_info = get_course_info() self.dashboard_page = DashboardPageExtended(self.browser) self.instructor_dashboard = InstructorDashboardPageExtended( self.browser, get_course_key(course_info) ) self.course_page = CourseHomePageExtended( self.browser, get_course_key(course_info) ) self.dashboard_page.visit() self.dashboard_page.select_course(get_course_display_name()) self.course_page.wait_for_page() self.instructor_dashboard.visit()
class AnalyticsTest(WebAppTest):
random_line_split
test_instructor_dashboard.py
""" End to end tests for Instructor Dashboard. """ from bok_choy.web_app_test import WebAppTest from regression.pages.lms.course_page_lms import CourseHomePageExtended from regression.pages.lms.dashboard_lms import DashboardPageExtended from regression.pages.lms.instructor_dashboard import InstructorDashboardPageExtended from regression.pages.lms.utils import get_course_key from regression.tests.helpers.api_clients import LmsLoginApi from regression.tests.helpers.utils import get_course_display_name, get_course_info class AnalyticsTest(WebAppTest):
self.dashboard_page.visit() self.dashboard_page.select_course(get_course_display_name()) self.course_page.wait_for_page() self.instructor_dashboard.visit()
""" Regression tests on Analytics on Instructor Dashboard """ def setUp(self): super().setUp() login_api = LmsLoginApi() login_api.authenticate(self.browser) course_info = get_course_info() self.dashboard_page = DashboardPageExtended(self.browser) self.instructor_dashboard = InstructorDashboardPageExtended( self.browser, get_course_key(course_info) ) self.course_page = CourseHomePageExtended( self.browser, get_course_key(course_info) )
identifier_body
test_instructor_dashboard.py
""" End to end tests for Instructor Dashboard. """ from bok_choy.web_app_test import WebAppTest from regression.pages.lms.course_page_lms import CourseHomePageExtended from regression.pages.lms.dashboard_lms import DashboardPageExtended from regression.pages.lms.instructor_dashboard import InstructorDashboardPageExtended from regression.pages.lms.utils import get_course_key from regression.tests.helpers.api_clients import LmsLoginApi from regression.tests.helpers.utils import get_course_display_name, get_course_info class AnalyticsTest(WebAppTest): """ Regression tests on Analytics on Instructor Dashboard """ def
(self): super().setUp() login_api = LmsLoginApi() login_api.authenticate(self.browser) course_info = get_course_info() self.dashboard_page = DashboardPageExtended(self.browser) self.instructor_dashboard = InstructorDashboardPageExtended( self.browser, get_course_key(course_info) ) self.course_page = CourseHomePageExtended( self.browser, get_course_key(course_info) ) self.dashboard_page.visit() self.dashboard_page.select_course(get_course_display_name()) self.course_page.wait_for_page() self.instructor_dashboard.visit()
setUp
identifier_name
reportcommon.py
def finish_run(): remove_run_file() display_detail("## Finished run") exit(0) def remove_run_file(): touchfile = '/Users/Shared/.com.github.munkireport.run' if os.path.exists(touchfile): os.remove(touchfile) def curl(url, values): options = dict() options["url"] = url options["method"] = "POST" options["content_type"] = "application/x-www-form-urlencoded" options["body"] = urlencode(values) options["logging_function"] = display_detail options["connection_timeout"] = 60 if pref("UseMunkiAdditionalHttpHeaders"): custom_headers = prefs.pref(constants.ADDITIONAL_HTTP_HEADERS_KEY) if custom_headers: options["additional_headers"] = dict() for header in custom_headers: m = re.search(r"^(?P<header_name>.*?): (?P<header_value>.*?)$", header) if m: options["additional_headers"][m.group("header_name")] = m.group( "header_value" ) else: raise CurlError( -1, "UseMunkiAdditionalHttpHeaders defined, " "but not found in Munki preferences", ) # Build Purl with initial settings connection = Purl.alloc().initWithOptions_(options) connection.start() try: while True: # if we did `while not connection.isDone()` we'd miss printing # messages if we exit the loop first if connection.isDone(): break except (KeyboardInterrupt, SystemExit): # safely kill the connection then re-raise connection.cancel() raise except Exception, err: # too general, I know # Let us out! ... Safely! Unexpectedly quit dialogs are annoying... connection.cancel() # Re-raise the error as a GurlError raise CurlError(-1, str(err)) if connection.error != None: # Gurl returned an error display.display_detail( "Download error %s: %s", connection.error.code(), connection.error.localizedDescription(), ) if connection.SSLerror: display_detail("SSL error detail: %s", str(connection.SSLerror)) display_detail("Headers: %s", connection.headers) raise CurlError( connection.error.code(), connection.error.localizedDescription() ) if connection.response != None and connection.status != 200: display.display_detail("Status: %s", connection.status) display.display_detail("Headers: %s", connection.headers) if connection.redirection != []: display.display_detail("Redirection: %s", connection.redirection) connection.headers["http_result_code"] = str(connection.status) description = NSHTTPURLResponse.localizedStringForStatusCode_(connection.status) connection.headers["http_result_description"] = description if str(connection.status).startswith("2"): return connection.get_response_data() else: # there was an HTTP error of some sort. raise CurlError( connection.status, "%s failed, HTTP returncode %s (%s)" % ( url, connection.status, connection.headers.get("http_result_description", "Failed"), ), ) def get_hardware_info(): """Uses system profiler to get hardware info for this machine.""" cmd = ["/usr/sbin/system_profiler", "SPHardwareDataType", "-xml"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, dummy_error) = proc.communicate() try: plist = FoundationPlist.readPlistFromString(output) # system_profiler xml is an array sp_dict = plist[0] items = sp_dict["_items"] sp_hardware_dict = items[0] return sp_hardware_dict except BaseException: return {} def get_long_username(username): try: long_name = pwd.getpwnam(username)[4] except: long_name = "" return long_name.decode("utf-8") def get_uid(username): try: uid = pwd.getpwnam(username)[2] except: uid = "" return uid def get_computername(): cmd = ["/usr/sbin/scutil", "--get", "ComputerName"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() output = output.strip() return output.decode("utf-8") def get_cpuinfo(): cmd = ["/usr/sbin/sysctl", "-n", "machdep.cpu.brand_string"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() output = output.strip() return output.decode("utf-8") def get_buildversion(): cmd = ["/usr/bin/sw_vers", "-buildVersion"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() output = output.strip() return output.decode("utf-8") def get_uptime(): cmd = ["/usr/sbin/sysctl", "-n", "kern.boottime"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() sec = int(re.sub(".*sec = (\d+),.*", "\\1", output)) up = int(time.time() - sec) return up if up > 0 else -1 def set_pref(pref_name, pref_value): """Sets a preference, See prefs.py for details.""" CFPreferencesSetValue( pref_name, pref_value, BUNDLE_ID, kCFPreferencesAnyUser, kCFPreferencesCurrentHost, ) CFPreferencesAppSynchronize(BUNDLE_ID) print "set pref" try: CFPreferencesSetValue( pref_name, pref_value, BUNDLE_ID, kCFPreferencesAnyUser, kCFPreferencesCurrentHost, ) CFPreferencesAppSynchronize(BUNDLE_ID) except Exception: pass def pref(pref_name): """Return a preference. See prefs.py for details """ pref_value = CFPreferencesCopyAppValue(pref_name, BUNDLE_ID) return pref_value def process(serial, items): """Process receives a list of items, checks if they need updating and updates them if necessary.""" # Sanitize serial serial = "".join([c for c in serial if c.isalnum()]) # Get prefs baseurl = pref("BaseUrl") or prefs.pref("SoftwareRepoURL") + "/report/" hashurl = baseurl + "index.php?/report/hash_check" checkurl = baseurl + "index.php?/report/check_in" # Get passphrase passphrase = pref("Passphrase") # Get hashes for all scripts for key, i in items.items(): if i.get("path"): i["hash"] = getmd5hash(i.get("path")) # Check dict check = {} for key, i in items.items(): if i.get("hash"): check[key] = {"hash": i.get("hash")} # Send hashes to server values = {"serial": serial, "items": serialize(check), "passphrase": passphrase} server_data = curl(hashurl, values) # = response.read() # Decode response try: result = unserialize(server_data) except Exception, e: display_error("Could not unserialize server data: %s" % str(e)) display_error("Request: %s" % str(values)) display_error("Response: %s" % str(server_data)) return -1 if result.get("error") != "": display_error("Server error: %s" % result["error"]) return -1 if result.get("info") != "": display_detail("Server info: %s" % result["info"]) # Retrieve hashes that need updating total_size = 0 for i in items.keys(): if i in result: if items[i].get("path"): try: f = open(items[i]["path"], "r") items[i]["data"] = f.read() except: display_warning("Can't open %s" % items[i]["path"]) del items[i] continue size = len(items[i]["data"]) display_detail("Need to update %s (%s)" % (i, sizeof_fmt(size))) total_size = total_size + size else: # delete items that don't have to be uploaded del items[i] # Send new files with hashes if len(items): display_detail("Sending items (%s)" % sizeof_fmt(total_size)) response = curl( checkurl, {"serial": serial, "items": serialize(items), "passphrase": passphrase}, ) display_detail(response) else:
"""Call display detail msg handler.""" display.display_detail("%s" % msg, *args)
identifier_body
reportcommon.py
options["additional_headers"] = dict() for header in custom_headers: m = re.search(r"^(?P<header_name>.*?): (?P<header_value>.*?)$", header) if m: options["additional_headers"][m.group("header_name")] = m.group( "header_value" ) else: raise CurlError( -1, "UseMunkiAdditionalHttpHeaders defined, " "but not found in Munki preferences", ) # Build Purl with initial settings connection = Purl.alloc().initWithOptions_(options) connection.start() try: while True: # if we did `while not connection.isDone()` we'd miss printing # messages if we exit the loop first if connection.isDone(): break except (KeyboardInterrupt, SystemExit): # safely kill the connection then re-raise connection.cancel() raise except Exception, err: # too general, I know # Let us out! ... Safely! Unexpectedly quit dialogs are annoying... connection.cancel() # Re-raise the error as a GurlError raise CurlError(-1, str(err)) if connection.error != None: # Gurl returned an error display.display_detail( "Download error %s: %s", connection.error.code(), connection.error.localizedDescription(), ) if connection.SSLerror: display_detail("SSL error detail: %s", str(connection.SSLerror)) display_detail("Headers: %s", connection.headers) raise CurlError( connection.error.code(), connection.error.localizedDescription() ) if connection.response != None and connection.status != 200: display.display_detail("Status: %s", connection.status) display.display_detail("Headers: %s", connection.headers) if connection.redirection != []: display.display_detail("Redirection: %s", connection.redirection) connection.headers["http_result_code"] = str(connection.status) description = NSHTTPURLResponse.localizedStringForStatusCode_(connection.status) connection.headers["http_result_description"] = description if str(connection.status).startswith("2"): return connection.get_response_data() else: # there was an HTTP error of some sort. raise CurlError( connection.status, "%s failed, HTTP returncode %s (%s)" % ( url, connection.status, connection.headers.get("http_result_description", "Failed"), ), ) def get_hardware_info(): """Uses system profiler to get hardware info for this machine.""" cmd = ["/usr/sbin/system_profiler", "SPHardwareDataType", "-xml"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, dummy_error) = proc.communicate() try: plist = FoundationPlist.readPlistFromString(output) # system_profiler xml is an array sp_dict = plist[0] items = sp_dict["_items"] sp_hardware_dict = items[0] return sp_hardware_dict except BaseException: return {} def get_long_username(username): try: long_name = pwd.getpwnam(username)[4] except: long_name = "" return long_name.decode("utf-8") def get_uid(username): try: uid = pwd.getpwnam(username)[2] except: uid = "" return uid def get_computername(): cmd = ["/usr/sbin/scutil", "--get", "ComputerName"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() output = output.strip() return output.decode("utf-8") def get_cpuinfo(): cmd = ["/usr/sbin/sysctl", "-n", "machdep.cpu.brand_string"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() output = output.strip() return output.decode("utf-8") def get_buildversion(): cmd = ["/usr/bin/sw_vers", "-buildVersion"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() output = output.strip() return output.decode("utf-8") def get_uptime(): cmd = ["/usr/sbin/sysctl", "-n", "kern.boottime"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() sec = int(re.sub(".*sec = (\d+),.*", "\\1", output)) up = int(time.time() - sec) return up if up > 0 else -1 def set_pref(pref_name, pref_value): """Sets a preference, See prefs.py for details.""" CFPreferencesSetValue( pref_name, pref_value, BUNDLE_ID, kCFPreferencesAnyUser, kCFPreferencesCurrentHost, ) CFPreferencesAppSynchronize(BUNDLE_ID) print "set pref" try: CFPreferencesSetValue( pref_name, pref_value, BUNDLE_ID, kCFPreferencesAnyUser, kCFPreferencesCurrentHost, ) CFPreferencesAppSynchronize(BUNDLE_ID) except Exception: pass def pref(pref_name): """Return a preference. See prefs.py for details """ pref_value = CFPreferencesCopyAppValue(pref_name, BUNDLE_ID) return pref_value def process(serial, items): """Process receives a list of items, checks if they need updating and updates them if necessary.""" # Sanitize serial serial = "".join([c for c in serial if c.isalnum()]) # Get prefs baseurl = pref("BaseUrl") or prefs.pref("SoftwareRepoURL") + "/report/" hashurl = baseurl + "index.php?/report/hash_check" checkurl = baseurl + "index.php?/report/check_in" # Get passphrase passphrase = pref("Passphrase") # Get hashes for all scripts for key, i in items.items(): if i.get("path"): i["hash"] = getmd5hash(i.get("path")) # Check dict check = {} for key, i in items.items(): if i.get("hash"): check[key] = {"hash": i.get("hash")} # Send hashes to server values = {"serial": serial, "items": serialize(check), "passphrase": passphrase} server_data = curl(hashurl, values) # = response.read() # Decode response try: result = unserialize(server_data) except Exception, e: display_error("Could not unserialize server data: %s" % str(e)) display_error("Request: %s" % str(values)) display_error("Response: %s" % str(server_data)) return -1 if result.get("error") != "": display_error("Server error: %s" % result["error"]) return -1 if result.get("info") != "": display_detail("Server info: %s" % result["info"]) # Retrieve hashes that need updating total_size = 0 for i in items.keys(): if i in result: if items[i].get("path"): try: f = open(items[i]["path"], "r") items[i]["data"] = f.read() except: display_warning("Can't open %s" % items[i]["path"]) del items[i] continue size = len(items[i]["data"]) display_detail("Need to update %s (%s)" % (i, sizeof_fmt(size))) total_size = total_size + size else: # delete items that don't have to be uploaded del items[i] # Send new files with hashes if len(items): display_detail("Sending items (%s)" % sizeof_fmt(total_size)) response = curl( checkurl, {"serial": serial, "items": serialize(items), "passphrase": passphrase}, ) display_detail(response) else: display_detail("No changes") def runExternalScriptWithTimeout( script, allow_insecure=False, script_args=(), timeout=30 ): """Run a script (e.g. preflight/postflight) and return its exit status. Args: script: string path to the script to execute. allow_insecure: bool skip the permissions check of executable. args: args to pass to the script. Returns: Tuple. (integer exit status from script, str stdout, str stderr). Raises: ScriptNotFoundError: the script was not found at the given path.
if not os.path.exists(script): raise ScriptNotFoundError("script does not exist: %s" % script) if not allow_insecure: try:
RunExternalScriptError: there was an error running the script. """ from munkilib import utils
random_line_split
reportcommon.py
options["additional_headers"] = dict() for header in custom_headers: m = re.search(r"^(?P<header_name>.*?): (?P<header_value>.*?)$", header) if m: options["additional_headers"][m.group("header_name")] = m.group( "header_value" ) else: raise CurlError( -1, "UseMunkiAdditionalHttpHeaders defined, " "but not found in Munki preferences", ) # Build Purl with initial settings connection = Purl.alloc().initWithOptions_(options) connection.start() try: while True: # if we did `while not connection.isDone()` we'd miss printing # messages if we exit the loop first if connection.isDone(): break except (KeyboardInterrupt, SystemExit): # safely kill the connection then re-raise connection.cancel() raise except Exception, err: # too general, I know # Let us out! ... Safely! Unexpectedly quit dialogs are annoying... connection.cancel() # Re-raise the error as a GurlError raise CurlError(-1, str(err)) if connection.error != None: # Gurl returned an error display.display_detail( "Download error %s: %s", connection.error.code(), connection.error.localizedDescription(), ) if connection.SSLerror: display_detail("SSL error detail: %s", str(connection.SSLerror)) display_detail("Headers: %s", connection.headers) raise CurlError( connection.error.code(), connection.error.localizedDescription() ) if connection.response != None and connection.status != 200: display.display_detail("Status: %s", connection.status) display.display_detail("Headers: %s", connection.headers) if connection.redirection != []: display.display_detail("Redirection: %s", connection.redirection) connection.headers["http_result_code"] = str(connection.status) description = NSHTTPURLResponse.localizedStringForStatusCode_(connection.status) connection.headers["http_result_description"] = description if str(connection.status).startswith("2"): return connection.get_response_data() else: # there was an HTTP error of some sort. raise CurlError( connection.status, "%s failed, HTTP returncode %s (%s)" % ( url, connection.status, connection.headers.get("http_result_description", "Failed"), ), ) def get_hardware_info(): """Uses system profiler to get hardware info for this machine.""" cmd = ["/usr/sbin/system_profiler", "SPHardwareDataType", "-xml"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, dummy_error) = proc.communicate() try: plist = FoundationPlist.readPlistFromString(output) # system_profiler xml is an array sp_dict = plist[0] items = sp_dict["_items"] sp_hardware_dict = items[0] return sp_hardware_dict except BaseException: return {} def get_long_username(username): try: long_name = pwd.getpwnam(username)[4] except: long_name = "" return long_name.decode("utf-8") def get_uid(username): try: uid = pwd.getpwnam(username)[2] except: uid = "" return uid def get_computername(): cmd = ["/usr/sbin/scutil", "--get", "ComputerName"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() output = output.strip() return output.decode("utf-8") def get_cpuinfo(): cmd = ["/usr/sbin/sysctl", "-n", "machdep.cpu.brand_string"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() output = output.strip() return output.decode("utf-8") def get_buildversion(): cmd = ["/usr/bin/sw_vers", "-buildVersion"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() output = output.strip() return output.decode("utf-8") def get_uptime(): cmd = ["/usr/sbin/sysctl", "-n", "kern.boottime"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() sec = int(re.sub(".*sec = (\d+),.*", "\\1", output)) up = int(time.time() - sec) return up if up > 0 else -1 def set_pref(pref_name, pref_value): """Sets a preference, See prefs.py for details.""" CFPreferencesSetValue( pref_name, pref_value, BUNDLE_ID, kCFPreferencesAnyUser, kCFPreferencesCurrentHost, ) CFPreferencesAppSynchronize(BUNDLE_ID) print "set pref" try: CFPreferencesSetValue( pref_name, pref_value, BUNDLE_ID, kCFPreferencesAnyUser, kCFPreferencesCurrentHost, ) CFPreferencesAppSynchronize(BUNDLE_ID) except Exception: pass def pref(pref_name): """Return a preference. See prefs.py for details """ pref_value = CFPreferencesCopyAppValue(pref_name, BUNDLE_ID) return pref_value def process(serial, items): """Process receives a list of items, checks if they need updating and updates them if necessary.""" # Sanitize serial serial = "".join([c for c in serial if c.isalnum()]) # Get prefs baseurl = pref("BaseUrl") or prefs.pref("SoftwareRepoURL") + "/report/" hashurl = baseurl + "index.php?/report/hash_check" checkurl = baseurl + "index.php?/report/check_in" # Get passphrase passphrase = pref("Passphrase") # Get hashes for all scripts for key, i in items.items(): if i.get("path"): i["hash"] = getmd5hash(i.get("path")) # Check dict check = {} for key, i in items.items(): if i.get("hash"): check[key] = {"hash": i.get("hash")} # Send hashes to server values = {"serial": serial, "items": serialize(check), "passphrase": passphrase} server_data = curl(hashurl, values) # = response.read() # Decode response try: result = unserialize(server_data) except Exception, e: display_error("Could not unserialize server data: %s" % str(e)) display_error("Request: %s" % str(values)) display_error("Response: %s" % str(server_data)) return -1 if result.get("error") != "": display_error("Server error: %s" % result["error"]) return -1 if result.get("info") != "": display_detail("Server info: %s" % result["info"]) # Retrieve hashes that need updating total_size = 0 for i in items.keys(): if i in result: if items[i].get("path"): try: f = open(items[i]["path"], "r") items[i]["data"] = f.read() except: display_warning("Can't open %s" % items[i]["path"]) del items[i] continue size = len(items[i]["data"]) display_detail("Need to update %s (%s)" % (i, sizeof_fmt(size))) total_size = total_size + size else: # delete items that don't have to be uploaded del items[i] # Send new files with hashes if len(items): display_detail("Sending items (%s)" % sizeof_fmt(total_size)) response = curl( checkurl, {"serial": serial, "items": serialize(items), "passphrase": passphrase}, ) display_detail(response) else: display_detail("No changes") def
( script, allow_insecure=False, script_args=(), timeout=30 ): """Run a script (e.g. preflight/postflight) and return its exit status. Args: script: string path to the script to execute. allow_insecure: bool skip the permissions check of executable. args: args to pass to the script. Returns: Tuple. (integer exit status from script, str stdout, str stderr). Raises: ScriptNotFoundError: the script was not found at the given path. RunExternalScriptError: there was an error running the script. """ from munkilib import utils if not os.path.exists(script): raise ScriptNotFoundError("script does not exist: %s" % script) if not allow_insecure: try:
runExternalScriptWithTimeout
identifier_name
reportcommon.py
) = proc.communicate() output = output.strip() return output.decode("utf-8") def get_uptime(): cmd = ["/usr/sbin/sysctl", "-n", "kern.boottime"] proc = subprocess.Popen( cmd, shell=False, bufsize=-1, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) (output, unused_error) = proc.communicate() sec = int(re.sub(".*sec = (\d+),.*", "\\1", output)) up = int(time.time() - sec) return up if up > 0 else -1 def set_pref(pref_name, pref_value): """Sets a preference, See prefs.py for details.""" CFPreferencesSetValue( pref_name, pref_value, BUNDLE_ID, kCFPreferencesAnyUser, kCFPreferencesCurrentHost, ) CFPreferencesAppSynchronize(BUNDLE_ID) print "set pref" try: CFPreferencesSetValue( pref_name, pref_value, BUNDLE_ID, kCFPreferencesAnyUser, kCFPreferencesCurrentHost, ) CFPreferencesAppSynchronize(BUNDLE_ID) except Exception: pass def pref(pref_name): """Return a preference. See prefs.py for details """ pref_value = CFPreferencesCopyAppValue(pref_name, BUNDLE_ID) return pref_value def process(serial, items): """Process receives a list of items, checks if they need updating and updates them if necessary.""" # Sanitize serial serial = "".join([c for c in serial if c.isalnum()]) # Get prefs baseurl = pref("BaseUrl") or prefs.pref("SoftwareRepoURL") + "/report/" hashurl = baseurl + "index.php?/report/hash_check" checkurl = baseurl + "index.php?/report/check_in" # Get passphrase passphrase = pref("Passphrase") # Get hashes for all scripts for key, i in items.items(): if i.get("path"): i["hash"] = getmd5hash(i.get("path")) # Check dict check = {} for key, i in items.items(): if i.get("hash"): check[key] = {"hash": i.get("hash")} # Send hashes to server values = {"serial": serial, "items": serialize(check), "passphrase": passphrase} server_data = curl(hashurl, values) # = response.read() # Decode response try: result = unserialize(server_data) except Exception, e: display_error("Could not unserialize server data: %s" % str(e)) display_error("Request: %s" % str(values)) display_error("Response: %s" % str(server_data)) return -1 if result.get("error") != "": display_error("Server error: %s" % result["error"]) return -1 if result.get("info") != "": display_detail("Server info: %s" % result["info"]) # Retrieve hashes that need updating total_size = 0 for i in items.keys(): if i in result: if items[i].get("path"): try: f = open(items[i]["path"], "r") items[i]["data"] = f.read() except: display_warning("Can't open %s" % items[i]["path"]) del items[i] continue size = len(items[i]["data"]) display_detail("Need to update %s (%s)" % (i, sizeof_fmt(size))) total_size = total_size + size else: # delete items that don't have to be uploaded del items[i] # Send new files with hashes if len(items): display_detail("Sending items (%s)" % sizeof_fmt(total_size)) response = curl( checkurl, {"serial": serial, "items": serialize(items), "passphrase": passphrase}, ) display_detail(response) else: display_detail("No changes") def runExternalScriptWithTimeout( script, allow_insecure=False, script_args=(), timeout=30 ): """Run a script (e.g. preflight/postflight) and return its exit status. Args: script: string path to the script to execute. allow_insecure: bool skip the permissions check of executable. args: args to pass to the script. Returns: Tuple. (integer exit status from script, str stdout, str stderr). Raises: ScriptNotFoundError: the script was not found at the given path. RunExternalScriptError: there was an error running the script. """ from munkilib import utils if not os.path.exists(script): raise ScriptNotFoundError("script does not exist: %s" % script) if not allow_insecure: try: utils.verifyFileOnlyWritableByMunkiAndRoot(script) except utils.VerifyFilePermissionsError, e: msg = ( "Skipping execution due to failed file permissions " "verification: %s\n%s" % (script, str(e)) ) raise utils.RunExternalScriptError(msg) if os.access(script, os.X_OK): cmd = [script] if script_args: cmd.extend(script_args) proc = subprocess.Popen( cmd, shell=False, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) while timeout > 0: if proc.poll() is not None: (stdout, stderr) = proc.communicate() return ( proc.returncode, stdout.decode("UTF-8", "replace"), stderr.decode("UTF-8", "replace"), ) time.sleep(0.1) timeout -= 0.1 else: try: proc.kill() except OSError, e: if e.errno != 3: raise raise utils.RunExternalScriptError("%s timed out" % script) return (0, None, None) else: raise utils.RunExternalScriptError("%s not executable" % script) def rundir(scriptdir, runtype, abort=False, submitscript=""): """Run scripts in directory scriptdir runtype is passed to the script if abort is True, a non-zero exit status will abort munki submitscript is put at the end of the scriptlist.""" if os.path.exists(scriptdir): from munkilib import utils # Get timeout for scripts scriptTimeOut = 30 if pref("scriptTimeOut"): scriptTimeOut = int(pref("scriptTimeOut")) display_detail("# Set custom script timeout to %s seconds" % scriptTimeOut) # Directory containing the scripts parentdir = os.path.basename(scriptdir) display_detail("# Executing scripts in %s" % parentdir) # Get all files in scriptdir files = os.listdir(scriptdir) # Sort files files.sort() # Find submit script and stick it on the end of the list if submitscript: try: sub = files.pop(files.index(submitscript)) files.append(sub) except Exception, e: display_error("%s not found in %s" % (submitscript, parentdir)) for script in files: # Skip files that start with a period if script.startswith("."): continue # Concatenate dir and filename scriptpath = os.path.join(scriptdir, script) # Skip directories if os.path.isdir(scriptpath): continue try: # Attempt to execute script display_detail("Running %s" % script) result, stdout, stderr = runExternalScriptWithTimeout( scriptpath, allow_insecure=False, script_args=[runtype], timeout=scriptTimeOut, ) if stdout: display_detail(stdout) if stderr: display_detail("%s Error: %s" % (script, stderr)) if result: if abort: display_detail("Aborted by %s" % script) exit(1) else: display_warning("%s return code: %d" % (script, result)) except utils.ScriptNotFoundError: pass # Script has disappeared - pass. except Exception, e: display_warning("%s: %s" % (script, str(e))) def sizeof_fmt(num): for unit in ["B", "KB", "MB", "GB", "TB", "PB", "EB", "ZB"]: if abs(num) < 1000.0: return "%.0f%s" % (num, unit) num /= 1000.0 return "%.1f%s" % (num, "YB") def gethash(filename, hash_function): """Calculates the hashvalue of the given file with the given hash_function. Args: filename: The file name to calculate the hash value of. hash_function: The hash function object to use, which was instantiated before calling this function, e.g. hashlib.md5(). Returns: The hashvalue of the given file as hex string. """ if not os.path.isfile(filename): return "NOT A FILE" fileref = open(filename, "rb") while 1:
chunk = fileref.read(2 ** 16) if not chunk: break hash_function.update(chunk)
conditional_block
regroup_gradient_stops.rs
// svgcleaner could help you to clean up your SVG files // from unnecessary data. // Copyright (C) 2012-2018 Evgeniy Reizner // // This program is free software; you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation; either version 2 of the License, or // (at your option) any later version. // // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // // You should have received a copy of the GNU General Public License along // with this program; if not, write to the Free Software Foundation, Inc., // 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. use svgdom::{ Document, ElementType, Node, }; use task::short::{EId, AId}; pub fn regroup_gradient_stops(doc: &mut Document)
if super::rm_dupl_defs::is_equal_stops(&node1, &node2) { join_nodes.push(node2.clone()); nodes.remove(i2 - 1); i2 -= 1; } } if !join_nodes.is_empty() { is_changed = true; let mut new_lg = doc.create_element(EId::LinearGradient); let new_id = gen_id(doc, "lg"); new_lg.set_id(new_id); while node1.has_children() { let mut c = node1.first_child().unwrap(); c.detach(); new_lg.append(&c); } node1.set_attribute((AId::XlinkHref, new_lg.clone())); node1.insert_before(&new_lg); for jn in &mut join_nodes { while jn.has_children() { let mut c = jn.first_child().unwrap(); c.remove(); } jn.set_attribute((AId::XlinkHref, new_lg.clone())); } join_nodes.clear(); } i1 += 1; } if is_changed { // We must resolve attributes for gradients created above. super::resolve_linear_gradient_attributes(doc); } } fn gen_id(doc: &Document, prefix: &str) -> String { let mut n = 1; let mut s = String::new(); loop { s.clear(); s.push_str(prefix); s.push_str(&n.to_string()); // TODO: very slow if !doc.descendants().any(|n| *n.id() == s) { break; } n += 1; } s } #[cfg(test)] mod tests { use super::*; use svgdom::{Document, ToStringWithOptions}; use task; macro_rules! test { ($name:ident, $in_text:expr, $out_text:expr) => ( #[test] fn $name() { let mut doc = Document::from_str($in_text).unwrap(); task::resolve_linear_gradient_attributes(&doc); task::resolve_radial_gradient_attributes(&doc); task::resolve_stop_attributes(&doc).unwrap(); regroup_gradient_stops(&mut doc); assert_eq_text!(doc.to_string_with_opt(&write_opt_for_tests!()), $out_text); } ) } test!(rm_1, "<svg> <linearGradient id='lg1' x1='50'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg2' x1='100'> <stop offset='0'/> <stop offset='1'/> </linearGradient> </svg>", "<svg> <linearGradient id='lg3'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg1' x1='50' xlink:href='#lg3'/> <linearGradient id='lg2' x1='100' xlink:href='#lg3'/> </svg> "); test!(rm_2, "<svg> <linearGradient id='lg1' x1='50'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg3' x1='50'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg2' x1='100'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg4' x1='100'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> </svg>", "<svg> <linearGradient id='lg5'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg1' x1='50' xlink:href='#lg5'/> <linearGradient id='lg6'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg3' x1='50' xlink:href='#lg6'/> <linearGradient id='lg2' x1='100' xlink:href='#lg5'/> <linearGradient id='lg4' x1='100' xlink:href='#lg6'/> </svg> "); }
{ let mut nodes: Vec<Node> = doc.descendants() .filter(|n| n.is_gradient()) .filter(|n| n.has_children()) .filter(|n| !n.has_attribute(AId::XlinkHref)) .collect(); let mut is_changed = false; let mut join_nodes = Vec::new(); // TODO: join with rm_dupl_defs::rm_loop let mut i1 = 0; while i1 < nodes.len() { let mut node1 = nodes[i1].clone(); let mut i2 = i1 + 1; while i2 < nodes.len() { let node2 = nodes[i2].clone(); i2 += 1;
identifier_body
regroup_gradient_stops.rs
// svgcleaner could help you to clean up your SVG files // from unnecessary data. // Copyright (C) 2012-2018 Evgeniy Reizner // // This program is free software; you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation; either version 2 of the License, or // (at your option) any later version. // // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // // You should have received a copy of the GNU General Public License along // with this program; if not, write to the Free Software Foundation, Inc., // 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. use svgdom::{ Document, ElementType, Node, }; use task::short::{EId, AId}; pub fn
(doc: &mut Document) { let mut nodes: Vec<Node> = doc.descendants() .filter(|n| n.is_gradient()) .filter(|n| n.has_children()) .filter(|n| !n.has_attribute(AId::XlinkHref)) .collect(); let mut is_changed = false; let mut join_nodes = Vec::new(); // TODO: join with rm_dupl_defs::rm_loop let mut i1 = 0; while i1 < nodes.len() { let mut node1 = nodes[i1].clone(); let mut i2 = i1 + 1; while i2 < nodes.len() { let node2 = nodes[i2].clone(); i2 += 1; if super::rm_dupl_defs::is_equal_stops(&node1, &node2) { join_nodes.push(node2.clone()); nodes.remove(i2 - 1); i2 -= 1; } } if !join_nodes.is_empty() { is_changed = true; let mut new_lg = doc.create_element(EId::LinearGradient); let new_id = gen_id(doc, "lg"); new_lg.set_id(new_id); while node1.has_children() { let mut c = node1.first_child().unwrap(); c.detach(); new_lg.append(&c); } node1.set_attribute((AId::XlinkHref, new_lg.clone())); node1.insert_before(&new_lg); for jn in &mut join_nodes { while jn.has_children() { let mut c = jn.first_child().unwrap(); c.remove(); } jn.set_attribute((AId::XlinkHref, new_lg.clone())); } join_nodes.clear(); } i1 += 1; } if is_changed { // We must resolve attributes for gradients created above. super::resolve_linear_gradient_attributes(doc); } } fn gen_id(doc: &Document, prefix: &str) -> String { let mut n = 1; let mut s = String::new(); loop { s.clear(); s.push_str(prefix); s.push_str(&n.to_string()); // TODO: very slow if !doc.descendants().any(|n| *n.id() == s) { break; } n += 1; } s } #[cfg(test)] mod tests { use super::*; use svgdom::{Document, ToStringWithOptions}; use task; macro_rules! test { ($name:ident, $in_text:expr, $out_text:expr) => ( #[test] fn $name() { let mut doc = Document::from_str($in_text).unwrap(); task::resolve_linear_gradient_attributes(&doc); task::resolve_radial_gradient_attributes(&doc); task::resolve_stop_attributes(&doc).unwrap(); regroup_gradient_stops(&mut doc); assert_eq_text!(doc.to_string_with_opt(&write_opt_for_tests!()), $out_text); } ) } test!(rm_1, "<svg> <linearGradient id='lg1' x1='50'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg2' x1='100'> <stop offset='0'/> <stop offset='1'/> </linearGradient> </svg>", "<svg> <linearGradient id='lg3'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg1' x1='50' xlink:href='#lg3'/> <linearGradient id='lg2' x1='100' xlink:href='#lg3'/> </svg> "); test!(rm_2, "<svg> <linearGradient id='lg1' x1='50'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg3' x1='50'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg2' x1='100'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg4' x1='100'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> </svg>", "<svg> <linearGradient id='lg5'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg1' x1='50' xlink:href='#lg5'/> <linearGradient id='lg6'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg3' x1='50' xlink:href='#lg6'/> <linearGradient id='lg2' x1='100' xlink:href='#lg5'/> <linearGradient id='lg4' x1='100' xlink:href='#lg6'/> </svg> "); }
regroup_gradient_stops
identifier_name
regroup_gradient_stops.rs
// svgcleaner could help you to clean up your SVG files // from unnecessary data. // Copyright (C) 2012-2018 Evgeniy Reizner // // This program is free software; you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation; either version 2 of the License, or // (at your option) any later version. // // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // // You should have received a copy of the GNU General Public License along // with this program; if not, write to the Free Software Foundation, Inc., // 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. use svgdom::{ Document, ElementType, Node, }; use task::short::{EId, AId}; pub fn regroup_gradient_stops(doc: &mut Document) { let mut nodes: Vec<Node> = doc.descendants() .filter(|n| n.is_gradient()) .filter(|n| n.has_children()) .filter(|n| !n.has_attribute(AId::XlinkHref)) .collect(); let mut is_changed = false; let mut join_nodes = Vec::new(); // TODO: join with rm_dupl_defs::rm_loop let mut i1 = 0; while i1 < nodes.len() { let mut node1 = nodes[i1].clone(); let mut i2 = i1 + 1; while i2 < nodes.len() { let node2 = nodes[i2].clone(); i2 += 1; if super::rm_dupl_defs::is_equal_stops(&node1, &node2) { join_nodes.push(node2.clone()); nodes.remove(i2 - 1); i2 -= 1; } } if !join_nodes.is_empty() { is_changed = true; let mut new_lg = doc.create_element(EId::LinearGradient); let new_id = gen_id(doc, "lg"); new_lg.set_id(new_id); while node1.has_children() { let mut c = node1.first_child().unwrap(); c.detach(); new_lg.append(&c); } node1.set_attribute((AId::XlinkHref, new_lg.clone())); node1.insert_before(&new_lg); for jn in &mut join_nodes { while jn.has_children() { let mut c = jn.first_child().unwrap(); c.remove(); } jn.set_attribute((AId::XlinkHref, new_lg.clone())); } join_nodes.clear(); } i1 += 1; } if is_changed { // We must resolve attributes for gradients created above. super::resolve_linear_gradient_attributes(doc); } } fn gen_id(doc: &Document, prefix: &str) -> String { let mut n = 1; let mut s = String::new(); loop { s.clear(); s.push_str(prefix); s.push_str(&n.to_string()); // TODO: very slow if !doc.descendants().any(|n| *n.id() == s)
n += 1; } s } #[cfg(test)] mod tests { use super::*; use svgdom::{Document, ToStringWithOptions}; use task; macro_rules! test { ($name:ident, $in_text:expr, $out_text:expr) => ( #[test] fn $name() { let mut doc = Document::from_str($in_text).unwrap(); task::resolve_linear_gradient_attributes(&doc); task::resolve_radial_gradient_attributes(&doc); task::resolve_stop_attributes(&doc).unwrap(); regroup_gradient_stops(&mut doc); assert_eq_text!(doc.to_string_with_opt(&write_opt_for_tests!()), $out_text); } ) } test!(rm_1, "<svg> <linearGradient id='lg1' x1='50'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg2' x1='100'> <stop offset='0'/> <stop offset='1'/> </linearGradient> </svg>", "<svg> <linearGradient id='lg3'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg1' x1='50' xlink:href='#lg3'/> <linearGradient id='lg2' x1='100' xlink:href='#lg3'/> </svg> "); test!(rm_2, "<svg> <linearGradient id='lg1' x1='50'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg3' x1='50'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg2' x1='100'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg4' x1='100'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> </svg>", "<svg> <linearGradient id='lg5'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg1' x1='50' xlink:href='#lg5'/> <linearGradient id='lg6'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg3' x1='50' xlink:href='#lg6'/> <linearGradient id='lg2' x1='100' xlink:href='#lg5'/> <linearGradient id='lg4' x1='100' xlink:href='#lg6'/> </svg> "); }
{ break; }
conditional_block
regroup_gradient_stops.rs
// svgcleaner could help you to clean up your SVG files // from unnecessary data. // Copyright (C) 2012-2018 Evgeniy Reizner // // This program is free software; you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation; either version 2 of the License, or // (at your option) any later version. // // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // // You should have received a copy of the GNU General Public License along // with this program; if not, write to the Free Software Foundation, Inc., // 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. use svgdom::{ Document, ElementType, Node, }; use task::short::{EId, AId}; pub fn regroup_gradient_stops(doc: &mut Document) { let mut nodes: Vec<Node> = doc.descendants() .filter(|n| n.is_gradient()) .filter(|n| n.has_children()) .filter(|n| !n.has_attribute(AId::XlinkHref)) .collect(); let mut is_changed = false; let mut join_nodes = Vec::new(); // TODO: join with rm_dupl_defs::rm_loop let mut i1 = 0; while i1 < nodes.len() { let mut node1 = nodes[i1].clone(); let mut i2 = i1 + 1; while i2 < nodes.len() { let node2 = nodes[i2].clone(); i2 += 1; if super::rm_dupl_defs::is_equal_stops(&node1, &node2) { join_nodes.push(node2.clone()); nodes.remove(i2 - 1); i2 -= 1; } } if !join_nodes.is_empty() { is_changed = true; let mut new_lg = doc.create_element(EId::LinearGradient); let new_id = gen_id(doc, "lg"); new_lg.set_id(new_id); while node1.has_children() { let mut c = node1.first_child().unwrap(); c.detach(); new_lg.append(&c); } node1.set_attribute((AId::XlinkHref, new_lg.clone())); node1.insert_before(&new_lg); for jn in &mut join_nodes { while jn.has_children() { let mut c = jn.first_child().unwrap(); c.remove(); } jn.set_attribute((AId::XlinkHref, new_lg.clone())); } join_nodes.clear(); } i1 += 1; } if is_changed { // We must resolve attributes for gradients created above. super::resolve_linear_gradient_attributes(doc); } } fn gen_id(doc: &Document, prefix: &str) -> String { let mut n = 1; let mut s = String::new(); loop { s.clear(); s.push_str(prefix); s.push_str(&n.to_string()); // TODO: very slow if !doc.descendants().any(|n| *n.id() == s) { break; } n += 1; } s } #[cfg(test)] mod tests { use super::*; use svgdom::{Document, ToStringWithOptions}; use task; macro_rules! test { ($name:ident, $in_text:expr, $out_text:expr) => ( #[test] fn $name() { let mut doc = Document::from_str($in_text).unwrap(); task::resolve_linear_gradient_attributes(&doc); task::resolve_radial_gradient_attributes(&doc); task::resolve_stop_attributes(&doc).unwrap();
) } test!(rm_1, "<svg> <linearGradient id='lg1' x1='50'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg2' x1='100'> <stop offset='0'/> <stop offset='1'/> </linearGradient> </svg>", "<svg> <linearGradient id='lg3'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg1' x1='50' xlink:href='#lg3'/> <linearGradient id='lg2' x1='100' xlink:href='#lg3'/> </svg> "); test!(rm_2, "<svg> <linearGradient id='lg1' x1='50'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg3' x1='50'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg2' x1='100'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg4' x1='100'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> </svg>", "<svg> <linearGradient id='lg5'> <stop offset='0'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg1' x1='50' xlink:href='#lg5'/> <linearGradient id='lg6'> <stop offset='0.5'/> <stop offset='1'/> </linearGradient> <linearGradient id='lg3' x1='50' xlink:href='#lg6'/> <linearGradient id='lg2' x1='100' xlink:href='#lg5'/> <linearGradient id='lg4' x1='100' xlink:href='#lg6'/> </svg> "); }
regroup_gradient_stops(&mut doc); assert_eq_text!(doc.to_string_with_opt(&write_opt_for_tests!()), $out_text); }
random_line_split
document_loader.rs
/* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ //! Tracking of pending loads in a document. //! https://html.spec.whatwg.org/multipage/#the-end use dom::bindings::js::JS; use dom::document::Document; use ipc_channel::ipc::IpcSender; use net_traits::{CoreResourceMsg, FetchResponseMsg, ResourceThreads, IpcSend}; use net_traits::request::RequestInit; use std::thread; use url::Url; #[derive(JSTraceable, PartialEq, Clone, Debug, HeapSizeOf)] pub enum LoadType { Image(Url), Script(Url), Subframe(Url), Stylesheet(Url), PageSource(Url), Media(Url), } impl LoadType { fn url(&self) -> &Url { match *self { LoadType::Image(ref url) | LoadType::Script(ref url) | LoadType::Subframe(ref url) | LoadType::Stylesheet(ref url) | LoadType::Media(ref url) | LoadType::PageSource(ref url) => url, } } } /// Canary value ensuring that manually added blocking loads (ie. ones that weren't /// created via DocumentLoader::fetch_async) are always removed by the time /// that the owner is destroyed. #[derive(JSTraceable, HeapSizeOf)] #[must_root] pub struct LoadBlocker { /// The document whose load event is blocked by this object existing. doc: JS<Document>, /// The load that is blocking the document's load event. load: Option<LoadType>, } impl LoadBlocker { /// Mark the document's load event as blocked on this new load. pub fn new(doc: &Document, load: LoadType) -> LoadBlocker { doc.mut_loader().add_blocking_load(load.clone()); LoadBlocker { doc: JS::from_ref(doc), load: Some(load), } } /// Remove this load from the associated document's list of blocking loads. pub fn terminate(blocker: &mut Option<LoadBlocker>) { if let Some(this) = blocker.as_mut() { this.doc.finish_load(this.load.take().unwrap()); } *blocker = None; } /// Return the url associated with this load. pub fn url(&self) -> Option<&Url> { self.load.as_ref().map(LoadType::url) } } impl Drop for LoadBlocker { fn drop(&mut self) { if !thread::panicking() { debug_assert!(self.load.is_none()); } } } #[derive(JSTraceable, HeapSizeOf)] pub struct DocumentLoader { resource_threads: ResourceThreads, blocking_loads: Vec<LoadType>, events_inhibited: bool, } impl DocumentLoader { pub fn new(existing: &DocumentLoader) -> DocumentLoader { DocumentLoader::new_with_threads(existing.resource_threads.clone(), None) } pub fn new_with_threads(resource_threads: ResourceThreads, initial_load: Option<Url>) -> DocumentLoader
/// Add a load to the list of blocking loads. fn add_blocking_load(&mut self, load: LoadType) { debug!("Adding blocking load {:?} ({}).", load, self.blocking_loads.len()); self.blocking_loads.push(load); } /// Initiate a new fetch. pub fn fetch_async(&mut self, load: LoadType, request: RequestInit, fetch_target: IpcSender<FetchResponseMsg>) { self.add_blocking_load(load); self.resource_threads.sender().send(CoreResourceMsg::Fetch(request, fetch_target)).unwrap(); } /// Mark an in-progress network request complete. pub fn finish_load(&mut self, load: &LoadType) { debug!("Removing blocking load {:?} ({}).", load, self.blocking_loads.len()); let idx = self.blocking_loads.iter().position(|unfinished| *unfinished == *load); self.blocking_loads.remove(idx.expect(&format!("unknown completed load {:?}", load))); } pub fn is_blocked(&self) -> bool { // TODO: Ensure that we report blocked if parsing is still ongoing. !self.blocking_loads.is_empty() } pub fn inhibit_events(&mut self) { self.events_inhibited = true; } pub fn events_inhibited(&self) -> bool { self.events_inhibited } pub fn resource_threads(&self) -> &ResourceThreads { &self.resource_threads } }
{ debug!("Initial blocking load {:?}.", initial_load); let initial_loads = initial_load.into_iter().map(LoadType::PageSource).collect(); DocumentLoader { resource_threads: resource_threads, blocking_loads: initial_loads, events_inhibited: false, } }
identifier_body
document_loader.rs
/* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ //! Tracking of pending loads in a document. //! https://html.spec.whatwg.org/multipage/#the-end use dom::bindings::js::JS; use dom::document::Document; use ipc_channel::ipc::IpcSender; use net_traits::{CoreResourceMsg, FetchResponseMsg, ResourceThreads, IpcSend}; use net_traits::request::RequestInit; use std::thread; use url::Url; #[derive(JSTraceable, PartialEq, Clone, Debug, HeapSizeOf)] pub enum LoadType { Image(Url), Script(Url), Subframe(Url), Stylesheet(Url), PageSource(Url), Media(Url), } impl LoadType { fn url(&self) -> &Url { match *self { LoadType::Image(ref url) | LoadType::Script(ref url) | LoadType::Subframe(ref url) | LoadType::Stylesheet(ref url) | LoadType::Media(ref url) | LoadType::PageSource(ref url) => url, } } } /// Canary value ensuring that manually added blocking loads (ie. ones that weren't /// created via DocumentLoader::fetch_async) are always removed by the time /// that the owner is destroyed. #[derive(JSTraceable, HeapSizeOf)] #[must_root] pub struct LoadBlocker { /// The document whose load event is blocked by this object existing. doc: JS<Document>, /// The load that is blocking the document's load event. load: Option<LoadType>, } impl LoadBlocker { /// Mark the document's load event as blocked on this new load. pub fn new(doc: &Document, load: LoadType) -> LoadBlocker { doc.mut_loader().add_blocking_load(load.clone()); LoadBlocker { doc: JS::from_ref(doc), load: Some(load), } } /// Remove this load from the associated document's list of blocking loads. pub fn terminate(blocker: &mut Option<LoadBlocker>) { if let Some(this) = blocker.as_mut() { this.doc.finish_load(this.load.take().unwrap()); } *blocker = None; } /// Return the url associated with this load. pub fn url(&self) -> Option<&Url> { self.load.as_ref().map(LoadType::url) } } impl Drop for LoadBlocker { fn drop(&mut self) { if !thread::panicking()
} } #[derive(JSTraceable, HeapSizeOf)] pub struct DocumentLoader { resource_threads: ResourceThreads, blocking_loads: Vec<LoadType>, events_inhibited: bool, } impl DocumentLoader { pub fn new(existing: &DocumentLoader) -> DocumentLoader { DocumentLoader::new_with_threads(existing.resource_threads.clone(), None) } pub fn new_with_threads(resource_threads: ResourceThreads, initial_load: Option<Url>) -> DocumentLoader { debug!("Initial blocking load {:?}.", initial_load); let initial_loads = initial_load.into_iter().map(LoadType::PageSource).collect(); DocumentLoader { resource_threads: resource_threads, blocking_loads: initial_loads, events_inhibited: false, } } /// Add a load to the list of blocking loads. fn add_blocking_load(&mut self, load: LoadType) { debug!("Adding blocking load {:?} ({}).", load, self.blocking_loads.len()); self.blocking_loads.push(load); } /// Initiate a new fetch. pub fn fetch_async(&mut self, load: LoadType, request: RequestInit, fetch_target: IpcSender<FetchResponseMsg>) { self.add_blocking_load(load); self.resource_threads.sender().send(CoreResourceMsg::Fetch(request, fetch_target)).unwrap(); } /// Mark an in-progress network request complete. pub fn finish_load(&mut self, load: &LoadType) { debug!("Removing blocking load {:?} ({}).", load, self.blocking_loads.len()); let idx = self.blocking_loads.iter().position(|unfinished| *unfinished == *load); self.blocking_loads.remove(idx.expect(&format!("unknown completed load {:?}", load))); } pub fn is_blocked(&self) -> bool { // TODO: Ensure that we report blocked if parsing is still ongoing. !self.blocking_loads.is_empty() } pub fn inhibit_events(&mut self) { self.events_inhibited = true; } pub fn events_inhibited(&self) -> bool { self.events_inhibited } pub fn resource_threads(&self) -> &ResourceThreads { &self.resource_threads } }
{ debug_assert!(self.load.is_none()); }
conditional_block
document_loader.rs
/* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ //! Tracking of pending loads in a document. //! https://html.spec.whatwg.org/multipage/#the-end use dom::bindings::js::JS; use dom::document::Document; use ipc_channel::ipc::IpcSender; use net_traits::{CoreResourceMsg, FetchResponseMsg, ResourceThreads, IpcSend}; use net_traits::request::RequestInit; use std::thread; use url::Url; #[derive(JSTraceable, PartialEq, Clone, Debug, HeapSizeOf)] pub enum LoadType { Image(Url), Script(Url), Subframe(Url), Stylesheet(Url), PageSource(Url), Media(Url), } impl LoadType { fn url(&self) -> &Url { match *self { LoadType::Image(ref url) | LoadType::Script(ref url) | LoadType::Subframe(ref url) | LoadType::Stylesheet(ref url) | LoadType::Media(ref url) | LoadType::PageSource(ref url) => url, } } } /// Canary value ensuring that manually added blocking loads (ie. ones that weren't /// created via DocumentLoader::fetch_async) are always removed by the time /// that the owner is destroyed. #[derive(JSTraceable, HeapSizeOf)] #[must_root] pub struct LoadBlocker { /// The document whose load event is blocked by this object existing. doc: JS<Document>, /// The load that is blocking the document's load event. load: Option<LoadType>, } impl LoadBlocker { /// Mark the document's load event as blocked on this new load.
doc: JS::from_ref(doc), load: Some(load), } } /// Remove this load from the associated document's list of blocking loads. pub fn terminate(blocker: &mut Option<LoadBlocker>) { if let Some(this) = blocker.as_mut() { this.doc.finish_load(this.load.take().unwrap()); } *blocker = None; } /// Return the url associated with this load. pub fn url(&self) -> Option<&Url> { self.load.as_ref().map(LoadType::url) } } impl Drop for LoadBlocker { fn drop(&mut self) { if !thread::panicking() { debug_assert!(self.load.is_none()); } } } #[derive(JSTraceable, HeapSizeOf)] pub struct DocumentLoader { resource_threads: ResourceThreads, blocking_loads: Vec<LoadType>, events_inhibited: bool, } impl DocumentLoader { pub fn new(existing: &DocumentLoader) -> DocumentLoader { DocumentLoader::new_with_threads(existing.resource_threads.clone(), None) } pub fn new_with_threads(resource_threads: ResourceThreads, initial_load: Option<Url>) -> DocumentLoader { debug!("Initial blocking load {:?}.", initial_load); let initial_loads = initial_load.into_iter().map(LoadType::PageSource).collect(); DocumentLoader { resource_threads: resource_threads, blocking_loads: initial_loads, events_inhibited: false, } } /// Add a load to the list of blocking loads. fn add_blocking_load(&mut self, load: LoadType) { debug!("Adding blocking load {:?} ({}).", load, self.blocking_loads.len()); self.blocking_loads.push(load); } /// Initiate a new fetch. pub fn fetch_async(&mut self, load: LoadType, request: RequestInit, fetch_target: IpcSender<FetchResponseMsg>) { self.add_blocking_load(load); self.resource_threads.sender().send(CoreResourceMsg::Fetch(request, fetch_target)).unwrap(); } /// Mark an in-progress network request complete. pub fn finish_load(&mut self, load: &LoadType) { debug!("Removing blocking load {:?} ({}).", load, self.blocking_loads.len()); let idx = self.blocking_loads.iter().position(|unfinished| *unfinished == *load); self.blocking_loads.remove(idx.expect(&format!("unknown completed load {:?}", load))); } pub fn is_blocked(&self) -> bool { // TODO: Ensure that we report blocked if parsing is still ongoing. !self.blocking_loads.is_empty() } pub fn inhibit_events(&mut self) { self.events_inhibited = true; } pub fn events_inhibited(&self) -> bool { self.events_inhibited } pub fn resource_threads(&self) -> &ResourceThreads { &self.resource_threads } }
pub fn new(doc: &Document, load: LoadType) -> LoadBlocker { doc.mut_loader().add_blocking_load(load.clone()); LoadBlocker {
random_line_split
document_loader.rs
/* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ //! Tracking of pending loads in a document. //! https://html.spec.whatwg.org/multipage/#the-end use dom::bindings::js::JS; use dom::document::Document; use ipc_channel::ipc::IpcSender; use net_traits::{CoreResourceMsg, FetchResponseMsg, ResourceThreads, IpcSend}; use net_traits::request::RequestInit; use std::thread; use url::Url; #[derive(JSTraceable, PartialEq, Clone, Debug, HeapSizeOf)] pub enum LoadType { Image(Url), Script(Url), Subframe(Url), Stylesheet(Url), PageSource(Url), Media(Url), } impl LoadType { fn url(&self) -> &Url { match *self { LoadType::Image(ref url) | LoadType::Script(ref url) | LoadType::Subframe(ref url) | LoadType::Stylesheet(ref url) | LoadType::Media(ref url) | LoadType::PageSource(ref url) => url, } } } /// Canary value ensuring that manually added blocking loads (ie. ones that weren't /// created via DocumentLoader::fetch_async) are always removed by the time /// that the owner is destroyed. #[derive(JSTraceable, HeapSizeOf)] #[must_root] pub struct LoadBlocker { /// The document whose load event is blocked by this object existing. doc: JS<Document>, /// The load that is blocking the document's load event. load: Option<LoadType>, } impl LoadBlocker { /// Mark the document's load event as blocked on this new load. pub fn new(doc: &Document, load: LoadType) -> LoadBlocker { doc.mut_loader().add_blocking_load(load.clone()); LoadBlocker { doc: JS::from_ref(doc), load: Some(load), } } /// Remove this load from the associated document's list of blocking loads. pub fn terminate(blocker: &mut Option<LoadBlocker>) { if let Some(this) = blocker.as_mut() { this.doc.finish_load(this.load.take().unwrap()); } *blocker = None; } /// Return the url associated with this load. pub fn url(&self) -> Option<&Url> { self.load.as_ref().map(LoadType::url) } } impl Drop for LoadBlocker { fn drop(&mut self) { if !thread::panicking() { debug_assert!(self.load.is_none()); } } } #[derive(JSTraceable, HeapSizeOf)] pub struct DocumentLoader { resource_threads: ResourceThreads, blocking_loads: Vec<LoadType>, events_inhibited: bool, } impl DocumentLoader { pub fn
(existing: &DocumentLoader) -> DocumentLoader { DocumentLoader::new_with_threads(existing.resource_threads.clone(), None) } pub fn new_with_threads(resource_threads: ResourceThreads, initial_load: Option<Url>) -> DocumentLoader { debug!("Initial blocking load {:?}.", initial_load); let initial_loads = initial_load.into_iter().map(LoadType::PageSource).collect(); DocumentLoader { resource_threads: resource_threads, blocking_loads: initial_loads, events_inhibited: false, } } /// Add a load to the list of blocking loads. fn add_blocking_load(&mut self, load: LoadType) { debug!("Adding blocking load {:?} ({}).", load, self.blocking_loads.len()); self.blocking_loads.push(load); } /// Initiate a new fetch. pub fn fetch_async(&mut self, load: LoadType, request: RequestInit, fetch_target: IpcSender<FetchResponseMsg>) { self.add_blocking_load(load); self.resource_threads.sender().send(CoreResourceMsg::Fetch(request, fetch_target)).unwrap(); } /// Mark an in-progress network request complete. pub fn finish_load(&mut self, load: &LoadType) { debug!("Removing blocking load {:?} ({}).", load, self.blocking_loads.len()); let idx = self.blocking_loads.iter().position(|unfinished| *unfinished == *load); self.blocking_loads.remove(idx.expect(&format!("unknown completed load {:?}", load))); } pub fn is_blocked(&self) -> bool { // TODO: Ensure that we report blocked if parsing is still ongoing. !self.blocking_loads.is_empty() } pub fn inhibit_events(&mut self) { self.events_inhibited = true; } pub fn events_inhibited(&self) -> bool { self.events_inhibited } pub fn resource_threads(&self) -> &ResourceThreads { &self.resource_threads } }
new
identifier_name
test_rowcount.py
from sqlalchemy import * from sqlalchemy.test import * class FoundRowsTest(TestBase, AssertsExecutionResults): """tests rowcount functionality""" __requires__ = ('sane_rowcount', ) @classmethod def
(cls): global employees_table, metadata metadata = MetaData(testing.db) employees_table = Table('employees', metadata, Column('employee_id', Integer, Sequence('employee_id_seq', optional=True), primary_key=True), Column('name', String(50)), Column('department', String(1)), ) metadata.create_all() def setup(self): global data data = [ ('Angela', 'A'), ('Andrew', 'A'), ('Anand', 'A'), ('Bob', 'B'), ('Bobette', 'B'), ('Buffy', 'B'), ('Charlie', 'C'), ('Cynthia', 'C'), ('Chris', 'C') ] i = employees_table.insert() i.execute(*[{'name':n, 'department':d} for n, d in data]) def teardown(self): employees_table.delete().execute() @classmethod def teardown_class(cls): metadata.drop_all() def testbasic(self): s = employees_table.select() r = s.execute().fetchall() assert len(r) == len(data) def test_update_rowcount1(self): # WHERE matches 3, 3 rows changed department = employees_table.c.department r = employees_table.update(department=='C').execute(department='Z') print "expecting 3, dialect reports %s" % r.rowcount assert r.rowcount == 3 def test_update_rowcount2(self): # WHERE matches 3, 0 rows changed department = employees_table.c.department r = employees_table.update(department=='C').execute(department='C') print "expecting 3, dialect reports %s" % r.rowcount assert r.rowcount == 3 def test_delete_rowcount(self): # WHERE matches 3, 3 rows deleted department = employees_table.c.department r = employees_table.delete(department=='C').execute() print "expecting 3, dialect reports %s" % r.rowcount assert r.rowcount == 3
setup_class
identifier_name
test_rowcount.py
from sqlalchemy import * from sqlalchemy.test import * class FoundRowsTest(TestBase, AssertsExecutionResults): """tests rowcount functionality""" __requires__ = ('sane_rowcount', ) @classmethod def setup_class(cls): global employees_table, metadata metadata = MetaData(testing.db) employees_table = Table('employees', metadata, Column('employee_id', Integer, Sequence('employee_id_seq', optional=True), primary_key=True), Column('name', String(50)), Column('department', String(1)), ) metadata.create_all() def setup(self): global data data = [ ('Angela', 'A'), ('Andrew', 'A'), ('Anand', 'A'), ('Bob', 'B'), ('Bobette', 'B'), ('Buffy', 'B'), ('Charlie', 'C'), ('Cynthia', 'C'), ('Chris', 'C') ] i = employees_table.insert() i.execute(*[{'name':n, 'department':d} for n, d in data]) def teardown(self): employees_table.delete().execute()
def testbasic(self): s = employees_table.select() r = s.execute().fetchall() assert len(r) == len(data) def test_update_rowcount1(self): # WHERE matches 3, 3 rows changed department = employees_table.c.department r = employees_table.update(department=='C').execute(department='Z') print "expecting 3, dialect reports %s" % r.rowcount assert r.rowcount == 3 def test_update_rowcount2(self): # WHERE matches 3, 0 rows changed department = employees_table.c.department r = employees_table.update(department=='C').execute(department='C') print "expecting 3, dialect reports %s" % r.rowcount assert r.rowcount == 3 def test_delete_rowcount(self): # WHERE matches 3, 3 rows deleted department = employees_table.c.department r = employees_table.delete(department=='C').execute() print "expecting 3, dialect reports %s" % r.rowcount assert r.rowcount == 3
@classmethod def teardown_class(cls): metadata.drop_all()
random_line_split
test_rowcount.py
from sqlalchemy import * from sqlalchemy.test import * class FoundRowsTest(TestBase, AssertsExecutionResults): """tests rowcount functionality""" __requires__ = ('sane_rowcount', ) @classmethod def setup_class(cls): global employees_table, metadata metadata = MetaData(testing.db) employees_table = Table('employees', metadata, Column('employee_id', Integer, Sequence('employee_id_seq', optional=True), primary_key=True), Column('name', String(50)), Column('department', String(1)), ) metadata.create_all() def setup(self): global data data = [ ('Angela', 'A'), ('Andrew', 'A'), ('Anand', 'A'), ('Bob', 'B'), ('Bobette', 'B'), ('Buffy', 'B'), ('Charlie', 'C'), ('Cynthia', 'C'), ('Chris', 'C') ] i = employees_table.insert() i.execute(*[{'name':n, 'department':d} for n, d in data]) def teardown(self): employees_table.delete().execute() @classmethod def teardown_class(cls): metadata.drop_all() def testbasic(self): s = employees_table.select() r = s.execute().fetchall() assert len(r) == len(data) def test_update_rowcount1(self): # WHERE matches 3, 3 rows changed
def test_update_rowcount2(self): # WHERE matches 3, 0 rows changed department = employees_table.c.department r = employees_table.update(department=='C').execute(department='C') print "expecting 3, dialect reports %s" % r.rowcount assert r.rowcount == 3 def test_delete_rowcount(self): # WHERE matches 3, 3 rows deleted department = employees_table.c.department r = employees_table.delete(department=='C').execute() print "expecting 3, dialect reports %s" % r.rowcount assert r.rowcount == 3
department = employees_table.c.department r = employees_table.update(department=='C').execute(department='Z') print "expecting 3, dialect reports %s" % r.rowcount assert r.rowcount == 3
identifier_body
lib.rs
// ================================================================= // // * WARNING * // // This file is generated! // // Changes made to this file will be overwritten. If changes are // required to the generated code, the service_crategen project // must be updated to generate the changes. // // ================================================================= #![doc(html_logo_url = "https://raw.githubusercontent.com/rusoto/rusoto/master/assets/logo-square.png")] //! <p>Using the Amazon Cognito User Pools API, you can create a user pool to manage directories and users. You can authenticate a user to obtain tokens related to user identity and access policies.</p> <p>This API reference provides information about user pools in Amazon Cognito User Pools.</p> <p>For more information, see the Amazon Cognito Documentation.</p> //! //! If you're using the service, you're probably looking for [CognitoIdentityProviderClient](struct.CognitoIdentityProviderClient.html) and [CognitoIdentityProvider](trait.CognitoIdentityProvider.html).
#[macro_use] extern crate serde_derive; extern crate serde_json; mod generated; mod custom; pub use generated::*; pub use custom::*;
extern crate futures; extern crate rusoto_core; extern crate serde;
random_line_split
common.js
"use strict"; var core_1 = require('@angular/core'); var NgTranscludeDirective = (function () { function NgTranscludeDirective(_viewRef)
Object.defineProperty(NgTranscludeDirective.prototype, "ngTransclude", { get: function () { return this._ngTransclude; }, set: function (templateRef) { this._ngTransclude = templateRef; if (templateRef) { this.viewRef.createEmbeddedView(templateRef); } }, enumerable: true, configurable: true }); NgTranscludeDirective.decorators = [ { type: core_1.Directive, args: [{ selector: '[ngTransclude]' },] }, ]; /** @nocollapse */ NgTranscludeDirective.ctorParameters = [ { type: core_1.ViewContainerRef, }, ]; NgTranscludeDirective.propDecorators = { 'ngTransclude': [{ type: core_1.Input },], }; return NgTranscludeDirective; }()); exports.NgTranscludeDirective = NgTranscludeDirective;
{ this._viewRef = _viewRef; this.viewRef = _viewRef; }
identifier_body
common.js
"use strict"; var core_1 = require('@angular/core'); var NgTranscludeDirective = (function () { function NgTranscludeDirective(_viewRef) { this._viewRef = _viewRef; this.viewRef = _viewRef; } Object.defineProperty(NgTranscludeDirective.prototype, "ngTransclude", { get: function () { return this._ngTransclude; }, set: function (templateRef) { this._ngTransclude = templateRef; if (templateRef)
}, enumerable: true, configurable: true }); NgTranscludeDirective.decorators = [ { type: core_1.Directive, args: [{ selector: '[ngTransclude]' },] }, ]; /** @nocollapse */ NgTranscludeDirective.ctorParameters = [ { type: core_1.ViewContainerRef, }, ]; NgTranscludeDirective.propDecorators = { 'ngTransclude': [{ type: core_1.Input },], }; return NgTranscludeDirective; }()); exports.NgTranscludeDirective = NgTranscludeDirective;
{ this.viewRef.createEmbeddedView(templateRef); }
conditional_block
common.js
"use strict"; var core_1 = require('@angular/core'); var NgTranscludeDirective = (function () { function
(_viewRef) { this._viewRef = _viewRef; this.viewRef = _viewRef; } Object.defineProperty(NgTranscludeDirective.prototype, "ngTransclude", { get: function () { return this._ngTransclude; }, set: function (templateRef) { this._ngTransclude = templateRef; if (templateRef) { this.viewRef.createEmbeddedView(templateRef); } }, enumerable: true, configurable: true }); NgTranscludeDirective.decorators = [ { type: core_1.Directive, args: [{ selector: '[ngTransclude]' },] }, ]; /** @nocollapse */ NgTranscludeDirective.ctorParameters = [ { type: core_1.ViewContainerRef, }, ]; NgTranscludeDirective.propDecorators = { 'ngTransclude': [{ type: core_1.Input },], }; return NgTranscludeDirective; }()); exports.NgTranscludeDirective = NgTranscludeDirective;
NgTranscludeDirective
identifier_name
common.js
"use strict"; var core_1 = require('@angular/core'); var NgTranscludeDirective = (function () { function NgTranscludeDirective(_viewRef) { this._viewRef = _viewRef; this.viewRef = _viewRef; } Object.defineProperty(NgTranscludeDirective.prototype, "ngTransclude", { get: function () { return this._ngTransclude; }, set: function (templateRef) { this._ngTransclude = templateRef; if (templateRef) { this.viewRef.createEmbeddedView(templateRef); } },
enumerable: true, configurable: true }); NgTranscludeDirective.decorators = [ { type: core_1.Directive, args: [{ selector: '[ngTransclude]' },] }, ]; /** @nocollapse */ NgTranscludeDirective.ctorParameters = [ { type: core_1.ViewContainerRef, }, ]; NgTranscludeDirective.propDecorators = { 'ngTransclude': [{ type: core_1.Input },], }; return NgTranscludeDirective; }()); exports.NgTranscludeDirective = NgTranscludeDirective;
random_line_split
mod.rs
// Copyright 2017, Igor Shaula // Licensed under the MIT License <LICENSE or // http://opensource.org/licenses/MIT>. This file // may not be copied, modified, or distributed // except according to those terms. use super::enums::*; use super::RegKey; use std::error::Error; use std::fmt; use std::io; use winapi::shared::minwindef::DWORD; macro_rules! read_value { ($s:ident) => { match mem::replace(&mut $s.f_name, None) { Some(ref s) => $s.key.get_value(s).map_err(DecoderError::IoError), None => Err(DecoderError::NoFieldName), } }; } macro_rules! parse_string { ($s:ident) => {{ let s: String = read_value!($s)?; s.parse() .map_err(|e| DecoderError::ParseError(format!("{:?}", e))) }}; } macro_rules! no_impl { ($e:expr) => { Err(DecoderError::DecodeNotImplemented($e.to_owned())) }; } #[cfg(feature = "serialization-serde")] mod serialization_serde; #[derive(Debug)] pub enum DecoderError { DecodeNotImplemented(String), DeserializerError(String), IoError(io::Error), ParseError(String), NoFieldName, } impl fmt::Display for DecoderError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{:?}", self) } } impl Error for DecoderError {} impl From<io::Error> for DecoderError { fn from(err: io::Error) -> DecoderError { DecoderError::IoError(err) } } pub type DecodeResult<T> = Result<T, DecoderError>; #[derive(Debug)] enum DecoderReadingState { WaitingForKey, WaitingForValue, } #[derive(Debug)] enum DecoderEnumerationState { EnumeratingKeys(DWORD), EnumeratingValues(DWORD), } #[derive(Debug)] pub struct Decoder { key: RegKey,
enumeration_state: DecoderEnumerationState, } const DECODER_SAM: DWORD = KEY_QUERY_VALUE | KEY_ENUMERATE_SUB_KEYS; impl Decoder { pub fn from_key(key: &RegKey) -> DecodeResult<Decoder> { key.open_subkey_with_flags("", DECODER_SAM) .map(Decoder::new) .map_err(DecoderError::IoError) } fn new(key: RegKey) -> Decoder { Decoder { key, f_name: None, reading_state: DecoderReadingState::WaitingForKey, enumeration_state: DecoderEnumerationState::EnumeratingKeys(0), } } }
f_name: Option<String>, reading_state: DecoderReadingState,
random_line_split
mod.rs
// Copyright 2017, Igor Shaula // Licensed under the MIT License <LICENSE or // http://opensource.org/licenses/MIT>. This file // may not be copied, modified, or distributed // except according to those terms. use super::enums::*; use super::RegKey; use std::error::Error; use std::fmt; use std::io; use winapi::shared::minwindef::DWORD; macro_rules! read_value { ($s:ident) => { match mem::replace(&mut $s.f_name, None) { Some(ref s) => $s.key.get_value(s).map_err(DecoderError::IoError), None => Err(DecoderError::NoFieldName), } }; } macro_rules! parse_string { ($s:ident) => {{ let s: String = read_value!($s)?; s.parse() .map_err(|e| DecoderError::ParseError(format!("{:?}", e))) }}; } macro_rules! no_impl { ($e:expr) => { Err(DecoderError::DecodeNotImplemented($e.to_owned())) }; } #[cfg(feature = "serialization-serde")] mod serialization_serde; #[derive(Debug)] pub enum DecoderError { DecodeNotImplemented(String), DeserializerError(String), IoError(io::Error), ParseError(String), NoFieldName, } impl fmt::Display for DecoderError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "{:?}", self) } } impl Error for DecoderError {} impl From<io::Error> for DecoderError { fn
(err: io::Error) -> DecoderError { DecoderError::IoError(err) } } pub type DecodeResult<T> = Result<T, DecoderError>; #[derive(Debug)] enum DecoderReadingState { WaitingForKey, WaitingForValue, } #[derive(Debug)] enum DecoderEnumerationState { EnumeratingKeys(DWORD), EnumeratingValues(DWORD), } #[derive(Debug)] pub struct Decoder { key: RegKey, f_name: Option<String>, reading_state: DecoderReadingState, enumeration_state: DecoderEnumerationState, } const DECODER_SAM: DWORD = KEY_QUERY_VALUE | KEY_ENUMERATE_SUB_KEYS; impl Decoder { pub fn from_key(key: &RegKey) -> DecodeResult<Decoder> { key.open_subkey_with_flags("", DECODER_SAM) .map(Decoder::new) .map_err(DecoderError::IoError) } fn new(key: RegKey) -> Decoder { Decoder { key, f_name: None, reading_state: DecoderReadingState::WaitingForKey, enumeration_state: DecoderEnumerationState::EnumeratingKeys(0), } } }
from
identifier_name
util.js
export const ATTR_ID = 'data-referid' export let info = { component: { amount: 0, mounts: 0, unmounts: 0 } } export let getId = () => Math.random().toString(36).substr(2) export let pipe = (fn1, fn2) => function(...args) { fn1.apply(this, args) return fn2.apply(this, args) } export let createCallbackStore = name => { let store = [] return { name, clear()
, push(item) { store.push(item) }, store } } export let wrapNative = (obj, method, fn) => { let nativeMethod = obj[method] let wrapper = function(...args) { fn.apply(this, args) return nativeMethod.apply(this, args) } obj[method] = wrapper return () => obj[method] = nativeMethod } if (!Object.assign) { Object.assign = (target, ...args) => { args.forEach(source => { for (let key in source) { if (!source.hasOwnProperty(key)) { continue } target[key] = source[key] } }) return target } }
{ while (store.length) { store.shift()() } }
identifier_body
util.js
export const ATTR_ID = 'data-referid' export let info = { component: { amount: 0, mounts: 0, unmounts: 0 } } export let getId = () => Math.random().toString(36).substr(2) export let pipe = (fn1, fn2) => function(...args) { fn1.apply(this, args) return fn2.apply(this, args) } export let createCallbackStore = name => { let store = [] return { name, clear() { while (store.length) { store.shift()() } },
(item) { store.push(item) }, store } } export let wrapNative = (obj, method, fn) => { let nativeMethod = obj[method] let wrapper = function(...args) { fn.apply(this, args) return nativeMethod.apply(this, args) } obj[method] = wrapper return () => obj[method] = nativeMethod } if (!Object.assign) { Object.assign = (target, ...args) => { args.forEach(source => { for (let key in source) { if (!source.hasOwnProperty(key)) { continue } target[key] = source[key] } }) return target } }
push
identifier_name
util.js
export const ATTR_ID = 'data-referid' export let info = { component: { amount: 0, mounts: 0, unmounts: 0 } } export let getId = () => Math.random().toString(36).substr(2) export let pipe = (fn1, fn2) => function(...args) { fn1.apply(this, args) return fn2.apply(this, args) } export let createCallbackStore = name => { let store = [] return { name, clear() { while (store.length) { store.shift()() } }, push(item) { store.push(item) }, store } } export let wrapNative = (obj, method, fn) => { let nativeMethod = obj[method] let wrapper = function(...args) { fn.apply(this, args) return nativeMethod.apply(this, args) } obj[method] = wrapper return () => obj[method] = nativeMethod } if (!Object.assign) { Object.assign = (target, ...args) => { args.forEach(source => { for (let key in source) { if (!source.hasOwnProperty(key))
target[key] = source[key] } }) return target } }
{ continue }
conditional_block
util.js
export const ATTR_ID = 'data-referid' export let info = { component: { amount: 0,
} export let getId = () => Math.random().toString(36).substr(2) export let pipe = (fn1, fn2) => function(...args) { fn1.apply(this, args) return fn2.apply(this, args) } export let createCallbackStore = name => { let store = [] return { name, clear() { while (store.length) { store.shift()() } }, push(item) { store.push(item) }, store } } export let wrapNative = (obj, method, fn) => { let nativeMethod = obj[method] let wrapper = function(...args) { fn.apply(this, args) return nativeMethod.apply(this, args) } obj[method] = wrapper return () => obj[method] = nativeMethod } if (!Object.assign) { Object.assign = (target, ...args) => { args.forEach(source => { for (let key in source) { if (!source.hasOwnProperty(key)) { continue } target[key] = source[key] } }) return target } }
mounts: 0, unmounts: 0 }
random_line_split
sensor.py
"""Bitcoin information service that uses blockchain.info.""" from datetime import timedelta import logging from blockchain import exchangerates, statistics import voluptuous as vol from homeassistant.components.sensor import PLATFORM_SCHEMA from homeassistant.const import ATTR_ATTRIBUTION, CONF_CURRENCY, CONF_DISPLAY_OPTIONS import homeassistant.helpers.config_validation as cv from homeassistant.helpers.entity import Entity _LOGGER = logging.getLogger(__name__) ATTRIBUTION = "Data provided by blockchain.info" DEFAULT_CURRENCY = "USD" ICON = "mdi:currency-btc" SCAN_INTERVAL = timedelta(minutes=5) OPTION_TYPES = { "exchangerate": ["Exchange rate (1 BTC)", None], "trade_volume_btc": ["Trade volume", "BTC"], "miners_revenue_usd": ["Miners revenue", "USD"], "btc_mined": ["Mined", "BTC"], "trade_volume_usd": ["Trade volume", "USD"], "difficulty": ["Difficulty", None], "minutes_between_blocks": ["Time between Blocks", "min"], "number_of_transactions": ["No. of Transactions", None], "hash_rate": ["Hash rate", "PH/s"], "timestamp": ["Timestamp", None], "mined_blocks": ["Mined Blocks", None], "blocks_size": ["Block size", None], "total_fees_btc": ["Total fees", "BTC"], "total_btc_sent": ["Total sent", "BTC"], "estimated_btc_sent": ["Estimated sent", "BTC"], "total_btc": ["Total", "BTC"], "total_blocks": ["Total Blocks", None], "next_retarget": ["Next retarget", None], "estimated_transaction_volume_usd": ["Est. Transaction volume", "USD"], "miners_revenue_btc": ["Miners revenue", "BTC"], "market_price_usd": ["Market price", "USD"], } PLATFORM_SCHEMA = PLATFORM_SCHEMA.extend( { vol.Required(CONF_DISPLAY_OPTIONS, default=[]): vol.All( cv.ensure_list, [vol.In(OPTION_TYPES)] ), vol.Optional(CONF_CURRENCY, default=DEFAULT_CURRENCY): cv.string, } ) def setup_platform(hass, config, add_entities, discovery_info=None): """Set up the Bitcoin sensors.""" currency = config.get(CONF_CURRENCY) if currency not in exchangerates.get_ticker(): _LOGGER.warning("Currency %s is not available. Using USD", currency) currency = DEFAULT_CURRENCY data = BitcoinData() dev = [] for variable in config[CONF_DISPLAY_OPTIONS]: dev.append(BitcoinSensor(data, variable, currency)) add_entities(dev, True) class BitcoinSensor(Entity): """Representation of a Bitcoin sensor.""" def __init__(self, data, option_type, currency): """Initialize the sensor.""" self.data = data self._name = OPTION_TYPES[option_type][0] self._unit_of_measurement = OPTION_TYPES[option_type][1] self._currency = currency self.type = option_type self._state = None @property def name(self): """Return the name of the sensor.""" return self._name @property def state(self): """Return the state of the sensor.""" return self._state @property def unit_of_measurement(self): """Return the unit the value is expressed in.""" return self._unit_of_measurement @property def icon(self): """Return the icon to use in the frontend, if any.""" return ICON @property def device_state_attributes(self): """Return the state attributes of the sensor.""" return {ATTR_ATTRIBUTION: ATTRIBUTION} def update(self): """Get the latest data and updates the states.""" self.data.update() stats = self.data.stats ticker = self.data.ticker if self.type == "exchangerate": self._state = ticker[self._currency].p15min self._unit_of_measurement = self._currency elif self.type == "trade_volume_btc": self._state = "{0:.1f}".format(stats.trade_volume_btc) elif self.type == "miners_revenue_usd": self._state = "{0:.0f}".format(stats.miners_revenue_usd) elif self.type == "btc_mined": self._state = "{}".format(stats.btc_mined * 0.00000001) elif self.type == "trade_volume_usd": self._state = "{0:.1f}".format(stats.trade_volume_usd) elif self.type == "difficulty": self._state = "{0:.0f}".format(stats.difficulty) elif self.type == "minutes_between_blocks": self._state = "{0:.2f}".format(stats.minutes_between_blocks) elif self.type == "number_of_transactions": self._state = "{}".format(stats.number_of_transactions) elif self.type == "hash_rate": self._state = "{0:.1f}".format(stats.hash_rate * 0.000001) elif self.type == "timestamp": self._state = stats.timestamp elif self.type == "mined_blocks": self._state = "{}".format(stats.mined_blocks) elif self.type == "blocks_size": self._state = "{0:.1f}".format(stats.blocks_size) elif self.type == "total_fees_btc": self._state = "{0:.2f}".format(stats.total_fees_btc * 0.00000001) elif self.type == "total_btc_sent":
elif self.type == "estimated_btc_sent": self._state = "{0:.2f}".format(stats.estimated_btc_sent * 0.00000001) elif self.type == "total_btc": self._state = "{0:.2f}".format(stats.total_btc * 0.00000001) elif self.type == "total_blocks": self._state = "{0:.0f}".format(stats.total_blocks) elif self.type == "next_retarget": self._state = "{0:.2f}".format(stats.next_retarget) elif self.type == "estimated_transaction_volume_usd": self._state = "{0:.2f}".format(stats.estimated_transaction_volume_usd) elif self.type == "miners_revenue_btc": self._state = "{0:.1f}".format(stats.miners_revenue_btc * 0.00000001) elif self.type == "market_price_usd": self._state = "{0:.2f}".format(stats.market_price_usd) class BitcoinData: """Get the latest data and update the states.""" def __init__(self): """Initialize the data object.""" self.stats = None self.ticker = None def update(self): """Get the latest data from blockchain.info.""" self.stats = statistics.get() self.ticker = exchangerates.get_ticker()
self._state = "{0:.2f}".format(stats.total_btc_sent * 0.00000001)
conditional_block
sensor.py
"""Bitcoin information service that uses blockchain.info.""" from datetime import timedelta import logging from blockchain import exchangerates, statistics import voluptuous as vol from homeassistant.components.sensor import PLATFORM_SCHEMA from homeassistant.const import ATTR_ATTRIBUTION, CONF_CURRENCY, CONF_DISPLAY_OPTIONS import homeassistant.helpers.config_validation as cv from homeassistant.helpers.entity import Entity _LOGGER = logging.getLogger(__name__) ATTRIBUTION = "Data provided by blockchain.info" DEFAULT_CURRENCY = "USD" ICON = "mdi:currency-btc" SCAN_INTERVAL = timedelta(minutes=5) OPTION_TYPES = { "exchangerate": ["Exchange rate (1 BTC)", None], "trade_volume_btc": ["Trade volume", "BTC"], "miners_revenue_usd": ["Miners revenue", "USD"], "btc_mined": ["Mined", "BTC"], "trade_volume_usd": ["Trade volume", "USD"], "difficulty": ["Difficulty", None], "minutes_between_blocks": ["Time between Blocks", "min"], "number_of_transactions": ["No. of Transactions", None], "hash_rate": ["Hash rate", "PH/s"], "timestamp": ["Timestamp", None], "mined_blocks": ["Mined Blocks", None], "blocks_size": ["Block size", None], "total_fees_btc": ["Total fees", "BTC"], "total_btc_sent": ["Total sent", "BTC"], "estimated_btc_sent": ["Estimated sent", "BTC"], "total_btc": ["Total", "BTC"], "total_blocks": ["Total Blocks", None], "next_retarget": ["Next retarget", None], "estimated_transaction_volume_usd": ["Est. Transaction volume", "USD"], "miners_revenue_btc": ["Miners revenue", "BTC"], "market_price_usd": ["Market price", "USD"], } PLATFORM_SCHEMA = PLATFORM_SCHEMA.extend( { vol.Required(CONF_DISPLAY_OPTIONS, default=[]): vol.All( cv.ensure_list, [vol.In(OPTION_TYPES)] ), vol.Optional(CONF_CURRENCY, default=DEFAULT_CURRENCY): cv.string, } ) def setup_platform(hass, config, add_entities, discovery_info=None): """Set up the Bitcoin sensors.""" currency = config.get(CONF_CURRENCY) if currency not in exchangerates.get_ticker(): _LOGGER.warning("Currency %s is not available. Using USD", currency) currency = DEFAULT_CURRENCY data = BitcoinData() dev = [] for variable in config[CONF_DISPLAY_OPTIONS]: dev.append(BitcoinSensor(data, variable, currency)) add_entities(dev, True) class BitcoinSensor(Entity): """Representation of a Bitcoin sensor.""" def __init__(self, data, option_type, currency):
@property def name(self): """Return the name of the sensor.""" return self._name @property def state(self): """Return the state of the sensor.""" return self._state @property def unit_of_measurement(self): """Return the unit the value is expressed in.""" return self._unit_of_measurement @property def icon(self): """Return the icon to use in the frontend, if any.""" return ICON @property def device_state_attributes(self): """Return the state attributes of the sensor.""" return {ATTR_ATTRIBUTION: ATTRIBUTION} def update(self): """Get the latest data and updates the states.""" self.data.update() stats = self.data.stats ticker = self.data.ticker if self.type == "exchangerate": self._state = ticker[self._currency].p15min self._unit_of_measurement = self._currency elif self.type == "trade_volume_btc": self._state = "{0:.1f}".format(stats.trade_volume_btc) elif self.type == "miners_revenue_usd": self._state = "{0:.0f}".format(stats.miners_revenue_usd) elif self.type == "btc_mined": self._state = "{}".format(stats.btc_mined * 0.00000001) elif self.type == "trade_volume_usd": self._state = "{0:.1f}".format(stats.trade_volume_usd) elif self.type == "difficulty": self._state = "{0:.0f}".format(stats.difficulty) elif self.type == "minutes_between_blocks": self._state = "{0:.2f}".format(stats.minutes_between_blocks) elif self.type == "number_of_transactions": self._state = "{}".format(stats.number_of_transactions) elif self.type == "hash_rate": self._state = "{0:.1f}".format(stats.hash_rate * 0.000001) elif self.type == "timestamp": self._state = stats.timestamp elif self.type == "mined_blocks": self._state = "{}".format(stats.mined_blocks) elif self.type == "blocks_size": self._state = "{0:.1f}".format(stats.blocks_size) elif self.type == "total_fees_btc": self._state = "{0:.2f}".format(stats.total_fees_btc * 0.00000001) elif self.type == "total_btc_sent": self._state = "{0:.2f}".format(stats.total_btc_sent * 0.00000001) elif self.type == "estimated_btc_sent": self._state = "{0:.2f}".format(stats.estimated_btc_sent * 0.00000001) elif self.type == "total_btc": self._state = "{0:.2f}".format(stats.total_btc * 0.00000001) elif self.type == "total_blocks": self._state = "{0:.0f}".format(stats.total_blocks) elif self.type == "next_retarget": self._state = "{0:.2f}".format(stats.next_retarget) elif self.type == "estimated_transaction_volume_usd": self._state = "{0:.2f}".format(stats.estimated_transaction_volume_usd) elif self.type == "miners_revenue_btc": self._state = "{0:.1f}".format(stats.miners_revenue_btc * 0.00000001) elif self.type == "market_price_usd": self._state = "{0:.2f}".format(stats.market_price_usd) class BitcoinData: """Get the latest data and update the states.""" def __init__(self): """Initialize the data object.""" self.stats = None self.ticker = None def update(self): """Get the latest data from blockchain.info.""" self.stats = statistics.get() self.ticker = exchangerates.get_ticker()
"""Initialize the sensor.""" self.data = data self._name = OPTION_TYPES[option_type][0] self._unit_of_measurement = OPTION_TYPES[option_type][1] self._currency = currency self.type = option_type self._state = None
identifier_body
sensor.py
"""Bitcoin information service that uses blockchain.info.""" from datetime import timedelta import logging from blockchain import exchangerates, statistics import voluptuous as vol from homeassistant.components.sensor import PLATFORM_SCHEMA from homeassistant.const import ATTR_ATTRIBUTION, CONF_CURRENCY, CONF_DISPLAY_OPTIONS import homeassistant.helpers.config_validation as cv from homeassistant.helpers.entity import Entity _LOGGER = logging.getLogger(__name__) ATTRIBUTION = "Data provided by blockchain.info" DEFAULT_CURRENCY = "USD" ICON = "mdi:currency-btc" SCAN_INTERVAL = timedelta(minutes=5) OPTION_TYPES = { "exchangerate": ["Exchange rate (1 BTC)", None], "trade_volume_btc": ["Trade volume", "BTC"], "miners_revenue_usd": ["Miners revenue", "USD"], "btc_mined": ["Mined", "BTC"], "trade_volume_usd": ["Trade volume", "USD"], "difficulty": ["Difficulty", None], "minutes_between_blocks": ["Time between Blocks", "min"], "number_of_transactions": ["No. of Transactions", None], "hash_rate": ["Hash rate", "PH/s"], "timestamp": ["Timestamp", None], "mined_blocks": ["Mined Blocks", None], "blocks_size": ["Block size", None], "total_fees_btc": ["Total fees", "BTC"], "total_btc_sent": ["Total sent", "BTC"], "estimated_btc_sent": ["Estimated sent", "BTC"], "total_btc": ["Total", "BTC"], "total_blocks": ["Total Blocks", None], "next_retarget": ["Next retarget", None], "estimated_transaction_volume_usd": ["Est. Transaction volume", "USD"], "miners_revenue_btc": ["Miners revenue", "BTC"], "market_price_usd": ["Market price", "USD"], } PLATFORM_SCHEMA = PLATFORM_SCHEMA.extend( { vol.Required(CONF_DISPLAY_OPTIONS, default=[]): vol.All( cv.ensure_list, [vol.In(OPTION_TYPES)] ), vol.Optional(CONF_CURRENCY, default=DEFAULT_CURRENCY): cv.string, } ) def setup_platform(hass, config, add_entities, discovery_info=None): """Set up the Bitcoin sensors.""" currency = config.get(CONF_CURRENCY) if currency not in exchangerates.get_ticker(): _LOGGER.warning("Currency %s is not available. Using USD", currency) currency = DEFAULT_CURRENCY data = BitcoinData() dev = [] for variable in config[CONF_DISPLAY_OPTIONS]: dev.append(BitcoinSensor(data, variable, currency)) add_entities(dev, True) class BitcoinSensor(Entity): """Representation of a Bitcoin sensor.""" def __init__(self, data, option_type, currency): """Initialize the sensor.""" self.data = data self._name = OPTION_TYPES[option_type][0] self._unit_of_measurement = OPTION_TYPES[option_type][1] self._currency = currency self.type = option_type self._state = None @property def name(self): """Return the name of the sensor.""" return self._name @property def state(self): """Return the state of the sensor.""" return self._state @property def unit_of_measurement(self): """Return the unit the value is expressed in.""" return self._unit_of_measurement @property def icon(self): """Return the icon to use in the frontend, if any.""" return ICON @property def device_state_attributes(self): """Return the state attributes of the sensor.""" return {ATTR_ATTRIBUTION: ATTRIBUTION} def update(self): """Get the latest data and updates the states.""" self.data.update() stats = self.data.stats ticker = self.data.ticker if self.type == "exchangerate": self._state = ticker[self._currency].p15min self._unit_of_measurement = self._currency elif self.type == "trade_volume_btc":
elif self.type == "trade_volume_usd": self._state = "{0:.1f}".format(stats.trade_volume_usd) elif self.type == "difficulty": self._state = "{0:.0f}".format(stats.difficulty) elif self.type == "minutes_between_blocks": self._state = "{0:.2f}".format(stats.minutes_between_blocks) elif self.type == "number_of_transactions": self._state = "{}".format(stats.number_of_transactions) elif self.type == "hash_rate": self._state = "{0:.1f}".format(stats.hash_rate * 0.000001) elif self.type == "timestamp": self._state = stats.timestamp elif self.type == "mined_blocks": self._state = "{}".format(stats.mined_blocks) elif self.type == "blocks_size": self._state = "{0:.1f}".format(stats.blocks_size) elif self.type == "total_fees_btc": self._state = "{0:.2f}".format(stats.total_fees_btc * 0.00000001) elif self.type == "total_btc_sent": self._state = "{0:.2f}".format(stats.total_btc_sent * 0.00000001) elif self.type == "estimated_btc_sent": self._state = "{0:.2f}".format(stats.estimated_btc_sent * 0.00000001) elif self.type == "total_btc": self._state = "{0:.2f}".format(stats.total_btc * 0.00000001) elif self.type == "total_blocks": self._state = "{0:.0f}".format(stats.total_blocks) elif self.type == "next_retarget": self._state = "{0:.2f}".format(stats.next_retarget) elif self.type == "estimated_transaction_volume_usd": self._state = "{0:.2f}".format(stats.estimated_transaction_volume_usd) elif self.type == "miners_revenue_btc": self._state = "{0:.1f}".format(stats.miners_revenue_btc * 0.00000001) elif self.type == "market_price_usd": self._state = "{0:.2f}".format(stats.market_price_usd) class BitcoinData: """Get the latest data and update the states.""" def __init__(self): """Initialize the data object.""" self.stats = None self.ticker = None def update(self): """Get the latest data from blockchain.info.""" self.stats = statistics.get() self.ticker = exchangerates.get_ticker()
self._state = "{0:.1f}".format(stats.trade_volume_btc) elif self.type == "miners_revenue_usd": self._state = "{0:.0f}".format(stats.miners_revenue_usd) elif self.type == "btc_mined": self._state = "{}".format(stats.btc_mined * 0.00000001)
random_line_split
sensor.py
"""Bitcoin information service that uses blockchain.info.""" from datetime import timedelta import logging from blockchain import exchangerates, statistics import voluptuous as vol from homeassistant.components.sensor import PLATFORM_SCHEMA from homeassistant.const import ATTR_ATTRIBUTION, CONF_CURRENCY, CONF_DISPLAY_OPTIONS import homeassistant.helpers.config_validation as cv from homeassistant.helpers.entity import Entity _LOGGER = logging.getLogger(__name__) ATTRIBUTION = "Data provided by blockchain.info" DEFAULT_CURRENCY = "USD" ICON = "mdi:currency-btc" SCAN_INTERVAL = timedelta(minutes=5) OPTION_TYPES = { "exchangerate": ["Exchange rate (1 BTC)", None], "trade_volume_btc": ["Trade volume", "BTC"], "miners_revenue_usd": ["Miners revenue", "USD"], "btc_mined": ["Mined", "BTC"], "trade_volume_usd": ["Trade volume", "USD"], "difficulty": ["Difficulty", None], "minutes_between_blocks": ["Time between Blocks", "min"], "number_of_transactions": ["No. of Transactions", None], "hash_rate": ["Hash rate", "PH/s"], "timestamp": ["Timestamp", None], "mined_blocks": ["Mined Blocks", None], "blocks_size": ["Block size", None], "total_fees_btc": ["Total fees", "BTC"], "total_btc_sent": ["Total sent", "BTC"], "estimated_btc_sent": ["Estimated sent", "BTC"], "total_btc": ["Total", "BTC"], "total_blocks": ["Total Blocks", None], "next_retarget": ["Next retarget", None], "estimated_transaction_volume_usd": ["Est. Transaction volume", "USD"], "miners_revenue_btc": ["Miners revenue", "BTC"], "market_price_usd": ["Market price", "USD"], } PLATFORM_SCHEMA = PLATFORM_SCHEMA.extend( { vol.Required(CONF_DISPLAY_OPTIONS, default=[]): vol.All( cv.ensure_list, [vol.In(OPTION_TYPES)] ), vol.Optional(CONF_CURRENCY, default=DEFAULT_CURRENCY): cv.string, } ) def
(hass, config, add_entities, discovery_info=None): """Set up the Bitcoin sensors.""" currency = config.get(CONF_CURRENCY) if currency not in exchangerates.get_ticker(): _LOGGER.warning("Currency %s is not available. Using USD", currency) currency = DEFAULT_CURRENCY data = BitcoinData() dev = [] for variable in config[CONF_DISPLAY_OPTIONS]: dev.append(BitcoinSensor(data, variable, currency)) add_entities(dev, True) class BitcoinSensor(Entity): """Representation of a Bitcoin sensor.""" def __init__(self, data, option_type, currency): """Initialize the sensor.""" self.data = data self._name = OPTION_TYPES[option_type][0] self._unit_of_measurement = OPTION_TYPES[option_type][1] self._currency = currency self.type = option_type self._state = None @property def name(self): """Return the name of the sensor.""" return self._name @property def state(self): """Return the state of the sensor.""" return self._state @property def unit_of_measurement(self): """Return the unit the value is expressed in.""" return self._unit_of_measurement @property def icon(self): """Return the icon to use in the frontend, if any.""" return ICON @property def device_state_attributes(self): """Return the state attributes of the sensor.""" return {ATTR_ATTRIBUTION: ATTRIBUTION} def update(self): """Get the latest data and updates the states.""" self.data.update() stats = self.data.stats ticker = self.data.ticker if self.type == "exchangerate": self._state = ticker[self._currency].p15min self._unit_of_measurement = self._currency elif self.type == "trade_volume_btc": self._state = "{0:.1f}".format(stats.trade_volume_btc) elif self.type == "miners_revenue_usd": self._state = "{0:.0f}".format(stats.miners_revenue_usd) elif self.type == "btc_mined": self._state = "{}".format(stats.btc_mined * 0.00000001) elif self.type == "trade_volume_usd": self._state = "{0:.1f}".format(stats.trade_volume_usd) elif self.type == "difficulty": self._state = "{0:.0f}".format(stats.difficulty) elif self.type == "minutes_between_blocks": self._state = "{0:.2f}".format(stats.minutes_between_blocks) elif self.type == "number_of_transactions": self._state = "{}".format(stats.number_of_transactions) elif self.type == "hash_rate": self._state = "{0:.1f}".format(stats.hash_rate * 0.000001) elif self.type == "timestamp": self._state = stats.timestamp elif self.type == "mined_blocks": self._state = "{}".format(stats.mined_blocks) elif self.type == "blocks_size": self._state = "{0:.1f}".format(stats.blocks_size) elif self.type == "total_fees_btc": self._state = "{0:.2f}".format(stats.total_fees_btc * 0.00000001) elif self.type == "total_btc_sent": self._state = "{0:.2f}".format(stats.total_btc_sent * 0.00000001) elif self.type == "estimated_btc_sent": self._state = "{0:.2f}".format(stats.estimated_btc_sent * 0.00000001) elif self.type == "total_btc": self._state = "{0:.2f}".format(stats.total_btc * 0.00000001) elif self.type == "total_blocks": self._state = "{0:.0f}".format(stats.total_blocks) elif self.type == "next_retarget": self._state = "{0:.2f}".format(stats.next_retarget) elif self.type == "estimated_transaction_volume_usd": self._state = "{0:.2f}".format(stats.estimated_transaction_volume_usd) elif self.type == "miners_revenue_btc": self._state = "{0:.1f}".format(stats.miners_revenue_btc * 0.00000001) elif self.type == "market_price_usd": self._state = "{0:.2f}".format(stats.market_price_usd) class BitcoinData: """Get the latest data and update the states.""" def __init__(self): """Initialize the data object.""" self.stats = None self.ticker = None def update(self): """Get the latest data from blockchain.info.""" self.stats = statistics.get() self.ticker = exchangerates.get_ticker()
setup_platform
identifier_name
main.rs
// CITA // Copyright 2016-2017 Cryptape Technologies LLC. // This program is free software: you can redistribute it // and/or modify it under the terms of the GNU General Public // License as published by the Free Software Foundation, // either version 3 of the License, or (at your option) any // later version. // This program is distributed in the hope that it will be // useful, but WITHOUT ANY WARRANTY; without even the implied // warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR // PURPOSE. See the GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with this program. If not, see <http://www.gnu.org/licenses/>. extern crate serde; extern crate serde_json; extern crate libproto; extern crate util; extern crate threadpool; extern crate rustc_serialize; extern crate protobuf; #[macro_use] extern crate log; extern crate clap; extern crate tx_pool; extern crate cita_crypto as crypto; extern crate proof; extern crate pubsub; extern crate engine_json; extern crate engine; extern crate parking_lot; extern crate cpuprofiler; extern crate cita_log; extern crate dotenv; pub mod core; use clap::App; use core::Spec; use core::handler; use cpuprofiler::PROFILER; use libproto::*; use log::LogLevelFilter; use pubsub::start_pubsub; use std::sync::mpsc::channel; use std::thread; use std::time::{Duration, Instant}; fn main() { dotenv::dotenv().ok(); // Always print backtrace on panic. ::std::env::set_var("RUST_BACKTRACE", "1"); cita_log::format(LogLevelFilter::Info); println!("CITA:consensus:poa"); let matches = App::new("authority_round") .version("0.1") .author("Cryptape") .about("CITA Block Chain Node powered by Rust") .args_from_usage("-c, --config=[FILE] 'Sets a custom config file'") .args_from_usage("--prof-start=[TIME] 'Sets profiling start time (second from app start)'") .args_from_usage("--prof-duration=[DURATION] 'Sets duration(second) of profiling'") .get_matches(); let mut config_path = "config"; if let Some(c) = matches.value_of("config") { trace!("Value for config: {}", c); config_path = c; } //start profiling let mut prof_start: u64 = 0; let mut prof_duration: u64 = 0; if let Some(start) = matches.value_of("prof-start") { trace!("Value for prof-start: {}", start); prof_start = start.parse::<u64>().unwrap(); } if let Some(duration) = matches.value_of("prof-duration") { trace!("Value for prof-duration: {}", duration); prof_duration = duration.parse::<u64>().unwrap(); } if prof_start != 0 && prof_duration != 0
let threadpool = threadpool::ThreadPool::new(2); let (tx, rx) = channel(); let (tx_sub, rx_sub) = channel(); let (tx_pub, rx_pub) = channel(); start_pubsub("consensus", vec!["net.tx", "jsonrpc.new_tx", "net.msg", "chain.status"], tx_sub, rx_pub); thread::spawn(move || loop { let (key, body) = rx_sub.recv().unwrap(); let tx = tx.clone(); handler::receive(&threadpool, &tx, key_to_id(&key), body); }); let spec = Spec::new_test_round(config_path); let engine = spec.engine; let ready = spec.rx; let process = engine.clone(); let tx_pub1 = tx_pub.clone(); thread::spawn(move || loop { let process = process.clone(); handler::process(process, &rx, tx_pub1.clone()); }); let seal = engine.clone(); let dur = engine.duration(); let mut old_height = 0; let mut new_height = 0; let tx_pub = tx_pub.clone(); thread::spawn(move || loop { let seal = seal.clone(); trace!("seal worker lock!"); loop { new_height = ready.recv().unwrap(); if new_height > old_height { old_height = new_height; break; } } trace!("seal worker go {}!!!", new_height); let now = Instant::now(); trace!("seal worker ready!"); handler::seal(seal, tx_pub.clone()); let elapsed = now.elapsed(); if let Some(dur1) = dur.checked_sub(elapsed) { trace!("seal worker sleep !!!!!{:?}", dur1); thread::sleep(dur1); } }); loop { thread::sleep(Duration::from_millis(10000)); } }
{ thread::spawn(move || { thread::sleep(Duration::new(prof_start, 0)); println!("******Profiling Start******"); PROFILER.lock().unwrap().start("./consensus_poa.profile").expect("Couldn't start"); thread::sleep(Duration::new(prof_duration, 0)); println!("******Profiling Stop******"); PROFILER.lock().unwrap().stop().unwrap(); }); }
conditional_block
main.rs
// CITA // Copyright 2016-2017 Cryptape Technologies LLC. // This program is free software: you can redistribute it // and/or modify it under the terms of the GNU General Public // License as published by the Free Software Foundation, // either version 3 of the License, or (at your option) any // later version. // This program is distributed in the hope that it will be // useful, but WITHOUT ANY WARRANTY; without even the implied // warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR // PURPOSE. See the GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with this program. If not, see <http://www.gnu.org/licenses/>. extern crate serde; extern crate serde_json; extern crate libproto; extern crate util; extern crate threadpool; extern crate rustc_serialize; extern crate protobuf; #[macro_use] extern crate log; extern crate clap; extern crate tx_pool; extern crate cita_crypto as crypto; extern crate proof; extern crate pubsub; extern crate engine_json; extern crate engine; extern crate parking_lot; extern crate cpuprofiler; extern crate cita_log; extern crate dotenv; pub mod core; use clap::App; use core::Spec; use core::handler; use cpuprofiler::PROFILER; use libproto::*; use log::LogLevelFilter; use pubsub::start_pubsub; use std::sync::mpsc::channel; use std::thread; use std::time::{Duration, Instant}; fn
() { dotenv::dotenv().ok(); // Always print backtrace on panic. ::std::env::set_var("RUST_BACKTRACE", "1"); cita_log::format(LogLevelFilter::Info); println!("CITA:consensus:poa"); let matches = App::new("authority_round") .version("0.1") .author("Cryptape") .about("CITA Block Chain Node powered by Rust") .args_from_usage("-c, --config=[FILE] 'Sets a custom config file'") .args_from_usage("--prof-start=[TIME] 'Sets profiling start time (second from app start)'") .args_from_usage("--prof-duration=[DURATION] 'Sets duration(second) of profiling'") .get_matches(); let mut config_path = "config"; if let Some(c) = matches.value_of("config") { trace!("Value for config: {}", c); config_path = c; } //start profiling let mut prof_start: u64 = 0; let mut prof_duration: u64 = 0; if let Some(start) = matches.value_of("prof-start") { trace!("Value for prof-start: {}", start); prof_start = start.parse::<u64>().unwrap(); } if let Some(duration) = matches.value_of("prof-duration") { trace!("Value for prof-duration: {}", duration); prof_duration = duration.parse::<u64>().unwrap(); } if prof_start != 0 && prof_duration != 0 { thread::spawn(move || { thread::sleep(Duration::new(prof_start, 0)); println!("******Profiling Start******"); PROFILER.lock().unwrap().start("./consensus_poa.profile").expect("Couldn't start"); thread::sleep(Duration::new(prof_duration, 0)); println!("******Profiling Stop******"); PROFILER.lock().unwrap().stop().unwrap(); }); } let threadpool = threadpool::ThreadPool::new(2); let (tx, rx) = channel(); let (tx_sub, rx_sub) = channel(); let (tx_pub, rx_pub) = channel(); start_pubsub("consensus", vec!["net.tx", "jsonrpc.new_tx", "net.msg", "chain.status"], tx_sub, rx_pub); thread::spawn(move || loop { let (key, body) = rx_sub.recv().unwrap(); let tx = tx.clone(); handler::receive(&threadpool, &tx, key_to_id(&key), body); }); let spec = Spec::new_test_round(config_path); let engine = spec.engine; let ready = spec.rx; let process = engine.clone(); let tx_pub1 = tx_pub.clone(); thread::spawn(move || loop { let process = process.clone(); handler::process(process, &rx, tx_pub1.clone()); }); let seal = engine.clone(); let dur = engine.duration(); let mut old_height = 0; let mut new_height = 0; let tx_pub = tx_pub.clone(); thread::spawn(move || loop { let seal = seal.clone(); trace!("seal worker lock!"); loop { new_height = ready.recv().unwrap(); if new_height > old_height { old_height = new_height; break; } } trace!("seal worker go {}!!!", new_height); let now = Instant::now(); trace!("seal worker ready!"); handler::seal(seal, tx_pub.clone()); let elapsed = now.elapsed(); if let Some(dur1) = dur.checked_sub(elapsed) { trace!("seal worker sleep !!!!!{:?}", dur1); thread::sleep(dur1); } }); loop { thread::sleep(Duration::from_millis(10000)); } }
main
identifier_name
main.rs
// CITA // Copyright 2016-2017 Cryptape Technologies LLC. // This program is free software: you can redistribute it // and/or modify it under the terms of the GNU General Public // License as published by the Free Software Foundation, // either version 3 of the License, or (at your option) any // later version. // This program is distributed in the hope that it will be // useful, but WITHOUT ANY WARRANTY; without even the implied // warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR // PURPOSE. See the GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with this program. If not, see <http://www.gnu.org/licenses/>. extern crate serde; extern crate serde_json; extern crate libproto; extern crate util; extern crate threadpool; extern crate rustc_serialize; extern crate protobuf; #[macro_use] extern crate log; extern crate clap; extern crate tx_pool; extern crate cita_crypto as crypto; extern crate proof; extern crate pubsub; extern crate engine_json; extern crate engine; extern crate parking_lot; extern crate cpuprofiler; extern crate cita_log; extern crate dotenv; pub mod core; use clap::App; use core::Spec; use core::handler; use cpuprofiler::PROFILER; use libproto::*; use log::LogLevelFilter; use pubsub::start_pubsub; use std::sync::mpsc::channel; use std::thread; use std::time::{Duration, Instant}; fn main() { dotenv::dotenv().ok(); // Always print backtrace on panic. ::std::env::set_var("RUST_BACKTRACE", "1"); cita_log::format(LogLevelFilter::Info); println!("CITA:consensus:poa"); let matches = App::new("authority_round") .version("0.1") .author("Cryptape") .about("CITA Block Chain Node powered by Rust") .args_from_usage("-c, --config=[FILE] 'Sets a custom config file'") .args_from_usage("--prof-start=[TIME] 'Sets profiling start time (second from app start)'") .args_from_usage("--prof-duration=[DURATION] 'Sets duration(second) of profiling'") .get_matches(); let mut config_path = "config"; if let Some(c) = matches.value_of("config") { trace!("Value for config: {}", c); config_path = c; } //start profiling let mut prof_start: u64 = 0; let mut prof_duration: u64 = 0; if let Some(start) = matches.value_of("prof-start") { trace!("Value for prof-start: {}", start); prof_start = start.parse::<u64>().unwrap(); } if let Some(duration) = matches.value_of("prof-duration") { trace!("Value for prof-duration: {}", duration); prof_duration = duration.parse::<u64>().unwrap(); } if prof_start != 0 && prof_duration != 0 { thread::spawn(move || { thread::sleep(Duration::new(prof_start, 0)); println!("******Profiling Start******"); PROFILER.lock().unwrap().start("./consensus_poa.profile").expect("Couldn't start"); thread::sleep(Duration::new(prof_duration, 0)); println!("******Profiling Stop******"); PROFILER.lock().unwrap().stop().unwrap(); }); } let threadpool = threadpool::ThreadPool::new(2); let (tx, rx) = channel(); let (tx_sub, rx_sub) = channel(); let (tx_pub, rx_pub) = channel(); start_pubsub("consensus", vec!["net.tx", "jsonrpc.new_tx", "net.msg", "chain.status"], tx_sub, rx_pub); thread::spawn(move || loop { let (key, body) = rx_sub.recv().unwrap(); let tx = tx.clone(); handler::receive(&threadpool, &tx, key_to_id(&key), body); }); let spec = Spec::new_test_round(config_path); let engine = spec.engine; let ready = spec.rx; let process = engine.clone(); let tx_pub1 = tx_pub.clone();
let seal = engine.clone(); let dur = engine.duration(); let mut old_height = 0; let mut new_height = 0; let tx_pub = tx_pub.clone(); thread::spawn(move || loop { let seal = seal.clone(); trace!("seal worker lock!"); loop { new_height = ready.recv().unwrap(); if new_height > old_height { old_height = new_height; break; } } trace!("seal worker go {}!!!", new_height); let now = Instant::now(); trace!("seal worker ready!"); handler::seal(seal, tx_pub.clone()); let elapsed = now.elapsed(); if let Some(dur1) = dur.checked_sub(elapsed) { trace!("seal worker sleep !!!!!{:?}", dur1); thread::sleep(dur1); } }); loop { thread::sleep(Duration::from_millis(10000)); } }
thread::spawn(move || loop { let process = process.clone(); handler::process(process, &rx, tx_pub1.clone()); });
random_line_split
main.rs
// CITA // Copyright 2016-2017 Cryptape Technologies LLC. // This program is free software: you can redistribute it // and/or modify it under the terms of the GNU General Public // License as published by the Free Software Foundation, // either version 3 of the License, or (at your option) any // later version. // This program is distributed in the hope that it will be // useful, but WITHOUT ANY WARRANTY; without even the implied // warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR // PURPOSE. See the GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with this program. If not, see <http://www.gnu.org/licenses/>. extern crate serde; extern crate serde_json; extern crate libproto; extern crate util; extern crate threadpool; extern crate rustc_serialize; extern crate protobuf; #[macro_use] extern crate log; extern crate clap; extern crate tx_pool; extern crate cita_crypto as crypto; extern crate proof; extern crate pubsub; extern crate engine_json; extern crate engine; extern crate parking_lot; extern crate cpuprofiler; extern crate cita_log; extern crate dotenv; pub mod core; use clap::App; use core::Spec; use core::handler; use cpuprofiler::PROFILER; use libproto::*; use log::LogLevelFilter; use pubsub::start_pubsub; use std::sync::mpsc::channel; use std::thread; use std::time::{Duration, Instant}; fn main()
config_path = c; } //start profiling let mut prof_start: u64 = 0; let mut prof_duration: u64 = 0; if let Some(start) = matches.value_of("prof-start") { trace!("Value for prof-start: {}", start); prof_start = start.parse::<u64>().unwrap(); } if let Some(duration) = matches.value_of("prof-duration") { trace!("Value for prof-duration: {}", duration); prof_duration = duration.parse::<u64>().unwrap(); } if prof_start != 0 && prof_duration != 0 { thread::spawn(move || { thread::sleep(Duration::new(prof_start, 0)); println!("******Profiling Start******"); PROFILER.lock().unwrap().start("./consensus_poa.profile").expect("Couldn't start"); thread::sleep(Duration::new(prof_duration, 0)); println!("******Profiling Stop******"); PROFILER.lock().unwrap().stop().unwrap(); }); } let threadpool = threadpool::ThreadPool::new(2); let (tx, rx) = channel(); let (tx_sub, rx_sub) = channel(); let (tx_pub, rx_pub) = channel(); start_pubsub("consensus", vec!["net.tx", "jsonrpc.new_tx", "net.msg", "chain.status"], tx_sub, rx_pub); thread::spawn(move || loop { let (key, body) = rx_sub.recv().unwrap(); let tx = tx.clone(); handler::receive(&threadpool, &tx, key_to_id(&key), body); }); let spec = Spec::new_test_round(config_path); let engine = spec.engine; let ready = spec.rx; let process = engine.clone(); let tx_pub1 = tx_pub.clone(); thread::spawn(move || loop { let process = process.clone(); handler::process(process, &rx, tx_pub1.clone()); }); let seal = engine.clone(); let dur = engine.duration(); let mut old_height = 0; let mut new_height = 0; let tx_pub = tx_pub.clone(); thread::spawn(move || loop { let seal = seal.clone(); trace!("seal worker lock!"); loop { new_height = ready.recv().unwrap(); if new_height > old_height { old_height = new_height; break; } } trace!("seal worker go {}!!!", new_height); let now = Instant::now(); trace!("seal worker ready!"); handler::seal(seal, tx_pub.clone()); let elapsed = now.elapsed(); if let Some(dur1) = dur.checked_sub(elapsed) { trace!("seal worker sleep !!!!!{:?}", dur1); thread::sleep(dur1); } }); loop { thread::sleep(Duration::from_millis(10000)); } }
{ dotenv::dotenv().ok(); // Always print backtrace on panic. ::std::env::set_var("RUST_BACKTRACE", "1"); cita_log::format(LogLevelFilter::Info); println!("CITA:consensus:poa"); let matches = App::new("authority_round") .version("0.1") .author("Cryptape") .about("CITA Block Chain Node powered by Rust") .args_from_usage("-c, --config=[FILE] 'Sets a custom config file'") .args_from_usage("--prof-start=[TIME] 'Sets profiling start time (second from app start)'") .args_from_usage("--prof-duration=[DURATION] 'Sets duration(second) of profiling'") .get_matches(); let mut config_path = "config"; if let Some(c) = matches.value_of("config") { trace!("Value for config: {}", c);
identifier_body
nccl_ops_test.py
# Copyright 2016 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the 'License'); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an 'AS IS' BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for nccl ops. See also the cc test for nccl_communicator.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from functools import partial import numpy as np from tensorflow.contrib import nccl from tensorflow.python.framework import errors from tensorflow.python.framework import ops from tensorflow.python.ops import array_ops from tensorflow.python.ops import gradients from tensorflow.python.platform import test def _DeviceTensors(tensors, devices): res = [] for t, d in zip(tensors, devices): with ops.device(d): res.append(array_ops.identity(t)) return res def _NcclAllReduce(nccl_fun, tensors, devices): return nccl_fun(_DeviceTensors(tensors, devices)) def _NcclReduce(nccl_fun, tensors, devices):
def _NcclBroadcast(tensors, devices): sender = np.random.randint(0, len(devices)) with ops.device(devices[sender]): tensor = array_ops.identity(tensors[0]) broadcast = nccl.broadcast(tensor) return _DeviceTensors([broadcast] * len(devices), devices) class NcclTestCase(test.TestCase): def _Test(self, nccl_reduce, numpy_fn, device_sets=(['/device:GPU:1', '/device:GPU:2', '/device:GPU:0'], ['/device:GPU:1', '/device:GPU:0'])): """Tests that nccl_reduce does the same as reduction with numpy_fn. Args: nccl_reduce: A function taking a list of tensors and a list of devices, and returns a list of reduced tensors and a list of ops to perform the reduction. numpy_fn: A function taking two tensors and returning the reduction of the two. device_sets: Tuple of virtual devices to run test on. """ if not test.is_gpu_available(): return # Test requires access to a GPU for dtype in [np.float32, np.int32, np.int64, np.float64]: # Create session inside outer loop to test use of # same communicator across multiple sessions. with self.test_session(use_gpu=True) as sess: for devices in device_sets: shape = (3, 4) random = (np.random.random_sample(shape) - .5) * 1024 tensors = [] for _ in devices: tensors.append(random.astype(dtype)) np_ans = tensors[0] for t in tensors[1:]: np_ans = numpy_fn(np_ans, t) reduce_tensors = nccl_reduce(tensors, devices) self.assertNotEmpty(reduce_tensors) # Test shape inference. for r in reduce_tensors: self.assertEqual(shape, r.get_shape()) result_tensors = [array_ops.identity(t) for t in reduce_tensors] # Test execution and results. for t in sess.run(result_tensors): self.assertAllClose(t, np_ans) def _TestGradient(self, nccl_reduce, numpy_fn): """Tests the gradient of nccl_reduce. Args: nccl_reduce: A function taking a list of tensors and a list of devices, and returns a list of reduced tensors and a list of ops to perform the reduction. numpy_fn: A function taking two tensors and returning the gradient of the reduction of the two. """ def _Gradient(tensors, devices): inputs = [array_ops.placeholder(t.dtype, t.shape) for t in tensors] reduce_tensors = nccl_reduce(inputs, devices) losses = _DeviceTensors(tensors, [t.device for t in reduce_tensors]) grads = gradients.gradients( reduce_tensors, inputs, losses, colocate_gradients_with_ops=True) return [g for g in grads if g is not None] self._Test(_Gradient, numpy_fn) class AllReduceTest(NcclTestCase): def testAllReduce(self): self._Test(partial(_NcclAllReduce, nccl.all_sum), lambda x, y: x + y) self._Test(partial(_NcclAllReduce, nccl.all_prod), lambda x, y: x * y) self._Test(partial(_NcclAllReduce, nccl.all_min), np.minimum) self._Test(partial(_NcclAllReduce, nccl.all_max), np.maximum) def testAllSumGrad(self): self._TestGradient( partial(_NcclAllReduce, nccl.all_sum), lambda x, y: x + y) def testErrors(self): with self.assertRaisesRegexp(ValueError, 'Device assignment required'): nccl.all_sum([array_ops.identity(np.random.random_sample((3, 4)))]) with self.assertRaisesRegexp(ValueError, 'Must pass >0 tensors'): nccl.all_sum([]) class SingleReduceTest(NcclTestCase): def testSum(self): self._Test(partial(_NcclReduce, nccl.reduce_sum), lambda x, y: x + y) def testSumGrad(self): self._TestGradient(partial(_NcclReduce, nccl.reduce_sum), lambda x, y: x) class BroadcastTest(NcclTestCase): def testBroadcast(self): self._Test(_NcclBroadcast, lambda x, y: x) def testBroadcastSingleDevice(self): # Broadcasts on a single device are removed completely during rewrite. self._Test(_NcclBroadcast, lambda x, y: x, (['/device:GPU:0', '/device:GPU:0'],)) def testBroadcastToCpuError(self): # Broadcasts to CPU is not supported. with self.assertRaisesRegexp( errors.NotFoundError, "No registered '_NcclBroadcastRecv' OpKernel for CPU devices"): self._Test(_NcclBroadcast, lambda x, y: x, (['/device:GPU:0', '/device:CPU:0'],)) class CombinedTest(NcclTestCase): """Test all-reduce vs. single-reduce plus broadcast in one session.run.""" def _Combined(self, tensors, devices): all_reduce_tensors = _NcclAllReduce(nccl.all_sum, tensors, devices) single_reduce_tensors = _NcclReduce(nccl.reduce_sum, tensors, devices) broadcast_tensors = _NcclBroadcast(single_reduce_tensors, devices) return all_reduce_tensors + broadcast_tensors def testCombined(self): self._Test(self._Combined, lambda x, y: x + y) if __name__ == '__main__': test.main()
receiver = np.random.randint(0, len(devices)) with ops.device(devices[receiver]): return [nccl_fun(_DeviceTensors(tensors, devices))]
identifier_body
nccl_ops_test.py
# Copyright 2016 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the 'License'); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an 'AS IS' BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for nccl ops. See also the cc test for nccl_communicator.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from functools import partial import numpy as np from tensorflow.contrib import nccl from tensorflow.python.framework import errors from tensorflow.python.framework import ops from tensorflow.python.ops import array_ops from tensorflow.python.ops import gradients from tensorflow.python.platform import test def
(tensors, devices): res = [] for t, d in zip(tensors, devices): with ops.device(d): res.append(array_ops.identity(t)) return res def _NcclAllReduce(nccl_fun, tensors, devices): return nccl_fun(_DeviceTensors(tensors, devices)) def _NcclReduce(nccl_fun, tensors, devices): receiver = np.random.randint(0, len(devices)) with ops.device(devices[receiver]): return [nccl_fun(_DeviceTensors(tensors, devices))] def _NcclBroadcast(tensors, devices): sender = np.random.randint(0, len(devices)) with ops.device(devices[sender]): tensor = array_ops.identity(tensors[0]) broadcast = nccl.broadcast(tensor) return _DeviceTensors([broadcast] * len(devices), devices) class NcclTestCase(test.TestCase): def _Test(self, nccl_reduce, numpy_fn, device_sets=(['/device:GPU:1', '/device:GPU:2', '/device:GPU:0'], ['/device:GPU:1', '/device:GPU:0'])): """Tests that nccl_reduce does the same as reduction with numpy_fn. Args: nccl_reduce: A function taking a list of tensors and a list of devices, and returns a list of reduced tensors and a list of ops to perform the reduction. numpy_fn: A function taking two tensors and returning the reduction of the two. device_sets: Tuple of virtual devices to run test on. """ if not test.is_gpu_available(): return # Test requires access to a GPU for dtype in [np.float32, np.int32, np.int64, np.float64]: # Create session inside outer loop to test use of # same communicator across multiple sessions. with self.test_session(use_gpu=True) as sess: for devices in device_sets: shape = (3, 4) random = (np.random.random_sample(shape) - .5) * 1024 tensors = [] for _ in devices: tensors.append(random.astype(dtype)) np_ans = tensors[0] for t in tensors[1:]: np_ans = numpy_fn(np_ans, t) reduce_tensors = nccl_reduce(tensors, devices) self.assertNotEmpty(reduce_tensors) # Test shape inference. for r in reduce_tensors: self.assertEqual(shape, r.get_shape()) result_tensors = [array_ops.identity(t) for t in reduce_tensors] # Test execution and results. for t in sess.run(result_tensors): self.assertAllClose(t, np_ans) def _TestGradient(self, nccl_reduce, numpy_fn): """Tests the gradient of nccl_reduce. Args: nccl_reduce: A function taking a list of tensors and a list of devices, and returns a list of reduced tensors and a list of ops to perform the reduction. numpy_fn: A function taking two tensors and returning the gradient of the reduction of the two. """ def _Gradient(tensors, devices): inputs = [array_ops.placeholder(t.dtype, t.shape) for t in tensors] reduce_tensors = nccl_reduce(inputs, devices) losses = _DeviceTensors(tensors, [t.device for t in reduce_tensors]) grads = gradients.gradients( reduce_tensors, inputs, losses, colocate_gradients_with_ops=True) return [g for g in grads if g is not None] self._Test(_Gradient, numpy_fn) class AllReduceTest(NcclTestCase): def testAllReduce(self): self._Test(partial(_NcclAllReduce, nccl.all_sum), lambda x, y: x + y) self._Test(partial(_NcclAllReduce, nccl.all_prod), lambda x, y: x * y) self._Test(partial(_NcclAllReduce, nccl.all_min), np.minimum) self._Test(partial(_NcclAllReduce, nccl.all_max), np.maximum) def testAllSumGrad(self): self._TestGradient( partial(_NcclAllReduce, nccl.all_sum), lambda x, y: x + y) def testErrors(self): with self.assertRaisesRegexp(ValueError, 'Device assignment required'): nccl.all_sum([array_ops.identity(np.random.random_sample((3, 4)))]) with self.assertRaisesRegexp(ValueError, 'Must pass >0 tensors'): nccl.all_sum([]) class SingleReduceTest(NcclTestCase): def testSum(self): self._Test(partial(_NcclReduce, nccl.reduce_sum), lambda x, y: x + y) def testSumGrad(self): self._TestGradient(partial(_NcclReduce, nccl.reduce_sum), lambda x, y: x) class BroadcastTest(NcclTestCase): def testBroadcast(self): self._Test(_NcclBroadcast, lambda x, y: x) def testBroadcastSingleDevice(self): # Broadcasts on a single device are removed completely during rewrite. self._Test(_NcclBroadcast, lambda x, y: x, (['/device:GPU:0', '/device:GPU:0'],)) def testBroadcastToCpuError(self): # Broadcasts to CPU is not supported. with self.assertRaisesRegexp( errors.NotFoundError, "No registered '_NcclBroadcastRecv' OpKernel for CPU devices"): self._Test(_NcclBroadcast, lambda x, y: x, (['/device:GPU:0', '/device:CPU:0'],)) class CombinedTest(NcclTestCase): """Test all-reduce vs. single-reduce plus broadcast in one session.run.""" def _Combined(self, tensors, devices): all_reduce_tensors = _NcclAllReduce(nccl.all_sum, tensors, devices) single_reduce_tensors = _NcclReduce(nccl.reduce_sum, tensors, devices) broadcast_tensors = _NcclBroadcast(single_reduce_tensors, devices) return all_reduce_tensors + broadcast_tensors def testCombined(self): self._Test(self._Combined, lambda x, y: x + y) if __name__ == '__main__': test.main()
_DeviceTensors
identifier_name
nccl_ops_test.py
# Copyright 2016 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the 'License'); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an 'AS IS' BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for nccl ops. See also the cc test for nccl_communicator.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from functools import partial import numpy as np from tensorflow.contrib import nccl from tensorflow.python.framework import errors from tensorflow.python.framework import ops from tensorflow.python.ops import array_ops from tensorflow.python.ops import gradients from tensorflow.python.platform import test def _DeviceTensors(tensors, devices): res = [] for t, d in zip(tensors, devices): with ops.device(d): res.append(array_ops.identity(t)) return res def _NcclAllReduce(nccl_fun, tensors, devices): return nccl_fun(_DeviceTensors(tensors, devices)) def _NcclReduce(nccl_fun, tensors, devices): receiver = np.random.randint(0, len(devices)) with ops.device(devices[receiver]): return [nccl_fun(_DeviceTensors(tensors, devices))] def _NcclBroadcast(tensors, devices): sender = np.random.randint(0, len(devices)) with ops.device(devices[sender]): tensor = array_ops.identity(tensors[0]) broadcast = nccl.broadcast(tensor) return _DeviceTensors([broadcast] * len(devices), devices) class NcclTestCase(test.TestCase): def _Test(self, nccl_reduce, numpy_fn, device_sets=(['/device:GPU:1', '/device:GPU:2', '/device:GPU:0'], ['/device:GPU:1', '/device:GPU:0'])): """Tests that nccl_reduce does the same as reduction with numpy_fn. Args: nccl_reduce: A function taking a list of tensors and a list of devices, and returns a list of reduced tensors and a list of ops to perform the reduction. numpy_fn: A function taking two tensors and returning the reduction of the two. device_sets: Tuple of virtual devices to run test on. """ if not test.is_gpu_available(): return # Test requires access to a GPU for dtype in [np.float32, np.int32, np.int64, np.float64]: # Create session inside outer loop to test use of # same communicator across multiple sessions. with self.test_session(use_gpu=True) as sess: for devices in device_sets: shape = (3, 4) random = (np.random.random_sample(shape) - .5) * 1024 tensors = [] for _ in devices:
np_ans = tensors[0] for t in tensors[1:]: np_ans = numpy_fn(np_ans, t) reduce_tensors = nccl_reduce(tensors, devices) self.assertNotEmpty(reduce_tensors) # Test shape inference. for r in reduce_tensors: self.assertEqual(shape, r.get_shape()) result_tensors = [array_ops.identity(t) for t in reduce_tensors] # Test execution and results. for t in sess.run(result_tensors): self.assertAllClose(t, np_ans) def _TestGradient(self, nccl_reduce, numpy_fn): """Tests the gradient of nccl_reduce. Args: nccl_reduce: A function taking a list of tensors and a list of devices, and returns a list of reduced tensors and a list of ops to perform the reduction. numpy_fn: A function taking two tensors and returning the gradient of the reduction of the two. """ def _Gradient(tensors, devices): inputs = [array_ops.placeholder(t.dtype, t.shape) for t in tensors] reduce_tensors = nccl_reduce(inputs, devices) losses = _DeviceTensors(tensors, [t.device for t in reduce_tensors]) grads = gradients.gradients( reduce_tensors, inputs, losses, colocate_gradients_with_ops=True) return [g for g in grads if g is not None] self._Test(_Gradient, numpy_fn) class AllReduceTest(NcclTestCase): def testAllReduce(self): self._Test(partial(_NcclAllReduce, nccl.all_sum), lambda x, y: x + y) self._Test(partial(_NcclAllReduce, nccl.all_prod), lambda x, y: x * y) self._Test(partial(_NcclAllReduce, nccl.all_min), np.minimum) self._Test(partial(_NcclAllReduce, nccl.all_max), np.maximum) def testAllSumGrad(self): self._TestGradient( partial(_NcclAllReduce, nccl.all_sum), lambda x, y: x + y) def testErrors(self): with self.assertRaisesRegexp(ValueError, 'Device assignment required'): nccl.all_sum([array_ops.identity(np.random.random_sample((3, 4)))]) with self.assertRaisesRegexp(ValueError, 'Must pass >0 tensors'): nccl.all_sum([]) class SingleReduceTest(NcclTestCase): def testSum(self): self._Test(partial(_NcclReduce, nccl.reduce_sum), lambda x, y: x + y) def testSumGrad(self): self._TestGradient(partial(_NcclReduce, nccl.reduce_sum), lambda x, y: x) class BroadcastTest(NcclTestCase): def testBroadcast(self): self._Test(_NcclBroadcast, lambda x, y: x) def testBroadcastSingleDevice(self): # Broadcasts on a single device are removed completely during rewrite. self._Test(_NcclBroadcast, lambda x, y: x, (['/device:GPU:0', '/device:GPU:0'],)) def testBroadcastToCpuError(self): # Broadcasts to CPU is not supported. with self.assertRaisesRegexp( errors.NotFoundError, "No registered '_NcclBroadcastRecv' OpKernel for CPU devices"): self._Test(_NcclBroadcast, lambda x, y: x, (['/device:GPU:0', '/device:CPU:0'],)) class CombinedTest(NcclTestCase): """Test all-reduce vs. single-reduce plus broadcast in one session.run.""" def _Combined(self, tensors, devices): all_reduce_tensors = _NcclAllReduce(nccl.all_sum, tensors, devices) single_reduce_tensors = _NcclReduce(nccl.reduce_sum, tensors, devices) broadcast_tensors = _NcclBroadcast(single_reduce_tensors, devices) return all_reduce_tensors + broadcast_tensors def testCombined(self): self._Test(self._Combined, lambda x, y: x + y) if __name__ == '__main__': test.main()
tensors.append(random.astype(dtype))
conditional_block
nccl_ops_test.py
# Copyright 2016 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the 'License'); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an 'AS IS' BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for nccl ops. See also the cc test for nccl_communicator.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from functools import partial import numpy as np from tensorflow.contrib import nccl from tensorflow.python.framework import errors from tensorflow.python.framework import ops from tensorflow.python.ops import array_ops from tensorflow.python.ops import gradients from tensorflow.python.platform import test def _DeviceTensors(tensors, devices): res = [] for t, d in zip(tensors, devices): with ops.device(d): res.append(array_ops.identity(t)) return res def _NcclAllReduce(nccl_fun, tensors, devices): return nccl_fun(_DeviceTensors(tensors, devices)) def _NcclReduce(nccl_fun, tensors, devices): receiver = np.random.randint(0, len(devices)) with ops.device(devices[receiver]): return [nccl_fun(_DeviceTensors(tensors, devices))] def _NcclBroadcast(tensors, devices): sender = np.random.randint(0, len(devices)) with ops.device(devices[sender]): tensor = array_ops.identity(tensors[0]) broadcast = nccl.broadcast(tensor) return _DeviceTensors([broadcast] * len(devices), devices) class NcclTestCase(test.TestCase): def _Test(self, nccl_reduce, numpy_fn, device_sets=(['/device:GPU:1', '/device:GPU:2', '/device:GPU:0'], ['/device:GPU:1', '/device:GPU:0'])): """Tests that nccl_reduce does the same as reduction with numpy_fn. Args: nccl_reduce: A function taking a list of tensors and a list of devices, and returns a list of reduced tensors and a list of ops to perform the reduction. numpy_fn: A function taking two tensors and returning the reduction of the two. device_sets: Tuple of virtual devices to run test on. """ if not test.is_gpu_available(): return # Test requires access to a GPU for dtype in [np.float32, np.int32, np.int64, np.float64]: # Create session inside outer loop to test use of # same communicator across multiple sessions.
for devices in device_sets: shape = (3, 4) random = (np.random.random_sample(shape) - .5) * 1024 tensors = [] for _ in devices: tensors.append(random.astype(dtype)) np_ans = tensors[0] for t in tensors[1:]: np_ans = numpy_fn(np_ans, t) reduce_tensors = nccl_reduce(tensors, devices) self.assertNotEmpty(reduce_tensors) # Test shape inference. for r in reduce_tensors: self.assertEqual(shape, r.get_shape()) result_tensors = [array_ops.identity(t) for t in reduce_tensors] # Test execution and results. for t in sess.run(result_tensors): self.assertAllClose(t, np_ans) def _TestGradient(self, nccl_reduce, numpy_fn): """Tests the gradient of nccl_reduce. Args: nccl_reduce: A function taking a list of tensors and a list of devices, and returns a list of reduced tensors and a list of ops to perform the reduction. numpy_fn: A function taking two tensors and returning the gradient of the reduction of the two. """ def _Gradient(tensors, devices): inputs = [array_ops.placeholder(t.dtype, t.shape) for t in tensors] reduce_tensors = nccl_reduce(inputs, devices) losses = _DeviceTensors(tensors, [t.device for t in reduce_tensors]) grads = gradients.gradients( reduce_tensors, inputs, losses, colocate_gradients_with_ops=True) return [g for g in grads if g is not None] self._Test(_Gradient, numpy_fn) class AllReduceTest(NcclTestCase): def testAllReduce(self): self._Test(partial(_NcclAllReduce, nccl.all_sum), lambda x, y: x + y) self._Test(partial(_NcclAllReduce, nccl.all_prod), lambda x, y: x * y) self._Test(partial(_NcclAllReduce, nccl.all_min), np.minimum) self._Test(partial(_NcclAllReduce, nccl.all_max), np.maximum) def testAllSumGrad(self): self._TestGradient( partial(_NcclAllReduce, nccl.all_sum), lambda x, y: x + y) def testErrors(self): with self.assertRaisesRegexp(ValueError, 'Device assignment required'): nccl.all_sum([array_ops.identity(np.random.random_sample((3, 4)))]) with self.assertRaisesRegexp(ValueError, 'Must pass >0 tensors'): nccl.all_sum([]) class SingleReduceTest(NcclTestCase): def testSum(self): self._Test(partial(_NcclReduce, nccl.reduce_sum), lambda x, y: x + y) def testSumGrad(self): self._TestGradient(partial(_NcclReduce, nccl.reduce_sum), lambda x, y: x) class BroadcastTest(NcclTestCase): def testBroadcast(self): self._Test(_NcclBroadcast, lambda x, y: x) def testBroadcastSingleDevice(self): # Broadcasts on a single device are removed completely during rewrite. self._Test(_NcclBroadcast, lambda x, y: x, (['/device:GPU:0', '/device:GPU:0'],)) def testBroadcastToCpuError(self): # Broadcasts to CPU is not supported. with self.assertRaisesRegexp( errors.NotFoundError, "No registered '_NcclBroadcastRecv' OpKernel for CPU devices"): self._Test(_NcclBroadcast, lambda x, y: x, (['/device:GPU:0', '/device:CPU:0'],)) class CombinedTest(NcclTestCase): """Test all-reduce vs. single-reduce plus broadcast in one session.run.""" def _Combined(self, tensors, devices): all_reduce_tensors = _NcclAllReduce(nccl.all_sum, tensors, devices) single_reduce_tensors = _NcclReduce(nccl.reduce_sum, tensors, devices) broadcast_tensors = _NcclBroadcast(single_reduce_tensors, devices) return all_reduce_tensors + broadcast_tensors def testCombined(self): self._Test(self._Combined, lambda x, y: x + y) if __name__ == '__main__': test.main()
with self.test_session(use_gpu=True) as sess:
random_line_split
runner.rs
use regex::Regex; use state::Cucumber; use event::request::{InvokeArgument, Request}; use event::response::{InvokeResponse, Response, StepMatchesResponse}; use definitions::registration::{CucumberRegistrar, SimpleStep}; use std::panic::{self, AssertUnwindSafe}; use std::str::FromStr; /// The step runner for [Cucumber state](../state/struct.Cucumber.html) /// /// The runner stands in for the Cucumber instance and provides an interface for /// [Request](../event/request/enum.Request.html) events to be translated into /// state changes and /// step invocations, along with a /// [Response](../event/response/enum.Response.html). These are typically /// supplied by a running /// [Server](../server/struct.Server.html), but may be supplied by a native /// Gherkin implementation /// later. /// /// Typically this struct will only be instantiated by the user, and then /// passed to a Server to /// maintain. /// #[allow(dead_code)] pub struct WorldRunner<World> { cuke: Cucumber<World>, world: World, } impl<World> WorldRunner<World> { #[allow(dead_code)] pub fn new(world: World) -> WorldRunner<World> { WorldRunner { cuke: Cucumber::new(), world: world, } } } /// An interface for implementers that can consume a /// [Request](../event/request/enum.Request.html) and yield a /// [Response](../event/response/enum.Response.html) /// /// This generally refers to [WorldRunner](./struct.WorldRunner.html) pub trait CommandRunner { fn execute_cmd(&mut self, req: Request) -> Response; } impl<T: Fn(Request) -> Response> CommandRunner for T { fn execute_cmd(&mut self, req: Request) -> Response { self(req) } } impl<World> CommandRunner for WorldRunner<World> { fn execute_cmd(&mut self, req: Request) -> Response { match req { Request::BeginScenario(params) => { self.cuke.tags = params.tags; Response::BeginScenario }, Request::Invoke(params) => { let step = self.cuke .step(u32::from_str(&params.id).unwrap()) .unwrap(); Response::Invoke(invoke_to_response(step, &self.cuke, &mut self.world, params.args)) }, Request::StepMatches(params) => { let matches = self.cuke.find_match(&params.name_to_match); if matches.len() == 0 { Response::StepMatches(StepMatchesResponse::NoMatch) } else { Response::StepMatches(StepMatchesResponse::Match(matches)) } }, Request::EndScenario(_) => { self.cuke.tags = Vec::new(); Response::EndScenario }, // TODO: For some reason, cucumber prints the ruby snippet too. Fix that Request::SnippetText(params) => { let text = format!(" // In a step registration block where cuke: &mut \ CucumberRegistrar<YourWorld>\n use cucumber::InvokeResponse;\n use \ cucumber::helpers::r;\n {}!(cuke, r(\"^{}$\"), Box::new(move |c, _, \ _| {{\n c.pending(\"TODO\")\n }}));", params.step_keyword, params.step_name); Response::SnippetText(text) }, } } } impl<World> CucumberRegistrar<World> for WorldRunner<World> { fn given(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>)
fn when(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>) { self.cuke.when(file, line, regex, step) } fn then(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>) { self.cuke.then(file, line, regex, step) } } pub fn invoke_to_response<World>(test_body: &SimpleStep<World>, cuke: &Cucumber<World>, world: &mut World, args: Vec<InvokeArgument>) -> InvokeResponse { let result = panic::catch_unwind(AssertUnwindSafe(|| test_body(cuke, world, args))); match result { Ok(()) => InvokeResponse::Success, Err(err) => { // Yoinked from rustc libstd, with InvokeResponse added as a possible cast let msg = match err.downcast_ref::<&'static str>() { Some(s) => *s, None => { match err.downcast_ref::<String>() { Some(s) => &s[..], None => { match err.downcast_ref::<InvokeResponse>() { Some(s) => return s.clone(), None => "Box<Any>", } }, } }, }; InvokeResponse::fail_from_str(msg) }, } }
{ self.cuke.given(file, line, regex, step) }
identifier_body
runner.rs
use regex::Regex; use state::Cucumber; use event::request::{InvokeArgument, Request}; use event::response::{InvokeResponse, Response, StepMatchesResponse}; use definitions::registration::{CucumberRegistrar, SimpleStep}; use std::panic::{self, AssertUnwindSafe}; use std::str::FromStr; /// The step runner for [Cucumber state](../state/struct.Cucumber.html) /// /// The runner stands in for the Cucumber instance and provides an interface for /// [Request](../event/request/enum.Request.html) events to be translated into /// state changes and /// step invocations, along with a /// [Response](../event/response/enum.Response.html). These are typically /// supplied by a running /// [Server](../server/struct.Server.html), but may be supplied by a native /// Gherkin implementation /// later. /// /// Typically this struct will only be instantiated by the user, and then /// passed to a Server to /// maintain. /// #[allow(dead_code)] pub struct WorldRunner<World> { cuke: Cucumber<World>, world: World, } impl<World> WorldRunner<World> { #[allow(dead_code)] pub fn new(world: World) -> WorldRunner<World> { WorldRunner { cuke: Cucumber::new(), world: world, } } } /// An interface for implementers that can consume a /// [Request](../event/request/enum.Request.html) and yield a /// [Response](../event/response/enum.Response.html) /// /// This generally refers to [WorldRunner](./struct.WorldRunner.html) pub trait CommandRunner { fn execute_cmd(&mut self, req: Request) -> Response; } impl<T: Fn(Request) -> Response> CommandRunner for T { fn execute_cmd(&mut self, req: Request) -> Response { self(req) } } impl<World> CommandRunner for WorldRunner<World> { fn execute_cmd(&mut self, req: Request) -> Response { match req { Request::BeginScenario(params) => { self.cuke.tags = params.tags; Response::BeginScenario }, Request::Invoke(params) => { let step = self.cuke .step(u32::from_str(&params.id).unwrap()) .unwrap(); Response::Invoke(invoke_to_response(step, &self.cuke, &mut self.world, params.args)) }, Request::StepMatches(params) => { let matches = self.cuke.find_match(&params.name_to_match); if matches.len() == 0 { Response::StepMatches(StepMatchesResponse::NoMatch) } else { Response::StepMatches(StepMatchesResponse::Match(matches)) } }, Request::EndScenario(_) => { self.cuke.tags = Vec::new(); Response::EndScenario }, // TODO: For some reason, cucumber prints the ruby snippet too. Fix that Request::SnippetText(params) => { let text = format!(" // In a step registration block where cuke: &mut \ CucumberRegistrar<YourWorld>\n use cucumber::InvokeResponse;\n use \ cucumber::helpers::r;\n {}!(cuke, r(\"^{}$\"), Box::new(move |c, _, \ _| {{\n c.pending(\"TODO\")\n }}));", params.step_keyword, params.step_name); Response::SnippetText(text) }, } } } impl<World> CucumberRegistrar<World> for WorldRunner<World> { fn given(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>) { self.cuke.given(file, line, regex, step) } fn when(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>) { self.cuke.when(file, line, regex, step) } fn then(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>) { self.cuke.then(file, line, regex, step) } } pub fn invoke_to_response<World>(test_body: &SimpleStep<World>, cuke: &Cucumber<World>, world: &mut World, args: Vec<InvokeArgument>) -> InvokeResponse { let result = panic::catch_unwind(AssertUnwindSafe(|| test_body(cuke, world, args))); match result { Ok(()) => InvokeResponse::Success, Err(err) => { // Yoinked from rustc libstd, with InvokeResponse added as a possible cast let msg = match err.downcast_ref::<&'static str>() { Some(s) => *s, None => { match err.downcast_ref::<String>() { Some(s) => &s[..], None => { match err.downcast_ref::<InvokeResponse>() { Some(s) => return s.clone(), None => "Box<Any>", } },
} }, }; InvokeResponse::fail_from_str(msg) }, } }
random_line_split
runner.rs
use regex::Regex; use state::Cucumber; use event::request::{InvokeArgument, Request}; use event::response::{InvokeResponse, Response, StepMatchesResponse}; use definitions::registration::{CucumberRegistrar, SimpleStep}; use std::panic::{self, AssertUnwindSafe}; use std::str::FromStr; /// The step runner for [Cucumber state](../state/struct.Cucumber.html) /// /// The runner stands in for the Cucumber instance and provides an interface for /// [Request](../event/request/enum.Request.html) events to be translated into /// state changes and /// step invocations, along with a /// [Response](../event/response/enum.Response.html). These are typically /// supplied by a running /// [Server](../server/struct.Server.html), but may be supplied by a native /// Gherkin implementation /// later. /// /// Typically this struct will only be instantiated by the user, and then /// passed to a Server to /// maintain. /// #[allow(dead_code)] pub struct WorldRunner<World> { cuke: Cucumber<World>, world: World, } impl<World> WorldRunner<World> { #[allow(dead_code)] pub fn new(world: World) -> WorldRunner<World> { WorldRunner { cuke: Cucumber::new(), world: world, } } } /// An interface for implementers that can consume a /// [Request](../event/request/enum.Request.html) and yield a /// [Response](../event/response/enum.Response.html) /// /// This generally refers to [WorldRunner](./struct.WorldRunner.html) pub trait CommandRunner { fn execute_cmd(&mut self, req: Request) -> Response; } impl<T: Fn(Request) -> Response> CommandRunner for T { fn execute_cmd(&mut self, req: Request) -> Response { self(req) } } impl<World> CommandRunner for WorldRunner<World> { fn execute_cmd(&mut self, req: Request) -> Response { match req { Request::BeginScenario(params) => { self.cuke.tags = params.tags; Response::BeginScenario }, Request::Invoke(params) => { let step = self.cuke .step(u32::from_str(&params.id).unwrap()) .unwrap(); Response::Invoke(invoke_to_response(step, &self.cuke, &mut self.world, params.args)) }, Request::StepMatches(params) =>
, Request::EndScenario(_) => { self.cuke.tags = Vec::new(); Response::EndScenario }, // TODO: For some reason, cucumber prints the ruby snippet too. Fix that Request::SnippetText(params) => { let text = format!(" // In a step registration block where cuke: &mut \ CucumberRegistrar<YourWorld>\n use cucumber::InvokeResponse;\n use \ cucumber::helpers::r;\n {}!(cuke, r(\"^{}$\"), Box::new(move |c, _, \ _| {{\n c.pending(\"TODO\")\n }}));", params.step_keyword, params.step_name); Response::SnippetText(text) }, } } } impl<World> CucumberRegistrar<World> for WorldRunner<World> { fn given(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>) { self.cuke.given(file, line, regex, step) } fn when(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>) { self.cuke.when(file, line, regex, step) } fn then(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>) { self.cuke.then(file, line, regex, step) } } pub fn invoke_to_response<World>(test_body: &SimpleStep<World>, cuke: &Cucumber<World>, world: &mut World, args: Vec<InvokeArgument>) -> InvokeResponse { let result = panic::catch_unwind(AssertUnwindSafe(|| test_body(cuke, world, args))); match result { Ok(()) => InvokeResponse::Success, Err(err) => { // Yoinked from rustc libstd, with InvokeResponse added as a possible cast let msg = match err.downcast_ref::<&'static str>() { Some(s) => *s, None => { match err.downcast_ref::<String>() { Some(s) => &s[..], None => { match err.downcast_ref::<InvokeResponse>() { Some(s) => return s.clone(), None => "Box<Any>", } }, } }, }; InvokeResponse::fail_from_str(msg) }, } }
{ let matches = self.cuke.find_match(&params.name_to_match); if matches.len() == 0 { Response::StepMatches(StepMatchesResponse::NoMatch) } else { Response::StepMatches(StepMatchesResponse::Match(matches)) } }
conditional_block
runner.rs
use regex::Regex; use state::Cucumber; use event::request::{InvokeArgument, Request}; use event::response::{InvokeResponse, Response, StepMatchesResponse}; use definitions::registration::{CucumberRegistrar, SimpleStep}; use std::panic::{self, AssertUnwindSafe}; use std::str::FromStr; /// The step runner for [Cucumber state](../state/struct.Cucumber.html) /// /// The runner stands in for the Cucumber instance and provides an interface for /// [Request](../event/request/enum.Request.html) events to be translated into /// state changes and /// step invocations, along with a /// [Response](../event/response/enum.Response.html). These are typically /// supplied by a running /// [Server](../server/struct.Server.html), but may be supplied by a native /// Gherkin implementation /// later. /// /// Typically this struct will only be instantiated by the user, and then /// passed to a Server to /// maintain. /// #[allow(dead_code)] pub struct WorldRunner<World> { cuke: Cucumber<World>, world: World, } impl<World> WorldRunner<World> { #[allow(dead_code)] pub fn new(world: World) -> WorldRunner<World> { WorldRunner { cuke: Cucumber::new(), world: world, } } } /// An interface for implementers that can consume a /// [Request](../event/request/enum.Request.html) and yield a /// [Response](../event/response/enum.Response.html) /// /// This generally refers to [WorldRunner](./struct.WorldRunner.html) pub trait CommandRunner { fn execute_cmd(&mut self, req: Request) -> Response; } impl<T: Fn(Request) -> Response> CommandRunner for T { fn execute_cmd(&mut self, req: Request) -> Response { self(req) } } impl<World> CommandRunner for WorldRunner<World> { fn execute_cmd(&mut self, req: Request) -> Response { match req { Request::BeginScenario(params) => { self.cuke.tags = params.tags; Response::BeginScenario }, Request::Invoke(params) => { let step = self.cuke .step(u32::from_str(&params.id).unwrap()) .unwrap(); Response::Invoke(invoke_to_response(step, &self.cuke, &mut self.world, params.args)) }, Request::StepMatches(params) => { let matches = self.cuke.find_match(&params.name_to_match); if matches.len() == 0 { Response::StepMatches(StepMatchesResponse::NoMatch) } else { Response::StepMatches(StepMatchesResponse::Match(matches)) } }, Request::EndScenario(_) => { self.cuke.tags = Vec::new(); Response::EndScenario }, // TODO: For some reason, cucumber prints the ruby snippet too. Fix that Request::SnippetText(params) => { let text = format!(" // In a step registration block where cuke: &mut \ CucumberRegistrar<YourWorld>\n use cucumber::InvokeResponse;\n use \ cucumber::helpers::r;\n {}!(cuke, r(\"^{}$\"), Box::new(move |c, _, \ _| {{\n c.pending(\"TODO\")\n }}));", params.step_keyword, params.step_name); Response::SnippetText(text) }, } } } impl<World> CucumberRegistrar<World> for WorldRunner<World> { fn given(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>) { self.cuke.given(file, line, regex, step) } fn
(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>) { self.cuke.when(file, line, regex, step) } fn then(&mut self, file: &str, line: u32, regex: Regex, step: SimpleStep<World>) { self.cuke.then(file, line, regex, step) } } pub fn invoke_to_response<World>(test_body: &SimpleStep<World>, cuke: &Cucumber<World>, world: &mut World, args: Vec<InvokeArgument>) -> InvokeResponse { let result = panic::catch_unwind(AssertUnwindSafe(|| test_body(cuke, world, args))); match result { Ok(()) => InvokeResponse::Success, Err(err) => { // Yoinked from rustc libstd, with InvokeResponse added as a possible cast let msg = match err.downcast_ref::<&'static str>() { Some(s) => *s, None => { match err.downcast_ref::<String>() { Some(s) => &s[..], None => { match err.downcast_ref::<InvokeResponse>() { Some(s) => return s.clone(), None => "Box<Any>", } }, } }, }; InvokeResponse::fail_from_str(msg) }, } }
when
identifier_name
counting.py
# python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """A simple, hierarchical distributed counter.""" import threading import time from typing import Dict, Mapping, Optional, Union from acme import core Number = Union[int, float] class Counter(core.Saveable): """A simple counter object that can periodically sync with a parent.""" def __init__(self, parent: Optional['Counter'] = None, prefix: str = '', time_delta: float = 1.0, return_only_prefixed: bool = False): """Initialize the counter. Args: parent: a Counter object to cache locally (or None for no caching). prefix: string prefix to use for all local counts. time_delta: time difference in seconds between syncing with the parent counter. return_only_prefixed: if True, and if `prefix` isn't empty, return counts restricted to the given `prefix` on each call to `increment` and `get_counts`. The `prefix` is stripped from returned count names. """ self._parent = parent self._prefix = prefix self._time_delta = time_delta # Hold local counts and we'll lock around that. # These are counts to be synced to the parent and the cache. self._counts = {} self._lock = threading.Lock() # We'll sync periodically (when the last sync was more than self._time_delta # seconds ago.) self._cache = {} self._last_sync_time = 0.0 self._return_only_prefixed = return_only_prefixed def increment(self, **counts: Number) -> Dict[str, Number]: """Increment a set of counters. Args: **counts: keyword arguments specifying count increments. Returns: The [name, value] mapping of all counters stored, i.e. this will also include counts that were not updated by this call to increment. """ with self._lock: for key, value in counts.items(): self._counts.setdefault(key, 0) self._counts[key] += value return self.get_counts() def get_counts(self) -> Dict[str, Number]: """Return all counts tracked by this counter.""" now = time.time() # TODO(b/144421838): use futures instead of blocking. if self._parent and (now - self._last_sync_time) > self._time_delta: with self._lock: counts = _prefix_keys(self._counts, self._prefix) # Reset the local counts, as they will be merged into the parent and the # cache. self._counts = {} self._cache = self._parent.increment(**counts) self._last_sync_time = now # Potentially prefix the keys in the counts dictionary. counts = _prefix_keys(self._counts, self._prefix) # If there's no prefix make a copy of the dictionary so we don't modify the # internal self._counts. if not self._prefix: counts = dict(counts) # Combine local counts with any parent counts. for key, value in self._cache.items():
if self._prefix and self._return_only_prefixed: counts = dict([(key[len(self._prefix) + 1:], value) for key, value in counts.items() if key.startswith(f'{self._prefix}_')]) return counts def save(self) -> Mapping[str, Mapping[str, Number]]: return {'counts': self._counts, 'cache': self._cache} def restore(self, state: Mapping[str, Mapping[str, Number]]): # Force a sync, if necessary, on the next get_counts call. self._last_sync_time = 0. self._counts = state['counts'] self._cache = state['cache'] def _prefix_keys(dictionary: Dict[str, Number], prefix: str): """Return a dictionary with prefixed keys. Args: dictionary: dictionary to return a copy of. prefix: string to use as the prefix. Returns: Return a copy of the given dictionary whose keys are replaced by "{prefix}_{key}". If the prefix is the empty string it returns the given dictionary unchanged. """ if prefix: dictionary = {f'{prefix}_{k}': v for k, v in dictionary.items()} return dictionary
counts[key] = counts.get(key, 0) + value
conditional_block
counting.py
# python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """A simple, hierarchical distributed counter.""" import threading import time from typing import Dict, Mapping, Optional, Union from acme import core Number = Union[int, float] class Counter(core.Saveable): """A simple counter object that can periodically sync with a parent.""" def __init__(self, parent: Optional['Counter'] = None, prefix: str = '', time_delta: float = 1.0, return_only_prefixed: bool = False): """Initialize the counter. Args: parent: a Counter object to cache locally (or None for no caching). prefix: string prefix to use for all local counts. time_delta: time difference in seconds between syncing with the parent counter. return_only_prefixed: if True, and if `prefix` isn't empty, return counts restricted to the given `prefix` on each call to `increment` and `get_counts`. The `prefix` is stripped from returned count names. """ self._parent = parent self._prefix = prefix self._time_delta = time_delta # Hold local counts and we'll lock around that. # These are counts to be synced to the parent and the cache. self._counts = {} self._lock = threading.Lock() # We'll sync periodically (when the last sync was more than self._time_delta # seconds ago.) self._cache = {} self._last_sync_time = 0.0 self._return_only_prefixed = return_only_prefixed def increment(self, **counts: Number) -> Dict[str, Number]: """Increment a set of counters. Args: **counts: keyword arguments specifying count increments. Returns: The [name, value] mapping of all counters stored, i.e. this will also include counts that were not updated by this call to increment. """ with self._lock: for key, value in counts.items(): self._counts.setdefault(key, 0) self._counts[key] += value return self.get_counts() def get_counts(self) -> Dict[str, Number]: """Return all counts tracked by this counter.""" now = time.time() # TODO(b/144421838): use futures instead of blocking. if self._parent and (now - self._last_sync_time) > self._time_delta: with self._lock: counts = _prefix_keys(self._counts, self._prefix) # Reset the local counts, as they will be merged into the parent and the # cache. self._counts = {} self._cache = self._parent.increment(**counts) self._last_sync_time = now # Potentially prefix the keys in the counts dictionary. counts = _prefix_keys(self._counts, self._prefix) # If there's no prefix make a copy of the dictionary so we don't modify the # internal self._counts. if not self._prefix: counts = dict(counts) # Combine local counts with any parent counts. for key, value in self._cache.items(): counts[key] = counts.get(key, 0) + value if self._prefix and self._return_only_prefixed: counts = dict([(key[len(self._prefix) + 1:], value) for key, value in counts.items() if key.startswith(f'{self._prefix}_')]) return counts def save(self) -> Mapping[str, Mapping[str, Number]]: return {'counts': self._counts, 'cache': self._cache} def restore(self, state: Mapping[str, Mapping[str, Number]]): # Force a sync, if necessary, on the next get_counts call. self._last_sync_time = 0. self._counts = state['counts'] self._cache = state['cache'] def _prefix_keys(dictionary: Dict[str, Number], prefix: str):
"""Return a dictionary with prefixed keys. Args: dictionary: dictionary to return a copy of. prefix: string to use as the prefix. Returns: Return a copy of the given dictionary whose keys are replaced by "{prefix}_{key}". If the prefix is the empty string it returns the given dictionary unchanged. """ if prefix: dictionary = {f'{prefix}_{k}': v for k, v in dictionary.items()} return dictionary
identifier_body
counting.py
# python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """A simple, hierarchical distributed counter.""" import threading import time from typing import Dict, Mapping, Optional, Union from acme import core Number = Union[int, float] class Counter(core.Saveable): """A simple counter object that can periodically sync with a parent.""" def __init__(self, parent: Optional['Counter'] = None, prefix: str = '', time_delta: float = 1.0, return_only_prefixed: bool = False): """Initialize the counter. Args: parent: a Counter object to cache locally (or None for no caching). prefix: string prefix to use for all local counts. time_delta: time difference in seconds between syncing with the parent counter. return_only_prefixed: if True, and if `prefix` isn't empty, return counts restricted to the given `prefix` on each call to `increment` and `get_counts`. The `prefix` is stripped from returned count names. """ self._parent = parent self._prefix = prefix self._time_delta = time_delta # Hold local counts and we'll lock around that. # These are counts to be synced to the parent and the cache. self._counts = {} self._lock = threading.Lock() # We'll sync periodically (when the last sync was more than self._time_delta # seconds ago.) self._cache = {} self._last_sync_time = 0.0 self._return_only_prefixed = return_only_prefixed def increment(self, **counts: Number) -> Dict[str, Number]: """Increment a set of counters. Args: **counts: keyword arguments specifying count increments. Returns: The [name, value] mapping of all counters stored, i.e. this will also include counts that were not updated by this call to increment. """ with self._lock: for key, value in counts.items(): self._counts.setdefault(key, 0) self._counts[key] += value return self.get_counts() def get_counts(self) -> Dict[str, Number]: """Return all counts tracked by this counter.""" now = time.time() # TODO(b/144421838): use futures instead of blocking. if self._parent and (now - self._last_sync_time) > self._time_delta: with self._lock: counts = _prefix_keys(self._counts, self._prefix) # Reset the local counts, as they will be merged into the parent and the # cache. self._counts = {} self._cache = self._parent.increment(**counts) self._last_sync_time = now # Potentially prefix the keys in the counts dictionary. counts = _prefix_keys(self._counts, self._prefix) # If there's no prefix make a copy of the dictionary so we don't modify the # internal self._counts. if not self._prefix: counts = dict(counts) # Combine local counts with any parent counts. for key, value in self._cache.items(): counts[key] = counts.get(key, 0) + value if self._prefix and self._return_only_prefixed: counts = dict([(key[len(self._prefix) + 1:], value) for key, value in counts.items() if key.startswith(f'{self._prefix}_')]) return counts def save(self) -> Mapping[str, Mapping[str, Number]]: return {'counts': self._counts, 'cache': self._cache} def restore(self, state: Mapping[str, Mapping[str, Number]]): # Force a sync, if necessary, on the next get_counts call. self._last_sync_time = 0. self._counts = state['counts'] self._cache = state['cache'] def
(dictionary: Dict[str, Number], prefix: str): """Return a dictionary with prefixed keys. Args: dictionary: dictionary to return a copy of. prefix: string to use as the prefix. Returns: Return a copy of the given dictionary whose keys are replaced by "{prefix}_{key}". If the prefix is the empty string it returns the given dictionary unchanged. """ if prefix: dictionary = {f'{prefix}_{k}': v for k, v in dictionary.items()} return dictionary
_prefix_keys
identifier_name
counting.py
# python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """A simple, hierarchical distributed counter.""" import threading import time from typing import Dict, Mapping, Optional, Union from acme import core Number = Union[int, float] class Counter(core.Saveable): """A simple counter object that can periodically sync with a parent.""" def __init__(self, parent: Optional['Counter'] = None, prefix: str = '', time_delta: float = 1.0, return_only_prefixed: bool = False): """Initialize the counter. Args: parent: a Counter object to cache locally (or None for no caching). prefix: string prefix to use for all local counts. time_delta: time difference in seconds between syncing with the parent counter. return_only_prefixed: if True, and if `prefix` isn't empty, return counts restricted to the given `prefix` on each call to `increment` and `get_counts`. The `prefix` is stripped from returned count names. """ self._parent = parent self._prefix = prefix self._time_delta = time_delta # Hold local counts and we'll lock around that. # These are counts to be synced to the parent and the cache. self._counts = {} self._lock = threading.Lock() # We'll sync periodically (when the last sync was more than self._time_delta # seconds ago.) self._cache = {} self._last_sync_time = 0.0 self._return_only_prefixed = return_only_prefixed def increment(self, **counts: Number) -> Dict[str, Number]: """Increment a set of counters. Args: **counts: keyword arguments specifying count increments. Returns: The [name, value] mapping of all counters stored, i.e. this will also include counts that were not updated by this call to increment. """ with self._lock: for key, value in counts.items(): self._counts.setdefault(key, 0) self._counts[key] += value return self.get_counts() def get_counts(self) -> Dict[str, Number]: """Return all counts tracked by this counter.""" now = time.time() # TODO(b/144421838): use futures instead of blocking. if self._parent and (now - self._last_sync_time) > self._time_delta:
with self._lock: counts = _prefix_keys(self._counts, self._prefix) # Reset the local counts, as they will be merged into the parent and the # cache. self._counts = {} self._cache = self._parent.increment(**counts) self._last_sync_time = now # Potentially prefix the keys in the counts dictionary. counts = _prefix_keys(self._counts, self._prefix) # If there's no prefix make a copy of the dictionary so we don't modify the # internal self._counts. if not self._prefix: counts = dict(counts) # Combine local counts with any parent counts. for key, value in self._cache.items(): counts[key] = counts.get(key, 0) + value if self._prefix and self._return_only_prefixed: counts = dict([(key[len(self._prefix) + 1:], value) for key, value in counts.items() if key.startswith(f'{self._prefix}_')]) return counts def save(self) -> Mapping[str, Mapping[str, Number]]: return {'counts': self._counts, 'cache': self._cache} def restore(self, state: Mapping[str, Mapping[str, Number]]): # Force a sync, if necessary, on the next get_counts call. self._last_sync_time = 0. self._counts = state['counts'] self._cache = state['cache'] def _prefix_keys(dictionary: Dict[str, Number], prefix: str): """Return a dictionary with prefixed keys. Args: dictionary: dictionary to return a copy of. prefix: string to use as the prefix. Returns: Return a copy of the given dictionary whose keys are replaced by "{prefix}_{key}". If the prefix is the empty string it returns the given dictionary unchanged. """ if prefix: dictionary = {f'{prefix}_{k}': v for k, v in dictionary.items()} return dictionary
random_line_split
edwards.js
, t); }; EdwardsCurve.prototype.pointFromX = function pointFromX(odd, x) { x = new bn(x, 16); if (!x.red) x = x.toRed(this.red); var x2 = x.redSqr(); var rhs = this.c2.redSub(this.a.redMul(x2)); var lhs = this.one.redSub(this.c2.redMul(this.d).redMul(x2)); var y = rhs.redMul(lhs.redInvm()).redSqrt(); var isOdd = y.fromRed().isOdd(); if (odd && !isOdd || !odd && isOdd) y = y.redNeg(); return this.point(x, y, curve.one); }; EdwardsCurve.prototype.validate = function validate(point) { if (point.isInfinity()) return true; // Curve: A * X^2 + Y^2 = C^2 * (1 + D * X^2 * Y^2) point.normalize(); var x2 = point.x.redSqr(); var y2 = point.y.redSqr(); var lhs = x2.redMul(this.a).redAdd(y2); var rhs = this.c2.redMul(this.one.redAdd(this.d.redMul(x2).redMul(y2))); return lhs.cmp(rhs) === 0; }; function Point(curve, x, y, z, t) { Base.BasePoint.call(this, curve, 'projective'); if (x === null && y === null && z === null) { this.x = this.curve.zero; this.y = this.curve.one; this.z = this.curve.one; this.t = this.curve.zero; this.zOne = true; } else { this.x = new bn(x, 16); this.y = new bn(y, 16); this.z = z ? new bn(z, 16) : this.curve.one; this.t = t && new bn(t, 16); if (!this.x.red) this.x = this.x.toRed(this.curve.red); if (!this.y.red) this.y = this.y.toRed(this.curve.red); if (!this.z.red) this.z = this.z.toRed(this.curve.red); if (this.t && !this.t.red) this.t = this.t.toRed(this.curve.red); this.zOne = this.z === this.curve.one; // Use extended coordinates if (this.curve.extended && !this.t) { this.t = this.x.redMul(this.y); if (!this.zOne) this.t = this.t.redMul(this.z.redInvm()); } } } inherits(Point, Base.BasePoint); EdwardsCurve.prototype.pointFromJSON = function pointFromJSON(obj) { return Point.fromJSON(this, obj); }; EdwardsCurve.prototype.point = function point(x, y, z, t) { return new Point(this, x, y, z, t); }; Point.fromJSON = function fromJSON(curve, obj) { return new Point(curve, obj[0], obj[1], obj[2]); }; Point.prototype.inspect = function inspect() { if (this.isInfinity()) return '<EC Point Infinity>'; return '<EC Point x: ' + this.x.fromRed().toString(16, 2) + ' y: ' + this.y.fromRed().toString(16, 2) + ' z: ' + this.z.fromRed().toString(16, 2) + '>'; }; Point.prototype.isInfinity = function isInfinity() { // XXX This code assumes that zero is always zero in red return this.x.cmpn(0) === 0 && this.y.cmp(this.z) === 0; }; Point.prototype._extDbl = function _extDbl() { // hyperelliptic.org/EFD/g1p/auto-twisted-extended-1.html
// B = Y1^2 var b = this.y.redSqr(); // C = 2 * Z1^2 var c = this.z.redSqr(); c = c.redIAdd(c); // D = a * A var d = this.curve._mulA(a); // E = (X1 + Y1)^2 - A - B var e = this.x.redAdd(this.y).redSqr().redISub(a).redISub(b); // G = D + B var g = d.redAdd(b); // F = G - C var f = g.redSub(c); // H = D - B var h = d.redSub(b); // X3 = E * F var nx = e.redMul(f); // Y3 = G * H var ny = g.redMul(h); // T3 = E * H var nt = e.redMul(h); // Z3 = F * G var nz = f.redMul(g); return this.curve.point(nx, ny, nz, nt); }; Point.prototype._projDbl = function _projDbl() { // hyperelliptic.org/EFD/g1p/auto-twisted-projective.html // #doubling-dbl-2008-bbjlp // #doubling-dbl-2007-bl // and others // Generally 3M + 4S or 2M + 4S // B = (X1 + Y1)^2 var b = this.x.redAdd(this.y).redSqr(); // C = X1^2 var c = this.x.redSqr(); // D = Y1^2 var d = this.y.redSqr(); var nx; var ny; var nz; if (this.curve.twisted) { // E = a * C var e = this.curve._mulA(c); // F = E + D var f = e.redAdd(d); if (this.zOne) { // X3 = (B - C - D) * (F - 2) nx = b.redSub(c).redSub(d).redMul(f.redSub(this.curve.two)); // Y3 = F * (E - D) ny = f.redMul(e.redSub(d)); // Z3 = F^2 - 2 * F nz = f.redSqr().redSub(f).redSub(f); } else { // H = Z1^2 var h = this.z.redSqr(); // J = F - 2 * H var j = f.redSub(h).redISub(h); // X3 = (B-C-D)*J nx = b.redSub(c).redISub(d).redMul(j); // Y3 = F * (E - D) ny = f.redMul(e.redSub(d)); // Z3 = F * J nz = f.redMul(j); } } else { // E = C + D var e = c.redAdd(d); // H = (c * Z1)^2 var h = this.curve._mulC(this.c.redMul(this.z)).redSqr(); // J = E - 2 * H var j = e.redSub(h).redSub(h); // X3 = c * (B - E) * J nx = this.curve._mulC(b.redISub(e)).redMul(j); // Y3 = c * E * (C - D) ny = this.curve._mulC(e).redMul(c.redISub(d)); // Z3 = E * J nz = e.redMul(j); } return this.curve.point(nx, ny, nz); }; Point.prototype.dbl = function dbl() { if (this.isInfinity()) return this; // Double in extended coordinates if (this.curve.extended) return this._extDbl(); else return this._projDbl(); }; Point.prototype._extAdd = function _extAdd(p) { // hyperelliptic.org/EFD/g1p/auto-twisted-extended-1.html // #addition-add-2008-hwcd-3 // 8M // A = (Y1 - X1) * (Y2 - X2) var a = this.y.redSub(this.x).redMul(p.y.redSub(p.x)); // B = (Y1 + X1) * (Y2 + X2) var b = this.y.redAdd(this.x).redMul(p.y.redAdd(p.x)); // C = T1 * k * T2 var c = this.t.redMul(this.curve.dd).redMul(p.t); // D = Z1 * 2 * Z2 var d = this.z.redMul(p.z.redAdd(p.z)); // E = B - A var e = b.redSub(a); // F = D - C var f = d.redSub(c); // G = D + C var g = d.red
// #doubling-dbl-2008-hwcd // 4M + 4S // A = X1^2 var a = this.x.redSqr();
random_line_split
edwards.js
(conf) { // NOTE: Important as we are creating point in Base.call() this.twisted = (conf.a | 0) !== 1; this.mOneA = this.twisted && (conf.a | 0) === -1; this.extended = this.mOneA; Base.call(this, 'edwards', conf); this.a = new bn(conf.a, 16).mod(this.red.m).toRed(this.red); this.c = new bn(conf.c, 16).toRed(this.red); this.c2 = this.c.redSqr(); this.d = new bn(conf.d, 16).toRed(this.red); this.dd = this.d.redAdd(this.d); assert(!this.twisted || this.c.fromRed().cmpn(1) === 0); this.oneC = (conf.c | 0) === 1; } inherits(EdwardsCurve, Base); module.exports = EdwardsCurve; EdwardsCurve.prototype._mulA = function _mulA(num) { if (this.mOneA) return num.redNeg(); else return this.a.redMul(num); }; EdwardsCurve.prototype._mulC = function _mulC(num) { if (this.oneC) return num; else return this.c.redMul(num); }; // Just for compatibility with Short curve EdwardsCurve.prototype.jpoint = function jpoint(x, y, z, t) { return this.point(x, y, z, t); }; EdwardsCurve.prototype.pointFromX = function pointFromX(odd, x) { x = new bn(x, 16); if (!x.red) x = x.toRed(this.red); var x2 = x.redSqr(); var rhs = this.c2.redSub(this.a.redMul(x2)); var lhs = this.one.redSub(this.c2.redMul(this.d).redMul(x2)); var y = rhs.redMul(lhs.redInvm()).redSqrt(); var isOdd = y.fromRed().isOdd(); if (odd && !isOdd || !odd && isOdd) y = y.redNeg(); return this.point(x, y, curve.one); }; EdwardsCurve.prototype.validate = function validate(point) { if (point.isInfinity()) return true; // Curve: A * X^2 + Y^2 = C^2 * (1 + D * X^2 * Y^2) point.normalize(); var x2 = point.x.redSqr(); var y2 = point.y.redSqr(); var lhs = x2.redMul(this.a).redAdd(y2); var rhs = this.c2.redMul(this.one.redAdd(this.d.redMul(x2).redMul(y2))); return lhs.cmp(rhs) === 0; }; function Point(curve, x, y, z, t) { Base.BasePoint.call(this, curve, 'projective'); if (x === null && y === null && z === null) { this.x = this.curve.zero; this.y = this.curve.one; this.z = this.curve.one; this.t = this.curve.zero; this.zOne = true; } else { this.x = new bn(x, 16); this.y = new bn(y, 16); this.z = z ? new bn(z, 16) : this.curve.one; this.t = t && new bn(t, 16); if (!this.x.red) this.x = this.x.toRed(this.curve.red); if (!this.y.red) this.y = this.y.toRed(this.curve.red); if (!this.z.red) this.z = this.z.toRed(this.curve.red); if (this.t && !this.t.red) this.t = this.t.toRed(this.curve.red); this.zOne = this.z === this.curve.one; // Use extended coordinates if (this.curve.extended && !this.t) { this.t = this.x.redMul(this.y); if (!this.zOne) this.t = this.t.redMul(this.z.redInvm()); } } } inherits(Point, Base.BasePoint); EdwardsCurve.prototype.pointFromJSON = function pointFromJSON(obj) { return Point.fromJSON(this, obj); }; EdwardsCurve.prototype.point = function point(x, y, z, t) { return new Point(this, x, y, z, t); }; Point.fromJSON = function fromJSON(curve, obj) { return new Point(curve, obj[0], obj[1], obj[2]); }; Point.prototype.inspect = function inspect() { if (this.isInfinity()) return '<EC Point Infinity>'; return '<EC Point x: ' + this.x.fromRed().toString(16, 2) + ' y: ' + this.y.fromRed().toString(16, 2) + ' z: ' + this.z.fromRed().toString(16, 2) + '>'; }; Point.prototype.isInfinity = function isInfinity() { // XXX This code assumes that zero is always zero in red return this.x.cmpn(0) === 0 && this.y.cmp(this.z) === 0; }; Point.prototype._extDbl = function _extDbl() { // hyperelliptic.org/EFD/g1p/auto-twisted-extended-1.html // #doubling-dbl-2008-hwcd // 4M + 4S // A = X1^2 var a = this.x.redSqr(); // B = Y1^2 var b = this.y.redSqr(); // C = 2 * Z1^2 var c = this.z.redSqr(); c = c.redIAdd(c); // D = a * A var d = this.curve._mulA(a); // E = (X1 + Y1)^2 - A - B var e = this.x.redAdd(this.y).redSqr().redISub(a).redISub(b); // G = D + B var g = d.redAdd(b); // F = G - C var f = g.redSub(c); // H = D - B var h = d.redSub(b); // X3 = E * F var nx = e.redMul(f); // Y3 = G * H var ny = g.redMul(h); // T3 = E * H var nt = e.redMul(h); // Z3 = F * G var nz = f.redMul(g); return this.curve.point(nx, ny, nz, nt); }; Point.prototype._projDbl = function _projDbl() { // hyperelliptic.org/EFD/g1p/auto-twisted-projective.html // #doubling-dbl-2008-bbjlp // #doubling-dbl-2007-bl // and others // Generally 3M + 4S or 2M + 4S // B = (X1 + Y1)^2 var b = this.x.redAdd(this.y).redSqr(); // C = X1^2 var c = this.x.redSqr(); // D = Y1^2 var d = this.y.redSqr(); var nx; var ny; var nz; if (this.curve.twisted) { // E = a * C var e = this.curve._mulA(c); // F = E + D var f = e.redAdd(d); if (this.zOne) { // X3 = (B - C - D) * (F - 2) nx = b.redSub(c).redSub(d).redMul(f.redSub(this.curve.two)); // Y3 = F * (E - D) ny = f.redMul(e.redSub(d)); // Z3 = F^2 - 2 * F nz = f.redSqr().redSub(f).redSub(f); } else { // H = Z1^2 var h = this.z.redSqr(); // J = F - 2 * H var j = f.redSub(h).redISub(h); // X3 = (B-C-D)*J nx = b.redSub(c).redISub(d).redMul(j); // Y3 = F * (E - D) ny = f.redMul(e.redSub(d)); // Z3 = F * J nz = f.redMul(j); } } else { // E = C + D var e = c.redAdd(d); // H = (c * Z1)^2 var h = this.curve._mulC(this.c.redMul(this.z)).redSqr(); // J = E - 2 * H var j = e.redSub(h).redSub(h); // X3 = c * (B - E) * J nx = this.curve._mulC(b.redISub(e)).redMul(j); // Y3 = c * E * (C - D) ny = this.curve._mulC(e).redMul(c.redISub(d));
EdwardsCurve
identifier_name
edwards.js
t); }; EdwardsCurve.prototype.pointFromX = function pointFromX(odd, x) { x = new bn(x, 16); if (!x.red) x = x.toRed(this.red); var x2 = x.redSqr(); var rhs = this.c2.redSub(this.a.redMul(x2)); var lhs = this.one.redSub(this.c2.redMul(this.d).redMul(x2)); var y = rhs.redMul(lhs.redInvm()).redSqrt(); var isOdd = y.fromRed().isOdd(); if (odd && !isOdd || !odd && isOdd) y = y.redNeg(); return this.point(x, y, curve.one); }; EdwardsCurve.prototype.validate = function validate(point) { if (point.isInfinity()) return true; // Curve: A * X^2 + Y^2 = C^2 * (1 + D * X^2 * Y^2) point.normalize(); var x2 = point.x.redSqr(); var y2 = point.y.redSqr(); var lhs = x2.redMul(this.a).redAdd(y2); var rhs = this.c2.redMul(this.one.redAdd(this.d.redMul(x2).redMul(y2))); return lhs.cmp(rhs) === 0; }; function Point(curve, x, y, z, t)
this.t = this.t.toRed(this.curve.red); this.zOne = this.z === this.curve.one; // Use extended coordinates if (this.curve.extended && !this.t) { this.t = this.x.redMul(this.y); if (!this.zOne) this.t = this.t.redMul(this.z.redInvm()); } } } inherits(Point, Base.BasePoint); EdwardsCurve.prototype.pointFromJSON = function pointFromJSON(obj) { return Point.fromJSON(this, obj); }; EdwardsCurve.prototype.point = function point(x, y, z, t) { return new Point(this, x, y, z, t); }; Point.fromJSON = function fromJSON(curve, obj) { return new Point(curve, obj[0], obj[1], obj[2]); }; Point.prototype.inspect = function inspect() { if (this.isInfinity()) return '<EC Point Infinity>'; return '<EC Point x: ' + this.x.fromRed().toString(16, 2) + ' y: ' + this.y.fromRed().toString(16, 2) + ' z: ' + this.z.fromRed().toString(16, 2) + '>'; }; Point.prototype.isInfinity = function isInfinity() { // XXX This code assumes that zero is always zero in red return this.x.cmpn(0) === 0 && this.y.cmp(this.z) === 0; }; Point.prototype._extDbl = function _extDbl() { // hyperelliptic.org/EFD/g1p/auto-twisted-extended-1.html // #doubling-dbl-2008-hwcd // 4M + 4S // A = X1^2 var a = this.x.redSqr(); // B = Y1^2 var b = this.y.redSqr(); // C = 2 * Z1^2 var c = this.z.redSqr(); c = c.redIAdd(c); // D = a * A var d = this.curve._mulA(a); // E = (X1 + Y1)^2 - A - B var e = this.x.redAdd(this.y).redSqr().redISub(a).redISub(b); // G = D + B var g = d.redAdd(b); // F = G - C var f = g.redSub(c); // H = D - B var h = d.redSub(b); // X3 = E * F var nx = e.redMul(f); // Y3 = G * H var ny = g.redMul(h); // T3 = E * H var nt = e.redMul(h); // Z3 = F * G var nz = f.redMul(g); return this.curve.point(nx, ny, nz, nt); }; Point.prototype._projDbl = function _projDbl() { // hyperelliptic.org/EFD/g1p/auto-twisted-projective.html // #doubling-dbl-2008-bbjlp // #doubling-dbl-2007-bl // and others // Generally 3M + 4S or 2M + 4S // B = (X1 + Y1)^2 var b = this.x.redAdd(this.y).redSqr(); // C = X1^2 var c = this.x.redSqr(); // D = Y1^2 var d = this.y.redSqr(); var nx; var ny; var nz; if (this.curve.twisted) { // E = a * C var e = this.curve._mulA(c); // F = E + D var f = e.redAdd(d); if (this.zOne) { // X3 = (B - C - D) * (F - 2) nx = b.redSub(c).redSub(d).redMul(f.redSub(this.curve.two)); // Y3 = F * (E - D) ny = f.redMul(e.redSub(d)); // Z3 = F^2 - 2 * F nz = f.redSqr().redSub(f).redSub(f); } else { // H = Z1^2 var h = this.z.redSqr(); // J = F - 2 * H var j = f.redSub(h).redISub(h); // X3 = (B-C-D)*J nx = b.redSub(c).redISub(d).redMul(j); // Y3 = F * (E - D) ny = f.redMul(e.redSub(d)); // Z3 = F * J nz = f.redMul(j); } } else { // E = C + D var e = c.redAdd(d); // H = (c * Z1)^2 var h = this.curve._mulC(this.c.redMul(this.z)).redSqr(); // J = E - 2 * H var j = e.redSub(h).redSub(h); // X3 = c * (B - E) * J nx = this.curve._mulC(b.redISub(e)).redMul(j); // Y3 = c * E * (C - D) ny = this.curve._mulC(e).redMul(c.redISub(d)); // Z3 = E * J nz = e.redMul(j); } return this.curve.point(nx, ny, nz); }; Point.prototype.dbl = function dbl() { if (this.isInfinity()) return this; // Double in extended coordinates if (this.curve.extended) return this._extDbl(); else return this._projDbl(); }; Point.prototype._extAdd = function _extAdd(p) { // hyperelliptic.org/EFD/g1p/auto-twisted-extended-1.html // #addition-add-2008-hwcd-3 // 8M // A = (Y1 - X1) * (Y2 - X2) var a = this.y.redSub(this.x).redMul(p.y.redSub(p.x)); // B = (Y1 + X1) * (Y2 + X2) var b = this.y.redAdd(this.x).redMul(p.y.redAdd(p.x)); // C = T1 * k * T2 var c = this.t.redMul(this.curve.dd).redMul(p.t); // D = Z1 * 2 * Z2 var d = this.z.redMul(p.z.redAdd(p.z)); // E = B - A var e = b.redSub(a); // F = D - C var f = d.redSub(c); // G = D + C var g = d
{ Base.BasePoint.call(this, curve, 'projective'); if (x === null && y === null && z === null) { this.x = this.curve.zero; this.y = this.curve.one; this.z = this.curve.one; this.t = this.curve.zero; this.zOne = true; } else { this.x = new bn(x, 16); this.y = new bn(y, 16); this.z = z ? new bn(z, 16) : this.curve.one; this.t = t && new bn(t, 16); if (!this.x.red) this.x = this.x.toRed(this.curve.red); if (!this.y.red) this.y = this.y.toRed(this.curve.red); if (!this.z.red) this.z = this.z.toRed(this.curve.red); if (this.t && !this.t.red)
identifier_body
IncreaseStat.ts
/// <reference path="../Dialogs/DialogAction.ts" /> @DialogActionClass class IncreaseStat extends ActionClass { public Display(id: number, values: string[], updateFunction?: string): string { var html = ""; html += this.Label("Stat"); html += this.OptionList(id, 0, world.Stats.map(c => c.Name).sort(), values[0], updateFunction); html += this.Label("Quantity"); html += this.Input(id, 1, values[1] = (values[1] || values[1] == "" ? values[1] : "1"), updateFunction); return html; } public Execute(values: string[], env?: CodeEnvironement): void {
} world.Player.SetStat(values[0], world.Player.GetStat(values[0]) + val); } }
if (!this.Execute.caller) { play.devTools = true; world.Player.InformServer(); return; } if (!values[0]) throw "The action 'Increase Stat' requires a name."; if (!env) env = new CodeEnvironement(); var val = 0; try { val = CodeParser.ExecuteStatement(values[1], env.variables).GetNumber(); } catch (ex) { throw "The expression used in 'Increase Stat' for the quantity is invalid.";
identifier_body
IncreaseStat.ts
/// <reference path="../Dialogs/DialogAction.ts" /> @DialogActionClass class IncreaseStat extends ActionClass { public Display(id: number, values: string[], updateFunction?: string): string { var html = ""; html += this.Label("Stat"); html += this.OptionList(id, 0, world.Stats.map(c => c.Name).sort(), values[0], updateFunction); html += this.Label("Quantity"); html += this.Input(id, 1, values[1] = (values[1] || values[1] == "" ? values[1] : "1"), updateFunction); return html; } public Execute(values: string[], env?: CodeEnvironement): void { if (!this.Execute.caller) { play.devTools = true; world.Player.InformServer(); return; } if (!values[0]) throw "The action 'Increase Stat' requires a name.";
try { val = CodeParser.ExecuteStatement(values[1], env.variables).GetNumber(); } catch (ex) { throw "The expression used in 'Increase Stat' for the quantity is invalid."; } world.Player.SetStat(values[0], world.Player.GetStat(values[0]) + val); } }
if (!env) env = new CodeEnvironement(); var val = 0;
random_line_split
IncreaseStat.ts
/// <reference path="../Dialogs/DialogAction.ts" /> @DialogActionClass class IncreaseStat extends ActionClass { public Display(id: number, values: string[], updateFunction?: string): string { var html = ""; html += this.Label("Stat"); html += this.OptionList(id, 0, world.Stats.map(c => c.Name).sort(), values[0], updateFunction); html += this.Label("Quantity"); html += this.Input(id, 1, values[1] = (values[1] || values[1] == "" ? values[1] : "1"), updateFunction); return html; } public Ex
alues: string[], env?: CodeEnvironement): void { if (!this.Execute.caller) { play.devTools = true; world.Player.InformServer(); return; } if (!values[0]) throw "The action 'Increase Stat' requires a name."; if (!env) env = new CodeEnvironement(); var val = 0; try { val = CodeParser.ExecuteStatement(values[1], env.variables).GetNumber(); } catch (ex) { throw "The expression used in 'Increase Stat' for the quantity is invalid."; } world.Player.SetStat(values[0], world.Player.GetStat(values[0]) + val); } }
ecute(v
identifier_name
0009_auto_20160517_2016.py
# -*- coding: utf-8 -*- # Generated by Django 1.9.1 on 2016-05-17 20:16 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('samaritan', '0008_member_baptismal_place'), ] operations = [
migrations.RemoveField( model_name='membership', name='type', ), migrations.RemoveField( model_name='member', name='membership', ), migrations.AddField( model_name='member', name='membership_date', field=models.DateField(blank=True, null=True), ), migrations.AddField( model_name='member', name='type', field=models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='samaritan.MembershipType'), ), migrations.DeleteModel( name='Membership', ), ]
random_line_split
0009_auto_20160517_2016.py
# -*- coding: utf-8 -*- # Generated by Django 1.9.1 on 2016-05-17 20:16 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration):
name='type', field=models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='samaritan.MembershipType'), ), migrations.DeleteModel( name='Membership', ), ]
dependencies = [ ('samaritan', '0008_member_baptismal_place'), ] operations = [ migrations.RemoveField( model_name='membership', name='type', ), migrations.RemoveField( model_name='member', name='membership', ), migrations.AddField( model_name='member', name='membership_date', field=models.DateField(blank=True, null=True), ), migrations.AddField( model_name='member',
identifier_body
0009_auto_20160517_2016.py
# -*- coding: utf-8 -*- # Generated by Django 1.9.1 on 2016-05-17 20:16 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class
(migrations.Migration): dependencies = [ ('samaritan', '0008_member_baptismal_place'), ] operations = [ migrations.RemoveField( model_name='membership', name='type', ), migrations.RemoveField( model_name='member', name='membership', ), migrations.AddField( model_name='member', name='membership_date', field=models.DateField(blank=True, null=True), ), migrations.AddField( model_name='member', name='type', field=models.ForeignKey(null=True, on_delete=django.db.models.deletion.CASCADE, to='samaritan.MembershipType'), ), migrations.DeleteModel( name='Membership', ), ]
Migration
identifier_name
be2bill.py
# -*- coding: utf-8 -*- from decimal import Decimal from be2bill_sdk import Be2BillForm from cartridge_external_payment.providers.base import PaymentProvider class Be2BillProvider(PaymentProvider): def get_start_payment_form(self, request, order): total = Decimal(order.total * 100).quantize(Decimal('0')) fullname = order.billing_detail_first_name + ' ' + \ order.billing_detail_last_name client_ident = "{} ({})".format(fullname, order.billing_detail_email) return Be2BillForm(operation_type="payment", client_ident=client_ident, description="X", order_id=order.id, amount=total, client_email=order.billing_detail_email, card_full_name=fullname, #Save cart id for notification extra_data=request.cart.id) def get_order_id(self, notification_request):
def get_transaction_id(self, notification_request): return notification_request.GET.get('TRANSACTIONID', None) def get_cart_id(self, notification_request): raise notification_request.GET.get('EXTRADATA', None)
return notification_request.GET.get('ORDERID', None)
identifier_body
be2bill.py
# -*- coding: utf-8 -*- from decimal import Decimal from be2bill_sdk import Be2BillForm from cartridge_external_payment.providers.base import PaymentProvider class Be2BillProvider(PaymentProvider): def get_start_payment_form(self, request, order): total = Decimal(order.total * 100).quantize(Decimal('0')) fullname = order.billing_detail_first_name + ' ' + \ order.billing_detail_last_name client_ident = "{} ({})".format(fullname, order.billing_detail_email) return Be2BillForm(operation_type="payment", client_ident=client_ident, description="X", order_id=order.id, amount=total, client_email=order.billing_detail_email, card_full_name=fullname, #Save cart id for notification extra_data=request.cart.id) def get_order_id(self, notification_request): return notification_request.GET.get('ORDERID', None) def get_transaction_id(self, notification_request): return notification_request.GET.get('TRANSACTIONID', None)
def get_cart_id(self, notification_request): raise notification_request.GET.get('EXTRADATA', None)
random_line_split
be2bill.py
# -*- coding: utf-8 -*- from decimal import Decimal from be2bill_sdk import Be2BillForm from cartridge_external_payment.providers.base import PaymentProvider class Be2BillProvider(PaymentProvider): def
(self, request, order): total = Decimal(order.total * 100).quantize(Decimal('0')) fullname = order.billing_detail_first_name + ' ' + \ order.billing_detail_last_name client_ident = "{} ({})".format(fullname, order.billing_detail_email) return Be2BillForm(operation_type="payment", client_ident=client_ident, description="X", order_id=order.id, amount=total, client_email=order.billing_detail_email, card_full_name=fullname, #Save cart id for notification extra_data=request.cart.id) def get_order_id(self, notification_request): return notification_request.GET.get('ORDERID', None) def get_transaction_id(self, notification_request): return notification_request.GET.get('TRANSACTIONID', None) def get_cart_id(self, notification_request): raise notification_request.GET.get('EXTRADATA', None)
get_start_payment_form
identifier_name
quick.rs
//! An example using the Builder pattern API to configure the logger at run-time based on command //! line arguments. //! //! The default output is `module::path: message`, and the "tag", which is the text to the left of //! the colon, is colorized. This example allows the user to dynamically change the output based //! on command line arguments. //! //! The [clap](https://crates.io/crates/clap) argument parser is used in this example, but loggerv //! works with any argument parser. extern crate ansi_term; #[macro_use] extern crate log; extern crate loggerv; extern crate clap; use clap::{Arg, App}; fn main()
warn!("This too is always printed to stderr"); info!("This is optional info printed to stdout"); // for ./app -v or higher debug!("This is optional debug printed to stdout"); // for ./app -vv or higher trace!("This is optional trace printed to stdout"); // for ./app -vvv }
{ // Add the following line near the beginning of the main function for an application to enable // colorized output on Windows 10. // // Based on documentation for the ansi_term crate, Windows 10 supports ANSI escape characters, // but it must be enabled first using the `ansi_term::enable_ansi_support()` function. It is // conditionally compiled and only exists for Windows builds. To avoid build errors on // non-windows platforms, a cfg guard should be put in place. #[cfg(windows)] ansi_term::enable_ansi_support().unwrap(); let args = App::new("app") .arg(Arg::with_name("v") .short("v") .multiple(true) .help("Sets the level of verbosity")) .get_matches(); loggerv::init_with_verbosity(args.occurrences_of("v")).unwrap(); error!("This is always printed to stderr");
identifier_body
quick.rs
//! An example using the Builder pattern API to configure the logger at run-time based on command //! line arguments. //! //! The default output is `module::path: message`, and the "tag", which is the text to the left of //! the colon, is colorized. This example allows the user to dynamically change the output based //! on command line arguments. //! //! The [clap](https://crates.io/crates/clap) argument parser is used in this example, but loggerv //! works with any argument parser. extern crate ansi_term; #[macro_use] extern crate log; extern crate loggerv; extern crate clap; use clap::{Arg, App}; fn
() { // Add the following line near the beginning of the main function for an application to enable // colorized output on Windows 10. // // Based on documentation for the ansi_term crate, Windows 10 supports ANSI escape characters, // but it must be enabled first using the `ansi_term::enable_ansi_support()` function. It is // conditionally compiled and only exists for Windows builds. To avoid build errors on // non-windows platforms, a cfg guard should be put in place. #[cfg(windows)] ansi_term::enable_ansi_support().unwrap(); let args = App::new("app") .arg(Arg::with_name("v") .short("v") .multiple(true) .help("Sets the level of verbosity")) .get_matches(); loggerv::init_with_verbosity(args.occurrences_of("v")).unwrap(); error!("This is always printed to stderr"); warn!("This too is always printed to stderr"); info!("This is optional info printed to stdout"); // for ./app -v or higher debug!("This is optional debug printed to stdout"); // for ./app -vv or higher trace!("This is optional trace printed to stdout"); // for ./app -vvv }
main
identifier_name
quick.rs
//! An example using the Builder pattern API to configure the logger at run-time based on command //! line arguments. //! //! The default output is `module::path: message`, and the "tag", which is the text to the left of //! the colon, is colorized. This example allows the user to dynamically change the output based //! on command line arguments. //! //! The [clap](https://crates.io/crates/clap) argument parser is used in this example, but loggerv //! works with any argument parser. extern crate ansi_term; #[macro_use] extern crate log; extern crate loggerv; extern crate clap;
// // Based on documentation for the ansi_term crate, Windows 10 supports ANSI escape characters, // but it must be enabled first using the `ansi_term::enable_ansi_support()` function. It is // conditionally compiled and only exists for Windows builds. To avoid build errors on // non-windows platforms, a cfg guard should be put in place. #[cfg(windows)] ansi_term::enable_ansi_support().unwrap(); let args = App::new("app") .arg(Arg::with_name("v") .short("v") .multiple(true) .help("Sets the level of verbosity")) .get_matches(); loggerv::init_with_verbosity(args.occurrences_of("v")).unwrap(); error!("This is always printed to stderr"); warn!("This too is always printed to stderr"); info!("This is optional info printed to stdout"); // for ./app -v or higher debug!("This is optional debug printed to stdout"); // for ./app -vv or higher trace!("This is optional trace printed to stdout"); // for ./app -vvv }
use clap::{Arg, App}; fn main() { // Add the following line near the beginning of the main function for an application to enable // colorized output on Windows 10.
random_line_split
httpInterceptor.spec.js
import { module, inject } from "angular-mocks"; import "../../../app/auth/auth.module"; (function () { 'use strict'; describe('HttpHeaderInterceptor', function () { var httpHeaderInterceptor, CookiesServiceMock; beforeEach(module('xr.auth')); beforeEach(module(function ($provide) { CookiesServiceMock = { get: function () {} }; $provide.value('$cookies', CookiesServiceMock); })); beforeEach(inject(function (_httpHeaderInterceptor_) { httpHeaderInterceptor = _httpHeaderInterceptor_;
it('should add Authorization header if token exists', function () { var token = '1234'; spyOn(CookiesServiceMock, 'get').and.returnValue(token); var config = httpHeaderInterceptor.request({ headers: {} }); expect(config.headers.Authorization).toBe('Bearer ' + token); }); it('should not add Authorization header if no token', function () { var config = httpHeaderInterceptor.request({ headers: {} }); expect(config.headers.Authorization).toBeUndefined(); }); }); }); })();
})); describe('request', function () {
random_line_split
CryptoProvider.ts
/* * Copyright (c) Microsoft Corporation. All rights reserved. * Licensed under the MIT License. */ import { ICrypto, PkceCodes } from "@azure/msal-common"; import { GuidGenerator } from "./GuidGenerator"; import { EncodingUtils } from "../utils/EncodingUtils"; import { PkceGenerator } from "./PkceGenerator"; import { HashUtils } from "./HashUtils"; /** * This class implements MSAL node's crypto interface, which allows it to perform base64 encoding and decoding, generating cryptographically random GUIDs and * implementing Proof Key for Code Exchange specs for the OAuth Authorization Code Flow using PKCE (rfc here: https://tools.ietf.org/html/rfc7636). * @public */ export class CryptoProvider implements ICrypto { private pkceGenerator: PkceGenerator; private hashUtils: HashUtils; constructor() { // Browser crypto needs to be validated first before any other classes can be set. this.pkceGenerator = new PkceGenerator(); this.hashUtils = new HashUtils(); } /** * Creates a new random GUID - used to populate state and nonce. * @returns string (GUID) */ createNewGuid(): string { return GuidGenerator.generateGuid(); } /** * Encodes input string to base64. * @param input - string to be encoded */ base64Encode(input: string): string
/** * Decodes input string from base64. * @param input - string to be decoded */ base64Decode(input: string): string { return EncodingUtils.base64Decode(input); } /** * Generates PKCE codes used in Authorization Code Flow. */ generatePkceCodes(): Promise<PkceCodes> { return this.pkceGenerator.generatePkceCodes(); } /** * Generates a keypair, stores it and returns a thumbprint - not yet implemented for node */ getPublicKeyThumbprint(): Promise<string> { throw new Error("Method not implemented."); } /** * Removes cryptographic keypair from key store matching the keyId passed in * @param kid */ removeTokenBindingKey(): Promise<boolean> { throw new Error("Method not implemented."); } /** * Removes all cryptographic keys from Keystore */ clearKeystore(): Promise<boolean> { throw new Error("Method not implemented."); } /** * Signs the given object as a jwt payload with private key retrieved by given kid - currently not implemented for node */ signJwt(): Promise<string> { throw new Error("Method not implemented."); } /** * Returns the SHA-256 hash of an input string */ async hashString(plainText: string): Promise<string> { return EncodingUtils.base64EncodeUrl( this.hashUtils.sha256(plainText).toString("base64"), "base64" ); } }
{ return EncodingUtils.base64Encode(input); }
identifier_body
CryptoProvider.ts
/* * Copyright (c) Microsoft Corporation. All rights reserved. * Licensed under the MIT License. */ import { ICrypto, PkceCodes } from "@azure/msal-common"; import { GuidGenerator } from "./GuidGenerator"; import { EncodingUtils } from "../utils/EncodingUtils"; import { PkceGenerator } from "./PkceGenerator"; import { HashUtils } from "./HashUtils"; /** * This class implements MSAL node's crypto interface, which allows it to perform base64 encoding and decoding, generating cryptographically random GUIDs and * implementing Proof Key for Code Exchange specs for the OAuth Authorization Code Flow using PKCE (rfc here: https://tools.ietf.org/html/rfc7636). * @public */ export class CryptoProvider implements ICrypto { private pkceGenerator: PkceGenerator; private hashUtils: HashUtils; constructor() { // Browser crypto needs to be validated first before any other classes can be set. this.pkceGenerator = new PkceGenerator(); this.hashUtils = new HashUtils(); } /** * Creates a new random GUID - used to populate state and nonce. * @returns string (GUID) */ createNewGuid(): string { return GuidGenerator.generateGuid(); } /** * Encodes input string to base64. * @param input - string to be encoded */ base64Encode(input: string): string { return EncodingUtils.base64Encode(input); } /** * Decodes input string from base64. * @param input - string to be decoded */
(input: string): string { return EncodingUtils.base64Decode(input); } /** * Generates PKCE codes used in Authorization Code Flow. */ generatePkceCodes(): Promise<PkceCodes> { return this.pkceGenerator.generatePkceCodes(); } /** * Generates a keypair, stores it and returns a thumbprint - not yet implemented for node */ getPublicKeyThumbprint(): Promise<string> { throw new Error("Method not implemented."); } /** * Removes cryptographic keypair from key store matching the keyId passed in * @param kid */ removeTokenBindingKey(): Promise<boolean> { throw new Error("Method not implemented."); } /** * Removes all cryptographic keys from Keystore */ clearKeystore(): Promise<boolean> { throw new Error("Method not implemented."); } /** * Signs the given object as a jwt payload with private key retrieved by given kid - currently not implemented for node */ signJwt(): Promise<string> { throw new Error("Method not implemented."); } /** * Returns the SHA-256 hash of an input string */ async hashString(plainText: string): Promise<string> { return EncodingUtils.base64EncodeUrl( this.hashUtils.sha256(plainText).toString("base64"), "base64" ); } }
base64Decode
identifier_name
CryptoProvider.ts
/* * Copyright (c) Microsoft Corporation. All rights reserved. * Licensed under the MIT License. */ import { ICrypto, PkceCodes } from "@azure/msal-common"; import { GuidGenerator } from "./GuidGenerator"; import { EncodingUtils } from "../utils/EncodingUtils"; import { PkceGenerator } from "./PkceGenerator"; import { HashUtils } from "./HashUtils"; /** * This class implements MSAL node's crypto interface, which allows it to perform base64 encoding and decoding, generating cryptographically random GUIDs and * implementing Proof Key for Code Exchange specs for the OAuth Authorization Code Flow using PKCE (rfc here: https://tools.ietf.org/html/rfc7636). * @public */ export class CryptoProvider implements ICrypto { private pkceGenerator: PkceGenerator; private hashUtils: HashUtils; constructor() { // Browser crypto needs to be validated first before any other classes can be set. this.pkceGenerator = new PkceGenerator(); this.hashUtils = new HashUtils(); } /** * Creates a new random GUID - used to populate state and nonce. * @returns string (GUID) */ createNewGuid(): string { return GuidGenerator.generateGuid(); } /** * Encodes input string to base64. * @param input - string to be encoded */ base64Encode(input: string): string { return EncodingUtils.base64Encode(input); } /** * Decodes input string from base64. * @param input - string to be decoded */ base64Decode(input: string): string { return EncodingUtils.base64Decode(input); } /** * Generates PKCE codes used in Authorization Code Flow. */ generatePkceCodes(): Promise<PkceCodes> { return this.pkceGenerator.generatePkceCodes(); } /** * Generates a keypair, stores it and returns a thumbprint - not yet implemented for node */ getPublicKeyThumbprint(): Promise<string> { throw new Error("Method not implemented."); }
throw new Error("Method not implemented."); } /** * Removes all cryptographic keys from Keystore */ clearKeystore(): Promise<boolean> { throw new Error("Method not implemented."); } /** * Signs the given object as a jwt payload with private key retrieved by given kid - currently not implemented for node */ signJwt(): Promise<string> { throw new Error("Method not implemented."); } /** * Returns the SHA-256 hash of an input string */ async hashString(plainText: string): Promise<string> { return EncodingUtils.base64EncodeUrl( this.hashUtils.sha256(plainText).toString("base64"), "base64" ); } }
/** * Removes cryptographic keypair from key store matching the keyId passed in * @param kid */ removeTokenBindingKey(): Promise<boolean> {
random_line_split
deserialization.rs
use alias::Alias; use config::Config; use consts::*; use itertools::Itertools; use family::Family; use glib::functions; use pango::Context; use pango::ContextExt; use pango::FontMapExt; use pango::FontFamilyExt; use range::Range; use sxd_document::dom::Comment; use sxd_document::dom::ChildOfElement; use sxd_document::dom::Element; use sxd_document::dom::Text; use sxd_document::parser; use std::cell::RefCell; use std::fs::File; use std::i32; use std::io::Read; use std::ops::Deref; use std::ops::DerefMut; pub fn list_families(context: &Context) -> Vec<RefCell<Family>> { match context.get_font_map() { Some(map) => { map.list_families() .iter() .filter_map(|x| x.get_name()) .filter(|x| !["Sans", "Serif", "Monospace"].contains(&x.as_str())) .map(|x| { RefCell::new(Family { name: x, stripped_ranges: vec![], }) }) .collect() } None => vec![], } } pub fn parse_or_default<'a>(families: &'a Vec<RefCell<Family>>) -> Config<'a> { let fc_config_path = functions::get_user_config_dir() .expect("$XDG_CONFIG_HOME not set!") .join("fontconfig/fonts.conf"); let config_parse = match File::open(fc_config_path.as_path()) { Ok(mut f) => { let mut buffer = String::new(); f.read_to_string(&mut buffer).expect( "Failed to parse your fonts.conf file", ); parser::parse(&buffer) } _ => parser::parse(DEFAULT_FONTS_CONF), }; let config_package = match config_parse { Ok(package) => package, Err((_, errors)) => panic!("Error parsing fonts.conf!\n{}", errors.iter().join("\n")), }; // scan matches collection let mut scan_matches: Vec<&'a RefCell<Family>> = vec![]; // aliases collection let mut aliases: Vec<Alias<'a>> = vec![]; { let doc = config_package.as_document(); let old_root_element = doc.root().children()[0].element().expect(INVALID_CONFIG); // rest of dom collection let new_root_element = doc.create_element(old_root_element.name()); for attr in old_root_element.attributes() { new_root_element.set_attribute_value(attr.name(), attr.value()); } // group children to correct collections for child in old_root_element.children() { match child { ChildOfElement::Comment(x) if is_typeholder_comment(x) => {} ChildOfElement::Element(x) if prev_is_typeholder_comment(x) => { if x.name().local_part() == "alias" { aliases.push(parse_alias(x, families)); } else if x.name().local_part() == "match" && x.attribute_value("target").unwrap_or("") == "scan" { match update_family(x, families) { Some(y) => scan_matches.push(y), _ => {} } } } x => new_root_element.append_child(x), } } // replace old_root_element with new_root_element doc.root().append_child(new_root_element); } Config { scan_matches: scan_matches, aliases: aliases, residue: config_package, } } fn prev_is_typeholder_comment(x: Element) -> bool { match x.preceding_siblings().last() { Some(y) => { match y.comment() { Some(z) => is_typeholder_comment(z), None => false, } } None => false, } } fn is_typeholder_comment(x: Comment) -> bool { x.text().starts_with(TYPEHOLDER_COMMENT_PREFIX) } fn update_family<'a>( e: Element, families: &'a Vec<RefCell<Family>>, ) -> Option<&'a RefCell<Family>> { let family_name = checked_text(checked_child_element( "string", checked_child_element("test", e), )).text(); let matched_family = families.iter().find(|x| x.borrow().name == family_name); if matched_family.is_some() { let nil_range_template = ("nil", "Custom"); let mut current_range_templates = nil_range_template; let charset_elem = checked_child_element( "charset", checked_child_element("minus", checked_child_element("edit", e)), ); let ranges = charset_elem .children() .into_iter() .group_by(|x| match x { &ChildOfElement::Comment(y) => { current_range_templates = y.text() .splitn(2, ',') .map(str::trim) .next_tuple::<(_, _)>() .expect(INVALID_CONFIG); current_range_templates } &ChildOfElement::Element(y) if y.name().local_part() == "range" => { current_range_templates } _ => nil_range_template, }) .into_iter() .map(|(k, group)| { ( k, group .filter_map(|child| child.element()) .filter(|elem| elem.name().local_part() == "range") .map(|range_elem| { children_element("int", range_elem) .map(|int_elem| { i32::from_str_radix(&checked_text(int_elem).text()[2..], 16) .expect(INVALID_CONFIG) }) .next_tuple::<(_, _)>() .expect(INVALID_CONFIG) }) .collect_vec(), ) }) .filter(|&(_, ref code_points)| !code_points.is_empty()) .coalesce(|mut r0, mut r1| if r0.0 == r1.0 { r0.1.append(&mut r1.1); Ok(r0) } else { Err((r0, r1)) }) .map(|(k, code_points)| match k.1 { "Block" => Range::Block { name: String::from(k.0), code_points: code_points[0], }, "Script" => Range::Script { name: String::from(k.0), code_points: code_points, }, _ => Range::Custom { name: String::from(k.0), code_points: code_points[0], }, }) .collect_vec(); matched_family .unwrap() .borrow_mut() .deref_mut() .stripped_ranges = ranges; } matched_family } fn parse_alias<'a>(e: Element, families: &'a Vec<RefCell<Family>>) -> Alias<'a> { let alias_name = checked_text(checked_child_element("family", e)).text(); let p_list = children_element("family", checked_child_element("prefer", e)) .filter_map(|x| { families.iter().find(|y| { y.borrow().deref().name == checked_text(x).text() }) }) .collect_vec(); Alias { name: String::from(alias_name), prefer_list: p_list, } } fn checked_child_element<'a: 'd, 'd>(name: &'a str, e: Element<'d>) -> Element<'d>
fn child_element<'a: 'd, 'd>(name: &'a str, e: Element<'d>) -> Option<Element<'d>> { children_element(name, e).next() } fn children_element<'a: 'd, 'd>( name: &'a str, e: Element<'d>, ) -> impl Iterator<Item = Element<'d>> + 'd { e.children() .into_iter() .filter_map(|x| x.element()) .filter(move |x| x.name().local_part() == name) } fn checked_text<'d>(e: Element<'d>) -> Text<'d> { text(e).expect(&format!("Element {} has no text!", e.name().local_part())) } fn text<'d>(e: Element<'d>) -> Option<Text<'d>> { e.children().into_iter().filter_map(|x| x.text()).next() }
{ child_element(name, e).expect(&format!( "Element {} has no {} child!", e.name().local_part(), name )) }
identifier_body
deserialization.rs
use alias::Alias; use config::Config; use consts::*; use itertools::Itertools; use family::Family; use glib::functions; use pango::Context; use pango::ContextExt; use pango::FontMapExt; use pango::FontFamilyExt; use range::Range; use sxd_document::dom::Comment; use sxd_document::dom::ChildOfElement; use sxd_document::dom::Element; use sxd_document::dom::Text; use sxd_document::parser; use std::cell::RefCell; use std::fs::File; use std::i32; use std::io::Read; use std::ops::Deref; use std::ops::DerefMut; pub fn list_families(context: &Context) -> Vec<RefCell<Family>> { match context.get_font_map() { Some(map) => { map.list_families() .iter() .filter_map(|x| x.get_name()) .filter(|x| !["Sans", "Serif", "Monospace"].contains(&x.as_str())) .map(|x| { RefCell::new(Family { name: x, stripped_ranges: vec![], }) }) .collect() } None => vec![], } } pub fn parse_or_default<'a>(families: &'a Vec<RefCell<Family>>) -> Config<'a> { let fc_config_path = functions::get_user_config_dir() .expect("$XDG_CONFIG_HOME not set!") .join("fontconfig/fonts.conf"); let config_parse = match File::open(fc_config_path.as_path()) { Ok(mut f) => { let mut buffer = String::new(); f.read_to_string(&mut buffer).expect( "Failed to parse your fonts.conf file", ); parser::parse(&buffer) } _ => parser::parse(DEFAULT_FONTS_CONF), }; let config_package = match config_parse { Ok(package) => package, Err((_, errors)) => panic!("Error parsing fonts.conf!\n{}", errors.iter().join("\n")), }; // scan matches collection let mut scan_matches: Vec<&'a RefCell<Family>> = vec![]; // aliases collection let mut aliases: Vec<Alias<'a>> = vec![]; { let doc = config_package.as_document(); let old_root_element = doc.root().children()[0].element().expect(INVALID_CONFIG); // rest of dom collection let new_root_element = doc.create_element(old_root_element.name()); for attr in old_root_element.attributes() { new_root_element.set_attribute_value(attr.name(), attr.value()); } // group children to correct collections for child in old_root_element.children() { match child { ChildOfElement::Comment(x) if is_typeholder_comment(x) => {} ChildOfElement::Element(x) if prev_is_typeholder_comment(x) => { if x.name().local_part() == "alias" { aliases.push(parse_alias(x, families)); } else if x.name().local_part() == "match" && x.attribute_value("target").unwrap_or("") == "scan" { match update_family(x, families) { Some(y) => scan_matches.push(y), _ => {} } } } x => new_root_element.append_child(x), } } // replace old_root_element with new_root_element doc.root().append_child(new_root_element); } Config { scan_matches: scan_matches, aliases: aliases, residue: config_package, } } fn prev_is_typeholder_comment(x: Element) -> bool { match x.preceding_siblings().last() { Some(y) => { match y.comment() { Some(z) => is_typeholder_comment(z), None => false, } } None => false, } } fn is_typeholder_comment(x: Comment) -> bool { x.text().starts_with(TYPEHOLDER_COMMENT_PREFIX) } fn update_family<'a>( e: Element, families: &'a Vec<RefCell<Family>>, ) -> Option<&'a RefCell<Family>> { let family_name = checked_text(checked_child_element( "string", checked_child_element("test", e), )).text(); let matched_family = families.iter().find(|x| x.borrow().name == family_name); if matched_family.is_some() { let nil_range_template = ("nil", "Custom"); let mut current_range_templates = nil_range_template; let charset_elem = checked_child_element( "charset", checked_child_element("minus", checked_child_element("edit", e)), ); let ranges = charset_elem .children() .into_iter() .group_by(|x| match x { &ChildOfElement::Comment(y) => { current_range_templates = y.text() .splitn(2, ',') .map(str::trim) .next_tuple::<(_, _)>() .expect(INVALID_CONFIG); current_range_templates } &ChildOfElement::Element(y) if y.name().local_part() == "range" => { current_range_templates } _ => nil_range_template, }) .into_iter() .map(|(k, group)| { ( k, group .filter_map(|child| child.element()) .filter(|elem| elem.name().local_part() == "range") .map(|range_elem| { children_element("int", range_elem) .map(|int_elem| { i32::from_str_radix(&checked_text(int_elem).text()[2..], 16) .expect(INVALID_CONFIG) }) .next_tuple::<(_, _)>() .expect(INVALID_CONFIG) }) .collect_vec(), ) }) .filter(|&(_, ref code_points)| !code_points.is_empty()) .coalesce(|mut r0, mut r1| if r0.0 == r1.0 { r0.1.append(&mut r1.1); Ok(r0) } else { Err((r0, r1)) }) .map(|(k, code_points)| match k.1 { "Block" => Range::Block { name: String::from(k.0), code_points: code_points[0], }, "Script" => Range::Script { name: String::from(k.0), code_points: code_points, }, _ => Range::Custom { name: String::from(k.0), code_points: code_points[0], }, })
.stripped_ranges = ranges; } matched_family } fn parse_alias<'a>(e: Element, families: &'a Vec<RefCell<Family>>) -> Alias<'a> { let alias_name = checked_text(checked_child_element("family", e)).text(); let p_list = children_element("family", checked_child_element("prefer", e)) .filter_map(|x| { families.iter().find(|y| { y.borrow().deref().name == checked_text(x).text() }) }) .collect_vec(); Alias { name: String::from(alias_name), prefer_list: p_list, } } fn checked_child_element<'a: 'd, 'd>(name: &'a str, e: Element<'d>) -> Element<'d> { child_element(name, e).expect(&format!( "Element {} has no {} child!", e.name().local_part(), name )) } fn child_element<'a: 'd, 'd>(name: &'a str, e: Element<'d>) -> Option<Element<'d>> { children_element(name, e).next() } fn children_element<'a: 'd, 'd>( name: &'a str, e: Element<'d>, ) -> impl Iterator<Item = Element<'d>> + 'd { e.children() .into_iter() .filter_map(|x| x.element()) .filter(move |x| x.name().local_part() == name) } fn checked_text<'d>(e: Element<'d>) -> Text<'d> { text(e).expect(&format!("Element {} has no text!", e.name().local_part())) } fn text<'d>(e: Element<'d>) -> Option<Text<'d>> { e.children().into_iter().filter_map(|x| x.text()).next() }
.collect_vec(); matched_family .unwrap() .borrow_mut() .deref_mut()
random_line_split
deserialization.rs
use alias::Alias; use config::Config; use consts::*; use itertools::Itertools; use family::Family; use glib::functions; use pango::Context; use pango::ContextExt; use pango::FontMapExt; use pango::FontFamilyExt; use range::Range; use sxd_document::dom::Comment; use sxd_document::dom::ChildOfElement; use sxd_document::dom::Element; use sxd_document::dom::Text; use sxd_document::parser; use std::cell::RefCell; use std::fs::File; use std::i32; use std::io::Read; use std::ops::Deref; use std::ops::DerefMut; pub fn
(context: &Context) -> Vec<RefCell<Family>> { match context.get_font_map() { Some(map) => { map.list_families() .iter() .filter_map(|x| x.get_name()) .filter(|x| !["Sans", "Serif", "Monospace"].contains(&x.as_str())) .map(|x| { RefCell::new(Family { name: x, stripped_ranges: vec![], }) }) .collect() } None => vec![], } } pub fn parse_or_default<'a>(families: &'a Vec<RefCell<Family>>) -> Config<'a> { let fc_config_path = functions::get_user_config_dir() .expect("$XDG_CONFIG_HOME not set!") .join("fontconfig/fonts.conf"); let config_parse = match File::open(fc_config_path.as_path()) { Ok(mut f) => { let mut buffer = String::new(); f.read_to_string(&mut buffer).expect( "Failed to parse your fonts.conf file", ); parser::parse(&buffer) } _ => parser::parse(DEFAULT_FONTS_CONF), }; let config_package = match config_parse { Ok(package) => package, Err((_, errors)) => panic!("Error parsing fonts.conf!\n{}", errors.iter().join("\n")), }; // scan matches collection let mut scan_matches: Vec<&'a RefCell<Family>> = vec![]; // aliases collection let mut aliases: Vec<Alias<'a>> = vec![]; { let doc = config_package.as_document(); let old_root_element = doc.root().children()[0].element().expect(INVALID_CONFIG); // rest of dom collection let new_root_element = doc.create_element(old_root_element.name()); for attr in old_root_element.attributes() { new_root_element.set_attribute_value(attr.name(), attr.value()); } // group children to correct collections for child in old_root_element.children() { match child { ChildOfElement::Comment(x) if is_typeholder_comment(x) => {} ChildOfElement::Element(x) if prev_is_typeholder_comment(x) => { if x.name().local_part() == "alias" { aliases.push(parse_alias(x, families)); } else if x.name().local_part() == "match" && x.attribute_value("target").unwrap_or("") == "scan" { match update_family(x, families) { Some(y) => scan_matches.push(y), _ => {} } } } x => new_root_element.append_child(x), } } // replace old_root_element with new_root_element doc.root().append_child(new_root_element); } Config { scan_matches: scan_matches, aliases: aliases, residue: config_package, } } fn prev_is_typeholder_comment(x: Element) -> bool { match x.preceding_siblings().last() { Some(y) => { match y.comment() { Some(z) => is_typeholder_comment(z), None => false, } } None => false, } } fn is_typeholder_comment(x: Comment) -> bool { x.text().starts_with(TYPEHOLDER_COMMENT_PREFIX) } fn update_family<'a>( e: Element, families: &'a Vec<RefCell<Family>>, ) -> Option<&'a RefCell<Family>> { let family_name = checked_text(checked_child_element( "string", checked_child_element("test", e), )).text(); let matched_family = families.iter().find(|x| x.borrow().name == family_name); if matched_family.is_some() { let nil_range_template = ("nil", "Custom"); let mut current_range_templates = nil_range_template; let charset_elem = checked_child_element( "charset", checked_child_element("minus", checked_child_element("edit", e)), ); let ranges = charset_elem .children() .into_iter() .group_by(|x| match x { &ChildOfElement::Comment(y) => { current_range_templates = y.text() .splitn(2, ',') .map(str::trim) .next_tuple::<(_, _)>() .expect(INVALID_CONFIG); current_range_templates } &ChildOfElement::Element(y) if y.name().local_part() == "range" => { current_range_templates } _ => nil_range_template, }) .into_iter() .map(|(k, group)| { ( k, group .filter_map(|child| child.element()) .filter(|elem| elem.name().local_part() == "range") .map(|range_elem| { children_element("int", range_elem) .map(|int_elem| { i32::from_str_radix(&checked_text(int_elem).text()[2..], 16) .expect(INVALID_CONFIG) }) .next_tuple::<(_, _)>() .expect(INVALID_CONFIG) }) .collect_vec(), ) }) .filter(|&(_, ref code_points)| !code_points.is_empty()) .coalesce(|mut r0, mut r1| if r0.0 == r1.0 { r0.1.append(&mut r1.1); Ok(r0) } else { Err((r0, r1)) }) .map(|(k, code_points)| match k.1 { "Block" => Range::Block { name: String::from(k.0), code_points: code_points[0], }, "Script" => Range::Script { name: String::from(k.0), code_points: code_points, }, _ => Range::Custom { name: String::from(k.0), code_points: code_points[0], }, }) .collect_vec(); matched_family .unwrap() .borrow_mut() .deref_mut() .stripped_ranges = ranges; } matched_family } fn parse_alias<'a>(e: Element, families: &'a Vec<RefCell<Family>>) -> Alias<'a> { let alias_name = checked_text(checked_child_element("family", e)).text(); let p_list = children_element("family", checked_child_element("prefer", e)) .filter_map(|x| { families.iter().find(|y| { y.borrow().deref().name == checked_text(x).text() }) }) .collect_vec(); Alias { name: String::from(alias_name), prefer_list: p_list, } } fn checked_child_element<'a: 'd, 'd>(name: &'a str, e: Element<'d>) -> Element<'d> { child_element(name, e).expect(&format!( "Element {} has no {} child!", e.name().local_part(), name )) } fn child_element<'a: 'd, 'd>(name: &'a str, e: Element<'d>) -> Option<Element<'d>> { children_element(name, e).next() } fn children_element<'a: 'd, 'd>( name: &'a str, e: Element<'d>, ) -> impl Iterator<Item = Element<'d>> + 'd { e.children() .into_iter() .filter_map(|x| x.element()) .filter(move |x| x.name().local_part() == name) } fn checked_text<'d>(e: Element<'d>) -> Text<'d> { text(e).expect(&format!("Element {} has no text!", e.name().local_part())) } fn text<'d>(e: Element<'d>) -> Option<Text<'d>> { e.children().into_iter().filter_map(|x| x.text()).next() }
list_families
identifier_name
email_test.py
import bountyfunding from bountyfunding.core.const import * from bountyfunding.core.data import clean_database from test import to_object from nose.tools import * USER = "bountyfunding" class Email_Test: def setup(self): self.app = bountyfunding.app.test_client() clean_database() def test_email(self): eq_(len(self.get_emails()), 0) r = self.app.post('/issues', data=dict(ref=1, status='READY', title='Title', link='/issue/1')) eq_(r.status_code, 200) r = self.app.post('/issue/1/sponsorships', data=dict(user=USER, amount=10)) eq_(r.status_code, 200)
r = self.app.get("/issue/1") eq_(r.status_code, 200) r = self.app.put('/issue/1', data=dict( status=IssueStatus.to_string(IssueStatus.STARTED))) eq_(r.status_code, 200) emails = self.get_emails() eq_(len(emails), 1) email = emails[0] eq_(email.recipient, USER) ok_(email.issue_id) ok_(email.body) r = self.app.delete("/email/%s" % email.id) eq_(r.status_code, 200) def get_emails(self): r = self.app.get("/emails") eq_(r.status_code, 200) return to_object(r).data
random_line_split
email_test.py
import bountyfunding from bountyfunding.core.const import * from bountyfunding.core.data import clean_database from test import to_object from nose.tools import * USER = "bountyfunding" class Email_Test: def setup(self): self.app = bountyfunding.app.test_client() clean_database() def
(self): eq_(len(self.get_emails()), 0) r = self.app.post('/issues', data=dict(ref=1, status='READY', title='Title', link='/issue/1')) eq_(r.status_code, 200) r = self.app.post('/issue/1/sponsorships', data=dict(user=USER, amount=10)) eq_(r.status_code, 200) r = self.app.get("/issue/1") eq_(r.status_code, 200) r = self.app.put('/issue/1', data=dict( status=IssueStatus.to_string(IssueStatus.STARTED))) eq_(r.status_code, 200) emails = self.get_emails() eq_(len(emails), 1) email = emails[0] eq_(email.recipient, USER) ok_(email.issue_id) ok_(email.body) r = self.app.delete("/email/%s" % email.id) eq_(r.status_code, 200) def get_emails(self): r = self.app.get("/emails") eq_(r.status_code, 200) return to_object(r).data
test_email
identifier_name
email_test.py
import bountyfunding from bountyfunding.core.const import * from bountyfunding.core.data import clean_database from test import to_object from nose.tools import * USER = "bountyfunding" class Email_Test:
eq_(r.status_code, 200) emails = self.get_emails() eq_(len(emails), 1) email = emails[0] eq_(email.recipient, USER) ok_(email.issue_id) ok_(email.body) r = self.app.delete("/email/%s" % email.id) eq_(r.status_code, 200) def get_emails(self): r = self.app.get("/emails") eq_(r.status_code, 200) return to_object(r).data
def setup(self): self.app = bountyfunding.app.test_client() clean_database() def test_email(self): eq_(len(self.get_emails()), 0) r = self.app.post('/issues', data=dict(ref=1, status='READY', title='Title', link='/issue/1')) eq_(r.status_code, 200) r = self.app.post('/issue/1/sponsorships', data=dict(user=USER, amount=10)) eq_(r.status_code, 200) r = self.app.get("/issue/1") eq_(r.status_code, 200) r = self.app.put('/issue/1', data=dict( status=IssueStatus.to_string(IssueStatus.STARTED)))
identifier_body