file_name
large_stringlengths
4
140
prefix
large_stringlengths
0
12.1k
suffix
large_stringlengths
0
12k
middle
large_stringlengths
0
7.51k
fim_type
large_stringclasses
4 values
url-path.ts
export class UrlPath { static ADMIN = 'admin'; static FILTERED_MAP = 'filteredMap'; static INSPECTOR = 'inspector'; static MAIN = 'main'; static REAL_TIME = 'realtime'; static SCATTER_FULL_SCREEN_MODE = 'scatterFullScreenMode'; static THREAD_DUMP = 'threadDump'; static TRANSACTION_DETAIL = 'transactionDetail'; static TRANSACTION_LIST = 'transactionList'; static TRANSACTION_VIEW = 'transactionView'; static BROWSER_NOT_SUPPORT = 'browserNotSupported'; static ERROR = 'error'; static CONFIG = 'config'; constructor() {} static
(): string[] { return [ UrlPath.CONFIG, UrlPath.ADMIN, UrlPath.ERROR, UrlPath.FILTERED_MAP, UrlPath.INSPECTOR, UrlPath.MAIN, UrlPath.REAL_TIME, UrlPath.SCATTER_FULL_SCREEN_MODE, UrlPath.THREAD_DUMP, UrlPath.TRANSACTION_DETAIL, UrlPath.TRANSACTION_LIST, UrlPath.TRANSACTION_VIEW ]; } } export default UrlPath;
getParamList
identifier_name
url-path.ts
export class UrlPath { static ADMIN = 'admin'; static FILTERED_MAP = 'filteredMap'; static INSPECTOR = 'inspector'; static MAIN = 'main'; static REAL_TIME = 'realtime'; static SCATTER_FULL_SCREEN_MODE = 'scatterFullScreenMode'; static THREAD_DUMP = 'threadDump'; static TRANSACTION_DETAIL = 'transactionDetail'; static TRANSACTION_LIST = 'transactionList'; static TRANSACTION_VIEW = 'transactionView'; static BROWSER_NOT_SUPPORT = 'browserNotSupported'; static ERROR = 'error'; static CONFIG = 'config'; constructor() {}
UrlPath.ADMIN, UrlPath.ERROR, UrlPath.FILTERED_MAP, UrlPath.INSPECTOR, UrlPath.MAIN, UrlPath.REAL_TIME, UrlPath.SCATTER_FULL_SCREEN_MODE, UrlPath.THREAD_DUMP, UrlPath.TRANSACTION_DETAIL, UrlPath.TRANSACTION_LIST, UrlPath.TRANSACTION_VIEW ]; } } export default UrlPath;
static getParamList(): string[] { return [ UrlPath.CONFIG,
random_line_split
url-path.ts
export class UrlPath { static ADMIN = 'admin'; static FILTERED_MAP = 'filteredMap'; static INSPECTOR = 'inspector'; static MAIN = 'main'; static REAL_TIME = 'realtime'; static SCATTER_FULL_SCREEN_MODE = 'scatterFullScreenMode'; static THREAD_DUMP = 'threadDump'; static TRANSACTION_DETAIL = 'transactionDetail'; static TRANSACTION_LIST = 'transactionList'; static TRANSACTION_VIEW = 'transactionView'; static BROWSER_NOT_SUPPORT = 'browserNotSupported'; static ERROR = 'error'; static CONFIG = 'config'; constructor()
static getParamList(): string[] { return [ UrlPath.CONFIG, UrlPath.ADMIN, UrlPath.ERROR, UrlPath.FILTERED_MAP, UrlPath.INSPECTOR, UrlPath.MAIN, UrlPath.REAL_TIME, UrlPath.SCATTER_FULL_SCREEN_MODE, UrlPath.THREAD_DUMP, UrlPath.TRANSACTION_DETAIL, UrlPath.TRANSACTION_LIST, UrlPath.TRANSACTION_VIEW ]; } } export default UrlPath;
{}
identifier_body
db.py
'payload' in kwargs: _request = kwargs['payload'].request_body _context = kwargs['payload'].context _network = kwargs['payload'].desired_state _orig = kwargs['payload'].states[0] else: _request = kwargs['request'] _context = kwargs['context'] _network = kwargs['network'] _orig = kwargs.get('original_network') @db_api.retry_if_session_inactive() def _do_ensure_external_network_default_value_callback( context, request, orig, network): is_default = request.get(api_const.IS_DEFAULT) if is_default is None: return if is_default: # ensure only one default external network at any given time pager = base_obj.Pager(limit=1) objs = net_obj.ExternalNetwork.get_objects(context, _pager=pager, is_default=True) if objs: if objs[0] and network['id'] != objs[0].network_id: raise exceptions.DefaultExternalNetworkExists( net_id=objs[0].network_id) if orig and orig.get(api_const.IS_DEFAULT) == is_default: return network[api_const.IS_DEFAULT] = is_default # Reflect the status of the is_default on the create/update request obj = net_obj.ExternalNetwork.get_object(context, network_id=network['id']) if obj: obj.is_default = is_default obj.update() _do_ensure_external_network_default_value_callback( _context, _request, _orig, _network) @resource_extend.has_resource_extenders class AutoAllocatedTopologyMixin(common_db_mixin.CommonDbMixin): def __new__(cls, *args, **kwargs): # NOTE(kevinbenton): we subscribe on object construction because # the tests blow away the callback manager for each run new = super(AutoAllocatedTopologyMixin, cls).__new__(cls, *args, **kwargs) registry.subscribe(_ensure_external_network_default_value_callback, resources.NETWORK, events.PRECOMMIT_UPDATE) registry.subscribe(_ensure_external_network_default_value_callback, resources.NETWORK, events.PRECOMMIT_CREATE) return new # TODO(armax): if a tenant modifies auto allocated resources under # the hood the behavior of the get_auto_allocated_topology API is # undetermined. Consider adding callbacks to deal with the following # situations: # - insert subnet -> plug router interface # - delete router -> remove the entire topology # - update subnet -> prevent operation # - update router gateway -> prevent operation # - ... @property def core_plugin(self): if not getattr(self, '_core_plugin', None): self._core_plugin = directory.get_plugin() return self._core_plugin @property def l3_plugin(self): if not getattr(self, '_l3_plugin', None): self._l3_plugin = directory.get_plugin(constants.L3) return self._l3_plugin @staticmethod @resource_extend.extends([net_def.COLLECTION_NAME]) def _extend_external_network_default(net_res, net_db): """Add is_default field to 'show' response.""" if net_db.external is not None: net_res[api_const.IS_DEFAULT] = net_db.external.is_default return net_res def get_auto_allocated_topology(self, context, tenant_id, fields=None): """Return tenant's network associated to auto-allocated topology. The topology will be provisioned upon return, if network is missing. """ fields = fields or [] tenant_id = self._validate(context, tenant_id) if CHECK_REQUIREMENTS in fields: # for dry-run requests, simply validates that subsequent # requests can be fulfilled based on a set of requirements # such as existence of default networks, pools, etc. return self._check_requirements(context, tenant_id) elif fields: raise n_exc.BadRequest(resource='auto_allocate', msg=_("Unrecognized field")) # Check for an existent topology network_id = self._get_auto_allocated_network(context, tenant_id) if network_id: return self._response(network_id, tenant_id, fields=fields) # See if we indeed have an external network to connect to, otherwise # we will fail fast default_external_network = self._get_default_external_network( context) # If we reach this point, then we got some work to do! network_id = self._build_topology( context, tenant_id, default_external_network) return self._response(network_id, tenant_id, fields=fields) def
(self, context, tenant_id): tenant_id = self._validate(context, tenant_id) topology = self._get_auto_allocated_topology(context, tenant_id) if topology: subnets = self.core_plugin.get_subnets( context, filters={'network_id': [topology['network_id']]}) self._cleanup( context, network_id=topology['network_id'], router_id=topology['router_id'], subnets=subnets) def _build_topology(self, context, tenant_id, default_external_network): """Build the network topology and returns its network UUID.""" try: subnets = self._provision_tenant_private_network( context, tenant_id) network_id = subnets[0]['network_id'] router = self._provision_external_connectivity( context, default_external_network, subnets, tenant_id) network_id = self._save( context, tenant_id, network_id, router['id'], subnets) return network_id except exceptions.UnknownProvisioningError as e: # Clean partially provisioned topologies, and reraise the # error. If it can be retried, so be it. LOG.error("Unknown error while provisioning topology for " "tenant %(tenant_id)s. Reason: %(reason)s", {'tenant_id': tenant_id, 'reason': e}) self._cleanup( context, network_id=e.network_id, router_id=e.router_id, subnets=e.subnets) raise e.error def _check_requirements(self, context, tenant_id): """Raise if requirements are not met.""" self._get_default_external_network(context) try: self._get_supported_subnetpools(context) except n_exc.NotFound: raise exceptions.AutoAllocationFailure( reason=_("No default subnetpools defined")) return {'id': 'dry-run=pass', 'tenant_id': tenant_id} def _validate(self, context, tenant_id): """Validate and return the tenant to be associated to the topology.""" if tenant_id == 'None': # NOTE(HenryG): the client might be sending us astray by # passing no tenant; this is really meant to be the tenant # issuing the request, therefore let's get it from the context tenant_id = context.tenant_id if not context.is_admin and tenant_id != context.tenant_id: raise n_exc.NotAuthorized() return tenant_id def _get_auto_allocated_topology(self, context, tenant_id): """Return the auto allocated topology record if present or None.""" return auto_allocate_obj.AutoAllocatedTopology.get_object( context, project_id=tenant_id) def _get_auto_allocated_network(self, context, tenant_id): """Get the auto allocated network for the tenant.""" network = self._get_auto_allocated_topology(context, tenant_id) if network: return network['network_id'] @staticmethod def _response(network_id, tenant_id, fields=None): """Build response for auto-allocated network.""" res = { 'id': network_id, 'tenant_id': tenant_id } return db_utils.resource_fields(res, fields) def _get_default_external_network(self, context): """Get the default external network for the deployment.""" default_external_networks = net_obj.ExternalNetwork.get_objects( context, is_default=True) if not default_external_networks: LOG.error("Unable to find default external network " "for deployment, please create/assign one to " "allow auto-allocation to work correctly.") raise exceptions.AutoAllocationFailure( reason=_("No default router:external network")) if len(default_external_networks) > 1: LOG.error("Multiple external default networks detected. " "Network %s is true 'default'.", default_external_networks[0]['network_id']) return default_external_networks[0].network_id def _get_supported_subnetpools(self, context): """Return the default subnet pools available.""" default_subnet_pools = [ self.core_plugin.get_default_subnetpool( context, ver) for ver in (4, 6) ] available_pools = [ s for s in default_subnet_pools if s ] if not available_pools: LOG.error("No default pools available") raise n_exc.NotFound() return available_pools def _provision_tenant_private_network(self, context, tenant_id): """Create a tenant private network/subnets.""" network = None try: network_args = { 'name': 'auto_allocated_network', 'admin_state_up': False, 'tenant_id': tenant_id, 'shared': False } network = p_utils.create_network( self.core_plugin, context, {'network': network_args}) subnets = [] for pool in self._get_supported
delete_auto_allocated_topology
identifier_name
db.py
'payload' in kwargs: _request = kwargs['payload'].request_body _context = kwargs['payload'].context _network = kwargs['payload'].desired_state _orig = kwargs['payload'].states[0] else: _request = kwargs['request'] _context = kwargs['context'] _network = kwargs['network'] _orig = kwargs.get('original_network') @db_api.retry_if_session_inactive() def _do_ensure_external_network_default_value_callback( context, request, orig, network): is_default = request.get(api_const.IS_DEFAULT) if is_default is None: return if is_default: # ensure only one default external network at any given time pager = base_obj.Pager(limit=1) objs = net_obj.ExternalNetwork.get_objects(context, _pager=pager, is_default=True) if objs: if objs[0] and network['id'] != objs[0].network_id: raise exceptions.DefaultExternalNetworkExists( net_id=objs[0].network_id) if orig and orig.get(api_const.IS_DEFAULT) == is_default: return network[api_const.IS_DEFAULT] = is_default # Reflect the status of the is_default on the create/update request obj = net_obj.ExternalNetwork.get_object(context, network_id=network['id']) if obj: obj.is_default = is_default obj.update() _do_ensure_external_network_default_value_callback( _context, _request, _orig, _network) @resource_extend.has_resource_extenders class AutoAllocatedTopologyMixin(common_db_mixin.CommonDbMixin): def __new__(cls, *args, **kwargs): # NOTE(kevinbenton): we subscribe on object construction because # the tests blow away the callback manager for each run new = super(AutoAllocatedTopologyMixin, cls).__new__(cls, *args, **kwargs) registry.subscribe(_ensure_external_network_default_value_callback, resources.NETWORK, events.PRECOMMIT_UPDATE) registry.subscribe(_ensure_external_network_default_value_callback, resources.NETWORK, events.PRECOMMIT_CREATE) return new # TODO(armax): if a tenant modifies auto allocated resources under # the hood the behavior of the get_auto_allocated_topology API is # undetermined. Consider adding callbacks to deal with the following # situations: # - insert subnet -> plug router interface # - delete router -> remove the entire topology # - update subnet -> prevent operation # - update router gateway -> prevent operation # - ... @property def core_plugin(self): if not getattr(self, '_core_plugin', None): self._core_plugin = directory.get_plugin() return self._core_plugin @property def l3_plugin(self): if not getattr(self, '_l3_plugin', None): self._l3_plugin = directory.get_plugin(constants.L3) return self._l3_plugin @staticmethod @resource_extend.extends([net_def.COLLECTION_NAME]) def _extend_external_network_default(net_res, net_db): """Add is_default field to 'show' response.""" if net_db.external is not None: net_res[api_const.IS_DEFAULT] = net_db.external.is_default return net_res def get_auto_allocated_topology(self, context, tenant_id, fields=None): """Return tenant's network associated to auto-allocated topology. The topology will be provisioned upon return, if network is missing. """ fields = fields or [] tenant_id = self._validate(context, tenant_id) if CHECK_REQUIREMENTS in fields: # for dry-run requests, simply validates that subsequent # requests can be fulfilled based on a set of requirements # such as existence of default networks, pools, etc. return self._check_requirements(context, tenant_id) elif fields: raise n_exc.BadRequest(resource='auto_allocate', msg=_("Unrecognized field")) # Check for an existent topology network_id = self._get_auto_allocated_network(context, tenant_id) if network_id: return self._response(network_id, tenant_id, fields=fields) # See if we indeed have an external network to connect to, otherwise # we will fail fast default_external_network = self._get_default_external_network( context) # If we reach this point, then we got some work to do! network_id = self._build_topology( context, tenant_id, default_external_network) return self._response(network_id, tenant_id, fields=fields) def delete_auto_allocated_topology(self, context, tenant_id): tenant_id = self._validate(context, tenant_id) topology = self._get_auto_allocated_topology(context, tenant_id) if topology: subnets = self.core_plugin.get_subnets( context, filters={'network_id': [topology['network_id']]}) self._cleanup( context, network_id=topology['network_id'], router_id=topology['router_id'], subnets=subnets) def _build_topology(self, context, tenant_id, default_external_network): """Build the network topology and returns its network UUID.""" try: subnets = self._provision_tenant_private_network( context, tenant_id) network_id = subnets[0]['network_id'] router = self._provision_external_connectivity( context, default_external_network, subnets, tenant_id) network_id = self._save( context, tenant_id, network_id, router['id'], subnets) return network_id except exceptions.UnknownProvisioningError as e: # Clean partially provisioned topologies, and reraise the # error. If it can be retried, so be it. LOG.error("Unknown error while provisioning topology for " "tenant %(tenant_id)s. Reason: %(reason)s", {'tenant_id': tenant_id, 'reason': e}) self._cleanup( context, network_id=e.network_id, router_id=e.router_id, subnets=e.subnets) raise e.error def _check_requirements(self, context, tenant_id): """Raise if requirements are not met.""" self._get_default_external_network(context) try: self._get_supported_subnetpools(context) except n_exc.NotFound: raise exceptions.AutoAllocationFailure( reason=_("No default subnetpools defined")) return {'id': 'dry-run=pass', 'tenant_id': tenant_id} def _validate(self, context, tenant_id): """Validate and return the tenant to be associated to the topology.""" if tenant_id == 'None': # NOTE(HenryG): the client might be sending us astray by # passing no tenant; this is really meant to be the tenant # issuing the request, therefore let's get it from the context tenant_id = context.tenant_id if not context.is_admin and tenant_id != context.tenant_id: raise n_exc.NotAuthorized() return tenant_id def _get_auto_allocated_topology(self, context, tenant_id): """Return the auto allocated topology record if present or None.""" return auto_allocate_obj.AutoAllocatedTopology.get_object( context, project_id=tenant_id) def _get_auto_allocated_network(self, context, tenant_id): """Get the auto allocated network for the tenant.""" network = self._get_auto_allocated_topology(context, tenant_id) if network: return network['network_id'] @staticmethod def _response(network_id, tenant_id, fields=None): """Build response for auto-allocated network.""" res = { 'id': network_id, 'tenant_id': tenant_id } return db_utils.resource_fields(res, fields) def _get_default_external_network(self, context): """Get the default external network for the deployment.""" default_external_networks = net_obj.ExternalNetwork.get_objects( context, is_default=True) if not default_external_networks:
if len(default_external_networks) > 1: LOG.error("Multiple external default networks detected. " "Network %s is true 'default'.", default_external_networks[0]['network_id']) return default_external_networks[0].network_id def _get_supported_subnetpools(self, context): """Return the default subnet pools available.""" default_subnet_pools = [ self.core_plugin.get_default_subnetpool( context, ver) for ver in (4, 6) ] available_pools = [ s for s in default_subnet_pools if s ] if not available_pools: LOG.error("No default pools available") raise n_exc.NotFound() return available_pools def _provision_tenant_private_network(self, context, tenant_id): """Create a tenant private network/subnets.""" network = None try: network_args = { 'name': 'auto_allocated_network', 'admin_state_up': False, 'tenant_id': tenant_id, 'shared': False } network = p_utils.create_network( self.core_plugin, context, {'network': network_args}) subnets = [] for pool in self._get_supported
LOG.error("Unable to find default external network " "for deployment, please create/assign one to " "allow auto-allocation to work correctly.") raise exceptions.AutoAllocationFailure( reason=_("No default router:external network"))
conditional_block
db.py
if 'payload' in kwargs: _request = kwargs['payload'].request_body _context = kwargs['payload'].context _network = kwargs['payload'].desired_state _orig = kwargs['payload'].states[0] else: _request = kwargs['request'] _context = kwargs['context'] _network = kwargs['network'] _orig = kwargs.get('original_network') @db_api.retry_if_session_inactive() def _do_ensure_external_network_default_value_callback( context, request, orig, network): is_default = request.get(api_const.IS_DEFAULT) if is_default is None: return if is_default: # ensure only one default external network at any given time pager = base_obj.Pager(limit=1) objs = net_obj.ExternalNetwork.get_objects(context, _pager=pager, is_default=True) if objs: if objs[0] and network['id'] != objs[0].network_id: raise exceptions.DefaultExternalNetworkExists( net_id=objs[0].network_id) if orig and orig.get(api_const.IS_DEFAULT) == is_default: return network[api_const.IS_DEFAULT] = is_default # Reflect the status of the is_default on the create/update request obj = net_obj.ExternalNetwork.get_object(context, network_id=network['id']) if obj: obj.is_default = is_default obj.update() _do_ensure_external_network_default_value_callback( _context, _request, _orig, _network) @resource_extend.has_resource_extenders class AutoAllocatedTopologyMixin(common_db_mixin.CommonDbMixin): def __new__(cls, *args, **kwargs): # NOTE(kevinbenton): we subscribe on object construction because # the tests blow away the callback manager for each run new = super(AutoAllocatedTopologyMixin, cls).__new__(cls, *args, **kwargs) registry.subscribe(_ensure_external_network_default_value_callback, resources.NETWORK, events.PRECOMMIT_UPDATE) registry.subscribe(_ensure_external_network_default_value_callback, resources.NETWORK, events.PRECOMMIT_CREATE) return new # TODO(armax): if a tenant modifies auto allocated resources under # the hood the behavior of the get_auto_allocated_topology API is # undetermined. Consider adding callbacks to deal with the following # situations: # - insert subnet -> plug router interface # - delete router -> remove the entire topology # - update subnet -> prevent operation # - update router gateway -> prevent operation # - ... @property def core_plugin(self): if not getattr(self, '_core_plugin', None): self._core_plugin = directory.get_plugin() return self._core_plugin @property def l3_plugin(self): if not getattr(self, '_l3_plugin', None): self._l3_plugin = directory.get_plugin(constants.L3) return self._l3_plugin @staticmethod @resource_extend.extends([net_def.COLLECTION_NAME]) def _extend_external_network_default(net_res, net_db): """Add is_default field to 'show' response.""" if net_db.external is not None: net_res[api_const.IS_DEFAULT] = net_db.external.is_default return net_res def get_auto_allocated_topology(self, context, tenant_id, fields=None): """Return tenant's network associated to auto-allocated topology. The topology will be provisioned upon return, if network is missing. """ fields = fields or [] tenant_id = self._validate(context, tenant_id) if CHECK_REQUIREMENTS in fields: # for dry-run requests, simply validates that subsequent # requests can be fulfilled based on a set of requirements # such as existence of default networks, pools, etc. return self._check_requirements(context, tenant_id) elif fields: raise n_exc.BadRequest(resource='auto_allocate', msg=_("Unrecognized field")) # Check for an existent topology network_id = self._get_auto_allocated_network(context, tenant_id) if network_id: return self._response(network_id, tenant_id, fields=fields) # See if we indeed have an external network to connect to, otherwise # we will fail fast default_external_network = self._get_default_external_network( context) # If we reach this point, then we got some work to do! network_id = self._build_topology( context, tenant_id, default_external_network) return self._response(network_id, tenant_id, fields=fields) def delete_auto_allocated_topology(self, context, tenant_id): tenant_id = self._validate(context, tenant_id) topology = self._get_auto_allocated_topology(context, tenant_id) if topology: subnets = self.core_plugin.get_subnets( context, filters={'network_id': [topology['network_id']]}) self._cleanup( context, network_id=topology['network_id'], router_id=topology['router_id'], subnets=subnets) def _build_topology(self, context, tenant_id, default_external_network): """Build the network topology and returns its network UUID.""" try: subnets = self._provision_tenant_private_network( context, tenant_id) network_id = subnets[0]['network_id'] router = self._provision_external_connectivity( context, default_external_network, subnets, tenant_id) network_id = self._save( context, tenant_id, network_id, router['id'], subnets) return network_id except exceptions.UnknownProvisioningError as e: # Clean partially provisioned topologies, and reraise the # error. If it can be retried, so be it. LOG.error("Unknown error while provisioning topology for " "tenant %(tenant_id)s. Reason: %(reason)s", {'tenant_id': tenant_id, 'reason': e}) self._cleanup( context, network_id=e.network_id, router_id=e.router_id, subnets=e.subnets) raise e.error def _check_requirements(self, context, tenant_id): """Raise if requirements are not met.""" self._get_default_external_network(context) try: self._get_supported_subnetpools(context) except n_exc.NotFound: raise exceptions.AutoAllocationFailure( reason=_("No default subnetpools defined")) return {'id': 'dry-run=pass', 'tenant_id': tenant_id} def _validate(self, context, tenant_id): """Validate and return the tenant to be associated to the topology.""" if tenant_id == 'None': # NOTE(HenryG): the client might be sending us astray by # passing no tenant; this is really meant to be the tenant # issuing the request, therefore let's get it from the context tenant_id = context.tenant_id if not context.is_admin and tenant_id != context.tenant_id: raise n_exc.NotAuthorized() return tenant_id def _get_auto_allocated_topology(self, context, tenant_id): """Return the auto allocated topology record if present or None.""" return auto_allocate_obj.AutoAllocatedTopology.get_object( context, project_id=tenant_id) def _get_auto_allocated_network(self, context, tenant_id): """Get the auto allocated network for the tenant.""" network = self._get_auto_allocated_topology(context, tenant_id) if network: return network['network_id'] @staticmethod def _response(network_id, tenant_id, fields=None): """Build response for auto-allocated network.""" res = { 'id': network_id, 'tenant_id': tenant_id } return db_utils.resource_fields(res, fields) def _get_default_external_network(self, context): """Get the default external network for the deployment.""" default_external_networks = net_obj.ExternalNetwork.get_objects( context, is_default=True) if not default_external_networks: LOG.error("Unable to find default external network " "for deployment, please create/assign one to " "allow auto-allocation to work correctly.") raise exceptions.AutoAllocationFailure( reason=_("No default router:external network")) if len(default_external_networks) > 1: LOG.error("Multiple external default networks detected. " "Network %s is true 'default'.", default_external_networks[0]['network_id']) return default_external_networks[0].network_id def _get_supported_subnetpools(self, context):
def _provision_tenant_private_network(self, context, tenant_id): """Create a tenant private network/subnets.""" network = None try: network_args = { 'name': 'auto_allocated_network', 'admin_state_up': False, 'tenant_id': tenant_id, 'shared': False } network = p_utils.create_network( self.core_plugin, context, {'network': network_args}) subnets = [] for pool in self._get_supported
"""Return the default subnet pools available.""" default_subnet_pools = [ self.core_plugin.get_default_subnetpool( context, ver) for ver in (4, 6) ] available_pools = [ s for s in default_subnet_pools if s ] if not available_pools: LOG.error("No default pools available") raise n_exc.NotFound() return available_pools
identifier_body
db.py
fields=fields) def delete_auto_allocated_topology(self, context, tenant_id): tenant_id = self._validate(context, tenant_id) topology = self._get_auto_allocated_topology(context, tenant_id) if topology: subnets = self.core_plugin.get_subnets( context, filters={'network_id': [topology['network_id']]}) self._cleanup( context, network_id=topology['network_id'], router_id=topology['router_id'], subnets=subnets) def _build_topology(self, context, tenant_id, default_external_network): """Build the network topology and returns its network UUID.""" try: subnets = self._provision_tenant_private_network( context, tenant_id) network_id = subnets[0]['network_id'] router = self._provision_external_connectivity( context, default_external_network, subnets, tenant_id) network_id = self._save( context, tenant_id, network_id, router['id'], subnets) return network_id except exceptions.UnknownProvisioningError as e: # Clean partially provisioned topologies, and reraise the # error. If it can be retried, so be it. LOG.error("Unknown error while provisioning topology for " "tenant %(tenant_id)s. Reason: %(reason)s", {'tenant_id': tenant_id, 'reason': e}) self._cleanup( context, network_id=e.network_id, router_id=e.router_id, subnets=e.subnets) raise e.error def _check_requirements(self, context, tenant_id): """Raise if requirements are not met.""" self._get_default_external_network(context) try: self._get_supported_subnetpools(context) except n_exc.NotFound: raise exceptions.AutoAllocationFailure( reason=_("No default subnetpools defined")) return {'id': 'dry-run=pass', 'tenant_id': tenant_id} def _validate(self, context, tenant_id): """Validate and return the tenant to be associated to the topology.""" if tenant_id == 'None': # NOTE(HenryG): the client might be sending us astray by # passing no tenant; this is really meant to be the tenant # issuing the request, therefore let's get it from the context tenant_id = context.tenant_id if not context.is_admin and tenant_id != context.tenant_id: raise n_exc.NotAuthorized() return tenant_id def _get_auto_allocated_topology(self, context, tenant_id): """Return the auto allocated topology record if present or None.""" return auto_allocate_obj.AutoAllocatedTopology.get_object( context, project_id=tenant_id) def _get_auto_allocated_network(self, context, tenant_id): """Get the auto allocated network for the tenant.""" network = self._get_auto_allocated_topology(context, tenant_id) if network: return network['network_id'] @staticmethod def _response(network_id, tenant_id, fields=None): """Build response for auto-allocated network.""" res = { 'id': network_id, 'tenant_id': tenant_id } return db_utils.resource_fields(res, fields) def _get_default_external_network(self, context): """Get the default external network for the deployment.""" default_external_networks = net_obj.ExternalNetwork.get_objects( context, is_default=True) if not default_external_networks: LOG.error("Unable to find default external network " "for deployment, please create/assign one to " "allow auto-allocation to work correctly.") raise exceptions.AutoAllocationFailure( reason=_("No default router:external network")) if len(default_external_networks) > 1: LOG.error("Multiple external default networks detected. " "Network %s is true 'default'.", default_external_networks[0]['network_id']) return default_external_networks[0].network_id def _get_supported_subnetpools(self, context): """Return the default subnet pools available.""" default_subnet_pools = [ self.core_plugin.get_default_subnetpool( context, ver) for ver in (4, 6) ] available_pools = [ s for s in default_subnet_pools if s ] if not available_pools: LOG.error("No default pools available") raise n_exc.NotFound() return available_pools def _provision_tenant_private_network(self, context, tenant_id): """Create a tenant private network/subnets.""" network = None try: network_args = { 'name': 'auto_allocated_network', 'admin_state_up': False, 'tenant_id': tenant_id, 'shared': False } network = p_utils.create_network( self.core_plugin, context, {'network': network_args}) subnets = [] for pool in self._get_supported_subnetpools(context): subnet_args = { 'name': 'auto_allocated_subnet_v%s' % pool['ip_version'], 'network_id': network['id'], 'tenant_id': tenant_id, 'ip_version': pool['ip_version'], 'subnetpool_id': pool['id'], } subnets.append(p_utils.create_subnet( self.core_plugin, context, {'subnet': subnet_args})) return subnets except (c_exc.SubnetAllocationError, ValueError, n_exc.BadRequest, n_exc.NotFound) as e: LOG.error("Unable to auto allocate topology for tenant " "%(tenant_id)s due to missing or unmet " "requirements. Reason: %(reason)s", {'tenant_id': tenant_id, 'reason': e}) if network: self._cleanup(context, network['id']) raise exceptions.AutoAllocationFailure( reason=_("Unable to provide tenant private network")) except Exception as e: network_id = network['id'] if network else None raise exceptions.UnknownProvisioningError(e, network_id=network_id) def _provision_external_connectivity( self, context, default_external_network, subnets, tenant_id): """Uplink tenant subnet(s) to external network.""" router_args = { 'name': 'auto_allocated_router', l3_apidef.EXTERNAL_GW_INFO: { 'network_id': default_external_network}, 'tenant_id': tenant_id, 'admin_state_up': True } router = None attached_subnets = [] try: router = self.l3_plugin.create_router( context, {'router': router_args}) for subnet in subnets: self.l3_plugin.add_router_interface( context, router['id'], {'subnet_id': subnet['id']}) attached_subnets.append(subnet) return router except n_exc.BadRequest as e: LOG.error("Unable to auto allocate topology for tenant " "%(tenant_id)s because of router errors. " "Reason: %(reason)s", {'tenant_id': tenant_id, 'reason': e}) router_id = router['id'] if router else None self._cleanup(context, network_id=subnets[0]['network_id'], router_id=router_id, subnets=attached_subnets) raise exceptions.AutoAllocationFailure( reason=_("Unable to provide external connectivity")) except Exception as e: router_id = router['id'] if router else None raise exceptions.UnknownProvisioningError( e, network_id=subnets[0]['network_id'], router_id=router_id, subnets=subnets) def _save(self, context, tenant_id, network_id, router_id, subnets): """Save auto-allocated topology, or revert in case of DB errors.""" try: auto_allocate_obj.AutoAllocatedTopology( context, project_id=tenant_id, network_id=network_id, router_id=router_id).create() self.core_plugin.update_network( context, network_id, {'network': {'admin_state_up': True}}) except obj_exc.NeutronDbObjectDuplicateEntry: LOG.debug("Multiple auto-allocated networks detected for " "tenant %s. Attempting clean up for network %s " "and router %s.", tenant_id, network_id, router_id) self._cleanup( context, network_id=network_id, router_id=router_id, subnets=subnets) network_id = self._get_auto_allocated_network(context, tenant_id) except Exception as e: raise exceptions.UnknownProvisioningError( e, network_id=network_id, router_id=router_id, subnets=subnets) return network_id def _cleanup(self, context, network_id=None, router_id=None, subnets=None): """Clean up auto allocated resources.""" # Concurrent attempts to delete the topology may interleave and # cause some operations to fail with NotFound exceptions. Rather # than fail partially, the exceptions should be ignored and the # cleanup should proceed uninterrupted. if router_id: for subnet in subnets or []: ignore_notfound( self.l3_plugin.remove_router_interface, context, router_id, {'subnet_id': subnet['id']}) ignore_notfound(self.l3_plugin.delete_router, context, router_id) if network_id: ignore_notfound( self.core_plugin.delete_network, context, network_id) def ignore_notfound(func, *args, **kwargs): """Call the given function and pass if a `NotFound` exception is raised.""" try:
return func(*args, **kwargs)
random_line_split
index.tsx
import React from 'react'; import { parse } from 'qs';
import { withStyles } from '../../helpers/withStylesHelper'; import { setActiveFilterButton } from '../../actions/searchFilter'; import { SearchActions } from '../../actions/actionTypes'; import FilterButton, { FILTER_BUTTON_TYPE } from '../filterButton'; import { AppState } from '../../reducers'; import SearchQueryManager from '../../helpers/searchQueryManager'; import ActionTicketManager from '../../helpers/actionTicketManager'; const s = require('./sortingDropdown.scss'); const SORTING_TYPES: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS[] = ['RELEVANCE', 'NEWEST_FIRST', 'MOST_CITATIONS']; interface SortingDropdownProps { dispatch: Dispatch<SearchActions>; } function trackSorting(sortOption: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS) { ActionTicketManager.trackTicket({ pageType: 'searchResult', actionType: 'fire', actionArea: 'sortBar', actionTag: 'paperSorting', actionLabel: sortOption, }); } function getSortText(sortOption: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS) { switch (sortOption) { case 'RELEVANCE': { return 'Relevance'; } case 'NEWEST_FIRST': { return 'Newest'; } case 'MOST_CITATIONS': { return 'Most Citations'; } default: return 'Relevance'; } } const SortingDropdown: React.FC< SortingDropdownProps & ReturnType<typeof mapStateToProps> & RouteComponentProps > = React.memo(props => { const anchorEl = React.useRef(null); const queryParams = parse(location.search, { ignoreQueryPrefix: true }); const filter = SearchQueryManager.objectifyPaperFilter(queryParams.filter); function getNextLocation(sortOption: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS) { return { pathname: '/search', search: SearchQueryManager.stringifyPapersQuery({ query: props.query, page: 1, sort: sortOption, filter, }), }; } const sortingButtons = SORTING_TYPES.map(types => { return ( <Button key={types} elementType="link" to={getNextLocation(types)} variant="text" color="black" onClick={() => { trackSorting(types); props.dispatch(setActiveFilterButton(null)); }} fullWidth > <span style={{ textAlign: 'left' }}>{getSortText(types)}</span> </Button> ); }); return ( <div ref={anchorEl}> <FilterButton onClick={() => { if (props.isActive) { props.dispatch(setActiveFilterButton(null)); } else { props.dispatch(setActiveFilterButton(FILTER_BUTTON_TYPE.SORTING)); } }} content={getSortText(props.sorting)} isActive={props.isActive} selected={false} /> <Popover onClose={() => { if (props.isActive) { props.dispatch(setActiveFilterButton(null)); } }} anchorOrigin={{ vertical: 'bottom', horizontal: 'left', }} transformOrigin={{ vertical: 'top', horizontal: 'left', }} elevation={0} transitionDuration={150} classes={{ paper: s.dropBoxWrapper, }} open={props.isActive} anchorEl={anchorEl.current} > {sortingButtons} </Popover> </div> ); }); function mapStateToProps(state: AppState) { return { query: state.articleSearch.searchInput, sorting: state.searchFilterState.sorting, isActive: state.searchFilterState.activeButton === FILTER_BUTTON_TYPE.SORTING, }; } export default withRouter(connect(mapStateToProps)(withStyles<typeof SortingDropdown>(s)(SortingDropdown)));
import { Dispatch } from 'redux'; import { connect } from 'react-redux'; import { RouteComponentProps, withRouter } from 'react-router-dom'; import Popover from '@material-ui/core/Popover'; import { Button } from '@pluto_network/pluto-design-elements';
random_line_split
index.tsx
import React from 'react'; import { parse } from 'qs'; import { Dispatch } from 'redux'; import { connect } from 'react-redux'; import { RouteComponentProps, withRouter } from 'react-router-dom'; import Popover from '@material-ui/core/Popover'; import { Button } from '@pluto_network/pluto-design-elements'; import { withStyles } from '../../helpers/withStylesHelper'; import { setActiveFilterButton } from '../../actions/searchFilter'; import { SearchActions } from '../../actions/actionTypes'; import FilterButton, { FILTER_BUTTON_TYPE } from '../filterButton'; import { AppState } from '../../reducers'; import SearchQueryManager from '../../helpers/searchQueryManager'; import ActionTicketManager from '../../helpers/actionTicketManager'; const s = require('./sortingDropdown.scss'); const SORTING_TYPES: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS[] = ['RELEVANCE', 'NEWEST_FIRST', 'MOST_CITATIONS']; interface SortingDropdownProps { dispatch: Dispatch<SearchActions>; } function trackSorting(sortOption: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS) { ActionTicketManager.trackTicket({ pageType: 'searchResult', actionType: 'fire', actionArea: 'sortBar', actionTag: 'paperSorting', actionLabel: sortOption, }); } function getSortText(sortOption: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS) { switch (sortOption) { case 'RELEVANCE': { return 'Relevance'; } case 'NEWEST_FIRST': { return 'Newest'; } case 'MOST_CITATIONS': { return 'Most Citations'; } default: return 'Relevance'; } } const SortingDropdown: React.FC< SortingDropdownProps & ReturnType<typeof mapStateToProps> & RouteComponentProps > = React.memo(props => { const anchorEl = React.useRef(null); const queryParams = parse(location.search, { ignoreQueryPrefix: true }); const filter = SearchQueryManager.objectifyPaperFilter(queryParams.filter); function
(sortOption: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS) { return { pathname: '/search', search: SearchQueryManager.stringifyPapersQuery({ query: props.query, page: 1, sort: sortOption, filter, }), }; } const sortingButtons = SORTING_TYPES.map(types => { return ( <Button key={types} elementType="link" to={getNextLocation(types)} variant="text" color="black" onClick={() => { trackSorting(types); props.dispatch(setActiveFilterButton(null)); }} fullWidth > <span style={{ textAlign: 'left' }}>{getSortText(types)}</span> </Button> ); }); return ( <div ref={anchorEl}> <FilterButton onClick={() => { if (props.isActive) { props.dispatch(setActiveFilterButton(null)); } else { props.dispatch(setActiveFilterButton(FILTER_BUTTON_TYPE.SORTING)); } }} content={getSortText(props.sorting)} isActive={props.isActive} selected={false} /> <Popover onClose={() => { if (props.isActive) { props.dispatch(setActiveFilterButton(null)); } }} anchorOrigin={{ vertical: 'bottom', horizontal: 'left', }} transformOrigin={{ vertical: 'top', horizontal: 'left', }} elevation={0} transitionDuration={150} classes={{ paper: s.dropBoxWrapper, }} open={props.isActive} anchorEl={anchorEl.current} > {sortingButtons} </Popover> </div> ); }); function mapStateToProps(state: AppState) { return { query: state.articleSearch.searchInput, sorting: state.searchFilterState.sorting, isActive: state.searchFilterState.activeButton === FILTER_BUTTON_TYPE.SORTING, }; } export default withRouter(connect(mapStateToProps)(withStyles<typeof SortingDropdown>(s)(SortingDropdown)));
getNextLocation
identifier_name
index.tsx
import React from 'react'; import { parse } from 'qs'; import { Dispatch } from 'redux'; import { connect } from 'react-redux'; import { RouteComponentProps, withRouter } from 'react-router-dom'; import Popover from '@material-ui/core/Popover'; import { Button } from '@pluto_network/pluto-design-elements'; import { withStyles } from '../../helpers/withStylesHelper'; import { setActiveFilterButton } from '../../actions/searchFilter'; import { SearchActions } from '../../actions/actionTypes'; import FilterButton, { FILTER_BUTTON_TYPE } from '../filterButton'; import { AppState } from '../../reducers'; import SearchQueryManager from '../../helpers/searchQueryManager'; import ActionTicketManager from '../../helpers/actionTicketManager'; const s = require('./sortingDropdown.scss'); const SORTING_TYPES: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS[] = ['RELEVANCE', 'NEWEST_FIRST', 'MOST_CITATIONS']; interface SortingDropdownProps { dispatch: Dispatch<SearchActions>; } function trackSorting(sortOption: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS) { ActionTicketManager.trackTicket({ pageType: 'searchResult', actionType: 'fire', actionArea: 'sortBar', actionTag: 'paperSorting', actionLabel: sortOption, }); } function getSortText(sortOption: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS)
const SortingDropdown: React.FC< SortingDropdownProps & ReturnType<typeof mapStateToProps> & RouteComponentProps > = React.memo(props => { const anchorEl = React.useRef(null); const queryParams = parse(location.search, { ignoreQueryPrefix: true }); const filter = SearchQueryManager.objectifyPaperFilter(queryParams.filter); function getNextLocation(sortOption: Scinapse.ArticleSearch.SEARCH_SORT_OPTIONS) { return { pathname: '/search', search: SearchQueryManager.stringifyPapersQuery({ query: props.query, page: 1, sort: sortOption, filter, }), }; } const sortingButtons = SORTING_TYPES.map(types => { return ( <Button key={types} elementType="link" to={getNextLocation(types)} variant="text" color="black" onClick={() => { trackSorting(types); props.dispatch(setActiveFilterButton(null)); }} fullWidth > <span style={{ textAlign: 'left' }}>{getSortText(types)}</span> </Button> ); }); return ( <div ref={anchorEl}> <FilterButton onClick={() => { if (props.isActive) { props.dispatch(setActiveFilterButton(null)); } else { props.dispatch(setActiveFilterButton(FILTER_BUTTON_TYPE.SORTING)); } }} content={getSortText(props.sorting)} isActive={props.isActive} selected={false} /> <Popover onClose={() => { if (props.isActive) { props.dispatch(setActiveFilterButton(null)); } }} anchorOrigin={{ vertical: 'bottom', horizontal: 'left', }} transformOrigin={{ vertical: 'top', horizontal: 'left', }} elevation={0} transitionDuration={150} classes={{ paper: s.dropBoxWrapper, }} open={props.isActive} anchorEl={anchorEl.current} > {sortingButtons} </Popover> </div> ); }); function mapStateToProps(state: AppState) { return { query: state.articleSearch.searchInput, sorting: state.searchFilterState.sorting, isActive: state.searchFilterState.activeButton === FILTER_BUTTON_TYPE.SORTING, }; } export default withRouter(connect(mapStateToProps)(withStyles<typeof SortingDropdown>(s)(SortingDropdown)));
{ switch (sortOption) { case 'RELEVANCE': { return 'Relevance'; } case 'NEWEST_FIRST': { return 'Newest'; } case 'MOST_CITATIONS': { return 'Most Citations'; } default: return 'Relevance'; } }
identifier_body
roman_numerals.py
# https://www.codeeval.com/open_challenges/106/ import sys from collections import namedtuple test_cases = open(sys.argv[1], 'r') # test_cases = open('roman_numerals.txt', 'r') test_lines = (line.rstrip() for line in test_cases) def num_to_components(num): num_comp = namedtuple('NumComponents', ('thousands', 'hundreds', 'tens', 'singles')) thousands = int(num/1000) hundreds = int((num - 1000*thousands)/100) tens = int((num - 1000*thousands - 100*hundreds)/10) singles = int(num - 1000*thousands - 100*hundreds - 10*tens) return num_comp(thousands=thousands, hundreds=hundreds, tens=tens, singles=singles) def to_roman(num_components):
r_tens = 'XC' elif num_components.tens == 5: r_tens = 'L' elif num_components.tens <= 3: r_tens = 'X'*num_components.tens elif num_components.tens in range(6, 9): r_tens = 'L' + 'X' * (num_components.tens - 5) # for singles if num_components.singles == 4: r_singles = 'IV' elif num_components.singles == 9: r_singles = 'IX' elif num_components.singles == 5: r_singles = 'V' elif num_components.singles <= 3: r_singles = 'I'*num_components.singles elif num_components.singles in range(6, 9): r_singles = 'V' + 'I' * (num_components.singles - 5) roman_num = r_thousands + r_hundreds + r_tens + r_singles print(roman_num) if __name__ == '__main__': for test in test_lines: components = num_to_components(int(test)) to_roman(components) test_cases.close()
r_thousands = 'M'*num_components.thousands r_hundreds = '' r_tens = '' r_singles = '' # for hundreds if num_components.hundreds == 4: r_hundreds = 'CD' elif num_components.hundreds == 9: r_hundreds = 'CM' elif num_components.hundreds == 5: r_hundreds = 'D' elif num_components.hundreds <= 3: r_hundreds = 'C'*num_components.hundreds elif num_components.hundreds in range(6, 9): r_hundreds = 'D' + 'C' * (num_components.hundreds - 5) # for Tens if num_components.tens == 4: r_tens = 'XL' elif num_components.tens == 9:
identifier_body
roman_numerals.py
# https://www.codeeval.com/open_challenges/106/ import sys from collections import namedtuple test_cases = open(sys.argv[1], 'r') # test_cases = open('roman_numerals.txt', 'r') test_lines = (line.rstrip() for line in test_cases) def num_to_components(num): num_comp = namedtuple('NumComponents', ('thousands', 'hundreds', 'tens', 'singles')) thousands = int(num/1000) hundreds = int((num - 1000*thousands)/100) tens = int((num - 1000*thousands - 100*hundreds)/10) singles = int(num - 1000*thousands - 100*hundreds - 10*tens) return num_comp(thousands=thousands, hundreds=hundreds, tens=tens, singles=singles) def to_roman(num_components): r_thousands = 'M'*num_components.thousands r_hundreds = '' r_tens = '' r_singles = '' # for hundreds if num_components.hundreds == 4: r_hundreds = 'CD' elif num_components.hundreds == 9: r_hundreds = 'CM' elif num_components.hundreds == 5: r_hundreds = 'D' elif num_components.hundreds <= 3: r_hundreds = 'C'*num_components.hundreds elif num_components.hundreds in range(6, 9): r_hundreds = 'D' + 'C' * (num_components.hundreds - 5) # for Tens if num_components.tens == 4: r_tens = 'XL' elif num_components.tens == 9:
elif num_components.tens in range(6, 9): r_tens = 'L' + 'X' * (num_components.tens - 5) # for singles if num_components.singles == 4: r_singles = 'IV' elif num_components.singles == 9: r_singles = 'IX' elif num_components.singles == 5: r_singles = 'V' elif num_components.singles <= 3: r_singles = 'I'*num_components.singles elif num_components.singles in range(6, 9): r_singles = 'V' + 'I' * (num_components.singles - 5) roman_num = r_thousands + r_hundreds + r_tens + r_singles print(roman_num) if __name__ == '__main__': for test in test_lines: components = num_to_components(int(test)) to_roman(components) test_cases.close()
r_tens = 'XC' elif num_components.tens == 5: r_tens = 'L' elif num_components.tens <= 3: r_tens = 'X'*num_components.tens
random_line_split
roman_numerals.py
# https://www.codeeval.com/open_challenges/106/ import sys from collections import namedtuple test_cases = open(sys.argv[1], 'r') # test_cases = open('roman_numerals.txt', 'r') test_lines = (line.rstrip() for line in test_cases) def num_to_components(num): num_comp = namedtuple('NumComponents', ('thousands', 'hundreds', 'tens', 'singles')) thousands = int(num/1000) hundreds = int((num - 1000*thousands)/100) tens = int((num - 1000*thousands - 100*hundreds)/10) singles = int(num - 1000*thousands - 100*hundreds - 10*tens) return num_comp(thousands=thousands, hundreds=hundreds, tens=tens, singles=singles) def
(num_components): r_thousands = 'M'*num_components.thousands r_hundreds = '' r_tens = '' r_singles = '' # for hundreds if num_components.hundreds == 4: r_hundreds = 'CD' elif num_components.hundreds == 9: r_hundreds = 'CM' elif num_components.hundreds == 5: r_hundreds = 'D' elif num_components.hundreds <= 3: r_hundreds = 'C'*num_components.hundreds elif num_components.hundreds in range(6, 9): r_hundreds = 'D' + 'C' * (num_components.hundreds - 5) # for Tens if num_components.tens == 4: r_tens = 'XL' elif num_components.tens == 9: r_tens = 'XC' elif num_components.tens == 5: r_tens = 'L' elif num_components.tens <= 3: r_tens = 'X'*num_components.tens elif num_components.tens in range(6, 9): r_tens = 'L' + 'X' * (num_components.tens - 5) # for singles if num_components.singles == 4: r_singles = 'IV' elif num_components.singles == 9: r_singles = 'IX' elif num_components.singles == 5: r_singles = 'V' elif num_components.singles <= 3: r_singles = 'I'*num_components.singles elif num_components.singles in range(6, 9): r_singles = 'V' + 'I' * (num_components.singles - 5) roman_num = r_thousands + r_hundreds + r_tens + r_singles print(roman_num) if __name__ == '__main__': for test in test_lines: components = num_to_components(int(test)) to_roman(components) test_cases.close()
to_roman
identifier_name
roman_numerals.py
# https://www.codeeval.com/open_challenges/106/ import sys from collections import namedtuple test_cases = open(sys.argv[1], 'r') # test_cases = open('roman_numerals.txt', 'r') test_lines = (line.rstrip() for line in test_cases) def num_to_components(num): num_comp = namedtuple('NumComponents', ('thousands', 'hundreds', 'tens', 'singles')) thousands = int(num/1000) hundreds = int((num - 1000*thousands)/100) tens = int((num - 1000*thousands - 100*hundreds)/10) singles = int(num - 1000*thousands - 100*hundreds - 10*tens) return num_comp(thousands=thousands, hundreds=hundreds, tens=tens, singles=singles) def to_roman(num_components): r_thousands = 'M'*num_components.thousands r_hundreds = '' r_tens = '' r_singles = '' # for hundreds if num_components.hundreds == 4: r_hundreds = 'CD' elif num_components.hundreds == 9: r_hundreds = 'CM' elif num_components.hundreds == 5: r_hundreds = 'D' elif num_components.hundreds <= 3: r_hundreds = 'C'*num_components.hundreds elif num_components.hundreds in range(6, 9): r_hundreds = 'D' + 'C' * (num_components.hundreds - 5) # for Tens if num_components.tens == 4: r_tens = 'XL' elif num_components.tens == 9: r_tens = 'XC' elif num_components.tens == 5: r_tens = 'L' elif num_components.tens <= 3: r_tens = 'X'*num_components.tens elif num_components.tens in range(6, 9): r_tens = 'L' + 'X' * (num_components.tens - 5) # for singles if num_components.singles == 4: r_singles = 'IV' elif num_components.singles == 9: r_singles = 'IX' elif num_components.singles == 5: r_singles = 'V' elif num_components.singles <= 3: r_singles = 'I'*num_components.singles elif num_components.singles in range(6, 9): r_singles = 'V' + 'I' * (num_components.singles - 5) roman_num = r_thousands + r_hundreds + r_tens + r_singles print(roman_num) if __name__ == '__main__': for test in test_lines:
test_cases.close()
components = num_to_components(int(test)) to_roman(components)
conditional_block
KIconDialog.py
# encoding: utf-8 # module PyKDE4.kio # from /usr/lib/python3/dist-packages/PyKDE4/kio.cpython-34m-x86_64-linux-gnu.so # by generator 1.135 # no doc # imports import PyKDE4.kdeui as __PyKDE4_kdeui import PyQt4.QtCore as __PyQt4_QtCore import PyQt4.QtGui as __PyQt4_QtGui class KIconDialog(__PyKDE4_kdeui.KDialog): # no doc def getIcon(self, *args, **kwargs): # real signature unknown pass def iconSize(self, *args, **kwargs): # real signature unknown pass def newIconName(self, *args, **kwargs): # real signature unknown pass def openDialog(self, *args, **kwargs): # real signature unknown pass def setCustomLocation(self, *args, **kwargs): # real signature unknown pass def setIconSize(self, *args, **kwargs): # real signature unknown pass def setStrictIconSize(self, *args, **kwargs): # real signature unknown pass def setup(self, *args, **kwargs): # real signature unknown pass def showDialog(self, *args, **kwargs): # real signature unknown pass def slotOk(self, *args, **kwargs): # real signature unknown
def strictIconSize(self, *args, **kwargs): # real signature unknown pass def __init__(self, *args, **kwargs): # real signature unknown pass
pass
identifier_body
KIconDialog.py
# encoding: utf-8 # module PyKDE4.kio # from /usr/lib/python3/dist-packages/PyKDE4/kio.cpython-34m-x86_64-linux-gnu.so # by generator 1.135 # no doc # imports import PyKDE4.kdeui as __PyKDE4_kdeui import PyQt4.QtCore as __PyQt4_QtCore import PyQt4.QtGui as __PyQt4_QtGui class
(__PyKDE4_kdeui.KDialog): # no doc def getIcon(self, *args, **kwargs): # real signature unknown pass def iconSize(self, *args, **kwargs): # real signature unknown pass def newIconName(self, *args, **kwargs): # real signature unknown pass def openDialog(self, *args, **kwargs): # real signature unknown pass def setCustomLocation(self, *args, **kwargs): # real signature unknown pass def setIconSize(self, *args, **kwargs): # real signature unknown pass def setStrictIconSize(self, *args, **kwargs): # real signature unknown pass def setup(self, *args, **kwargs): # real signature unknown pass def showDialog(self, *args, **kwargs): # real signature unknown pass def slotOk(self, *args, **kwargs): # real signature unknown pass def strictIconSize(self, *args, **kwargs): # real signature unknown pass def __init__(self, *args, **kwargs): # real signature unknown pass
KIconDialog
identifier_name
KIconDialog.py
# encoding: utf-8 # module PyKDE4.kio
# by generator 1.135 # no doc # imports import PyKDE4.kdeui as __PyKDE4_kdeui import PyQt4.QtCore as __PyQt4_QtCore import PyQt4.QtGui as __PyQt4_QtGui class KIconDialog(__PyKDE4_kdeui.KDialog): # no doc def getIcon(self, *args, **kwargs): # real signature unknown pass def iconSize(self, *args, **kwargs): # real signature unknown pass def newIconName(self, *args, **kwargs): # real signature unknown pass def openDialog(self, *args, **kwargs): # real signature unknown pass def setCustomLocation(self, *args, **kwargs): # real signature unknown pass def setIconSize(self, *args, **kwargs): # real signature unknown pass def setStrictIconSize(self, *args, **kwargs): # real signature unknown pass def setup(self, *args, **kwargs): # real signature unknown pass def showDialog(self, *args, **kwargs): # real signature unknown pass def slotOk(self, *args, **kwargs): # real signature unknown pass def strictIconSize(self, *args, **kwargs): # real signature unknown pass def __init__(self, *args, **kwargs): # real signature unknown pass
# from /usr/lib/python3/dist-packages/PyKDE4/kio.cpython-34m-x86_64-linux-gnu.so
random_line_split
karma.conf.js
// Karma configuration // Generated on Thu Jul 03 2014 13:23:26 GMT+0530 (India Standard Time) module.exports = function(config) { config.set({ // base path that will be used to resolve all patterns (eg. files, exclude) basePath: '', // frameworks to use // available frameworks: https://npmjs.org/browse/keyword/karma-adapter frameworks: ['jasmine'], // list of files / patterns to load in the browser files: [ 'dist/gainda.js', 'test/spec/*.js' ], // list of files to exclude exclude: [ ], // preprocess matching files before serving them to the browser // available preprocessors: https://npmjs.org/browse/keyword/karma-preprocessor preprocessors: { }, // test results reporter to use // possible values: 'dots', 'progress' // available reporters: https://npmjs.org/browse/keyword/karma-reporter reporters: ['progress'], // web server port port: 9876, // enable / disable colors in the output (reporters and logs) colors: true, // level of logging // possible values: config.LOG_DISABLE || config.LOG_ERROR || config.LOG_WARN || config.LOG_INFO || config.LOG_DEBUG logLevel: config.LOG_INFO, // enable / disable watching file and executing tests whenever any file changes autoWatch: true,
// Continuous Integration mode // if true, Karma captures browsers, runs the tests and exits singleRun: false }); };
// start these browsers // available browser launchers: https://npmjs.org/browse/keyword/karma-launcher browsers: ['Chrome'],
random_line_split
TypeTextAtCefTest.ts
import { Keys } from '@ephox/agar'; import { describe, it } from '@ephox/bedrock-client'; import { TinyAssertions, TinyContentActions, TinyHooks, TinySelections } from '@ephox/mcagar'; import Editor from 'tinymce/core/api/Editor'; import Theme from 'tinymce/themes/silver/Theme'; describe('browser.tinymce.core.keyboard.TypeTextAtCef', () => { const hook = TinyHooks.bddSetupLight<Editor>({ add_unload_trigger: false, base_url: '/project/tinymce/js/tinymce' }, [ Theme ], true); it('Type text before cef inline element', () => { const editor = hook.editor(); editor.setContent('<p><span contenteditable="false">a</span></p>'); TinySelections.select(editor, 'p', [ 1 ]); TinyContentActions.keystroke(editor, Keys.left()); TinyContentActions.type(editor, 'bc');
it('Type after cef inline element', () => { const editor = hook.editor(); editor.setContent('<p><span contenteditable="false">a</span></p>'); TinySelections.select(editor, 'p', [ 1 ]); TinyContentActions.keystroke(editor, Keys.right()); TinyContentActions.type(editor, 'bc'); TinyAssertions.assertCursor(editor, [ 0, 1 ], 3); TinyAssertions.assertContent(editor, '<p><span contenteditable="false">a</span>bc</p>'); }); it('Type between cef inline elements', () => { const editor = hook.editor(); editor.setContent('<p><span contenteditable="false">a</span>&nbsp;<span contenteditable="false">b</span></p>'); TinySelections.select(editor, 'p', [ 3 ]); TinyContentActions.keystroke(editor, Keys.left()); TinyContentActions.keystroke(editor, Keys.left()); TinyContentActions.type(editor, 'bc'); TinyAssertions.assertSelection(editor, [ 0, 1 ], 3, [ 0, 1 ], 3); TinyAssertions.assertContent(editor, '<p><span contenteditable="false">a</span>bc&nbsp;<span contenteditable="false">b</span></p>'); }); });
TinyAssertions.assertCursor(editor, [ 0, 0 ], 2); TinyAssertions.assertContent(editor, '<p>bc<span contenteditable="false">a</span></p>'); });
random_line_split
process.rs
(arr: &[u8]) -> i32 { let a = arr[0] as u32; let b = arr[1] as u32; let c = arr[2] as u32; let d = arr[3] as u32; ((a << 24) | (b << 16) | (c << 8) | (d << 0)) as i32 } let p = Process{ pid: pid }; drop(output); let mut bytes = [0; 8]; return match input.read(&mut bytes) { Ok(8) => { assert!(combine(CLOEXEC_MSG_FOOTER) == combine(&bytes[4.. 8]), "Validation on the CLOEXEC pipe failed: {:?}", bytes); let errno = combine(&bytes[0.. 4]); assert!(p.wait(0).is_ok(), "wait(0) should either return Ok or panic"); Err(super::decode_error(errno)) } Err(ref e) if e.kind == EndOfFile => Ok(p), Err(e) => { assert!(p.wait(0).is_ok(), "wait(0) should either return Ok or panic"); panic!("the CLOEXEC pipe failed: {:?}", e) }, Ok(..) => { // pipe I/O up to PIPE_BUF bytes should be atomic assert!(p.wait(0).is_ok(), "wait(0) should either return Ok or panic"); panic!("short read on the CLOEXEC pipe") } }; } // And at this point we've reached a special time in the life of the // child. The child must now be considered hamstrung and unable to // do anything other than syscalls really. Consider the following // scenario: // // 1. Thread A of process 1 grabs the malloc() mutex // 2. Thread B of process 1 forks(), creating thread C // 3. Thread C of process 2 then attempts to malloc() // 4. The memory of process 2 is the same as the memory of // process 1, so the mutex is locked. // // This situation looks a lot like deadlock, right? It turns out // that this is what pthread_atfork() takes care of, which is // presumably implemented across platforms. The first thing that // threads to *before* forking is to do things like grab the malloc // mutex, and then after the fork they unlock it. // // Despite this information, libnative's spawn has been witnessed to // deadlock on both OSX and FreeBSD. I'm not entirely sure why, but // all collected backtraces point at malloc/free traffic in the // child spawned process. // // For this reason, the block of code below should contain 0 // invocations of either malloc of free (or their related friends). // // As an example of not having malloc/free traffic, we don't close // this file descriptor by dropping the FileDesc (which contains an // allocation). Instead we just close it manually. This will never // have the drop glue anyway because this code never returns (the // child will either exec() or invoke libc::exit) let _ = libc::close(input.fd()); fn fail(output: &mut FileDesc) -> ! { let errno = sys::os::errno() as u32; let bytes = [ (errno >> 24) as u8, (errno >> 16) as u8, (errno >> 8) as u8, (errno >> 0) as u8, CLOEXEC_MSG_FOOTER[0], CLOEXEC_MSG_FOOTER[1], CLOEXEC_MSG_FOOTER[2], CLOEXEC_MSG_FOOTER[3] ]; // pipe I/O up to PIPE_BUF bytes should be atomic assert!(output.write(&bytes).is_ok()); unsafe { libc::_exit(1) } } rustrt::rust_unset_sigprocmask(); // If a stdio file descriptor is set to be ignored (via a -1 file // descriptor), then we don't actually close it, but rather open // up /dev/null into that file descriptor. Otherwise, the first file // descriptor opened up in the child would be numbered as one of the // stdio file descriptors, which is likely to wreak havoc. let setup = |src: Option<P>, dst: c_int| { let src = match src { None => { let flags = if dst == libc::STDIN_FILENO { libc::O_RDONLY } else { libc::O_RDWR }; libc::open(devnull.as_ptr() as *const _, flags, 0) } Some(obj) => { let fd = obj.as_inner().fd(); // Leak the memory and the file descriptor. We're in the // child now an all our resources are going to be // cleaned up very soon mem::forget(obj); fd } }; src != -1 && retry(|| dup2(src, dst)) != -1 }; if !setup(in_fd, libc::STDIN_FILENO) { fail(&mut output) } if !setup(out_fd, libc::STDOUT_FILENO) { fail(&mut output) } if !setup(err_fd, libc::STDERR_FILENO) { fail(&mut output) } // close all other fds for fd in (3..getdtablesize()).rev() { if fd != output.fd() { let _ = close(fd as c_int); } } match cfg.gid() { Some(u) => { if libc::setgid(u as libc::gid_t) != 0 { fail(&mut output); } } None => {} } match cfg.uid() { Some(u) => { // When dropping privileges from root, the `setgroups` call // will remove any extraneous groups. If we don't call this, // then even though our uid has dropped, we may still have // groups that enable us to do super-user things. This will // fail if we aren't root, so don't bother checking the // return value, this is just done as an optimistic // privilege dropping function. extern { fn setgroups(ngroups: libc::c_int, ptr: *const libc::c_void) -> libc::c_int; } let _ = setgroups(0, ptr::null()); if libc::setuid(u as libc::uid_t) != 0 { fail(&mut output); } } None => {} } if cfg.detach() { // Don't check the error of setsid because it fails if we're the // process leader already. We just forked so it shouldn't return // error, but ignore it anyway. let _ = libc::setsid(); } if !dirp.is_null() && chdir(dirp) == -1 { fail(&mut output); } if !envp.is_null() { *sys::os::environ() = envp as *const _; } let _ = execvp(*argv, argv as *mut _); fail(&mut output); }) }) } pub fn wait(&self, deadline: u64) -> IoResult<ProcessExit> { use cmp; use sync::mpsc::TryRecvError; static mut WRITE_FD: libc::c_int = 0; let mut status = 0 as c_int; if deadline == 0 { return match retry(|| unsafe { c::waitpid(self.pid, &mut status, 0) }) { -1 => panic!("unknown waitpid error: {:?}", super::last_error()), _ => Ok(translate_status(status)), } } // On unix, wait() and its friends have no timeout parameters, so there is // no way to time out a thread in wait(). From some googling and some // thinking, it appears that there are a few ways to handle timeouts in // wait(), but the only real reasonable one for a multi-threaded program is // to listen for SIGCHLD. // // With this in mind, the waiting mechanism with a timeout barely uses // waitpid() at all. There are a few times that waitpid() is invoked with // WNOHANG, but otherwise all the necessary blocking is done by waiting for // a SIGCHLD to arrive (and that blocking has a timeout). Note, however, // that waitpid() is still used to actually reap the child. // // Signal handling is super tricky in general, and this is no exception. Due // to the async nature of SIGCHLD, we use the self-pipe trick to transmit // data out of the signal handler to the rest of the application. The first // idea would be to have each thread waiting with a timeout to read this // output file descriptor, but a write() is akin to a signal(), not a // broadcast(), so it would only wake up one thread, and possibly the wrong // thread
combine
identifier_name
process.rs
pub fn spawn<K, V, C, P>(cfg: &C, in_fd: Option<P>, out_fd: Option<P>, err_fd: Option<P>) -> IoResult<Process> where C: ProcessConfig<K, V>, P: AsInner<FileDesc>, K: BytesContainer + Eq + Hash, V: BytesContainer { use libc::funcs::posix88::unistd::{fork, dup2, close, chdir, execvp}; mod rustrt { extern { pub fn rust_unset_sigprocmask(); } } unsafe fn set_cloexec(fd: c_int) { let ret = c::ioctl(fd, c::FIOCLEX); assert_eq!(ret, 0); } #[cfg(all(target_os = "android", target_arch = "aarch64"))] unsafe fn getdtablesize() -> c_int { libc::sysconf(libc::consts::os::sysconf::_SC_OPEN_MAX) as c_int } #[cfg(not(all(target_os = "android", target_arch = "aarch64")))] unsafe fn getdtablesize() -> c_int { libc::funcs::bsd44::getdtablesize() } let dirp = cfg.cwd().map(|c| c.as_ptr()).unwrap_or(ptr::null()); // temporary until unboxed closures land let cfg = unsafe { mem::transmute::<&ProcessConfig<K,V>,&'static ProcessConfig<K,V>>(cfg) }; with_envp(cfg.env(), move|envp: *const c_void| { with_argv(cfg.program(), cfg.args(), move|argv: *const *const libc::c_char| unsafe { let (input, mut output) = try!(sys::os::pipe()); // We may use this in the child, so perform allocations before the // fork let devnull = b"/dev/null\0"; set_cloexec(output.fd()); let pid = fork(); if pid < 0 { return Err(super::last_error()) } else if pid > 0 { #[inline] fn combine(arr: &[u8]) -> i32 { let a = arr[0] as u32; let b = arr[1] as u32; let c = arr[2] as u32; let d = arr[3] as u32; ((a << 24) | (b << 16) | (c << 8) | (d << 0)) as i32 } let p = Process{ pid: pid }; drop(output); let mut bytes = [0; 8]; return match input.read(&mut bytes) { Ok(8) => { assert!(combine(CLOEXEC_MSG_FOOTER) == combine(&bytes[4.. 8]), "Validation on the CLOEXEC pipe failed: {:?}", bytes); let errno = combine(&bytes[0.. 4]); assert!(p.wait(0).is_ok(), "wait(0) should either return Ok or panic"); Err(super::decode_error(errno)) } Err(ref e) if e.kind == EndOfFile => Ok(p), Err(e) => { assert!(p.wait(0).is_ok(), "wait(0) should either return Ok or panic"); panic!("the CLOEXEC pipe failed: {:?}", e) }, Ok(..) => { // pipe I/O up to PIPE_BUF bytes should be atomic assert!(p.wait(0).is_ok(), "wait(0) should either return Ok or panic"); panic!("short read on the CLOEXEC pipe") } }; } // And at this point we've reached a special time in the life of the // child. The child must now be considered hamstrung and unable to // do anything other than syscalls really. Consider the following // scenario: // // 1. Thread A of process 1 grabs the malloc() mutex // 2. Thread B of process 1 forks(), creating thread C // 3. Thread C of process 2 then attempts to malloc() // 4. The memory of process 2 is the same as the memory of // process 1, so the mutex is locked. // // This situation looks a lot like deadlock, right? It turns out // that this is what pthread_atfork() takes care of, which is // presumably implemented across platforms. The first thing that // threads to *before* forking is to do things like grab the malloc // mutex, and then after the fork they unlock it. // // Despite this information, libnative's spawn has been witnessed to // deadlock on both OSX and FreeBSD. I'm not entirely sure why, but // all collected backtraces point at malloc/free traffic in the // child spawned process. // // For this reason, the block of code below should contain 0 // invocations of either malloc of free (or their related friends). // // As an example of not having malloc/free traffic, we don't close // this file descriptor by dropping the FileDesc (which contains an // allocation). Instead we just close it manually. This will never // have the drop glue anyway because this code never returns (the // child will either exec() or invoke libc::exit) let _ = libc::close(input.fd()); fn fail(output: &mut FileDesc) -> ! { let errno = sys::os::errno() as u32; let bytes = [ (errno >> 24) as u8, (errno >> 16) as u8, (errno >> 8) as u8, (errno >> 0) as u8, CLOEXEC_MSG_FOOTER[0], CLOEXEC_MSG_FOOTER[1], CLOEXEC_MSG_FOOTER[2], CLOEXEC_MSG_FOOTER[3] ]; // pipe I/O up to PIPE_BUF bytes should be atomic assert!(output.write(&bytes).is_ok()); unsafe { libc::_exit(1) } } rustrt::rust_unset_sigprocmask(); // If a stdio file descriptor is set to be ignored (via a -1 file // descriptor), then we don't actually close it, but rather open // up /dev/null into that file descriptor. Otherwise, the first file // descriptor opened up in the child would be numbered as one of the // stdio file descriptors, which is likely to wreak havoc. let setup = |src: Option<P>, dst: c_int| { let src = match src { None => { let flags = if dst == libc::STDIN_FILENO { libc::O_RDONLY } else { libc::O_RDWR }; libc::open(devnull.as_ptr() as *const _, flags, 0) } Some(obj) => { let fd = obj.as_inner().fd(); // Leak the memory and the file descriptor. We're in the // child now an all our resources are going to be // cleaned up very soon mem::forget(obj); fd } }; src != -1 && retry(|| dup2(src, dst)) != -1 }; if !setup(in_fd, libc::STDIN_FILENO) { fail(&mut output) } if !setup(out_fd, libc::STDOUT_FILENO) { fail(&mut output) } if !setup(err_fd, libc::STDERR_FILENO) { fail(&mut output) } // close all other fds for fd in (3..getdtablesize()).rev() { if fd != output.fd() { let _ = close(fd as c_int); } } match cfg.gid() { Some(u) => { if libc::setgid(u as libc::gid_t) != 0 { fail(&mut output); } } None => {} } match cfg.uid() { Some(u) => { // When dropping privileges from root, the `setgroups` call // will remove any extraneous groups. If we don't call this, // then even though our uid has dropped, we may still have // groups that enable us to do super-user things. This will // fail if we aren't root, so don't bother checking the // return value, this is just done as an optimistic // privilege dropping function. extern { fn setgroups(ngroups: libc::c_int, ptr: *const libc::c_void) -> libc::c_int; } let _ = setgroups(0, ptr::null()); if libc::setuid(u as libc::uid_t) != 0 { fail(&mut output); } } None => {} } if cfg.detach() { // Don't check the error of setsid because it fails if we're the // process leader already. We just forked so it shouldn't return // error, but ignore it anyway. let _ =
{ let r = libc::funcs::posix88::signal::kill(pid, signal as c_int); mkerr_libc(r) }
identifier_body
process.rs
// error, but ignore it anyway. let _ = libc::setsid(); } if !dirp.is_null() && chdir(dirp) == -1 { fail(&mut output); } if !envp.is_null() { *sys::os::environ() = envp as *const _; } let _ = execvp(*argv, argv as *mut _); fail(&mut output); }) }) } pub fn wait(&self, deadline: u64) -> IoResult<ProcessExit> { use cmp; use sync::mpsc::TryRecvError; static mut WRITE_FD: libc::c_int = 0; let mut status = 0 as c_int; if deadline == 0 { return match retry(|| unsafe { c::waitpid(self.pid, &mut status, 0) }) { -1 => panic!("unknown waitpid error: {:?}", super::last_error()), _ => Ok(translate_status(status)), } } // On unix, wait() and its friends have no timeout parameters, so there is // no way to time out a thread in wait(). From some googling and some // thinking, it appears that there are a few ways to handle timeouts in // wait(), but the only real reasonable one for a multi-threaded program is // to listen for SIGCHLD. // // With this in mind, the waiting mechanism with a timeout barely uses // waitpid() at all. There are a few times that waitpid() is invoked with // WNOHANG, but otherwise all the necessary blocking is done by waiting for // a SIGCHLD to arrive (and that blocking has a timeout). Note, however, // that waitpid() is still used to actually reap the child. // // Signal handling is super tricky in general, and this is no exception. Due // to the async nature of SIGCHLD, we use the self-pipe trick to transmit // data out of the signal handler to the rest of the application. The first // idea would be to have each thread waiting with a timeout to read this // output file descriptor, but a write() is akin to a signal(), not a // broadcast(), so it would only wake up one thread, and possibly the wrong // thread. Hence a helper thread is used. // // The helper thread here is responsible for farming requests for a // waitpid() with a timeout, and then processing all of the wait requests. // By guaranteeing that only this helper thread is reading half of the // self-pipe, we're sure that we'll never lose a SIGCHLD. This helper thread // is also responsible for select() to wait for incoming messages or // incoming SIGCHLD messages, along with passing an appropriate timeout to // select() to wake things up as necessary. // // The ordering of the following statements is also very purposeful. First, // we must be guaranteed that the helper thread is booted and available to // receive SIGCHLD signals, and then we must also ensure that we do a // nonblocking waitpid() at least once before we go ask the sigchld helper. // This prevents the race where the child exits, we boot the helper, and // then we ask for the child's exit status (never seeing a sigchld). // // The actual communication between the helper thread and this thread is // quite simple, just a channel moving data around. HELPER.boot(register_sigchld, waitpid_helper); match self.try_wait() { Some(ret) => return Ok(ret), None => {} } let (tx, rx) = channel(); HELPER.send(NewChild(self.pid, tx, deadline)); return match rx.recv() { Ok(e) => Ok(e), Err(..) => Err(timeout("wait timed out")), }; // Register a new SIGCHLD handler, returning the reading half of the // self-pipe plus the old handler registered (return value of sigaction). // // Be sure to set up the self-pipe first because as soon as we register a // handler we're going to start receiving signals. fn register_sigchld() -> (libc::c_int, c::sigaction) { unsafe { let mut pipes = [0; 2]; assert_eq!(libc::pipe(pipes.as_mut_ptr()), 0); set_nonblocking(pipes[0], true); set_nonblocking(pipes[1], true); WRITE_FD = pipes[1]; let mut old: c::sigaction = mem::zeroed(); let mut new: c::sigaction = mem::zeroed(); new.sa_handler = sigchld_handler; new.sa_flags = c::SA_NOCLDSTOP; assert_eq!(c::sigaction(c::SIGCHLD, &new, &mut old), 0); (pipes[0], old) } } // Helper thread for processing SIGCHLD messages fn waitpid_helper(input: libc::c_int, messages: Receiver<Req>, (read_fd, old): (libc::c_int, c::sigaction)) { set_nonblocking(input, true); let mut set: c::fd_set = unsafe { mem::zeroed() }; let mut tv: libc::timeval; let mut active = Vec::<(libc::pid_t, Sender<ProcessExit>, u64)>::new(); let max = cmp::max(input, read_fd) + 1; 'outer: loop { // Figure out the timeout of our syscall-to-happen. If we're waiting // for some processes, then they'll have a timeout, otherwise we // wait indefinitely for a message to arrive. // // FIXME: sure would be nice to not have to scan the entire array let min = active.iter().map(|a| a.2).enumerate().min_by(|p| { p.1 }); let (p, idx) = match min { Some((idx, deadline)) => { let now = sys::timer::now(); let ms = if now < deadline {deadline - now} else {0}; tv = ms_to_timeval(ms); (&mut tv as *mut _, idx) } None => (ptr::null_mut(), -1), }; // Wait for something to happen c::fd_set(&mut set, input); c::fd_set(&mut set, read_fd); match unsafe { c::select(max, &mut set, ptr::null_mut(), ptr::null_mut(), p) } { // interrupted, retry -1 if os::errno() == libc::EINTR as i32 => continue, // We read something, break out and process 1 | 2 => {} // Timeout, the pending request is removed 0 => { drop(active.remove(idx)); continue } n => panic!("error in select {:?} ({:?})", os::errno(), n), } // Process any pending messages if drain(input) { loop { match messages.try_recv() { Ok(NewChild(pid, tx, deadline)) => { active.push((pid, tx, deadline)); } // Once we've been disconnected it means the main // thread is exiting (at_exit has run). We could // still have active waiter for other threads, so // we're just going to drop them all on the floor. // This means that they won't receive a "you're // done" message in which case they'll be considered // as timed out, but more generally errors will // start propagating. Err(TryRecvError::Disconnected) => { break 'outer; } Err(TryRecvError::Empty) => break, } } } // If a child exited (somehow received SIGCHLD), then poll all // children to see if any of them exited. // // We also attempt to be responsible netizens when dealing with // SIGCHLD by invoking any previous SIGCHLD handler instead of just // ignoring any previous SIGCHLD handler. Note that we don't provide // a 1:1 mapping of our handler invocations to the previous handler // invocations because we drain the `read_fd` entirely. This is // probably OK because the kernel is already allowed to coalesce // simultaneous signals, we're just doing some extra coalescing. // // Another point of note is that this likely runs the signal handler // on a different thread than the one that received the signal. I // *think* this is ok at this time. // // The main reason for doing this is to allow stdtest to run native // tests as well. Both libgreen and libnative are running around // with process timeouts, but libgreen should get there first // (currently libuv doesn't handle old signal handlers). if drain(read_fd) { let i: uint = unsafe { mem::transmute(old.sa_handler) }; if i != 0 { assert!(old.sa_flags & c::SA_SIGINFO == 0); (old.sa_handler)(c::SIGCHLD); }
random_line_split
process.rs
"wait(0) should either return Ok or panic"); panic!("the CLOEXEC pipe failed: {:?}", e) }, Ok(..) => { // pipe I/O up to PIPE_BUF bytes should be atomic assert!(p.wait(0).is_ok(), "wait(0) should either return Ok or panic"); panic!("short read on the CLOEXEC pipe") } }; } // And at this point we've reached a special time in the life of the // child. The child must now be considered hamstrung and unable to // do anything other than syscalls really. Consider the following // scenario: // // 1. Thread A of process 1 grabs the malloc() mutex // 2. Thread B of process 1 forks(), creating thread C // 3. Thread C of process 2 then attempts to malloc() // 4. The memory of process 2 is the same as the memory of // process 1, so the mutex is locked. // // This situation looks a lot like deadlock, right? It turns out // that this is what pthread_atfork() takes care of, which is // presumably implemented across platforms. The first thing that // threads to *before* forking is to do things like grab the malloc // mutex, and then after the fork they unlock it. // // Despite this information, libnative's spawn has been witnessed to // deadlock on both OSX and FreeBSD. I'm not entirely sure why, but // all collected backtraces point at malloc/free traffic in the // child spawned process. // // For this reason, the block of code below should contain 0 // invocations of either malloc of free (or their related friends). // // As an example of not having malloc/free traffic, we don't close // this file descriptor by dropping the FileDesc (which contains an // allocation). Instead we just close it manually. This will never // have the drop glue anyway because this code never returns (the // child will either exec() or invoke libc::exit) let _ = libc::close(input.fd()); fn fail(output: &mut FileDesc) -> ! { let errno = sys::os::errno() as u32; let bytes = [ (errno >> 24) as u8, (errno >> 16) as u8, (errno >> 8) as u8, (errno >> 0) as u8, CLOEXEC_MSG_FOOTER[0], CLOEXEC_MSG_FOOTER[1], CLOEXEC_MSG_FOOTER[2], CLOEXEC_MSG_FOOTER[3] ]; // pipe I/O up to PIPE_BUF bytes should be atomic assert!(output.write(&bytes).is_ok()); unsafe { libc::_exit(1) } } rustrt::rust_unset_sigprocmask(); // If a stdio file descriptor is set to be ignored (via a -1 file // descriptor), then we don't actually close it, but rather open // up /dev/null into that file descriptor. Otherwise, the first file // descriptor opened up in the child would be numbered as one of the // stdio file descriptors, which is likely to wreak havoc. let setup = |src: Option<P>, dst: c_int| { let src = match src { None => { let flags = if dst == libc::STDIN_FILENO { libc::O_RDONLY } else { libc::O_RDWR }; libc::open(devnull.as_ptr() as *const _, flags, 0) } Some(obj) => { let fd = obj.as_inner().fd(); // Leak the memory and the file descriptor. We're in the // child now an all our resources are going to be // cleaned up very soon mem::forget(obj); fd } }; src != -1 && retry(|| dup2(src, dst)) != -1 }; if !setup(in_fd, libc::STDIN_FILENO) { fail(&mut output) } if !setup(out_fd, libc::STDOUT_FILENO) { fail(&mut output) } if !setup(err_fd, libc::STDERR_FILENO) { fail(&mut output) } // close all other fds for fd in (3..getdtablesize()).rev() { if fd != output.fd() { let _ = close(fd as c_int); } } match cfg.gid() { Some(u) =>
None => {} } match cfg.uid() { Some(u) => { // When dropping privileges from root, the `setgroups` call // will remove any extraneous groups. If we don't call this, // then even though our uid has dropped, we may still have // groups that enable us to do super-user things. This will // fail if we aren't root, so don't bother checking the // return value, this is just done as an optimistic // privilege dropping function. extern { fn setgroups(ngroups: libc::c_int, ptr: *const libc::c_void) -> libc::c_int; } let _ = setgroups(0, ptr::null()); if libc::setuid(u as libc::uid_t) != 0 { fail(&mut output); } } None => {} } if cfg.detach() { // Don't check the error of setsid because it fails if we're the // process leader already. We just forked so it shouldn't return // error, but ignore it anyway. let _ = libc::setsid(); } if !dirp.is_null() && chdir(dirp) == -1 { fail(&mut output); } if !envp.is_null() { *sys::os::environ() = envp as *const _; } let _ = execvp(*argv, argv as *mut _); fail(&mut output); }) }) } pub fn wait(&self, deadline: u64) -> IoResult<ProcessExit> { use cmp; use sync::mpsc::TryRecvError; static mut WRITE_FD: libc::c_int = 0; let mut status = 0 as c_int; if deadline == 0 { return match retry(|| unsafe { c::waitpid(self.pid, &mut status, 0) }) { -1 => panic!("unknown waitpid error: {:?}", super::last_error()), _ => Ok(translate_status(status)), } } // On unix, wait() and its friends have no timeout parameters, so there is // no way to time out a thread in wait(). From some googling and some // thinking, it appears that there are a few ways to handle timeouts in // wait(), but the only real reasonable one for a multi-threaded program is // to listen for SIGCHLD. // // With this in mind, the waiting mechanism with a timeout barely uses // waitpid() at all. There are a few times that waitpid() is invoked with // WNOHANG, but otherwise all the necessary blocking is done by waiting for // a SIGCHLD to arrive (and that blocking has a timeout). Note, however, // that waitpid() is still used to actually reap the child. // // Signal handling is super tricky in general, and this is no exception. Due // to the async nature of SIGCHLD, we use the self-pipe trick to transmit // data out of the signal handler to the rest of the application. The first // idea would be to have each thread waiting with a timeout to read this // output file descriptor, but a write() is akin to a signal(), not a // broadcast(), so it would only wake up one thread, and possibly the wrong // thread. Hence a helper thread is used. // // The helper thread here is responsible for farming requests for a // waitpid() with a timeout, and then processing all of the wait requests. // By guaranteeing that only this helper thread is reading half of the // self-pipe, we're sure that we'll never lose a SIGCHLD. This helper thread // is also responsible for select() to wait for incoming messages or // incoming SIGCHLD messages, along with passing an appropriate timeout to // select() to wake things up as necessary. // // The ordering of the following statements is also very purposeful. First, // we must be guaranteed that the helper thread is booted and available to // receive SIGCHLD signals, and then we must also ensure that we do a // nonblocking waitpid() at least once before we go ask the sigchld helper. // This prevents the race where the child exits, we boot the helper, and // then we ask for the child's exit status (never seeing a sigchld). // // The actual communication between the helper thread and this thread is // quite simple,
{ if libc::setgid(u as libc::gid_t) != 0 { fail(&mut output); } }
conditional_block
fs.rs
use std::{iter}; use std::collections::{HashMap, HashSet, VecDeque}; use prelude::*; pub struct Filesystem { pub volume: Box<Volume>, pub superblock: Superblock, pub superblock_bytes: Vec<u8>, pub superblock_dirty: bool, pub groups: Vec<Group>, pub inode_cache: HashMap<u64, Inode>, pub dirty_inos: HashSet<u64>, pub reused_inos: HashSet<u64>, pub cache_queue: VecDeque<u64>, } pub struct Group { pub idx: u64, pub desc: GroupDesc, pub block_bitmap: Vec<u8>, pub inode_bitmap: Vec<u8>, pub dirty: bool, } pub const ROOT_INO: u64 = 2; impl Filesystem { pub fn block_size(&self) -> u64 { 1024 << self.superblock.log_block_size } pub fn
(&self) -> u64 { let a = self.superblock.blocks_count as u64; let b = self.superblock.blocks_per_group as u64; (a + b - 1) / b } } pub fn mount_fs(mut volume: Box<Volume>) -> Result<Filesystem> { let mut superblock_bytes = make_buffer(1024); try!(volume.read(1024, &mut superblock_bytes[..])); let superblock = try!(decode_superblock(&superblock_bytes[..], true)); let mut fs = Filesystem { volume: volume, superblock: superblock, superblock_bytes: superblock_bytes, superblock_dirty: false, groups: Vec::new(), inode_cache: HashMap::new(), dirty_inos: HashSet::new(), reused_inos: HashSet::new(), cache_queue: VecDeque::new(), }; for group_idx in 0..fs.group_count() { let group = try!(read_group(&mut fs, group_idx)); fs.groups.push(group); } try!(flush_superblock(&mut fs, false)); Ok(fs) } pub fn flush_fs(fs: &mut Filesystem) -> Result<()> { let dirty_inos = fs.dirty_inos.clone(); for dirty_ino in dirty_inos { try!(flush_ino(fs, dirty_ino)); } for group_idx in 0..fs.group_count() { try!(flush_group(fs, group_idx)); } flush_superblock(fs, true) } fn flush_superblock(fs: &mut Filesystem, clean: bool) -> Result<()> { let state = if clean { 1 } else { 2 }; fs.superblock_dirty = fs.superblock_dirty || fs.superblock.state != state; fs.superblock.state = state; if fs.superblock_dirty { try!(encode_superblock(&fs.superblock, &mut fs.superblock_bytes[..])); try!(fs.volume.write(1024, &fs.superblock_bytes[..])); fs.superblock_dirty = false; } Ok(()) } pub fn make_buffer(size: u64) -> Vec<u8> { iter::repeat(0).take(size as usize).collect() }
group_count
identifier_name
fs.rs
use std::{iter}; use std::collections::{HashMap, HashSet, VecDeque}; use prelude::*; pub struct Filesystem { pub volume: Box<Volume>, pub superblock: Superblock, pub superblock_bytes: Vec<u8>, pub superblock_dirty: bool, pub groups: Vec<Group>, pub inode_cache: HashMap<u64, Inode>, pub dirty_inos: HashSet<u64>, pub reused_inos: HashSet<u64>, pub cache_queue: VecDeque<u64>, } pub struct Group { pub idx: u64, pub desc: GroupDesc, pub block_bitmap: Vec<u8>, pub inode_bitmap: Vec<u8>, pub dirty: bool, } pub const ROOT_INO: u64 = 2; impl Filesystem { pub fn block_size(&self) -> u64 { 1024 << self.superblock.log_block_size } pub fn group_count(&self) -> u64 { let a = self.superblock.blocks_count as u64; let b = self.superblock.blocks_per_group as u64; (a + b - 1) / b } } pub fn mount_fs(mut volume: Box<Volume>) -> Result<Filesystem> { let mut superblock_bytes = make_buffer(1024); try!(volume.read(1024, &mut superblock_bytes[..])); let superblock = try!(decode_superblock(&superblock_bytes[..], true)); let mut fs = Filesystem { volume: volume, superblock: superblock, superblock_bytes: superblock_bytes, superblock_dirty: false, groups: Vec::new(), inode_cache: HashMap::new(), dirty_inos: HashSet::new(), reused_inos: HashSet::new(), cache_queue: VecDeque::new(), }; for group_idx in 0..fs.group_count() { let group = try!(read_group(&mut fs, group_idx)); fs.groups.push(group); } try!(flush_superblock(&mut fs, false)); Ok(fs) } pub fn flush_fs(fs: &mut Filesystem) -> Result<()> { let dirty_inos = fs.dirty_inos.clone(); for dirty_ino in dirty_inos { try!(flush_ino(fs, dirty_ino)); } for group_idx in 0..fs.group_count() { try!(flush_group(fs, group_idx)); } flush_superblock(fs, true) } fn flush_superblock(fs: &mut Filesystem, clean: bool) -> Result<()> { let state = if clean { 1 } else { 2 }; fs.superblock_dirty = fs.superblock_dirty || fs.superblock.state != state; fs.superblock.state = state; if fs.superblock_dirty
Ok(()) } pub fn make_buffer(size: u64) -> Vec<u8> { iter::repeat(0).take(size as usize).collect() }
{ try!(encode_superblock(&fs.superblock, &mut fs.superblock_bytes[..])); try!(fs.volume.write(1024, &fs.superblock_bytes[..])); fs.superblock_dirty = false; }
conditional_block
fs.rs
use std::{iter}; use std::collections::{HashMap, HashSet, VecDeque}; use prelude::*; pub struct Filesystem { pub volume: Box<Volume>, pub superblock: Superblock, pub superblock_bytes: Vec<u8>, pub superblock_dirty: bool, pub groups: Vec<Group>, pub inode_cache: HashMap<u64, Inode>, pub dirty_inos: HashSet<u64>, pub reused_inos: HashSet<u64>, pub cache_queue: VecDeque<u64>, } pub struct Group { pub idx: u64, pub desc: GroupDesc, pub block_bitmap: Vec<u8>, pub inode_bitmap: Vec<u8>, pub dirty: bool, } pub const ROOT_INO: u64 = 2; impl Filesystem { pub fn block_size(&self) -> u64 { 1024 << self.superblock.log_block_size } pub fn group_count(&self) -> u64 { let a = self.superblock.blocks_count as u64; let b = self.superblock.blocks_per_group as u64; (a + b - 1) / b } } pub fn mount_fs(mut volume: Box<Volume>) -> Result<Filesystem>
} try!(flush_superblock(&mut fs, false)); Ok(fs) } pub fn flush_fs(fs: &mut Filesystem) -> Result<()> { let dirty_inos = fs.dirty_inos.clone(); for dirty_ino in dirty_inos { try!(flush_ino(fs, dirty_ino)); } for group_idx in 0..fs.group_count() { try!(flush_group(fs, group_idx)); } flush_superblock(fs, true) } fn flush_superblock(fs: &mut Filesystem, clean: bool) -> Result<()> { let state = if clean { 1 } else { 2 }; fs.superblock_dirty = fs.superblock_dirty || fs.superblock.state != state; fs.superblock.state = state; if fs.superblock_dirty { try!(encode_superblock(&fs.superblock, &mut fs.superblock_bytes[..])); try!(fs.volume.write(1024, &fs.superblock_bytes[..])); fs.superblock_dirty = false; } Ok(()) } pub fn make_buffer(size: u64) -> Vec<u8> { iter::repeat(0).take(size as usize).collect() }
{ let mut superblock_bytes = make_buffer(1024); try!(volume.read(1024, &mut superblock_bytes[..])); let superblock = try!(decode_superblock(&superblock_bytes[..], true)); let mut fs = Filesystem { volume: volume, superblock: superblock, superblock_bytes: superblock_bytes, superblock_dirty: false, groups: Vec::new(), inode_cache: HashMap::new(), dirty_inos: HashSet::new(), reused_inos: HashSet::new(), cache_queue: VecDeque::new(), }; for group_idx in 0..fs.group_count() { let group = try!(read_group(&mut fs, group_idx)); fs.groups.push(group);
identifier_body
fs.rs
use std::{iter}; use std::collections::{HashMap, HashSet, VecDeque}; use prelude::*; pub struct Filesystem { pub volume: Box<Volume>, pub superblock: Superblock, pub superblock_bytes: Vec<u8>, pub superblock_dirty: bool, pub groups: Vec<Group>, pub inode_cache: HashMap<u64, Inode>, pub dirty_inos: HashSet<u64>, pub reused_inos: HashSet<u64>, pub cache_queue: VecDeque<u64>, } pub struct Group { pub idx: u64, pub desc: GroupDesc, pub block_bitmap: Vec<u8>, pub inode_bitmap: Vec<u8>, pub dirty: bool, } pub const ROOT_INO: u64 = 2; impl Filesystem { pub fn block_size(&self) -> u64 { 1024 << self.superblock.log_block_size } pub fn group_count(&self) -> u64 { let a = self.superblock.blocks_count as u64; let b = self.superblock.blocks_per_group as u64; (a + b - 1) / b } } pub fn mount_fs(mut volume: Box<Volume>) -> Result<Filesystem> { let mut superblock_bytes = make_buffer(1024); try!(volume.read(1024, &mut superblock_bytes[..])); let superblock = try!(decode_superblock(&superblock_bytes[..], true)); let mut fs = Filesystem { volume: volume, superblock: superblock, superblock_bytes: superblock_bytes, superblock_dirty: false, groups: Vec::new(), inode_cache: HashMap::new(), dirty_inos: HashSet::new(), reused_inos: HashSet::new(), cache_queue: VecDeque::new(), }; for group_idx in 0..fs.group_count() { let group = try!(read_group(&mut fs, group_idx)); fs.groups.push(group); } try!(flush_superblock(&mut fs, false)); Ok(fs) } pub fn flush_fs(fs: &mut Filesystem) -> Result<()> { let dirty_inos = fs.dirty_inos.clone(); for dirty_ino in dirty_inos {
try!(flush_group(fs, group_idx)); } flush_superblock(fs, true) } fn flush_superblock(fs: &mut Filesystem, clean: bool) -> Result<()> { let state = if clean { 1 } else { 2 }; fs.superblock_dirty = fs.superblock_dirty || fs.superblock.state != state; fs.superblock.state = state; if fs.superblock_dirty { try!(encode_superblock(&fs.superblock, &mut fs.superblock_bytes[..])); try!(fs.volume.write(1024, &fs.superblock_bytes[..])); fs.superblock_dirty = false; } Ok(()) } pub fn make_buffer(size: u64) -> Vec<u8> { iter::repeat(0).take(size as usize).collect() }
try!(flush_ino(fs, dirty_ino)); } for group_idx in 0..fs.group_count() {
random_line_split
enrollmentTermsApi.js
// // Copyright (C) 2016 - present Instructure, Inc. // // This file is part of Canvas. // // Canvas is free software: you can redistribute it and/or modify it under // the terms of the GNU Affero General Public License as published by the Free // Software Foundation, version 3 of the License. // // Canvas is distributed in the hope that it will be useful, but WITHOUT ANY // WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR // A PARTICULAR PURPOSE. See the GNU Affero General Public License for more // details. // // You should have received a copy of the GNU Affero General Public License along // with this program. If not, see <http://www.gnu.org/licenses/>. import _ from 'underscore' import Depaginate from 'jsx/shared/CheatDepaginator' const listUrl = () => ENV.ENROLLMENT_TERMS_URL const deserializeTerms = termGroups => _.flatten( _.map(termGroups, group => _.map(group.enrollment_terms, (term) => { const groupID = term.grading_period_group_id const newGroupID = _.isNumber(groupID) ? groupID.toString() : groupID return { id: term.id.toString(), name: term.name, startAt: term.start_at ? new Date(term.start_at) : null, endAt: term.end_at ? new Date(term.end_at) : null, createdAt: term.created_at ? new Date(term.created_at) : null, gradingPeriodGroupId: newGroupID, } }) ) ) export default { list (terms)
}
{ return new Promise((resolve, reject) => { Depaginate(listUrl()) .then(response => resolve(deserializeTerms(response))) .fail(error => reject(error)) }) }
identifier_body
enrollmentTermsApi.js
// // Copyright (C) 2016 - present Instructure, Inc. // // This file is part of Canvas. // // Canvas is free software: you can redistribute it and/or modify it under // the terms of the GNU Affero General Public License as published by the Free // Software Foundation, version 3 of the License. // // Canvas is distributed in the hope that it will be useful, but WITHOUT ANY // WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR // A PARTICULAR PURPOSE. See the GNU Affero General Public License for more // details. // // You should have received a copy of the GNU Affero General Public License along // with this program. If not, see <http://www.gnu.org/licenses/>. import _ from 'underscore' import Depaginate from 'jsx/shared/CheatDepaginator' const listUrl = () => ENV.ENROLLMENT_TERMS_URL const deserializeTerms = termGroups => _.flatten( _.map(termGroups, group => _.map(group.enrollment_terms, (term) => { const groupID = term.grading_period_group_id const newGroupID = _.isNumber(groupID) ? groupID.toString() : groupID return { id: term.id.toString(), name: term.name, startAt: term.start_at ? new Date(term.start_at) : null, endAt: term.end_at ? new Date(term.end_at) : null, createdAt: term.created_at ? new Date(term.created_at) : null, gradingPeriodGroupId: newGroupID, } }) ) ) export default {
(terms) { return new Promise((resolve, reject) => { Depaginate(listUrl()) .then(response => resolve(deserializeTerms(response))) .fail(error => reject(error)) }) } }
list
identifier_name
enrollmentTermsApi.js
// // Copyright (C) 2016 - present Instructure, Inc. // // This file is part of Canvas. // // Canvas is free software: you can redistribute it and/or modify it under // the terms of the GNU Affero General Public License as published by the Free // Software Foundation, version 3 of the License. // // Canvas is distributed in the hope that it will be useful, but WITHOUT ANY // WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR // A PARTICULAR PURPOSE. See the GNU Affero General Public License for more // details. // // You should have received a copy of the GNU Affero General Public License along // with this program. If not, see <http://www.gnu.org/licenses/>. import _ from 'underscore' import Depaginate from 'jsx/shared/CheatDepaginator' const listUrl = () => ENV.ENROLLMENT_TERMS_URL const deserializeTerms = termGroups => _.flatten( _.map(termGroups, group => _.map(group.enrollment_terms, (term) => { const groupID = term.grading_period_group_id const newGroupID = _.isNumber(groupID) ? groupID.toString() : groupID return { id: term.id.toString(), name: term.name, startAt: term.start_at ? new Date(term.start_at) : null, endAt: term.end_at ? new Date(term.end_at) : null, createdAt: term.created_at ? new Date(term.created_at) : null, gradingPeriodGroupId: newGroupID, } }) ) )
return new Promise((resolve, reject) => { Depaginate(listUrl()) .then(response => resolve(deserializeTerms(response))) .fail(error => reject(error)) }) } }
export default { list (terms) {
random_line_split
yaml_generator.py
YAML_OUTPUT = """terminals: %s non-terminals: %s eof-marker: %s error-marker: %s start-symbol: %s productions: %s table: %s""" YAML_OUTPUT_NO_TABLE = """terminals: %s non-terminals: %s eof-marker: %s error-marker: %s start-symbol: %s productions: %s""" class YamlGenerator(object): """docstring for yaml_generator""" def __init__(self, grammar): self.grammar = grammar def print_yaml(self, ll1_table = None): def convert_list_str(a_list): return "[%s]" % (", ".join(a_list)) def convert_dict_str(a_dict): return "{%s}" % ", ".join(["%s: %s" % (key, value) for key, value in a_dict.items()]) def convert_dict_dict_str(a_dict): return "\n %s" % ("\n ".join(["%s: %s" % (key, convert_dict_str(value)) for key, value in a_dict.items()])) def convert_dict_list_str(a_dict): return "{%s}" % (", \n ".join(["%s: %s" % (key, convert_list_str(value)) for key, value in a_dict.items()])) def convert_dict_dict_list_str(a_dict): return "\n %s" % ("\n ".join(["%s: %s" % (key, convert_dict_list_str(value)) for key, value in a_dict.items()])) if ll1_table: return YAML_OUTPUT % (convert_list_str(list(self.grammar.term)), convert_list_str(list(self.grammar.non_term)), EOF, ERROR_MARKER, self.grammar.goal, convert_dict_dict_list_str(self.convert_production()), convert_dict_dict_str(ll1_table)) else: return YAML_OUTPUT_NO_TABLE % (convert_list_str(list(self.grammar.term)), convert_list_str(list(self.grammar.non_term)), EOF, ERROR_MARKER, self.grammar.goal, convert_dict_dict_list_str(self.convert_production())) def convert_production(self): return {idx : {production.left_hand.lexeme : [item.lexeme for item in production.right_hand if item.lexeme is not EPSILON]} for idx, production in enumerate(self.grammar.production)}
from ll1_symbols import *
random_line_split
yaml_generator.py
from ll1_symbols import * YAML_OUTPUT = """terminals: %s non-terminals: %s eof-marker: %s error-marker: %s start-symbol: %s productions: %s table: %s""" YAML_OUTPUT_NO_TABLE = """terminals: %s non-terminals: %s eof-marker: %s error-marker: %s start-symbol: %s productions: %s""" class YamlGenerator(object): """docstring for yaml_generator""" def __init__(self, grammar): self.grammar = grammar def print_yaml(self, ll1_table = None): def convert_list_str(a_list): return "[%s]" % (", ".join(a_list)) def convert_dict_str(a_dict):
def convert_dict_dict_str(a_dict): return "\n %s" % ("\n ".join(["%s: %s" % (key, convert_dict_str(value)) for key, value in a_dict.items()])) def convert_dict_list_str(a_dict): return "{%s}" % (", \n ".join(["%s: %s" % (key, convert_list_str(value)) for key, value in a_dict.items()])) def convert_dict_dict_list_str(a_dict): return "\n %s" % ("\n ".join(["%s: %s" % (key, convert_dict_list_str(value)) for key, value in a_dict.items()])) if ll1_table: return YAML_OUTPUT % (convert_list_str(list(self.grammar.term)), convert_list_str(list(self.grammar.non_term)), EOF, ERROR_MARKER, self.grammar.goal, convert_dict_dict_list_str(self.convert_production()), convert_dict_dict_str(ll1_table)) else: return YAML_OUTPUT_NO_TABLE % (convert_list_str(list(self.grammar.term)), convert_list_str(list(self.grammar.non_term)), EOF, ERROR_MARKER, self.grammar.goal, convert_dict_dict_list_str(self.convert_production())) def convert_production(self): return {idx : {production.left_hand.lexeme : [item.lexeme for item in production.right_hand if item.lexeme is not EPSILON]} for idx, production in enumerate(self.grammar.production)}
return "{%s}" % ", ".join(["%s: %s" % (key, value) for key, value in a_dict.items()])
identifier_body
yaml_generator.py
from ll1_symbols import * YAML_OUTPUT = """terminals: %s non-terminals: %s eof-marker: %s error-marker: %s start-symbol: %s productions: %s table: %s""" YAML_OUTPUT_NO_TABLE = """terminals: %s non-terminals: %s eof-marker: %s error-marker: %s start-symbol: %s productions: %s""" class YamlGenerator(object): """docstring for yaml_generator""" def __init__(self, grammar): self.grammar = grammar def print_yaml(self, ll1_table = None): def convert_list_str(a_list): return "[%s]" % (", ".join(a_list)) def convert_dict_str(a_dict): return "{%s}" % ", ".join(["%s: %s" % (key, value) for key, value in a_dict.items()]) def convert_dict_dict_str(a_dict): return "\n %s" % ("\n ".join(["%s: %s" % (key, convert_dict_str(value)) for key, value in a_dict.items()])) def convert_dict_list_str(a_dict): return "{%s}" % (", \n ".join(["%s: %s" % (key, convert_list_str(value)) for key, value in a_dict.items()])) def convert_dict_dict_list_str(a_dict): return "\n %s" % ("\n ".join(["%s: %s" % (key, convert_dict_list_str(value)) for key, value in a_dict.items()])) if ll1_table:
else: return YAML_OUTPUT_NO_TABLE % (convert_list_str(list(self.grammar.term)), convert_list_str(list(self.grammar.non_term)), EOF, ERROR_MARKER, self.grammar.goal, convert_dict_dict_list_str(self.convert_production())) def convert_production(self): return {idx : {production.left_hand.lexeme : [item.lexeme for item in production.right_hand if item.lexeme is not EPSILON]} for idx, production in enumerate(self.grammar.production)}
return YAML_OUTPUT % (convert_list_str(list(self.grammar.term)), convert_list_str(list(self.grammar.non_term)), EOF, ERROR_MARKER, self.grammar.goal, convert_dict_dict_list_str(self.convert_production()), convert_dict_dict_str(ll1_table))
conditional_block
yaml_generator.py
from ll1_symbols import * YAML_OUTPUT = """terminals: %s non-terminals: %s eof-marker: %s error-marker: %s start-symbol: %s productions: %s table: %s""" YAML_OUTPUT_NO_TABLE = """terminals: %s non-terminals: %s eof-marker: %s error-marker: %s start-symbol: %s productions: %s""" class YamlGenerator(object): """docstring for yaml_generator""" def __init__(self, grammar): self.grammar = grammar def print_yaml(self, ll1_table = None): def
(a_list): return "[%s]" % (", ".join(a_list)) def convert_dict_str(a_dict): return "{%s}" % ", ".join(["%s: %s" % (key, value) for key, value in a_dict.items()]) def convert_dict_dict_str(a_dict): return "\n %s" % ("\n ".join(["%s: %s" % (key, convert_dict_str(value)) for key, value in a_dict.items()])) def convert_dict_list_str(a_dict): return "{%s}" % (", \n ".join(["%s: %s" % (key, convert_list_str(value)) for key, value in a_dict.items()])) def convert_dict_dict_list_str(a_dict): return "\n %s" % ("\n ".join(["%s: %s" % (key, convert_dict_list_str(value)) for key, value in a_dict.items()])) if ll1_table: return YAML_OUTPUT % (convert_list_str(list(self.grammar.term)), convert_list_str(list(self.grammar.non_term)), EOF, ERROR_MARKER, self.grammar.goal, convert_dict_dict_list_str(self.convert_production()), convert_dict_dict_str(ll1_table)) else: return YAML_OUTPUT_NO_TABLE % (convert_list_str(list(self.grammar.term)), convert_list_str(list(self.grammar.non_term)), EOF, ERROR_MARKER, self.grammar.goal, convert_dict_dict_list_str(self.convert_production())) def convert_production(self): return {idx : {production.left_hand.lexeme : [item.lexeme for item in production.right_hand if item.lexeme is not EPSILON]} for idx, production in enumerate(self.grammar.production)}
convert_list_str
identifier_name
misc.rs
// Copyright 2015-2017 Parity Technologies (UK) Ltd. // This file is part of Parity. // Parity is free software: you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // Parity is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with Parity. If not, see <http://www.gnu.org/licenses/>. //! Diff misc. use Bytes; use rlp::RlpStream; use target_info::Target; include!(concat!(env!("OUT_DIR"), "/version.rs")); include!(concat!(env!("OUT_DIR"), "/rustc_version.rs")); #[cfg(feature = "final")] const THIS_TRACK: &'static str = "nightly"; // ^^^ should be reset to "stable" or "beta" according to the release branch. #[cfg(not(feature = "final"))] const THIS_TRACK: &'static str = "unstable"; // ^^^ This gets used when we're not building a final release; should stay as "unstable". /// Boolean type for clean/dirty status. #[derive(PartialEq, Eq, Clone, Copy, Debug)] pub enum Filth { /// Data has not been changed. Clean, /// Data has been changed. Dirty, } /// Get the platform identifier. pub fn platform() -> String { let env = Target::env(); let env_dash = if env.is_empty() { "" } else { "-" }; format!("{}-{}{}{}", Target::arch(), Target::os(), env_dash, env) } /// Get the standard version string for this software. pub fn version() -> String { let sha3 = short_sha(); let sha3_dash = if sha3.is_empty() { "" } else { "-" }; let commit_date = commit_date().replace("-", ""); let date_dash = if commit_date.is_empty() { "" } else { "-" }; format!("Parity/v{}-{}{}{}{}{}/{}/rustc{}", env!("CARGO_PKG_VERSION"), THIS_TRACK, sha3_dash, sha3, date_dash, commit_date, platform(), rustc_version()) } /// Get the standard version data for this software. pub fn version_data() -> Bytes { let mut s = RlpStream::new_list(4); let v = (env!("CARGO_PKG_VERSION_MAJOR") .parse::<u32>() .expect("Environment variables are known to be valid; qed") << 16) + (env!("CARGO_PKG_VERSION_MINOR") .parse::<u32>() .expect("Environment variables are known to be valid; qed") << 8) + env!("CARGO_PKG_VERSION_PATCH") .parse::<u32>() .expect("Environment variables are known to be valid; qed"); s.append(&v); s.append(&"Parity"); s.append(&rustc_version()); s.append(&&Target::os()[0..2]); s.out() } /// Provide raw information on the package. pub fn raw_package_info() -> (&'static str, &'static str, &'static str)
{ (THIS_TRACK, env!["CARGO_PKG_VERSION"], sha()) }
identifier_body
misc.rs
// Copyright 2015-2017 Parity Technologies (UK) Ltd. // This file is part of Parity. // Parity is free software: you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // Parity is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with Parity. If not, see <http://www.gnu.org/licenses/>. //! Diff misc. use Bytes; use rlp::RlpStream; use target_info::Target; include!(concat!(env!("OUT_DIR"), "/version.rs")); include!(concat!(env!("OUT_DIR"), "/rustc_version.rs")); #[cfg(feature = "final")] const THIS_TRACK: &'static str = "nightly"; // ^^^ should be reset to "stable" or "beta" according to the release branch. #[cfg(not(feature = "final"))] const THIS_TRACK: &'static str = "unstable"; // ^^^ This gets used when we're not building a final release; should stay as "unstable". /// Boolean type for clean/dirty status. #[derive(PartialEq, Eq, Clone, Copy, Debug)] pub enum Filth { /// Data has not been changed. Clean, /// Data has been changed. Dirty, } /// Get the platform identifier. pub fn platform() -> String { let env = Target::env(); let env_dash = if env.is_empty() { "" } else { "-" }; format!("{}-{}{}{}", Target::arch(), Target::os(), env_dash, env) } /// Get the standard version string for this software. pub fn version() -> String { let sha3 = short_sha(); let sha3_dash = if sha3.is_empty() { "" } else { "-" }; let commit_date = commit_date().replace("-", ""); let date_dash = if commit_date.is_empty() { "" } else
; format!("Parity/v{}-{}{}{}{}{}/{}/rustc{}", env!("CARGO_PKG_VERSION"), THIS_TRACK, sha3_dash, sha3, date_dash, commit_date, platform(), rustc_version()) } /// Get the standard version data for this software. pub fn version_data() -> Bytes { let mut s = RlpStream::new_list(4); let v = (env!("CARGO_PKG_VERSION_MAJOR") .parse::<u32>() .expect("Environment variables are known to be valid; qed") << 16) + (env!("CARGO_PKG_VERSION_MINOR") .parse::<u32>() .expect("Environment variables are known to be valid; qed") << 8) + env!("CARGO_PKG_VERSION_PATCH") .parse::<u32>() .expect("Environment variables are known to be valid; qed"); s.append(&v); s.append(&"Parity"); s.append(&rustc_version()); s.append(&&Target::os()[0..2]); s.out() } /// Provide raw information on the package. pub fn raw_package_info() -> (&'static str, &'static str, &'static str) { (THIS_TRACK, env!["CARGO_PKG_VERSION"], sha()) }
{ "-" }
conditional_block
misc.rs
// Copyright 2015-2017 Parity Technologies (UK) Ltd. // This file is part of Parity. // Parity is free software: you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // Parity is distributed in the hope that it will be useful,
// You should have received a copy of the GNU General Public License // along with Parity. If not, see <http://www.gnu.org/licenses/>. //! Diff misc. use Bytes; use rlp::RlpStream; use target_info::Target; include!(concat!(env!("OUT_DIR"), "/version.rs")); include!(concat!(env!("OUT_DIR"), "/rustc_version.rs")); #[cfg(feature = "final")] const THIS_TRACK: &'static str = "nightly"; // ^^^ should be reset to "stable" or "beta" according to the release branch. #[cfg(not(feature = "final"))] const THIS_TRACK: &'static str = "unstable"; // ^^^ This gets used when we're not building a final release; should stay as "unstable". /// Boolean type for clean/dirty status. #[derive(PartialEq, Eq, Clone, Copy, Debug)] pub enum Filth { /// Data has not been changed. Clean, /// Data has been changed. Dirty, } /// Get the platform identifier. pub fn platform() -> String { let env = Target::env(); let env_dash = if env.is_empty() { "" } else { "-" }; format!("{}-{}{}{}", Target::arch(), Target::os(), env_dash, env) } /// Get the standard version string for this software. pub fn version() -> String { let sha3 = short_sha(); let sha3_dash = if sha3.is_empty() { "" } else { "-" }; let commit_date = commit_date().replace("-", ""); let date_dash = if commit_date.is_empty() { "" } else { "-" }; format!("Parity/v{}-{}{}{}{}{}/{}/rustc{}", env!("CARGO_PKG_VERSION"), THIS_TRACK, sha3_dash, sha3, date_dash, commit_date, platform(), rustc_version()) } /// Get the standard version data for this software. pub fn version_data() -> Bytes { let mut s = RlpStream::new_list(4); let v = (env!("CARGO_PKG_VERSION_MAJOR") .parse::<u32>() .expect("Environment variables are known to be valid; qed") << 16) + (env!("CARGO_PKG_VERSION_MINOR") .parse::<u32>() .expect("Environment variables are known to be valid; qed") << 8) + env!("CARGO_PKG_VERSION_PATCH") .parse::<u32>() .expect("Environment variables are known to be valid; qed"); s.append(&v); s.append(&"Parity"); s.append(&rustc_version()); s.append(&&Target::os()[0..2]); s.out() } /// Provide raw information on the package. pub fn raw_package_info() -> (&'static str, &'static str, &'static str) { (THIS_TRACK, env!["CARGO_PKG_VERSION"], sha()) }
// but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details.
random_line_split
misc.rs
// Copyright 2015-2017 Parity Technologies (UK) Ltd. // This file is part of Parity. // Parity is free software: you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // Parity is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with Parity. If not, see <http://www.gnu.org/licenses/>. //! Diff misc. use Bytes; use rlp::RlpStream; use target_info::Target; include!(concat!(env!("OUT_DIR"), "/version.rs")); include!(concat!(env!("OUT_DIR"), "/rustc_version.rs")); #[cfg(feature = "final")] const THIS_TRACK: &'static str = "nightly"; // ^^^ should be reset to "stable" or "beta" according to the release branch. #[cfg(not(feature = "final"))] const THIS_TRACK: &'static str = "unstable"; // ^^^ This gets used when we're not building a final release; should stay as "unstable". /// Boolean type for clean/dirty status. #[derive(PartialEq, Eq, Clone, Copy, Debug)] pub enum Filth { /// Data has not been changed. Clean, /// Data has been changed. Dirty, } /// Get the platform identifier. pub fn platform() -> String { let env = Target::env(); let env_dash = if env.is_empty() { "" } else { "-" }; format!("{}-{}{}{}", Target::arch(), Target::os(), env_dash, env) } /// Get the standard version string for this software. pub fn version() -> String { let sha3 = short_sha(); let sha3_dash = if sha3.is_empty() { "" } else { "-" }; let commit_date = commit_date().replace("-", ""); let date_dash = if commit_date.is_empty() { "" } else { "-" }; format!("Parity/v{}-{}{}{}{}{}/{}/rustc{}", env!("CARGO_PKG_VERSION"), THIS_TRACK, sha3_dash, sha3, date_dash, commit_date, platform(), rustc_version()) } /// Get the standard version data for this software. pub fn version_data() -> Bytes { let mut s = RlpStream::new_list(4); let v = (env!("CARGO_PKG_VERSION_MAJOR") .parse::<u32>() .expect("Environment variables are known to be valid; qed") << 16) + (env!("CARGO_PKG_VERSION_MINOR") .parse::<u32>() .expect("Environment variables are known to be valid; qed") << 8) + env!("CARGO_PKG_VERSION_PATCH") .parse::<u32>() .expect("Environment variables are known to be valid; qed"); s.append(&v); s.append(&"Parity"); s.append(&rustc_version()); s.append(&&Target::os()[0..2]); s.out() } /// Provide raw information on the package. pub fn
() -> (&'static str, &'static str, &'static str) { (THIS_TRACK, env!["CARGO_PKG_VERSION"], sha()) }
raw_package_info
identifier_name
bp.py
import math import net SIGMOID = 0 TANH = 1 class bp: def __init__(self, net, learning_rate, momentum): self.type = net.getType() self.net = net self.lr = learning_rate self.m = momentum self.layer = net.getLayer() self.lc = [[[0]*max(self.layer)]*max(self.layer)]*len(self.layer) def _dfunc(self, y): if self.type==SIGMOID: return y * (1.0 - y) else: return 1.0 - y**2 def setLearningRate(self,x):
def setMomentum(self, x): self.m = x def backPropagate(self, input, target): if len(target)!=self.layer[-1]: print len(target) print self.layer[-1] raise ValueError('Wrong number of target values') self.net.process(input) nlayer = len(self.layer) delta = [] for i in range(0, nlayer): delta.append([0.0] * self.layer[i]) for i in range(0,self.layer[nlayer-1]): node = self.net.getNode(nlayer-1, i) error = target[i] - node delta[nlayer-1][i] = self._dfunc(node) * error for l in range(nlayer-2, 0, -1): for i in range(0, self.layer[l]): error = 0.0 for j in range(0, self.layer[l+1]): error = error + delta[l+1][j] * self.net.getWeight(l+1, i, j) delta[l][i] = self._dfunc(self.net.getNode(l,i)) * error for l in range(nlayer-2, -1, -1): for i in range(0, self.layer[l]): for j in range(0, self.layer[l+1]): change = delta[l+1][j] * self.net.getNode(l, i) w = self.net.getWeight(l+1, i, j) + self.lr * change + self.m * self.lc[l+1][i][j] self.net.setWeight(l+1, i, j, w) self.lc[l+1][i][j] = change for i in range(0, self.layer[l+1]): b = self.net.getBias(l+1, i) + delta[l+1][i] self.net.setBias(l+1, i, b) error = 0.0 for i in range(0, len(target)): error = error + 0.5 * (target[i] - self.net.getNode(nlayer-1, i))**2 return error
self.lr = x
identifier_body
bp.py
import math import net SIGMOID = 0 TANH = 1 class
: def __init__(self, net, learning_rate, momentum): self.type = net.getType() self.net = net self.lr = learning_rate self.m = momentum self.layer = net.getLayer() self.lc = [[[0]*max(self.layer)]*max(self.layer)]*len(self.layer) def _dfunc(self, y): if self.type==SIGMOID: return y * (1.0 - y) else: return 1.0 - y**2 def setLearningRate(self,x): self.lr = x def setMomentum(self, x): self.m = x def backPropagate(self, input, target): if len(target)!=self.layer[-1]: print len(target) print self.layer[-1] raise ValueError('Wrong number of target values') self.net.process(input) nlayer = len(self.layer) delta = [] for i in range(0, nlayer): delta.append([0.0] * self.layer[i]) for i in range(0,self.layer[nlayer-1]): node = self.net.getNode(nlayer-1, i) error = target[i] - node delta[nlayer-1][i] = self._dfunc(node) * error for l in range(nlayer-2, 0, -1): for i in range(0, self.layer[l]): error = 0.0 for j in range(0, self.layer[l+1]): error = error + delta[l+1][j] * self.net.getWeight(l+1, i, j) delta[l][i] = self._dfunc(self.net.getNode(l,i)) * error for l in range(nlayer-2, -1, -1): for i in range(0, self.layer[l]): for j in range(0, self.layer[l+1]): change = delta[l+1][j] * self.net.getNode(l, i) w = self.net.getWeight(l+1, i, j) + self.lr * change + self.m * self.lc[l+1][i][j] self.net.setWeight(l+1, i, j, w) self.lc[l+1][i][j] = change for i in range(0, self.layer[l+1]): b = self.net.getBias(l+1, i) + delta[l+1][i] self.net.setBias(l+1, i, b) error = 0.0 for i in range(0, len(target)): error = error + 0.5 * (target[i] - self.net.getNode(nlayer-1, i))**2 return error
bp
identifier_name
bp.py
import math import net SIGMOID = 0 TANH = 1 class bp: def __init__(self, net, learning_rate, momentum): self.type = net.getType() self.net = net
self.lc = [[[0]*max(self.layer)]*max(self.layer)]*len(self.layer) def _dfunc(self, y): if self.type==SIGMOID: return y * (1.0 - y) else: return 1.0 - y**2 def setLearningRate(self,x): self.lr = x def setMomentum(self, x): self.m = x def backPropagate(self, input, target): if len(target)!=self.layer[-1]: print len(target) print self.layer[-1] raise ValueError('Wrong number of target values') self.net.process(input) nlayer = len(self.layer) delta = [] for i in range(0, nlayer): delta.append([0.0] * self.layer[i]) for i in range(0,self.layer[nlayer-1]): node = self.net.getNode(nlayer-1, i) error = target[i] - node delta[nlayer-1][i] = self._dfunc(node) * error for l in range(nlayer-2, 0, -1): for i in range(0, self.layer[l]): error = 0.0 for j in range(0, self.layer[l+1]): error = error + delta[l+1][j] * self.net.getWeight(l+1, i, j) delta[l][i] = self._dfunc(self.net.getNode(l,i)) * error for l in range(nlayer-2, -1, -1): for i in range(0, self.layer[l]): for j in range(0, self.layer[l+1]): change = delta[l+1][j] * self.net.getNode(l, i) w = self.net.getWeight(l+1, i, j) + self.lr * change + self.m * self.lc[l+1][i][j] self.net.setWeight(l+1, i, j, w) self.lc[l+1][i][j] = change for i in range(0, self.layer[l+1]): b = self.net.getBias(l+1, i) + delta[l+1][i] self.net.setBias(l+1, i, b) error = 0.0 for i in range(0, len(target)): error = error + 0.5 * (target[i] - self.net.getNode(nlayer-1, i))**2 return error
self.lr = learning_rate self.m = momentum self.layer = net.getLayer()
random_line_split
bp.py
import math import net SIGMOID = 0 TANH = 1 class bp: def __init__(self, net, learning_rate, momentum): self.type = net.getType() self.net = net self.lr = learning_rate self.m = momentum self.layer = net.getLayer() self.lc = [[[0]*max(self.layer)]*max(self.layer)]*len(self.layer) def _dfunc(self, y): if self.type==SIGMOID: return y * (1.0 - y) else: return 1.0 - y**2 def setLearningRate(self,x): self.lr = x def setMomentum(self, x): self.m = x def backPropagate(self, input, target): if len(target)!=self.layer[-1]: print len(target) print self.layer[-1] raise ValueError('Wrong number of target values') self.net.process(input) nlayer = len(self.layer) delta = [] for i in range(0, nlayer): delta.append([0.0] * self.layer[i]) for i in range(0,self.layer[nlayer-1]): node = self.net.getNode(nlayer-1, i) error = target[i] - node delta[nlayer-1][i] = self._dfunc(node) * error for l in range(nlayer-2, 0, -1): for i in range(0, self.layer[l]): error = 0.0 for j in range(0, self.layer[l+1]): error = error + delta[l+1][j] * self.net.getWeight(l+1, i, j) delta[l][i] = self._dfunc(self.net.getNode(l,i)) * error for l in range(nlayer-2, -1, -1):
error = 0.0 for i in range(0, len(target)): error = error + 0.5 * (target[i] - self.net.getNode(nlayer-1, i))**2 return error
for i in range(0, self.layer[l]): for j in range(0, self.layer[l+1]): change = delta[l+1][j] * self.net.getNode(l, i) w = self.net.getWeight(l+1, i, j) + self.lr * change + self.m * self.lc[l+1][i][j] self.net.setWeight(l+1, i, j, w) self.lc[l+1][i][j] = change for i in range(0, self.layer[l+1]): b = self.net.getBias(l+1, i) + delta[l+1][i] self.net.setBias(l+1, i, b)
conditional_block
conf.py
# # Note that not all possible configuration values are present in this # autogenerated file. # # All configuration values have a default; values that are commented out # serve to show the default. import sys import os # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. #sys.path.insert(0, os.path.abspath('.')) sys.path.insert(0, os.path.abspath('../')) sys.path.insert(0, os.path.abspath('../totalopenstation')) # -- General configuration ------------------------------------------------ # If your documentation needs a minimal Sphinx version, state it here. #needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = [ 'sphinx.ext.autodoc', 'sphinx.ext.napoleon', ] # Add any paths that contain templates here, relative to this directory. templates_path = ['_templates'] # The suffix of source filenames. source_suffix = '.rst' # The encoding of source files. #source_encoding = 'utf-8-sig' # The master toctree document. master_doc = 'index' # General information about the project. project = 'Total Open Station' copyright = '2015-2020, Stefano Costa, Damien Gaignon and Luca Bianconi' author = 'Stefano Costa' # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. version = '0.5' # The full version, including alpha/beta/rc tags. release = '0.5.2' # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. #language = None # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: #today = '' # Else, today_fmt is used as the format for a strftime call. #today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. exclude_patterns = ['_build', 'global.rst', ] # The reST default role (used for this markup: `text`) to use for all # documents. #default_role = None # If true, '()' will be appended to :func: etc. cross-reference text. #add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). #add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. #show_authors = False # The name of the Pygments (syntax highlighting) style to use.
# A list of ignored prefixes for module index sorting. #modindex_common_prefix = [] # If true, keep warnings as "system message" paragraphs in the built documents. #keep_warnings = False rst_prolog = """ .. include:: /global.rst """ # -- Options for HTML output ---------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. html_theme = 'alabaster' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. html_theme_options = { 'github_user': 'totalopenstation', 'github_repo': 'totalopenstation', 'github_type': 'star', 'github_count': 'true', 'github_button': True, 'description': 'Download and export field survey data from your total station' } # Add any paths that contain custom themes here, relative to this directory. #html_theme_path = [] # The name for this set of Sphinx documents. If None, it defaults to # "<project> v<release> documentation". #html_title = None # A shorter title for the navigation bar. Default is the same as html_title. #html_short_title = None # The name of an image file (relative to this directory) to place at the top # of the sidebar. #html_logo = None # The name of an image file (within the static path) to use as favicon of the # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. html_favicon = "tops.ico" # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ['_static'] # Add any extra paths that contain custom files (such as robots.txt or # .htaccess) here, relative to this directory. These files are copied # directly to the root of the documentation. #html_extra_path = [] # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. #html_last_updated_fmt = '%b %d, %Y' # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. html_use_smartypants = True # Custom sidebar templates, maps document names to template names. html_sidebars = { '**': [ 'about.html', 'navigation.html', 'relations.html', 'searchbox.html', 'donate.html', ] } # Additional templates that should be rendered to pages, maps page names to # template names. #html_additional_pages = {} # If false, no module index is generated. #html_domain_indices = True # If false, no index is generated. #html_use_index = True # If true, the index is split into individual pages for each letter. #html_split_index = False # If true, links to the reST sources are added to the pages. #html_show_sourcelink = True # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. #html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. #html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. #html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). #html_file_suffix = None # Output file base name for HTML help builder. htmlhelp_basename = 'TotalOpenStationdoc' # -- Options for LaTeX output --------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). 'papersize': 'a4paper', # The font size ('10pt', '11pt' or '12pt'). 'pointsize': '12pt', # Additional stuff for the LaTeX preamble. #'preamble': '', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, # author, documentclass [howto, manual, or own class]). latex_documents = [ ('index', 'TotalOpenStation.tex', 'Total Open Station Documentation', 'Stefano Costa, Damien Gaignon, Luca Bianconi', 'manual'), ] # The name of an image file (relative to this directory) to place at the top of # the title page. #latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. #latex_use_parts = False # If true, show page references after internal links. #latex_show_pagerefs = False # If true, show URL addresses after external links. #latex_show_urls = False # Documents to append as an appendix to all manuals. #latex_appendices = [] # If false, no module index is generated. #latex_domain_indices = True # -- Options for manual page output --------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [ ('totalopenstation-cli-parser', 'totalopenstation-cli-parser', 'Total Open Station command line converter', ['Stefano Costa, Luca Bianconi'], 1), ('totalopenstation-cli-connector', 'totalopenstation-cli-connector', 'Total Open Station command line downloader', ['Stefano Costa, Luca Bianconi'], 1) ] # If true, show URL addresses after external links. #man_show_urls = False # -- Options for Texinfo output ------------------------------------------- # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ ('index', 'TotalOpenStation', 'Total Open Station Documentation', 'Stef
pygments_style = 'sphinx'
random_line_split
gdocsbackend.py
# -*- Mode:Python; indent-tabs-mode:nil; tab-width:4 -*- # # Copyright 2011 Carlos Abalde <carlos.abalde@gmail.com> # # This file is part of duplicity. # # Duplicity is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the # Free Software Foundation; either version 2 of the License, or (at your # option) any later version. # # Duplicity is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with duplicity; if not, write to the Free Software Foundation, # Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA import os.path import string import urllib import duplicity.backend from duplicity.errors import BackendException class GDocsBackend(duplicity.backend.Backend): """Connect to remote store using Google Google Documents List API""" ROOT_FOLDER_ID = 'folder%3Aroot' BACKUP_DOCUMENT_TYPE = 'application/binary' def __init__(self, parsed_url):
# Fetch destination folder entry (and crete hierarchy if required). folder_names = string.split(parsed_url.path[1:], '/') parent_folder = None parent_folder_id = GDocsBackend.ROOT_FOLDER_ID for folder_name in folder_names: entries = self._fetch_entries(parent_folder_id, 'folder', folder_name) if entries is not None: if len(entries) == 1: parent_folder = entries[0] elif len(entries) == 0: folder = gdata.docs.data.Resource(type='folder', title=folder_name) parent_folder = self.client.create_resource(folder, collection=parent_folder) else: parent_folder = None if parent_folder: parent_folder_id = parent_folder.resource_id.text else: raise BackendException("Error while creating destination folder '%s'." % folder_name) else: raise BackendException("Error while fetching destination folder '%s'." % folder_name) self.folder = parent_folder def _put(self, source_path, remote_filename): self._delete(remote_filename) # Set uploader instance. Note that resumable uploads are required in order to # enable uploads for all file types. # (see http://googleappsdeveloper.blogspot.com/2011/05/upload-all-file-types-to-any-google.html) file = source_path.open() uploader = gdata.client.ResumableUploader( self.client, file, GDocsBackend.BACKUP_DOCUMENT_TYPE, os.path.getsize(file.name), chunk_size=gdata.client.ResumableUploader.DEFAULT_CHUNK_SIZE, desired_class=gdata.docs.data.Resource) if uploader: # Chunked upload. entry = gdata.docs.data.Resource(title=atom.data.Title(text=remote_filename)) uri = self.folder.get_resumable_create_media_link().href + '?convert=false' entry = uploader.UploadFile(uri, entry=entry) if not entry: raise BackendException("Failed to upload file '%s' to remote folder '%s'" % (source_path.get_filename(), self.folder.title.text)) else: raise BackendException("Failed to initialize upload of file '%s' to remote folder '%s'" % (source_path.get_filename(), self.folder.title.text)) assert not file.close() def _get(self, remote_filename, local_path): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE, remote_filename) if len(entries) == 1: entry = entries[0] self.client.DownloadResource(entry, local_path.name) else: raise BackendException("Failed to find file '%s' in remote folder '%s'" % (remote_filename, self.folder.title.text)) def _list(self): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE) return [entry.title.text for entry in entries] def _delete(self, filename): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE, filename) for entry in entries: self.client.delete(entry.get_edit_link().href + '?delete=true', force=True) def _authorize(self, email, password, captcha_token=None, captcha_response=None): try: self.client.client_login(email, password, source='duplicity $version', service='writely', captcha_token=captcha_token, captcha_response=captcha_response) except gdata.client.CaptchaChallenge as challenge: print('A captcha challenge in required. Please visit ' + challenge.captcha_url) answer = None while not answer: answer = raw_input('Answer to the challenge? ') self._authorize(email, password, challenge.captcha_token, answer) except gdata.client.BadAuthentication: raise BackendException( 'Invalid user credentials given. Be aware that accounts ' 'that use 2-step verification require creating an application specific ' 'access code for using this Duplicity backend. Follow the instruction in ' 'http://www.google.com/support/accounts/bin/static.py?page=guide.cs&guide=1056283&topic=1056286 ' 'and create your application-specific password to run duplicity backups.') def _fetch_entries(self, folder_id, type, title=None): # Build URI. uri = '/feeds/default/private/full/%s/contents' % folder_id if type == 'folder': uri += '/-/folder?showfolders=true' elif type == GDocsBackend.BACKUP_DOCUMENT_TYPE: uri += '?showfolders=false' else: uri += '?showfolders=true' if title: uri += '&title=' + urllib.quote(title) + '&title-exact=true' # Fetch entries. entries = self.client.get_all_resources(uri=uri) # When filtering by entry title, API is returning (don't know why) documents in other # folders (apart from folder_id) matching the title, so some extra filtering is required. if title: result = [] for entry in entries: resource_type = entry.get_resource_type() if (not type) \ or (type == 'folder' and resource_type == 'folder') \ or (type == GDocsBackend.BACKUP_DOCUMENT_TYPE and resource_type != 'folder'): if folder_id != GDocsBackend.ROOT_FOLDER_ID: for link in entry.in_collections(): folder_entry = self.client.get_entry(link.href, None, None, desired_class=gdata.docs.data.Resource) if folder_entry and (folder_entry.resource_id.text == folder_id): result.append(entry) elif len(entry.in_collections()) == 0: result.append(entry) else: result = entries # Done! return result """ gdata is an alternate way to access gdocs, currently 05/2015 lacking OAuth support """ duplicity.backend.register_backend('gdata+gdocs', GDocsBackend) duplicity.backend.uses_netloc.extend(['gdata+gdocs'])
duplicity.backend.Backend.__init__(self, parsed_url) # Import Google Data APIs libraries. try: global atom global gdata import atom.data import gdata.client import gdata.docs.client import gdata.docs.data except ImportError as e: raise BackendException("""\ Google Docs backend requires Google Data APIs Python Client Library (see http://code.google.com/p/gdata-python-client/). Exception: %s""" % str(e)) # Setup client instance. self.client = gdata.docs.client.DocsClient(source='duplicity $version') self.client.ssl = True self.client.http_client.debug = False self._authorize(parsed_url.username + '@' + parsed_url.hostname, self.get_password())
identifier_body
gdocsbackend.py
# -*- Mode:Python; indent-tabs-mode:nil; tab-width:4 -*- # # Copyright 2011 Carlos Abalde <carlos.abalde@gmail.com> # # This file is part of duplicity. # # Duplicity is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the # Free Software Foundation; either version 2 of the License, or (at your # option) any later version. # # Duplicity is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with duplicity; if not, write to the Free Software Foundation, # Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA import os.path import string import urllib import duplicity.backend from duplicity.errors import BackendException class GDocsBackend(duplicity.backend.Backend): """Connect to remote store using Google Google Documents List API""" ROOT_FOLDER_ID = 'folder%3Aroot' BACKUP_DOCUMENT_TYPE = 'application/binary' def __init__(self, parsed_url): duplicity.backend.Backend.__init__(self, parsed_url) # Import Google Data APIs libraries. try: global atom global gdata import atom.data import gdata.client import gdata.docs.client import gdata.docs.data except ImportError as e: raise BackendException("""\ Google Docs backend requires Google Data APIs Python Client Library (see http://code.google.com/p/gdata-python-client/). Exception: %s""" % str(e)) # Setup client instance. self.client = gdata.docs.client.DocsClient(source='duplicity $version') self.client.ssl = True self.client.http_client.debug = False self._authorize(parsed_url.username + '@' + parsed_url.hostname, self.get_password()) # Fetch destination folder entry (and crete hierarchy if required). folder_names = string.split(parsed_url.path[1:], '/') parent_folder = None parent_folder_id = GDocsBackend.ROOT_FOLDER_ID for folder_name in folder_names: entries = self._fetch_entries(parent_folder_id, 'folder', folder_name) if entries is not None: if len(entries) == 1: parent_folder = entries[0] elif len(entries) == 0: folder = gdata.docs.data.Resource(type='folder', title=folder_name) parent_folder = self.client.create_resource(folder, collection=parent_folder) else: parent_folder = None if parent_folder: parent_folder_id = parent_folder.resource_id.text else: raise BackendException("Error while creating destination folder '%s'." % folder_name) else: raise BackendException("Error while fetching destination folder '%s'." % folder_name) self.folder = parent_folder def _put(self, source_path, remote_filename): self._delete(remote_filename) # Set uploader instance. Note that resumable uploads are required in order to # enable uploads for all file types. # (see http://googleappsdeveloper.blogspot.com/2011/05/upload-all-file-types-to-any-google.html) file = source_path.open() uploader = gdata.client.ResumableUploader( self.client, file, GDocsBackend.BACKUP_DOCUMENT_TYPE, os.path.getsize(file.name), chunk_size=gdata.client.ResumableUploader.DEFAULT_CHUNK_SIZE, desired_class=gdata.docs.data.Resource) if uploader: # Chunked upload. entry = gdata.docs.data.Resource(title=atom.data.Title(text=remote_filename)) uri = self.folder.get_resumable_create_media_link().href + '?convert=false' entry = uploader.UploadFile(uri, entry=entry) if not entry: raise BackendException("Failed to upload file '%s' to remote folder '%s'" % (source_path.get_filename(), self.folder.title.text)) else: raise BackendException("Failed to initialize upload of file '%s' to remote folder '%s'" % (source_path.get_filename(), self.folder.title.text)) assert not file.close() def
(self, remote_filename, local_path): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE, remote_filename) if len(entries) == 1: entry = entries[0] self.client.DownloadResource(entry, local_path.name) else: raise BackendException("Failed to find file '%s' in remote folder '%s'" % (remote_filename, self.folder.title.text)) def _list(self): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE) return [entry.title.text for entry in entries] def _delete(self, filename): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE, filename) for entry in entries: self.client.delete(entry.get_edit_link().href + '?delete=true', force=True) def _authorize(self, email, password, captcha_token=None, captcha_response=None): try: self.client.client_login(email, password, source='duplicity $version', service='writely', captcha_token=captcha_token, captcha_response=captcha_response) except gdata.client.CaptchaChallenge as challenge: print('A captcha challenge in required. Please visit ' + challenge.captcha_url) answer = None while not answer: answer = raw_input('Answer to the challenge? ') self._authorize(email, password, challenge.captcha_token, answer) except gdata.client.BadAuthentication: raise BackendException( 'Invalid user credentials given. Be aware that accounts ' 'that use 2-step verification require creating an application specific ' 'access code for using this Duplicity backend. Follow the instruction in ' 'http://www.google.com/support/accounts/bin/static.py?page=guide.cs&guide=1056283&topic=1056286 ' 'and create your application-specific password to run duplicity backups.') def _fetch_entries(self, folder_id, type, title=None): # Build URI. uri = '/feeds/default/private/full/%s/contents' % folder_id if type == 'folder': uri += '/-/folder?showfolders=true' elif type == GDocsBackend.BACKUP_DOCUMENT_TYPE: uri += '?showfolders=false' else: uri += '?showfolders=true' if title: uri += '&title=' + urllib.quote(title) + '&title-exact=true' # Fetch entries. entries = self.client.get_all_resources(uri=uri) # When filtering by entry title, API is returning (don't know why) documents in other # folders (apart from folder_id) matching the title, so some extra filtering is required. if title: result = [] for entry in entries: resource_type = entry.get_resource_type() if (not type) \ or (type == 'folder' and resource_type == 'folder') \ or (type == GDocsBackend.BACKUP_DOCUMENT_TYPE and resource_type != 'folder'): if folder_id != GDocsBackend.ROOT_FOLDER_ID: for link in entry.in_collections(): folder_entry = self.client.get_entry(link.href, None, None, desired_class=gdata.docs.data.Resource) if folder_entry and (folder_entry.resource_id.text == folder_id): result.append(entry) elif len(entry.in_collections()) == 0: result.append(entry) else: result = entries # Done! return result """ gdata is an alternate way to access gdocs, currently 05/2015 lacking OAuth support """ duplicity.backend.register_backend('gdata+gdocs', GDocsBackend) duplicity.backend.uses_netloc.extend(['gdata+gdocs'])
_get
identifier_name
gdocsbackend.py
# -*- Mode:Python; indent-tabs-mode:nil; tab-width:4 -*- # # Copyright 2011 Carlos Abalde <carlos.abalde@gmail.com> # # This file is part of duplicity. # # Duplicity is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the # Free Software Foundation; either version 2 of the License, or (at your # option) any later version. # # Duplicity is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with duplicity; if not, write to the Free Software Foundation, # Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA import os.path import string import urllib import duplicity.backend from duplicity.errors import BackendException class GDocsBackend(duplicity.backend.Backend): """Connect to remote store using Google Google Documents List API""" ROOT_FOLDER_ID = 'folder%3Aroot' BACKUP_DOCUMENT_TYPE = 'application/binary' def __init__(self, parsed_url): duplicity.backend.Backend.__init__(self, parsed_url) # Import Google Data APIs libraries. try: global atom global gdata import atom.data import gdata.client import gdata.docs.client import gdata.docs.data except ImportError as e: raise BackendException("""\ Google Docs backend requires Google Data APIs Python Client Library (see http://code.google.com/p/gdata-python-client/). Exception: %s""" % str(e)) # Setup client instance. self.client = gdata.docs.client.DocsClient(source='duplicity $version') self.client.ssl = True self.client.http_client.debug = False self._authorize(parsed_url.username + '@' + parsed_url.hostname, self.get_password()) # Fetch destination folder entry (and crete hierarchy if required). folder_names = string.split(parsed_url.path[1:], '/') parent_folder = None parent_folder_id = GDocsBackend.ROOT_FOLDER_ID for folder_name in folder_names: entries = self._fetch_entries(parent_folder_id, 'folder', folder_name) if entries is not None: if len(entries) == 1: parent_folder = entries[0] elif len(entries) == 0: folder = gdata.docs.data.Resource(type='folder', title=folder_name) parent_folder = self.client.create_resource(folder, collection=parent_folder) else: parent_folder = None if parent_folder: parent_folder_id = parent_folder.resource_id.text else: raise BackendException("Error while creating destination folder '%s'." % folder_name) else: raise BackendException("Error while fetching destination folder '%s'." % folder_name) self.folder = parent_folder def _put(self, source_path, remote_filename): self._delete(remote_filename) # Set uploader instance. Note that resumable uploads are required in order to # enable uploads for all file types. # (see http://googleappsdeveloper.blogspot.com/2011/05/upload-all-file-types-to-any-google.html) file = source_path.open() uploader = gdata.client.ResumableUploader( self.client, file, GDocsBackend.BACKUP_DOCUMENT_TYPE, os.path.getsize(file.name), chunk_size=gdata.client.ResumableUploader.DEFAULT_CHUNK_SIZE, desired_class=gdata.docs.data.Resource) if uploader: # Chunked upload. entry = gdata.docs.data.Resource(title=atom.data.Title(text=remote_filename)) uri = self.folder.get_resumable_create_media_link().href + '?convert=false' entry = uploader.UploadFile(uri, entry=entry) if not entry: raise BackendException("Failed to upload file '%s' to remote folder '%s'" % (source_path.get_filename(), self.folder.title.text)) else: raise BackendException("Failed to initialize upload of file '%s' to remote folder '%s'" % (source_path.get_filename(), self.folder.title.text)) assert not file.close() def _get(self, remote_filename, local_path): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE, remote_filename) if len(entries) == 1: entry = entries[0] self.client.DownloadResource(entry, local_path.name) else: raise BackendException("Failed to find file '%s' in remote folder '%s'" % (remote_filename, self.folder.title.text)) def _list(self): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE) return [entry.title.text for entry in entries] def _delete(self, filename): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE, filename) for entry in entries: self.client.delete(entry.get_edit_link().href + '?delete=true', force=True) def _authorize(self, email, password, captcha_token=None, captcha_response=None): try: self.client.client_login(email, password, source='duplicity $version', service='writely', captcha_token=captcha_token, captcha_response=captcha_response) except gdata.client.CaptchaChallenge as challenge: print('A captcha challenge in required. Please visit ' + challenge.captcha_url) answer = None while not answer: answer = raw_input('Answer to the challenge? ') self._authorize(email, password, challenge.captcha_token, answer) except gdata.client.BadAuthentication: raise BackendException( 'Invalid user credentials given. Be aware that accounts ' 'that use 2-step verification require creating an application specific ' 'access code for using this Duplicity backend. Follow the instruction in ' 'http://www.google.com/support/accounts/bin/static.py?page=guide.cs&guide=1056283&topic=1056286 ' 'and create your application-specific password to run duplicity backups.') def _fetch_entries(self, folder_id, type, title=None): # Build URI. uri = '/feeds/default/private/full/%s/contents' % folder_id if type == 'folder': uri += '/-/folder?showfolders=true' elif type == GDocsBackend.BACKUP_DOCUMENT_TYPE: uri += '?showfolders=false' else: uri += '?showfolders=true' if title: uri += '&title=' + urllib.quote(title) + '&title-exact=true' # Fetch entries. entries = self.client.get_all_resources(uri=uri) # When filtering by entry title, API is returning (don't know why) documents in other # folders (apart from folder_id) matching the title, so some extra filtering is required. if title: result = [] for entry in entries: resource_type = entry.get_resource_type() if (not type) \ or (type == 'folder' and resource_type == 'folder') \ or (type == GDocsBackend.BACKUP_DOCUMENT_TYPE and resource_type != 'folder'): if folder_id != GDocsBackend.ROOT_FOLDER_ID:
elif len(entry.in_collections()) == 0: result.append(entry) else: result = entries # Done! return result """ gdata is an alternate way to access gdocs, currently 05/2015 lacking OAuth support """ duplicity.backend.register_backend('gdata+gdocs', GDocsBackend) duplicity.backend.uses_netloc.extend(['gdata+gdocs'])
for link in entry.in_collections(): folder_entry = self.client.get_entry(link.href, None, None, desired_class=gdata.docs.data.Resource) if folder_entry and (folder_entry.resource_id.text == folder_id): result.append(entry)
conditional_block
gdocsbackend.py
# -*- Mode:Python; indent-tabs-mode:nil; tab-width:4 -*- # # Copyright 2011 Carlos Abalde <carlos.abalde@gmail.com> # # This file is part of duplicity. # # Duplicity is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the # Free Software Foundation; either version 2 of the License, or (at your # option) any later version. # # Duplicity is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with duplicity; if not, write to the Free Software Foundation, # Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA import os.path import string import urllib import duplicity.backend from duplicity.errors import BackendException class GDocsBackend(duplicity.backend.Backend): """Connect to remote store using Google Google Documents List API""" ROOT_FOLDER_ID = 'folder%3Aroot' BACKUP_DOCUMENT_TYPE = 'application/binary' def __init__(self, parsed_url): duplicity.backend.Backend.__init__(self, parsed_url) # Import Google Data APIs libraries. try: global atom global gdata import atom.data import gdata.client import gdata.docs.client import gdata.docs.data except ImportError as e: raise BackendException("""\ Google Docs backend requires Google Data APIs Python Client Library (see http://code.google.com/p/gdata-python-client/). Exception: %s""" % str(e)) # Setup client instance. self.client = gdata.docs.client.DocsClient(source='duplicity $version') self.client.ssl = True self.client.http_client.debug = False self._authorize(parsed_url.username + '@' + parsed_url.hostname, self.get_password()) # Fetch destination folder entry (and crete hierarchy if required). folder_names = string.split(parsed_url.path[1:], '/') parent_folder = None parent_folder_id = GDocsBackend.ROOT_FOLDER_ID for folder_name in folder_names: entries = self._fetch_entries(parent_folder_id, 'folder', folder_name) if entries is not None: if len(entries) == 1: parent_folder = entries[0] elif len(entries) == 0:
folder = gdata.docs.data.Resource(type='folder', title=folder_name) parent_folder = self.client.create_resource(folder, collection=parent_folder) else: parent_folder = None if parent_folder: parent_folder_id = parent_folder.resource_id.text else: raise BackendException("Error while creating destination folder '%s'." % folder_name) else: raise BackendException("Error while fetching destination folder '%s'." % folder_name) self.folder = parent_folder def _put(self, source_path, remote_filename): self._delete(remote_filename) # Set uploader instance. Note that resumable uploads are required in order to # enable uploads for all file types. # (see http://googleappsdeveloper.blogspot.com/2011/05/upload-all-file-types-to-any-google.html) file = source_path.open() uploader = gdata.client.ResumableUploader( self.client, file, GDocsBackend.BACKUP_DOCUMENT_TYPE, os.path.getsize(file.name), chunk_size=gdata.client.ResumableUploader.DEFAULT_CHUNK_SIZE, desired_class=gdata.docs.data.Resource) if uploader: # Chunked upload. entry = gdata.docs.data.Resource(title=atom.data.Title(text=remote_filename)) uri = self.folder.get_resumable_create_media_link().href + '?convert=false' entry = uploader.UploadFile(uri, entry=entry) if not entry: raise BackendException("Failed to upload file '%s' to remote folder '%s'" % (source_path.get_filename(), self.folder.title.text)) else: raise BackendException("Failed to initialize upload of file '%s' to remote folder '%s'" % (source_path.get_filename(), self.folder.title.text)) assert not file.close() def _get(self, remote_filename, local_path): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE, remote_filename) if len(entries) == 1: entry = entries[0] self.client.DownloadResource(entry, local_path.name) else: raise BackendException("Failed to find file '%s' in remote folder '%s'" % (remote_filename, self.folder.title.text)) def _list(self): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE) return [entry.title.text for entry in entries] def _delete(self, filename): entries = self._fetch_entries(self.folder.resource_id.text, GDocsBackend.BACKUP_DOCUMENT_TYPE, filename) for entry in entries: self.client.delete(entry.get_edit_link().href + '?delete=true', force=True) def _authorize(self, email, password, captcha_token=None, captcha_response=None): try: self.client.client_login(email, password, source='duplicity $version', service='writely', captcha_token=captcha_token, captcha_response=captcha_response) except gdata.client.CaptchaChallenge as challenge: print('A captcha challenge in required. Please visit ' + challenge.captcha_url) answer = None while not answer: answer = raw_input('Answer to the challenge? ') self._authorize(email, password, challenge.captcha_token, answer) except gdata.client.BadAuthentication: raise BackendException( 'Invalid user credentials given. Be aware that accounts ' 'that use 2-step verification require creating an application specific ' 'access code for using this Duplicity backend. Follow the instruction in ' 'http://www.google.com/support/accounts/bin/static.py?page=guide.cs&guide=1056283&topic=1056286 ' 'and create your application-specific password to run duplicity backups.') def _fetch_entries(self, folder_id, type, title=None): # Build URI. uri = '/feeds/default/private/full/%s/contents' % folder_id if type == 'folder': uri += '/-/folder?showfolders=true' elif type == GDocsBackend.BACKUP_DOCUMENT_TYPE: uri += '?showfolders=false' else: uri += '?showfolders=true' if title: uri += '&title=' + urllib.quote(title) + '&title-exact=true' # Fetch entries. entries = self.client.get_all_resources(uri=uri) # When filtering by entry title, API is returning (don't know why) documents in other # folders (apart from folder_id) matching the title, so some extra filtering is required. if title: result = [] for entry in entries: resource_type = entry.get_resource_type() if (not type) \ or (type == 'folder' and resource_type == 'folder') \ or (type == GDocsBackend.BACKUP_DOCUMENT_TYPE and resource_type != 'folder'): if folder_id != GDocsBackend.ROOT_FOLDER_ID: for link in entry.in_collections(): folder_entry = self.client.get_entry(link.href, None, None, desired_class=gdata.docs.data.Resource) if folder_entry and (folder_entry.resource_id.text == folder_id): result.append(entry) elif len(entry.in_collections()) == 0: result.append(entry) else: result = entries # Done! return result """ gdata is an alternate way to access gdocs, currently 05/2015 lacking OAuth support """ duplicity.backend.register_backend('gdata+gdocs', GDocsBackend) duplicity.backend.uses_netloc.extend(['gdata+gdocs'])
random_line_split
color.js
[0-9a-f]{3}){1,2}$/i; /** * @see goog.color.rgbColorRe_ * @const * @type {RegExp} * @private */ ol.color.rgbColorRe_ = /^(?:rgb)?\((0|[1-9]\d{0,2}),\s?(0|[1-9]\d{0,2}),\s?(0|[1-9]\d{0,2})\)$/i; /** * @see goog.color.alpha.rgbaColorRe_ * @const * @type {RegExp} * @private */ ol.color.rgbaColorRe_ = /^(?:rgba)?\((0|[1-9]\d{0,2}),\s?(0|[1-9]\d{0,2}),\s?(0|[1-9]\d{0,2}),\s?(0|1|0\.\d{0,10})\)$/i; /** * @param {ol.Color} dst Destination. * @param {ol.Color} src Source. * @param {ol.Color=} opt_color Color. * @return {ol.Color} Color. */ ol.color.blend = function(dst, src, opt_color) { // http://en.wikipedia.org/wiki/Alpha_compositing // FIXME do we need to scale by 255? var out = opt_color ? opt_color : []; var dstA = dst[3]; var srcA = src[3]; if (dstA == 1) { out[0] = (src[0] * srcA + dst[0] * (1 - srcA) + 0.5) | 0; out[1] = (src[1] * srcA + dst[1] * (1 - srcA) + 0.5) | 0; out[2] = (src[2] * srcA + dst[2] * (1 - srcA) + 0.5) | 0; out[3] = 1; } else if (srcA === 0) { out[0] = dst[0]; out[1] = dst[1]; out[2] = dst[2]; out[3] = dstA; } else { var outA = srcA + dstA * (1 - srcA); if (outA === 0) { out[0] = 0; out[1] = 0; out[2] = 0; out[3] = 0; } else { out[0] = ((src[0] * srcA + dst[0] * dstA * (1 - srcA)) / outA + 0.5) | 0; out[1] = ((src[1] * srcA + dst[1] * dstA * (1 - srcA)) / outA + 0.5) | 0; out[2] = ((src[2] * srcA + dst[2] * dstA * (1 - srcA)) / outA + 0.5) | 0; out[3] = outA; } } goog.asserts.assert(ol.color.isValid(out), 'Output color of blend should be a valid color'); return out; }; /** * Return the color as an array. This function maintains a cache of calculated * arrays which means the result should not be modified. * @param {ol.Color|string} color Color. * @return {ol.Color} Color. * @api */ ol.color.asArray = function(color) { if (goog.isArray(color)) { return color; } else { goog.asserts.assert(goog.isString(color), 'Color should be a string'); return ol.color.fromString(color); } }; /** * Return the color as an rgba string. * @param {ol.Color|string} color Color. * @return {string} Rgba string. * @api */ ol.color.asString = function(color) { if (goog.isString(color)) { return color; } else { goog.asserts.assert(goog.isArray(color), 'Color should be an array'); return ol.color.toString(color); } }; /** * @param {ol.Color} color1 Color1. * @param {ol.Color} color2 Color2. * @return {boolean} Equals. */ ol.color.equals = function(color1, color2) { return color1 === color2 || ( color1[0] == color2[0] && color1[1] == color2[1] && color1[2] == color2[2] && color1[3] == color2[3]); }; /** * @param {string} s String. * @return {ol.Color} Color. */ ol.color.fromString = ( /** * @return {function(string): ol.Color} */ function() { // We maintain a small cache of parsed strings. To provide cheap LRU-like // semantics, whenever the cache grows too large we simply delete an // arbitrary 25% of the entries. /** * @const * @type {number} */ var MAX_CACHE_SIZE = 1024; /** * @type {Object.<string, ol.Color>} */ var cache = {}; /** * @type {number} */ var cacheSize = 0; return ( /** * @param {string} s String. * @return {ol.Color} Color. */ function(s) { var color; if (cache.hasOwnProperty(s)) { color = cache[s]; } else { if (cacheSize >= MAX_CACHE_SIZE) { var i = 0; var key; for (key in cache) { if ((i++ & 3) === 0) { delete cache[key]; --cacheSize; } } } color = ol.color.fromStringInternal_(s); cache[s] = color; ++cacheSize; } return color; }); })(); /** * @param {string} s String. * @private * @return {ol.Color} Color. */ ol.color.fromStringInternal_ = function(s) { var isHex = false; if (ol.ENABLE_NAMED_COLORS && goog.color.names.hasOwnProperty(s)) { // goog.color.names does not have a type declaration, so add a typecast s = /** @type {string} */ (goog.color.names[s]); isHex = true; } var r, g, b, a, color, match; if (isHex || (match = ol.color.hexColorRe_.exec(s)))
else if ((match = ol.color.rgbaColorRe_.exec(s))) { // rgba() r = Number(match[1]); g = Number(match[2]); b = Number(match[3]); a = Number(match[4]); color = [r, g, b, a]; return ol.color.normalize(color, color); } else if ((match = ol.color.rgbColorRe_.exec(s))) { // rgb() r = Number(match[1]); g = Number(match[2]); b = Number(match[3]); color = [r, g, b, 1]; return ol.color.normalize(color, color); } else { goog.asserts.fail(s + ' is not a valid color'); } }; /** * @param {ol.Color} color Color. * @return {boolean} Is valid. */ ol.color.isValid = function(color) { return 0 <= color[0] && color[0] < 256 && 0 <= color[1] && color[1] < 256 && 0 <= color[2] && color[2] < 256 && 0 <= color[3] && color[3] <= 1; }; /** * @param {ol.Color} color Color. * @param {ol.Color=} opt_color Color. * @return {ol.Color} Clamped color. */ ol.color.normalize = function(color, opt_color) { var result = opt_color || []; result[0] = ol.math.clamp((color[0] + 0.5) | 0, 0, 255); result[1]
{ // hex var n = s.length - 1; // number of hex digits goog.asserts.assert(n == 3 || n == 6, 'Color string length should be 3 or 6'); var d = n == 3 ? 1 : 2; // number of digits per channel r = parseInt(s.substr(1 + 0 * d, d), 16); g = parseInt(s.substr(1 + 1 * d, d), 16); b = parseInt(s.substr(1 + 2 * d, d), 16); if (d == 1) { r = (r << 4) + r; g = (g << 4) + g; b = (b << 4) + b; } a = 1; color = [r, g, b, a]; goog.asserts.assert(ol.color.isValid(color), 'Color should be a valid color'); return color; }
conditional_block
color.js
[0-9a-f]{3}){1,2}$/i; /** * @see goog.color.rgbColorRe_ * @const * @type {RegExp} * @private */ ol.color.rgbColorRe_ = /^(?:rgb)?\((0|[1-9]\d{0,2}),\s?(0|[1-9]\d{0,2}),\s?(0|[1-9]\d{0,2})\)$/i; /** * @see goog.color.alpha.rgbaColorRe_ * @const * @type {RegExp} * @private */ ol.color.rgbaColorRe_ = /^(?:rgba)?\((0|[1-9]\d{0,2}),\s?(0|[1-9]\d{0,2}),\s?(0|[1-9]\d{0,2}),\s?(0|1|0\.\d{0,10})\)$/i; /** * @param {ol.Color} dst Destination. * @param {ol.Color} src Source. * @param {ol.Color=} opt_color Color. * @return {ol.Color} Color. */ ol.color.blend = function(dst, src, opt_color) { // http://en.wikipedia.org/wiki/Alpha_compositing // FIXME do we need to scale by 255? var out = opt_color ? opt_color : []; var dstA = dst[3]; var srcA = src[3]; if (dstA == 1) { out[0] = (src[0] * srcA + dst[0] * (1 - srcA) + 0.5) | 0; out[1] = (src[1] * srcA + dst[1] * (1 - srcA) + 0.5) | 0; out[2] = (src[2] * srcA + dst[2] * (1 - srcA) + 0.5) | 0; out[3] = 1; } else if (srcA === 0) { out[0] = dst[0]; out[1] = dst[1]; out[2] = dst[2]; out[3] = dstA; } else { var outA = srcA + dstA * (1 - srcA); if (outA === 0) { out[0] = 0; out[1] = 0; out[2] = 0; out[3] = 0; } else { out[0] = ((src[0] * srcA + dst[0] * dstA * (1 - srcA)) / outA + 0.5) | 0; out[1] = ((src[1] * srcA + dst[1] * dstA * (1 - srcA)) / outA + 0.5) | 0; out[2] = ((src[2] * srcA + dst[2] * dstA * (1 - srcA)) / outA + 0.5) | 0; out[3] = outA; } } goog.asserts.assert(ol.color.isValid(out), 'Output color of blend should be a valid color'); return out; }; /** * Return the color as an array. This function maintains a cache of calculated * arrays which means the result should not be modified. * @param {ol.Color|string} color Color. * @return {ol.Color} Color. * @api */ ol.color.asArray = function(color) { if (goog.isArray(color)) { return color; } else { goog.asserts.assert(goog.isString(color), 'Color should be a string'); return ol.color.fromString(color); } }; /** * Return the color as an rgba string. * @param {ol.Color|string} color Color. * @return {string} Rgba string. * @api */ ol.color.asString = function(color) { if (goog.isString(color)) { return color; } else { goog.asserts.assert(goog.isArray(color), 'Color should be an array'); return ol.color.toString(color); } }; /** * @param {ol.Color} color1 Color1. * @param {ol.Color} color2 Color2. * @return {boolean} Equals. */ ol.color.equals = function(color1, color2) { return color1 === color2 || ( color1[0] == color2[0] && color1[1] == color2[1] && color1[2] == color2[2] && color1[3] == color2[3]); }; /** * @param {string} s String. * @return {ol.Color} Color. */ ol.color.fromString = ( /** * @return {function(string): ol.Color} */ function() { // We maintain a small cache of parsed strings. To provide cheap LRU-like // semantics, whenever the cache grows too large we simply delete an // arbitrary 25% of the entries. /** * @const * @type {number} */ var MAX_CACHE_SIZE = 1024; /** * @type {Object.<string, ol.Color>} */ var cache = {}; /** * @type {number} */ var cacheSize = 0; return ( /** * @param {string} s String. * @return {ol.Color} Color. */ function(s) { var color; if (cache.hasOwnProperty(s)) { color = cache[s]; } else { if (cacheSize >= MAX_CACHE_SIZE) { var i = 0; var key; for (key in cache) { if ((i++ & 3) === 0) { delete cache[key]; --cacheSize; } }
cache[s] = color; ++cacheSize; } return color; }); })(); /** * @param {string} s String. * @private * @return {ol.Color} Color. */ ol.color.fromStringInternal_ = function(s) { var isHex = false; if (ol.ENABLE_NAMED_COLORS && goog.color.names.hasOwnProperty(s)) { // goog.color.names does not have a type declaration, so add a typecast s = /** @type {string} */ (goog.color.names[s]); isHex = true; } var r, g, b, a, color, match; if (isHex || (match = ol.color.hexColorRe_.exec(s))) { // hex var n = s.length - 1; // number of hex digits goog.asserts.assert(n == 3 || n == 6, 'Color string length should be 3 or 6'); var d = n == 3 ? 1 : 2; // number of digits per channel r = parseInt(s.substr(1 + 0 * d, d), 16); g = parseInt(s.substr(1 + 1 * d, d), 16); b = parseInt(s.substr(1 + 2 * d, d), 16); if (d == 1) { r = (r << 4) + r; g = (g << 4) + g; b = (b << 4) + b; } a = 1; color = [r, g, b, a]; goog.asserts.assert(ol.color.isValid(color), 'Color should be a valid color'); return color; } else if ((match = ol.color.rgbaColorRe_.exec(s))) { // rgba() r = Number(match[1]); g = Number(match[2]); b = Number(match[3]); a = Number(match[4]); color = [r, g, b, a]; return ol.color.normalize(color, color); } else if ((match = ol.color.rgbColorRe_.exec(s))) { // rgb() r = Number(match[1]); g = Number(match[2]); b = Number(match[3]); color = [r, g, b, 1]; return ol.color.normalize(color, color); } else { goog.asserts.fail(s + ' is not a valid color'); } }; /** * @param {ol.Color} color Color. * @return {boolean} Is valid. */ ol.color.isValid = function(color) { return 0 <= color[0] && color[0] < 256 && 0 <= color[1] && color[1] < 256 && 0 <= color[2] && color[2] < 256 && 0 <= color[3] && color[3] <= 1; }; /** * @param {ol.Color} color Color. * @param {ol.Color=} opt_color Color. * @return {ol.Color} Clamped color. */ ol.color.normalize = function(color, opt_color) { var result = opt_color || []; result[0] = ol.math.clamp((color[0] + 0.5) | 0, 0, 255); result[1] =
} color = ol.color.fromStringInternal_(s);
random_line_split
server.js
} }, function(next) { // TWITTER fs.readFile("twitter-credentials.json", function(error, data) { if (error) { console.error("twitter-credentials.json is missing or couldn't be read. I need this."); process.exit(); return; } var twitterDetails = JSON.parse(data); twitterClient = new twitter(twitterDetails); next(); }); }, function(next) { // // FRONTEND (UI) ROUTES // server.app.get("/", function(req, res) { res.render("index"); }); server.app.get("/twitter/", function(req, res) { var twitter = "jadaradix"; res.render( "twitter", { "twitter": twitter } ); }); server.app.get("/twitter/*/", function(req, res) { var twitter = req.params[0]; res.render( "twitter", { "twitter": twitter } ); }); server.app.get("/process/", function(req, res) { if (!req.query.hasOwnProperty("name") || !req.query.hasOwnProperty("topics")) { res.redirect(302, "/"); return; } var name = _s.humanize(req.query.name) + " Museum"; var topics = req.query.topics.split(","); var museums = server.db("museums"); //insert function doesIdExist(id) { var r = museums.find({ id: id }); return (r.value() ? true : false); } function generateId() { var id = _.times(16, function(n) { return _.random(0, 10); }).join(""); return id; } var id = generateId(); while (doesIdExist(id)) { var id = generateId(); } var museum = { "id": id, "name": name, "isSpontaneous": true, "topics": topics } museums.push(museum); server.db.save(); res.redirect(302, "/museum/" + id); }); server.app.get("/process/*/", function(req, res) { var id = req.params[0]; var museums = server.db("museums"); var r = museums.find({ id: id }); var rValue = r.value(); if (!rValue) { res.redirect(302, "/twitter/" + id); return; } res.render( "process", { "twitter": id } ); }); server.app.get("/museum/*/", function(req, res) { var id = req.params[0]; var museums = server.db("museums"); var r = museums.find({ id: id }); var rValue = r.value(); if (!rValue || (!(("topics" in rValue)) || rValue.topics.length == 0)) { res.redirect(302, "/twitter/" + id); return; } res.render( "museum", { "title": rValue.name, "id": id } ); }); server.app.get("/favicon.ico", function(req, res) { res.redirect(301, "/static/favicon.ico"); }); // // BACKEND (API) ROUTES // server.app.get("/api/reset", function(req, res) { async.waterfall([ function(next) { server.resetDatabase(next); }, function(next) { server.jsonSuccess("The database was reset.", res); } ]); }); server.app.get("/api/twitter/*", function(req, res) { var screenName = req.params[0]; if (!screenName.length) { server.jsonError("The Twitter screen name was empty.", res); return; } async.waterfall([ function(next) { // Account twitterClient.get( 'users/show', { "screen_name": screenName, }, function(error, data, raw) { if (error) { var errorText; switch(error.statusCode) { case 404: errorText = "That Twitter account doesn't exist."; break; default: errorText = "The twitter account couldn't be accessed; that's all."; break; } server.jsonError(errorText, res); return; } var newData = { "id": screenName, "name": data.name + "'" + (!_s.endsWith(data.name.toLowerCase(), "s") ? "s" : "") + " Museum", "isSpontaneous": false, "twitter": { "account": { "screenName": data.screen_name, "name": data.name, "location": data.location, "description": data.description, "language": data.lang, "picture": data.profile_image_url } } }; next(null, newData); } ); }, function(passedData, next) { twitterClient.get( 'statuses/user_timeline', { "screen_name": screenName, "trim_user": true, "count": 200, "include_rts": true }, function(error, data, raw) { if (error) { var errorText; switch(error.statusCode) { case 401: errorText = "That Twitter account is probably using &lsquo;protected tweets&rsquo;."; break; default: errorText = "The tweets couldn't be retrieved; that's all."; break; } server.jsonError(errorText, res); return; } var tweets = _.map(data, twinglish.cleanTweet); // var tweets = data; //some tweets may have been removed (annulled) tweets = _.filter(tweets, function(tweet) { return (!(tweet == null || tweet == undefined)); //sorry }); passedData["twitter"]["tweets"] = tweets; var tweetTexts = _.map(tweets, function(tweet) { return tweet.text; }); var textStuff = tweetTexts.join(" "); // from http://codereview.stackexchange.com/questions/63947/count-frequency-of-words-in-a-string-returning-list-of-unique-words-sorted-by-fr function getFrequency2(string, cutOff) { var cleanString = string.replace(/[\.,-\/#!$%\^&\*;:{}=\-_`~()]/g,""), words = cleanString.split(' '), frequencies = {}, word, frequency, i; for( i=0; i<words.length; i++ ) { word = words[i]; frequencies[word] = frequencies[word] || 0; frequencies[word]++; } words = Object.keys( frequencies ); return words.sort(function (a,b) { return frequencies[b] -frequencies[a];}).slice(0,cutOff); } var stuff = getFrequency2(textStuff, 100); var stuff2 = stuff.slice(95, 100); console.log(stuff2); passedData["topics"] = stuff2; next(null, passedData); } ); }, function(passedData, next) { if ("metrics" in req.query) { passedData["twitter"]["metrics"] = _.map(twetrics.metrics, function(metric) { return metric.method(passedData["twitter"]["tweets"]); }); } next(null, passedData); }, function(passedData, next) { var format = (("format" in req.query) ? req.query.format : "json"); var data; switch(format) { case "store": var museums = server.db("museums"); var r = museums.remove({ id: screenName }); // if (r["__wrapped__"].length == 0) { // //did not exist in the first place // return; // } museums.push(passedData); server.db.save(); server.jsonSuccess("The Twitter data was stored.", res); return; break; case "json": data = passedData; break; case "prettyjson": data = JSON.stringify(passedData, undefined, 2); break; case "csv": //return only tweets data = csv.fromObjects( passedData["twitter"]["tweets"], ["when", "text", "isRetweet"], true ); break; default: data = []; break; } res.send(data); } ]); }); server.app.get("/api/process/*", function(req, res) { var museums = server.db("museums"); var r = museums.find({ id: req.params[0] }).value(); if (!r) { server.jsonError("There's no data for this screen name. Stop hacking.", res); return; } res.send(r); }); server.app.get("/api/museum/*", function(req, res) { var museums = server.db("museums"); var r = museums.find({ id: req.params[0] }).value(); if (!r)
{ server.jsonError("There's no data for this screen name. Stop hacking.", res); return; }
conditional_block
server.js
": __dirname, "viewPath": "pages", "logging": true, "loggingFormat": ":method :url -> HTTP :status; :response-time ms", "staticDirectories": ["css", "fonts", "js", "static"] }); server.dbInit = function() { var _self = this; _self.db = low( "db.json", { "autosave": false } ); } server.resetDatabase = function(callback) { var _self = this; var blankData = fs.readFileSync("db-blank.json"); fs.writeFileSync("db.json", blankData); _self.dbInit(); callback(); }; server.jsonError = function(text, res) { res.send({ "error": text }); }; server.jsonSuccess = function(text, res) { res.send({ "ok": text }); }; next(); }, function(next) { // DATABASE if (!fs.existsSync("db.json")) { server.resetDatabase(next); } else { server.dbInit(); next(); } }, function(next) { // TWITTER fs.readFile("twitter-credentials.json", function(error, data) { if (error) { console.error("twitter-credentials.json is missing or couldn't be read. I need this."); process.exit(); return; } var twitterDetails = JSON.parse(data); twitterClient = new twitter(twitterDetails); next(); }); }, function(next) { // // FRONTEND (UI) ROUTES // server.app.get("/", function(req, res) { res.render("index"); }); server.app.get("/twitter/", function(req, res) { var twitter = "jadaradix"; res.render( "twitter", { "twitter": twitter } ); }); server.app.get("/twitter/*/", function(req, res) { var twitter = req.params[0]; res.render( "twitter", { "twitter": twitter } ); }); server.app.get("/process/", function(req, res) { if (!req.query.hasOwnProperty("name") || !req.query.hasOwnProperty("topics")) { res.redirect(302, "/"); return; } var name = _s.humanize(req.query.name) + " Museum"; var topics = req.query.topics.split(","); var museums = server.db("museums"); //insert function doesIdExist(id) { var r = museums.find({ id: id }); return (r.value() ? true : false); } function generateId() { var id = _.times(16, function(n) { return _.random(0, 10); }).join(""); return id; } var id = generateId(); while (doesIdExist(id)) { var id = generateId(); } var museum = { "id": id, "name": name, "isSpontaneous": true, "topics": topics } museums.push(museum); server.db.save();
var museums = server.db("museums"); var r = museums.find({ id: id }); var rValue = r.value(); if (!rValue) { res.redirect(302, "/twitter/" + id); return; } res.render( "process", { "twitter": id } ); }); server.app.get("/museum/*/", function(req, res) { var id = req.params[0]; var museums = server.db("museums"); var r = museums.find({ id: id }); var rValue = r.value(); if (!rValue || (!(("topics" in rValue)) || rValue.topics.length == 0)) { res.redirect(302, "/twitter/" + id); return; } res.render( "museum", { "title": rValue.name, "id": id } ); }); server.app.get("/favicon.ico", function(req, res) { res.redirect(301, "/static/favicon.ico"); }); // // BACKEND (API) ROUTES // server.app.get("/api/reset", function(req, res) { async.waterfall([ function(next) { server.resetDatabase(next); }, function(next) { server.jsonSuccess("The database was reset.", res); } ]); }); server.app.get("/api/twitter/*", function(req, res) { var screenName = req.params[0]; if (!screenName.length) { server.jsonError("The Twitter screen name was empty.", res); return; } async.waterfall([ function(next) { // Account twitterClient.get( 'users/show', { "screen_name": screenName, }, function(error, data, raw) { if (error) { var errorText; switch(error.statusCode) { case 404: errorText = "That Twitter account doesn't exist."; break; default: errorText = "The twitter account couldn't be accessed; that's all."; break; } server.jsonError(errorText, res); return; } var newData = { "id": screenName, "name": data.name + "'" + (!_s.endsWith(data.name.toLowerCase(), "s") ? "s" : "") + " Museum", "isSpontaneous": false, "twitter": { "account": { "screenName": data.screen_name, "name": data.name, "location": data.location, "description": data.description, "language": data.lang, "picture": data.profile_image_url } } }; next(null, newData); } ); }, function(passedData, next) { twitterClient.get( 'statuses/user_timeline', { "screen_name": screenName, "trim_user": true, "count": 200, "include_rts": true }, function(error, data, raw) { if (error) { var errorText; switch(error.statusCode) { case 401: errorText = "That Twitter account is probably using &lsquo;protected tweets&rsquo;."; break; default: errorText = "The tweets couldn't be retrieved; that's all."; break; } server.jsonError(errorText, res); return; } var tweets = _.map(data, twinglish.cleanTweet); // var tweets = data; //some tweets may have been removed (annulled) tweets = _.filter(tweets, function(tweet) { return (!(tweet == null || tweet == undefined)); //sorry }); passedData["twitter"]["tweets"] = tweets; var tweetTexts = _.map(tweets, function(tweet) { return tweet.text; }); var textStuff = tweetTexts.join(" "); // from http://codereview.stackexchange.com/questions/63947/count-frequency-of-words-in-a-string-returning-list-of-unique-words-sorted-by-fr function getFrequency2(string, cutOff) { var cleanString = string.replace(/[\.,-\/#!$%\^&\*;:{}=\-_`~()]/g,""), words = cleanString.split(' '), frequencies = {}, word, frequency, i; for( i=0; i<words.length; i++ ) { word = words[i]; frequencies[word] = frequencies[word] || 0; frequencies[word]++; } words = Object.keys( frequencies ); return words.sort(function (a,b) { return frequencies[b] -frequencies[a];}).slice(0,cutOff); } var stuff = getFrequency2(textStuff, 100); var stuff2 = stuff.slice(95, 100); console.log(stuff2); passedData["topics"] = stuff2; next(null, passedData); } ); }, function(passedData, next) { if ("metrics" in req.query) { passedData["twitter"]["metrics"] = _.map(twetrics.metrics, function(metric) { return metric.method(passedData["twitter"]["tweets"]); }); } next(null, passedData); }, function(passedData, next) { var format = (("format" in req.query) ? req.query.format : "json"); var data; switch(format) { case "store": var museums = server.db("museums"); var r = museums.remove({ id: screenName }); // if (r["__wrapped__"].length == 0) { // //did not exist in the first place // return; // } museums.push(passedData); server.db.save(); server.jsonSuccess("The Twitter data was stored.", res); return; break; case "json": data = passedData; break; case "prettyjson": data = JSON.stringify
res.redirect(302, "/museum/" + id); }); server.app.get("/process/*/", function(req, res) { var id = req.params[0];
random_line_split
server.js
": __dirname, "viewPath": "pages", "logging": true, "loggingFormat": ":method :url -> HTTP :status; :response-time ms", "staticDirectories": ["css", "fonts", "js", "static"] }); server.dbInit = function() { var _self = this; _self.db = low( "db.json", { "autosave": false } ); } server.resetDatabase = function(callback) { var _self = this; var blankData = fs.readFileSync("db-blank.json"); fs.writeFileSync("db.json", blankData); _self.dbInit(); callback(); }; server.jsonError = function(text, res) { res.send({ "error": text }); }; server.jsonSuccess = function(text, res) { res.send({ "ok": text }); }; next(); }, function(next) { // DATABASE if (!fs.existsSync("db.json")) { server.resetDatabase(next); } else { server.dbInit(); next(); } }, function(next) { // TWITTER fs.readFile("twitter-credentials.json", function(error, data) { if (error) { console.error("twitter-credentials.json is missing or couldn't be read. I need this."); process.exit(); return; } var twitterDetails = JSON.parse(data); twitterClient = new twitter(twitterDetails); next(); }); }, function(next) { // // FRONTEND (UI) ROUTES // server.app.get("/", function(req, res) { res.render("index"); }); server.app.get("/twitter/", function(req, res) { var twitter = "jadaradix"; res.render( "twitter", { "twitter": twitter } ); }); server.app.get("/twitter/*/", function(req, res) { var twitter = req.params[0]; res.render( "twitter", { "twitter": twitter } ); }); server.app.get("/process/", function(req, res) { if (!req.query.hasOwnProperty("name") || !req.query.hasOwnProperty("topics")) { res.redirect(302, "/"); return; } var name = _s.humanize(req.query.name) + " Museum"; var topics = req.query.topics.split(","); var museums = server.db("museums"); //insert function doesIdExist(id)
function generateId() { var id = _.times(16, function(n) { return _.random(0, 10); }).join(""); return id; } var id = generateId(); while (doesIdExist(id)) { var id = generateId(); } var museum = { "id": id, "name": name, "isSpontaneous": true, "topics": topics } museums.push(museum); server.db.save(); res.redirect(302, "/museum/" + id); }); server.app.get("/process/*/", function(req, res) { var id = req.params[0]; var museums = server.db("museums"); var r = museums.find({ id: id }); var rValue = r.value(); if (!rValue) { res.redirect(302, "/twitter/" + id); return; } res.render( "process", { "twitter": id } ); }); server.app.get("/museum/*/", function(req, res) { var id = req.params[0]; var museums = server.db("museums"); var r = museums.find({ id: id }); var rValue = r.value(); if (!rValue || (!(("topics" in rValue)) || rValue.topics.length == 0)) { res.redirect(302, "/twitter/" + id); return; } res.render( "museum", { "title": rValue.name, "id": id } ); }); server.app.get("/favicon.ico", function(req, res) { res.redirect(301, "/static/favicon.ico"); }); // // BACKEND (API) ROUTES // server.app.get("/api/reset", function(req, res) { async.waterfall([ function(next) { server.resetDatabase(next); }, function(next) { server.jsonSuccess("The database was reset.", res); } ]); }); server.app.get("/api/twitter/*", function(req, res) { var screenName = req.params[0]; if (!screenName.length) { server.jsonError("The Twitter screen name was empty.", res); return; } async.waterfall([ function(next) { // Account twitterClient.get( 'users/show', { "screen_name": screenName, }, function(error, data, raw) { if (error) { var errorText; switch(error.statusCode) { case 404: errorText = "That Twitter account doesn't exist."; break; default: errorText = "The twitter account couldn't be accessed; that's all."; break; } server.jsonError(errorText, res); return; } var newData = { "id": screenName, "name": data.name + "'" + (!_s.endsWith(data.name.toLowerCase(), "s") ? "s" : "") + " Museum", "isSpontaneous": false, "twitter": { "account": { "screenName": data.screen_name, "name": data.name, "location": data.location, "description": data.description, "language": data.lang, "picture": data.profile_image_url } } }; next(null, newData); } ); }, function(passedData, next) { twitterClient.get( 'statuses/user_timeline', { "screen_name": screenName, "trim_user": true, "count": 200, "include_rts": true }, function(error, data, raw) { if (error) { var errorText; switch(error.statusCode) { case 401: errorText = "That Twitter account is probably using &lsquo;protected tweets&rsquo;."; break; default: errorText = "The tweets couldn't be retrieved; that's all."; break; } server.jsonError(errorText, res); return; } var tweets = _.map(data, twinglish.cleanTweet); // var tweets = data; //some tweets may have been removed (annulled) tweets = _.filter(tweets, function(tweet) { return (!(tweet == null || tweet == undefined)); //sorry }); passedData["twitter"]["tweets"] = tweets; var tweetTexts = _.map(tweets, function(tweet) { return tweet.text; }); var textStuff = tweetTexts.join(" "); // from http://codereview.stackexchange.com/questions/63947/count-frequency-of-words-in-a-string-returning-list-of-unique-words-sorted-by-fr function getFrequency2(string, cutOff) { var cleanString = string.replace(/[\.,-\/#!$%\^&\*;:{}=\-_`~()]/g,""), words = cleanString.split(' '), frequencies = {}, word, frequency, i; for( i=0; i<words.length; i++ ) { word = words[i]; frequencies[word] = frequencies[word] || 0; frequencies[word]++; } words = Object.keys( frequencies ); return words.sort(function (a,b) { return frequencies[b] -frequencies[a];}).slice(0,cutOff); } var stuff = getFrequency2(textStuff, 100); var stuff2 = stuff.slice(95, 100); console.log(stuff2); passedData["topics"] = stuff2; next(null, passedData); } ); }, function(passedData, next) { if ("metrics" in req.query) { passedData["twitter"]["metrics"] = _.map(twetrics.metrics, function(metric) { return metric.method(passedData["twitter"]["tweets"]); }); } next(null, passedData); }, function(passedData, next) { var format = (("format" in req.query) ? req.query.format : "json"); var data; switch(format) { case "store": var museums = server.db("museums"); var r = museums.remove({ id: screenName }); // if (r["__wrapped__"].length == 0) { // //did not exist in the first place // return; // } museums.push(passedData); server.db.save(); server.jsonSuccess("The Twitter data was stored.", res); return; break; case "json": data = passedData; break; case "prettyjson": data =
{ var r = museums.find({ id: id }); return (r.value() ? true : false); }
identifier_body
server.js
": __dirname, "viewPath": "pages", "logging": true, "loggingFormat": ":method :url -> HTTP :status; :response-time ms", "staticDirectories": ["css", "fonts", "js", "static"] }); server.dbInit = function() { var _self = this; _self.db = low( "db.json", { "autosave": false } ); } server.resetDatabase = function(callback) { var _self = this; var blankData = fs.readFileSync("db-blank.json"); fs.writeFileSync("db.json", blankData); _self.dbInit(); callback(); }; server.jsonError = function(text, res) { res.send({ "error": text }); }; server.jsonSuccess = function(text, res) { res.send({ "ok": text }); }; next(); }, function(next) { // DATABASE if (!fs.existsSync("db.json")) { server.resetDatabase(next); } else { server.dbInit(); next(); } }, function(next) { // TWITTER fs.readFile("twitter-credentials.json", function(error, data) { if (error) { console.error("twitter-credentials.json is missing or couldn't be read. I need this."); process.exit(); return; } var twitterDetails = JSON.parse(data); twitterClient = new twitter(twitterDetails); next(); }); }, function(next) { // // FRONTEND (UI) ROUTES // server.app.get("/", function(req, res) { res.render("index"); }); server.app.get("/twitter/", function(req, res) { var twitter = "jadaradix"; res.render( "twitter", { "twitter": twitter } ); }); server.app.get("/twitter/*/", function(req, res) { var twitter = req.params[0]; res.render( "twitter", { "twitter": twitter } ); }); server.app.get("/process/", function(req, res) { if (!req.query.hasOwnProperty("name") || !req.query.hasOwnProperty("topics")) { res.redirect(302, "/"); return; } var name = _s.humanize(req.query.name) + " Museum"; var topics = req.query.topics.split(","); var museums = server.db("museums"); //insert function doesIdExist(id) { var r = museums.find({ id: id }); return (r.value() ? true : false); } function generateId() { var id = _.times(16, function(n) { return _.random(0, 10); }).join(""); return id; } var id = generateId(); while (doesIdExist(id)) { var id = generateId(); } var museum = { "id": id, "name": name, "isSpontaneous": true, "topics": topics } museums.push(museum); server.db.save(); res.redirect(302, "/museum/" + id); }); server.app.get("/process/*/", function(req, res) { var id = req.params[0]; var museums = server.db("museums"); var r = museums.find({ id: id }); var rValue = r.value(); if (!rValue) { res.redirect(302, "/twitter/" + id); return; } res.render( "process", { "twitter": id } ); }); server.app.get("/museum/*/", function(req, res) { var id = req.params[0]; var museums = server.db("museums"); var r = museums.find({ id: id }); var rValue = r.value(); if (!rValue || (!(("topics" in rValue)) || rValue.topics.length == 0)) { res.redirect(302, "/twitter/" + id); return; } res.render( "museum", { "title": rValue.name, "id": id } ); }); server.app.get("/favicon.ico", function(req, res) { res.redirect(301, "/static/favicon.ico"); }); // // BACKEND (API) ROUTES // server.app.get("/api/reset", function(req, res) { async.waterfall([ function(next) { server.resetDatabase(next); }, function(next) { server.jsonSuccess("The database was reset.", res); } ]); }); server.app.get("/api/twitter/*", function(req, res) { var screenName = req.params[0]; if (!screenName.length) { server.jsonError("The Twitter screen name was empty.", res); return; } async.waterfall([ function(next) { // Account twitterClient.get( 'users/show', { "screen_name": screenName, }, function(error, data, raw) { if (error) { var errorText; switch(error.statusCode) { case 404: errorText = "That Twitter account doesn't exist."; break; default: errorText = "The twitter account couldn't be accessed; that's all."; break; } server.jsonError(errorText, res); return; } var newData = { "id": screenName, "name": data.name + "'" + (!_s.endsWith(data.name.toLowerCase(), "s") ? "s" : "") + " Museum", "isSpontaneous": false, "twitter": { "account": { "screenName": data.screen_name, "name": data.name, "location": data.location, "description": data.description, "language": data.lang, "picture": data.profile_image_url } } }; next(null, newData); } ); }, function(passedData, next) { twitterClient.get( 'statuses/user_timeline', { "screen_name": screenName, "trim_user": true, "count": 200, "include_rts": true }, function(error, data, raw) { if (error) { var errorText; switch(error.statusCode) { case 401: errorText = "That Twitter account is probably using &lsquo;protected tweets&rsquo;."; break; default: errorText = "The tweets couldn't be retrieved; that's all."; break; } server.jsonError(errorText, res); return; } var tweets = _.map(data, twinglish.cleanTweet); // var tweets = data; //some tweets may have been removed (annulled) tweets = _.filter(tweets, function(tweet) { return (!(tweet == null || tweet == undefined)); //sorry }); passedData["twitter"]["tweets"] = tweets; var tweetTexts = _.map(tweets, function(tweet) { return tweet.text; }); var textStuff = tweetTexts.join(" "); // from http://codereview.stackexchange.com/questions/63947/count-frequency-of-words-in-a-string-returning-list-of-unique-words-sorted-by-fr function
(string, cutOff) { var cleanString = string.replace(/[\.,-\/#!$%\^&\*;:{}=\-_`~()]/g,""), words = cleanString.split(' '), frequencies = {}, word, frequency, i; for( i=0; i<words.length; i++ ) { word = words[i]; frequencies[word] = frequencies[word] || 0; frequencies[word]++; } words = Object.keys( frequencies ); return words.sort(function (a,b) { return frequencies[b] -frequencies[a];}).slice(0,cutOff); } var stuff = getFrequency2(textStuff, 100); var stuff2 = stuff.slice(95, 100); console.log(stuff2); passedData["topics"] = stuff2; next(null, passedData); } ); }, function(passedData, next) { if ("metrics" in req.query) { passedData["twitter"]["metrics"] = _.map(twetrics.metrics, function(metric) { return metric.method(passedData["twitter"]["tweets"]); }); } next(null, passedData); }, function(passedData, next) { var format = (("format" in req.query) ? req.query.format : "json"); var data; switch(format) { case "store": var museums = server.db("museums"); var r = museums.remove({ id: screenName }); // if (r["__wrapped__"].length == 0) { // //did not exist in the first place // return; // } museums.push(passedData); server.db.save(); server.jsonSuccess("The Twitter data was stored.", res); return; break; case "json": data = passedData; break; case "prettyjson": data = JSON
getFrequency2
identifier_name
datasource-view.spec.tsx
/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ import { shallow } from 'enzyme'; import React from 'react'; import { Capabilities } from '../../utils'; import { DatasourcesView } from './datasource-view'; describe('DatasourcesView', () => { it('matches snapshot', () => { const dataSourceView = shallow( <DatasourcesView goToQuery={() => {}} goToTask={() => null} goToSegments={() => {}} capabilities={Capabilities.FULL} />, ); expect(dataSourceView).toMatchSnapshot(); }); });
random_line_split
windowactivatable.py
# -*- coding: UTF-8 -*- # Gedit External Tools plugin # Copyright (C) 2005-2006 Steve Frécinaux <steve@istique.net> # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA __all__ = ('ExternalToolsPlugin', 'Manager', 'OutputPanel', 'Capture', 'UniqueById') from gi.repository import GLib, Gio, GObject, Gtk, Gedit, PeasGtk from .manager import Manager from .library import ToolLibrary from .outputpanel import OutputPanel from .capture import Capture from .functions import * class ToolMenu(object): def __init__(self, library, window, panel, menu): super(ToolMenu, self).__init__() self._library = library self._window = window self._panel = panel self._menu = menu self._action_tools = {} self.update() def deactivate(self): self.remove() def remove(self): self._menu.remove_all() for name, tool in self._action_tools.items(): self._window.remove_action(name) if tool.shortcut: app = Gio.Application.get_default() app.remove_accelerator(tool.shortcut) self._action_tools = {} def _insert_directory(self, directory, menu): for d in sorted(directory.subdirs, key=lambda x: x.name.lower()): submenu = Gio.Menu() menu.append_submenu(d.name.replace('_', '__'), submenu) section = Gio.Menu() submenu.append_section(None, section) self._insert_directory(d, section) for tool in sorted(directory.tools, key=lambda x: x.name.lower()): action_name = 'external-tool_%X_%X' % (id(tool), id(tool.name)) self._action_tools[action_name] = tool action = Gio.SimpleAction(name=action_name) action.connect('activate', capture_menu_action, self._window, self._panel, tool) self._window.add_action(action) item = Gio.MenuItem.new(tool.name.replace('_', '__'), "win.%s" % action_name) item.set_attribute_value("hidden-when", GLib.Variant.new_string("action-disabled")) menu.append_item(item) if tool.shortcut: app = Gio.Application.get_default() app.add_accelerator(tool.shortcut, "win.%s" % action_name, None) def update(self): self.remove() self._insert_directory(self._library.tree, self._menu) self.filter(self._window.get_active_document()) def filter_language(self, language, item): if not item.languages: return True if not language and 'plain' in item.languages: return True if language and (language.get_id() in item.languages): return True else: return False def filter(self, document): if document is None: titled = False remote = False language = None else: titled = document.get_location() is not None remote = not document.is_local() language = document.get_language() states = { 'always': True, 'all' : document is not None, 'local': titled and not remote, 'remote': titled and remote, 'titled': titled, 'untitled': not titled, } for name, tool in self._action_tools.items(): action = self._window.lookup_action(name) if action: action.set_enabled(states[tool.applicability] and self.filter_language(language, tool)) # FIXME: restore the launch of the manager on configure using PeasGtk.Configurable class WindowActivatable(GObject.Object, Gedit.WindowActivatable): __gtype_name__ = "ExternalToolsWindowActivatable" window = GObject.property(type=Gedit.Window) def __init__(self): GObject.Object.__init__(self) self._manager = None self._manager_default_size = None self.menu = None def do_activate(self): # Ugly hack... we need to get access to the activatable to update the menuitems self.window._external_tools_window_activatable = self self._library = ToolLibrary() action = Gio.SimpleAction(name="manage_tools") action.connect("activate", lambda action, parameter: self.open_dialog()) self.window.add_action(action) self.gear_menu = self.extend_gear_menu("ext9") item = Gio.MenuItem.new(_("Manage _External Tools..."), "win.manage_tools") self.gear_menu.append_menu_item(item) external_tools_submenu = Gio.Menu() item = Gio.MenuItem.new_submenu(_("External _Tools"), external_tools_submenu) self.gear_menu.append_menu_item(item) external_tools_submenu_section = Gio.Menu() external_tools_submenu.append_section(None, external_tools_submenu_section) # Create output console self._output_buffer = OutputPanel(self.plugin_info.get_data_dir(), self.window) self.menu = ToolMenu(self._library, self.window, self._output_buffer, external_tools_submenu_section) bottom = self.window.get_bottom_panel() bottom.add_titled(self._output_buffer.panel, "GeditExternalToolsShellOutput", _("Tool Output")) def do_update_state(self): if self.menu is not None: self.menu.filter(self.window.get_active_document()) def do_deactivate(self): self.window._external_tools_window_activatable = None self.menu.deactivate() self.window.remove_action("manage_tools") bottom = self.window.get_bottom_panel() bottom.remove(self._output_buffer.panel) def open_dialog(self): if not self._manager: self._manager = Manager(self.plugin_info.get_data_dir()) if self._manager_default_size: self._manager.dialog.set_default_size(*self._manager_default_size) self._manager.dialog.connect('destroy', self.on_manager_destroy) self._manager.connect('tools-updated', self.on_manager_tools_updated) window = Gio.Application.get_default().get_active_window() self._manager.run(window) return self._manager.dialog def u
self, tool): if self._manager: self._manager.tool_changed(tool, True) def on_manager_destroy(self, dialog): self._manager_default_size = self._manager.get_final_size() self._manager = None def on_manager_tools_updated(self, manager): for window in Gio.Application.get_default().get_windows(): window._external_tools_window_activatable.menu.update() # ex:ts=4:et:
pdate_manager(
identifier_name
windowactivatable.py
# -*- coding: UTF-8 -*- # Gedit External Tools plugin # Copyright (C) 2005-2006 Steve Frécinaux <steve@istique.net> # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA __all__ = ('ExternalToolsPlugin', 'Manager', 'OutputPanel', 'Capture', 'UniqueById') from gi.repository import GLib, Gio, GObject, Gtk, Gedit, PeasGtk from .manager import Manager from .library import ToolLibrary from .outputpanel import OutputPanel from .capture import Capture from .functions import * class ToolMenu(object): def __init__(self, library, window, panel, menu): super(ToolMenu, self).__init__() self._library = library self._window = window self._panel = panel self._menu = menu self._action_tools = {} self.update()
def remove(self): self._menu.remove_all() for name, tool in self._action_tools.items(): self._window.remove_action(name) if tool.shortcut: app = Gio.Application.get_default() app.remove_accelerator(tool.shortcut) self._action_tools = {} def _insert_directory(self, directory, menu): for d in sorted(directory.subdirs, key=lambda x: x.name.lower()): submenu = Gio.Menu() menu.append_submenu(d.name.replace('_', '__'), submenu) section = Gio.Menu() submenu.append_section(None, section) self._insert_directory(d, section) for tool in sorted(directory.tools, key=lambda x: x.name.lower()): action_name = 'external-tool_%X_%X' % (id(tool), id(tool.name)) self._action_tools[action_name] = tool action = Gio.SimpleAction(name=action_name) action.connect('activate', capture_menu_action, self._window, self._panel, tool) self._window.add_action(action) item = Gio.MenuItem.new(tool.name.replace('_', '__'), "win.%s" % action_name) item.set_attribute_value("hidden-when", GLib.Variant.new_string("action-disabled")) menu.append_item(item) if tool.shortcut: app = Gio.Application.get_default() app.add_accelerator(tool.shortcut, "win.%s" % action_name, None) def update(self): self.remove() self._insert_directory(self._library.tree, self._menu) self.filter(self._window.get_active_document()) def filter_language(self, language, item): if not item.languages: return True if not language and 'plain' in item.languages: return True if language and (language.get_id() in item.languages): return True else: return False def filter(self, document): if document is None: titled = False remote = False language = None else: titled = document.get_location() is not None remote = not document.is_local() language = document.get_language() states = { 'always': True, 'all' : document is not None, 'local': titled and not remote, 'remote': titled and remote, 'titled': titled, 'untitled': not titled, } for name, tool in self._action_tools.items(): action = self._window.lookup_action(name) if action: action.set_enabled(states[tool.applicability] and self.filter_language(language, tool)) # FIXME: restore the launch of the manager on configure using PeasGtk.Configurable class WindowActivatable(GObject.Object, Gedit.WindowActivatable): __gtype_name__ = "ExternalToolsWindowActivatable" window = GObject.property(type=Gedit.Window) def __init__(self): GObject.Object.__init__(self) self._manager = None self._manager_default_size = None self.menu = None def do_activate(self): # Ugly hack... we need to get access to the activatable to update the menuitems self.window._external_tools_window_activatable = self self._library = ToolLibrary() action = Gio.SimpleAction(name="manage_tools") action.connect("activate", lambda action, parameter: self.open_dialog()) self.window.add_action(action) self.gear_menu = self.extend_gear_menu("ext9") item = Gio.MenuItem.new(_("Manage _External Tools..."), "win.manage_tools") self.gear_menu.append_menu_item(item) external_tools_submenu = Gio.Menu() item = Gio.MenuItem.new_submenu(_("External _Tools"), external_tools_submenu) self.gear_menu.append_menu_item(item) external_tools_submenu_section = Gio.Menu() external_tools_submenu.append_section(None, external_tools_submenu_section) # Create output console self._output_buffer = OutputPanel(self.plugin_info.get_data_dir(), self.window) self.menu = ToolMenu(self._library, self.window, self._output_buffer, external_tools_submenu_section) bottom = self.window.get_bottom_panel() bottom.add_titled(self._output_buffer.panel, "GeditExternalToolsShellOutput", _("Tool Output")) def do_update_state(self): if self.menu is not None: self.menu.filter(self.window.get_active_document()) def do_deactivate(self): self.window._external_tools_window_activatable = None self.menu.deactivate() self.window.remove_action("manage_tools") bottom = self.window.get_bottom_panel() bottom.remove(self._output_buffer.panel) def open_dialog(self): if not self._manager: self._manager = Manager(self.plugin_info.get_data_dir()) if self._manager_default_size: self._manager.dialog.set_default_size(*self._manager_default_size) self._manager.dialog.connect('destroy', self.on_manager_destroy) self._manager.connect('tools-updated', self.on_manager_tools_updated) window = Gio.Application.get_default().get_active_window() self._manager.run(window) return self._manager.dialog def update_manager(self, tool): if self._manager: self._manager.tool_changed(tool, True) def on_manager_destroy(self, dialog): self._manager_default_size = self._manager.get_final_size() self._manager = None def on_manager_tools_updated(self, manager): for window in Gio.Application.get_default().get_windows(): window._external_tools_window_activatable.menu.update() # ex:ts=4:et:
def deactivate(self): self.remove()
random_line_split
windowactivatable.py
# -*- coding: UTF-8 -*- # Gedit External Tools plugin # Copyright (C) 2005-2006 Steve Frécinaux <steve@istique.net> # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA __all__ = ('ExternalToolsPlugin', 'Manager', 'OutputPanel', 'Capture', 'UniqueById') from gi.repository import GLib, Gio, GObject, Gtk, Gedit, PeasGtk from .manager import Manager from .library import ToolLibrary from .outputpanel import OutputPanel from .capture import Capture from .functions import * class ToolMenu(object): def __init__(self, library, window, panel, menu): super(ToolMenu, self).__init__() self._library = library self._window = window self._panel = panel self._menu = menu self._action_tools = {} self.update() def deactivate(self): self.remove() def remove(self): self._menu.remove_all() for name, tool in self._action_tools.items(): self._window.remove_action(name) if tool.shortcut: app = Gio.Application.get_default() app.remove_accelerator(tool.shortcut) self._action_tools = {} def _insert_directory(self, directory, menu): for d in sorted(directory.subdirs, key=lambda x: x.name.lower()): submenu = Gio.Menu() menu.append_submenu(d.name.replace('_', '__'), submenu) section = Gio.Menu() submenu.append_section(None, section) self._insert_directory(d, section) for tool in sorted(directory.tools, key=lambda x: x.name.lower()): action_name = 'external-tool_%X_%X' % (id(tool), id(tool.name)) self._action_tools[action_name] = tool action = Gio.SimpleAction(name=action_name) action.connect('activate', capture_menu_action, self._window, self._panel, tool) self._window.add_action(action) item = Gio.MenuItem.new(tool.name.replace('_', '__'), "win.%s" % action_name) item.set_attribute_value("hidden-when", GLib.Variant.new_string("action-disabled")) menu.append_item(item) if tool.shortcut: app = Gio.Application.get_default() app.add_accelerator(tool.shortcut, "win.%s" % action_name, None) def update(self): self.remove() self._insert_directory(self._library.tree, self._menu) self.filter(self._window.get_active_document()) def filter_language(self, language, item): if not item.languages: return True if not language and 'plain' in item.languages: return True if language and (language.get_id() in item.languages): return True else: return False def filter(self, document): i
if action: action.set_enabled(states[tool.applicability] and self.filter_language(language, tool)) # FIXME: restore the launch of the manager on configure using PeasGtk.Configurable class WindowActivatable(GObject.Object, Gedit.WindowActivatable): __gtype_name__ = "ExternalToolsWindowActivatable" window = GObject.property(type=Gedit.Window) def __init__(self): GObject.Object.__init__(self) self._manager = None self._manager_default_size = None self.menu = None def do_activate(self): # Ugly hack... we need to get access to the activatable to update the menuitems self.window._external_tools_window_activatable = self self._library = ToolLibrary() action = Gio.SimpleAction(name="manage_tools") action.connect("activate", lambda action, parameter: self.open_dialog()) self.window.add_action(action) self.gear_menu = self.extend_gear_menu("ext9") item = Gio.MenuItem.new(_("Manage _External Tools..."), "win.manage_tools") self.gear_menu.append_menu_item(item) external_tools_submenu = Gio.Menu() item = Gio.MenuItem.new_submenu(_("External _Tools"), external_tools_submenu) self.gear_menu.append_menu_item(item) external_tools_submenu_section = Gio.Menu() external_tools_submenu.append_section(None, external_tools_submenu_section) # Create output console self._output_buffer = OutputPanel(self.plugin_info.get_data_dir(), self.window) self.menu = ToolMenu(self._library, self.window, self._output_buffer, external_tools_submenu_section) bottom = self.window.get_bottom_panel() bottom.add_titled(self._output_buffer.panel, "GeditExternalToolsShellOutput", _("Tool Output")) def do_update_state(self): if self.menu is not None: self.menu.filter(self.window.get_active_document()) def do_deactivate(self): self.window._external_tools_window_activatable = None self.menu.deactivate() self.window.remove_action("manage_tools") bottom = self.window.get_bottom_panel() bottom.remove(self._output_buffer.panel) def open_dialog(self): if not self._manager: self._manager = Manager(self.plugin_info.get_data_dir()) if self._manager_default_size: self._manager.dialog.set_default_size(*self._manager_default_size) self._manager.dialog.connect('destroy', self.on_manager_destroy) self._manager.connect('tools-updated', self.on_manager_tools_updated) window = Gio.Application.get_default().get_active_window() self._manager.run(window) return self._manager.dialog def update_manager(self, tool): if self._manager: self._manager.tool_changed(tool, True) def on_manager_destroy(self, dialog): self._manager_default_size = self._manager.get_final_size() self._manager = None def on_manager_tools_updated(self, manager): for window in Gio.Application.get_default().get_windows(): window._external_tools_window_activatable.menu.update() # ex:ts=4:et:
f document is None: titled = False remote = False language = None else: titled = document.get_location() is not None remote = not document.is_local() language = document.get_language() states = { 'always': True, 'all' : document is not None, 'local': titled and not remote, 'remote': titled and remote, 'titled': titled, 'untitled': not titled, } for name, tool in self._action_tools.items(): action = self._window.lookup_action(name)
identifier_body
windowactivatable.py
# -*- coding: UTF-8 -*- # Gedit External Tools plugin # Copyright (C) 2005-2006 Steve Frécinaux <steve@istique.net> # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA __all__ = ('ExternalToolsPlugin', 'Manager', 'OutputPanel', 'Capture', 'UniqueById') from gi.repository import GLib, Gio, GObject, Gtk, Gedit, PeasGtk from .manager import Manager from .library import ToolLibrary from .outputpanel import OutputPanel from .capture import Capture from .functions import * class ToolMenu(object): def __init__(self, library, window, panel, menu): super(ToolMenu, self).__init__() self._library = library self._window = window self._panel = panel self._menu = menu self._action_tools = {} self.update() def deactivate(self): self.remove() def remove(self): self._menu.remove_all() for name, tool in self._action_tools.items(): self._window.remove_action(name) if tool.shortcut: app = Gio.Application.get_default() app.remove_accelerator(tool.shortcut) self._action_tools = {} def _insert_directory(self, directory, menu): for d in sorted(directory.subdirs, key=lambda x: x.name.lower()): s
for tool in sorted(directory.tools, key=lambda x: x.name.lower()): action_name = 'external-tool_%X_%X' % (id(tool), id(tool.name)) self._action_tools[action_name] = tool action = Gio.SimpleAction(name=action_name) action.connect('activate', capture_menu_action, self._window, self._panel, tool) self._window.add_action(action) item = Gio.MenuItem.new(tool.name.replace('_', '__'), "win.%s" % action_name) item.set_attribute_value("hidden-when", GLib.Variant.new_string("action-disabled")) menu.append_item(item) if tool.shortcut: app = Gio.Application.get_default() app.add_accelerator(tool.shortcut, "win.%s" % action_name, None) def update(self): self.remove() self._insert_directory(self._library.tree, self._menu) self.filter(self._window.get_active_document()) def filter_language(self, language, item): if not item.languages: return True if not language and 'plain' in item.languages: return True if language and (language.get_id() in item.languages): return True else: return False def filter(self, document): if document is None: titled = False remote = False language = None else: titled = document.get_location() is not None remote = not document.is_local() language = document.get_language() states = { 'always': True, 'all' : document is not None, 'local': titled and not remote, 'remote': titled and remote, 'titled': titled, 'untitled': not titled, } for name, tool in self._action_tools.items(): action = self._window.lookup_action(name) if action: action.set_enabled(states[tool.applicability] and self.filter_language(language, tool)) # FIXME: restore the launch of the manager on configure using PeasGtk.Configurable class WindowActivatable(GObject.Object, Gedit.WindowActivatable): __gtype_name__ = "ExternalToolsWindowActivatable" window = GObject.property(type=Gedit.Window) def __init__(self): GObject.Object.__init__(self) self._manager = None self._manager_default_size = None self.menu = None def do_activate(self): # Ugly hack... we need to get access to the activatable to update the menuitems self.window._external_tools_window_activatable = self self._library = ToolLibrary() action = Gio.SimpleAction(name="manage_tools") action.connect("activate", lambda action, parameter: self.open_dialog()) self.window.add_action(action) self.gear_menu = self.extend_gear_menu("ext9") item = Gio.MenuItem.new(_("Manage _External Tools..."), "win.manage_tools") self.gear_menu.append_menu_item(item) external_tools_submenu = Gio.Menu() item = Gio.MenuItem.new_submenu(_("External _Tools"), external_tools_submenu) self.gear_menu.append_menu_item(item) external_tools_submenu_section = Gio.Menu() external_tools_submenu.append_section(None, external_tools_submenu_section) # Create output console self._output_buffer = OutputPanel(self.plugin_info.get_data_dir(), self.window) self.menu = ToolMenu(self._library, self.window, self._output_buffer, external_tools_submenu_section) bottom = self.window.get_bottom_panel() bottom.add_titled(self._output_buffer.panel, "GeditExternalToolsShellOutput", _("Tool Output")) def do_update_state(self): if self.menu is not None: self.menu.filter(self.window.get_active_document()) def do_deactivate(self): self.window._external_tools_window_activatable = None self.menu.deactivate() self.window.remove_action("manage_tools") bottom = self.window.get_bottom_panel() bottom.remove(self._output_buffer.panel) def open_dialog(self): if not self._manager: self._manager = Manager(self.plugin_info.get_data_dir()) if self._manager_default_size: self._manager.dialog.set_default_size(*self._manager_default_size) self._manager.dialog.connect('destroy', self.on_manager_destroy) self._manager.connect('tools-updated', self.on_manager_tools_updated) window = Gio.Application.get_default().get_active_window() self._manager.run(window) return self._manager.dialog def update_manager(self, tool): if self._manager: self._manager.tool_changed(tool, True) def on_manager_destroy(self, dialog): self._manager_default_size = self._manager.get_final_size() self._manager = None def on_manager_tools_updated(self, manager): for window in Gio.Application.get_default().get_windows(): window._external_tools_window_activatable.menu.update() # ex:ts=4:et:
ubmenu = Gio.Menu() menu.append_submenu(d.name.replace('_', '__'), submenu) section = Gio.Menu() submenu.append_section(None, section) self._insert_directory(d, section)
conditional_block
macro_parser.rs
step) //! //! Remaining input: `b` //! cur: [a $( a · )* a b] next: [a $( a )* a · b] //! Finish/Repeat (first item) //! next: [a $( a )* · a b] [a $( · a )* a b] //! //! - - - Advance over a `b`. - - - //! //! Remaining input: `` //! eof: [a $( a )* a b ·] pub use self::NamedMatch::*; pub use self::ParseResult::*; use self::TokenTreeOrTokenTreeVec::*; use ast; use ast::{TokenTree, Ident}; use ast::{TtDelimited, TtSequence, TtToken}; use codemap::{BytePos, mk_sp, Span}; use codemap; use parse::lexer::*; //resolve bug? use parse::ParseSess; use parse::attr::ParserAttr; use parse::parser::{LifetimeAndTypesWithoutColons, Parser}; use parse::token::{Eof, DocComment, MatchNt, SubstNt}; use parse::token::{Token, Nonterminal}; use parse::token; use print::pprust; use ptr::P; use std::mem; use std::rc::Rc; use std::collections::HashMap; use std::collections::hash_map::Entry::{Vacant, Occupied}; // To avoid costly uniqueness checks, we require that `MatchSeq` always has // a nonempty body. #[derive(Clone)] enum TokenTreeOrTokenTreeVec { Tt(ast::TokenTree), TtSeq(Rc<Vec<ast::TokenTree>>), } impl TokenTreeOrTokenTreeVec { fn len(&self) -> usize { match self { &TtSeq(ref v) => v.len(), &Tt(ref tt) => tt.len(), } } fn get_tt(&self, index: usize) -> TokenTree { match self { &TtSeq(ref v) => v[index].clone(), &Tt(ref tt) => tt.get_tt(index), } } } /// an unzipping of `TokenTree`s #[derive(Clone)] struct MatcherTtFrame { elts: TokenTreeOrTokenTreeVec, idx: usize, } #[derive(Clone)] pub struct MatcherPos { stack:
erTtFrame>, top_elts: TokenTreeOrTokenTreeVec, sep: Option<Token>, idx: usize, up: Option<Box<MatcherPos>>, matches: Vec<Vec<Rc<NamedMatch>>>, match_lo: usize, match_cur: usize, match_hi: usize, sp_lo: BytePos, } pub fn count_names(ms: &[TokenTree]) -> usize { ms.iter().fold(0, |count, elt| { count + match elt { &TtSequence(_, ref seq) => { seq.num_captures } &TtDelimited(_, ref delim) => { count_names(&delim.tts) } &TtToken(_, MatchNt(..)) => { 1 } &TtToken(_, _) => 0, } }) } pub fn initial_matcher_pos(ms: Rc<Vec<TokenTree>>, sep: Option<Token>, lo: BytePos) -> Box<MatcherPos> { let match_idx_hi = count_names(&ms[..]); let matches: Vec<_> = (0..match_idx_hi).map(|_| Vec::new()).collect(); Box::new(MatcherPos { stack: vec![], top_elts: TtSeq(ms), sep: sep, idx: 0, up: None, matches: matches, match_lo: 0, match_cur: 0, match_hi: match_idx_hi, sp_lo: lo }) } /// NamedMatch is a pattern-match result for a single token::MATCH_NONTERMINAL: /// so it is associated with a single ident in a parse, and all /// `MatchedNonterminal`s in the NamedMatch have the same nonterminal type /// (expr, item, etc). Each leaf in a single NamedMatch corresponds to a /// single token::MATCH_NONTERMINAL in the TokenTree that produced it. /// /// The in-memory structure of a particular NamedMatch represents the match /// that occurred when a particular subset of a matcher was applied to a /// particular token tree. /// /// The width of each MatchedSeq in the NamedMatch, and the identity of the /// `MatchedNonterminal`s, will depend on the token tree it was applied to: /// each MatchedSeq corresponds to a single TTSeq in the originating /// token tree. The depth of the NamedMatch structure will therefore depend /// only on the nesting depth of `ast::TTSeq`s in the originating /// token tree it was derived from. pub enum NamedMatch { MatchedSeq(Vec<Rc<NamedMatch>>, codemap::Span), MatchedNonterminal(Nonterminal) } pub fn nameize(p_s: &ParseSess, ms: &[TokenTree], res: &[Rc<NamedMatch>]) -> HashMap<Ident, Rc<NamedMatch>> { fn n_rec(p_s: &ParseSess, m: &TokenTree, res: &[Rc<NamedMatch>], ret_val: &mut HashMap<Ident, Rc<NamedMatch>>, idx: &mut usize) { match m { &TtSequence(_, ref seq) => { for next_m in &seq.tts { n_rec(p_s, next_m, res, ret_val, idx) } } &TtDelimited(_, ref delim) => { for next_m in &delim.tts { n_rec(p_s, next_m, res, ret_val, idx) } } &TtToken(sp, MatchNt(bind_name, _, _, _)) => { match ret_val.entry(bind_name) { Vacant(spot) => { spot.insert(res[*idx].clone()); *idx += 1; } Occupied(..) => { panic!(p_s.span_diagnostic .span_fatal(sp, &format!("duplicated bind name: {}", bind_name))) } } } &TtToken(_, SubstNt(..)) => panic!("Cannot fill in a NT"), &TtToken(_, _) => (), } } let mut ret_val = HashMap::new(); let mut idx = 0; for m in ms { n_rec(p_s, m, res, &mut ret_val, &mut idx) } ret_val } pub enum ParseResult<T> { Success(T), Failure(codemap::Span, String), Error(codemap::Span, String) } pub type NamedParseResult = ParseResult<HashMap<Ident, Rc<NamedMatch>>>; pub type PositionalParseResult = ParseResult<Vec<Rc<NamedMatch>>>; pub fn parse_or_else(sess: &ParseSess, cfg: ast::CrateConfig, rdr: TtReader, ms: Vec<TokenTree> ) -> HashMap<Ident, Rc<NamedMatch>> { match parse(sess, cfg, rdr, &ms[..]) { Success(m) => m, Failure(sp, str) => { panic!(sess.span_diagnostic.span_fatal(sp, &str[..])) } Error(sp, str) => { panic!(sess.span_diagnostic.span_fatal(sp, &str[..])) } } } /// Perform a token equality check, ignoring syntax context (that is, an /// unhygienic comparison) pub fn token_name_eq(t1 : &Token, t2 : &Token) -> bool { match (t1,t2) { (&token::Ident(id1,_),&token::Ident(id2,_)) | (&token::Lifetime(id1),&token::Lifetime(id2)) => id1.name == id2.name, _ => *t1 == *t2 } } pub fn parse(sess: &ParseSess, cfg: ast::CrateConfig, mut rdr: TtReader, ms: &[TokenTree]) -> NamedParseResult { let mut cur_eis = Vec::new(); cur_eis.push(initial_matcher_pos(Rc::new(ms.iter() .cloned() .collect()), None, rdr.peek().sp.lo)); loop { let mut bb_eis = Vec::new(); // black-box parsed by parser.rs let mut next_eis = Vec::new(); // or proceed normally let mut eof_eis = Vec::new(); let TokenAndSpan { tok, sp } = rdr.peek(); /* we append new items to this while we go */ loop { let mut ei = match cur_eis.pop() { None => break, /* for each Earley Item */ Some(ei) => ei, }; // When unzipped trees end, remove them while ei.idx >= ei.top_elts.len() { match ei.stack.pop() { Some(MatcherTtFrame { elts, idx }) => { ei.top_elts = elts; ei.idx = idx + 1; } None => break } } let idx = ei.idx; let len = ei.top_elts.len(); /* at end of sequence */ if idx >= len { // can't move out of `match`es, so: if ei
Vec<Match
identifier_name
macro_parser.rs
step) //! //! Remaining input: `b` //! cur: [a $( a · )* a b] next: [a $( a )* a · b] //! Finish/Repeat (first item) //! next: [a $( a )* · a b] [a $( · a )* a b] //! //! - - - Advance over a `b`. - - - //! //! Remaining input: `` //! eof: [a $( a )* a b ·] pub use self::NamedMatch::*; pub use self::ParseResult::*; use self::TokenTreeOrTokenTreeVec::*; use ast; use ast::{TokenTree, Ident}; use ast::{TtDelimited, TtSequence, TtToken}; use codemap::{BytePos, mk_sp, Span}; use codemap; use parse::lexer::*; //resolve bug? use parse::ParseSess; use parse::attr::ParserAttr; use parse::parser::{LifetimeAndTypesWithoutColons, Parser}; use parse::token::{Eof, DocComment, MatchNt, SubstNt}; use parse::token::{Token, Nonterminal}; use parse::token; use print::pprust; use ptr::P; use std::mem; use std::rc::Rc; use std::collections::HashMap; use std::collections::hash_map::Entry::{Vacant, Occupied}; // To avoid costly uniqueness checks, we require that `MatchSeq` always has // a nonempty body. #[derive(Clone)] enum TokenTreeOrTokenTreeVec { Tt(ast::TokenTree), TtSeq(Rc<Vec<ast::TokenTree>>), } impl TokenTreeOrTokenTreeVec { fn len(&self) -> usize { match self { &TtSeq(ref v) => v.len(), &Tt(ref tt) => tt.len(), } } fn get_tt(&self, index: usize) -> TokenTree { match self { &TtSeq(ref v) => v[index].clone(), &Tt(ref tt) => tt.get_tt(index), } } } /// an unzipping of `TokenTree`s #[derive(Clone)] struct MatcherTtFrame { elts: TokenTreeOrTokenTreeVec, idx: usize, } #[derive(Clone)] pub struct MatcherPos { stack: Vec<MatcherTtFrame>, top_elts: TokenTreeOrTokenTreeVec, sep: Option<Token>, idx: usize, up: Option<Box<MatcherPos>>, matches: Vec<Vec<Rc<NamedMatch>>>, match_lo: usize, match_cur: usize, match_hi: usize, sp_lo: BytePos, } pub fn count_names(ms: &[TokenTree]) -> usize { ms.iter().fold(0, |count, elt| { count + match elt { &TtSequence(_, ref seq) => { seq.num_captures } &TtDelimited(_, ref delim) => { count_names(&delim.tts) } &TtToken(_, MatchNt(..)) => { 1 } &TtToken(_, _) => 0, } }) } pub fn initial_matcher_pos(ms: Rc<Vec<TokenTree>>, sep: Option<Token>, lo: BytePos) -> Box<MatcherPos> { let match_idx_hi = count_names(&ms[..]); let matches: Vec<_> = (0..match_idx_hi).map(|_| Vec::new()).collect(); Box::new(MatcherPos { stack: vec![], top_elts: TtSeq(ms), sep: sep, idx: 0, up: None, matches: matches, match_lo: 0, match_cur: 0, match_hi: match_idx_hi, sp_lo: lo }) } /// NamedMatch is a pattern-match result for a single token::MATCH_NONTERMINAL: /// so it is associated with a single ident in a parse, and all /// `MatchedNonterminal`s in the NamedMatch have the same nonterminal type /// (expr, item, etc). Each leaf in a single NamedMatch corresponds to a /// single token::MATCH_NONTERMINAL in the TokenTree that produced it. /// /// The in-memory structure of a particular NamedMatch represents the match /// that occurred when a particular subset of a matcher was applied to a /// particular token tree. /// /// The width of each MatchedSeq in the NamedMatch, and the identity of the /// `MatchedNonterminal`s, will depend on the token tree it was applied to: /// each MatchedSeq corresponds to a single TTSeq in the originating /// token tree. The depth of the NamedMatch structure will therefore depend /// only on the nesting depth of `ast::TTSeq`s in the originating /// token tree it was derived from. pub enum NamedMatch { MatchedSeq(Vec<Rc<NamedMatch>>, codemap::Span), MatchedNonterminal(Nonterminal) } pub fn nameize(p_s: &ParseSess, ms: &[TokenTree], res: &[Rc<NamedMatch>]) -> HashMap<Ident, Rc<NamedMatch>> { fn n_rec(p_s: &ParseSess, m: &TokenTree, res: &[Rc<NamedMatch>], ret_val: &mut HashMap<Ident, Rc<NamedMatch>>, idx: &mut usize) { match m {
bind_name))) } } } &TtToken(_, SubstNt(..)) => panic!("Cannot fill in a NT"), &TtToken(_, _) => (), } } let mut ret_val = HashMap::new(); let mut idx = 0; for m in ms { n_rec(p_s, m, res, &mut ret_val, &mut idx) } ret_val } pub enum ParseResult<T> { Success(T), Failure(codemap::Span, String), Error(codemap::Span, String) } pub type NamedParseResult = ParseResult<HashMap<Ident, Rc<NamedMatch>>>; pub type PositionalParseResult = ParseResult<Vec<Rc<NamedMatch>>>; pub fn parse_or_else(sess: &ParseSess, cfg: ast::CrateConfig, rdr: TtReader, ms: Vec<TokenTree> ) -> HashMap<Ident, Rc<NamedMatch>> { match parse(sess, cfg, rdr, &ms[..]) { Success(m) => m, Failure(sp, str) => { panic!(sess.span_diagnostic.span_fatal(sp, &str[..])) } Error(sp, str) => { panic!(sess.span_diagnostic.span_fatal(sp, &str[..])) } } } /// Perform a token equality check, ignoring syntax context (that is, an /// unhygienic comparison) pub fn token_name_eq(t1 : &Token, t2 : &Token) -> bool { match (t1,t2) { (&token::Ident(id1,_),&token::Ident(id2,_)) | (&token::Lifetime(id1),&token::Lifetime(id2)) => id1.name == id2.name, _ => *t1 == *t2 } } pub fn parse(sess: &ParseSess, cfg: ast::CrateConfig, mut rdr: TtReader, ms: &[TokenTree]) -> NamedParseResult { let mut cur_eis = Vec::new(); cur_eis.push(initial_matcher_pos(Rc::new(ms.iter() .cloned() .collect()), None, rdr.peek().sp.lo)); loop { let mut bb_eis = Vec::new(); // black-box parsed by parser.rs let mut next_eis = Vec::new(); // or proceed normally let mut eof_eis = Vec::new(); let TokenAndSpan { tok, sp } = rdr.peek(); /* we append new items to this while we go */ loop { let mut ei = match cur_eis.pop() { None => break, /* for each Earley Item */ Some(ei) => ei, }; // When unzipped trees end, remove them while ei.idx >= ei.top_elts.len() { match ei.stack.pop() { Some(MatcherTtFrame { elts, idx }) => { ei.top_elts = elts; ei.idx = idx + 1; } None => break } } let idx = ei.idx; let len = ei.top_elts.len(); /* at end of sequence */ if idx >= len { // can't move out of `match`es, so: if ei
&TtSequence(_, ref seq) => { for next_m in &seq.tts { n_rec(p_s, next_m, res, ret_val, idx) } } &TtDelimited(_, ref delim) => { for next_m in &delim.tts { n_rec(p_s, next_m, res, ret_val, idx) } } &TtToken(sp, MatchNt(bind_name, _, _, _)) => { match ret_val.entry(bind_name) { Vacant(spot) => { spot.insert(res[*idx].clone()); *idx += 1; } Occupied(..) => { panic!(p_s.span_diagnostic .span_fatal(sp, &format!("duplicated bind name: {}",
identifier_body
macro_parser.rs
while ei.idx >= ei.top_elts.len() { match ei.stack.pop() { Some(MatcherTtFrame { elts, idx }) => { ei.top_elts = elts; ei.idx = idx + 1; } None => break } } let idx = ei.idx; let len = ei.top_elts.len(); /* at end of sequence */ if idx >= len { // can't move out of `match`es, so: if ei.up.is_some() { // hack: a matcher sequence is repeating iff it has a // parent (the top level is just a container) // disregard separator, try to go up // (remove this condition to make trailing seps ok) if idx == len { // pop from the matcher position let mut new_pos = ei.up.clone().unwrap(); // update matches (the MBE "parse tree") by appending // each tree as a subtree. // I bet this is a perf problem: we're preemptively // doing a lot of array work that will get thrown away // most of the time. // Only touch the binders we have actually bound for idx in ei.match_lo..ei.match_hi { let sub = (ei.matches[idx]).clone(); (&mut new_pos.matches[idx]) .push(Rc::new(MatchedSeq(sub, mk_sp(ei.sp_lo, sp.hi)))); } new_pos.match_cur = ei.match_hi; new_pos.idx += 1; cur_eis.push(new_pos); } // can we go around again? // the *_t vars are workarounds for the lack of unary move match ei.sep { Some(ref t) if idx == len => { // we need a separator // i'm conflicted about whether this should be hygienic.... // though in this case, if the separators are never legal // idents, it shouldn't matter. if token_name_eq(&tok, t) { //pass the separator let mut ei_t = ei.clone(); // ei_t.match_cur = ei_t.match_lo; ei_t.idx += 1; next_eis.push(ei_t); } } _ => { // we don't need a separator let mut ei_t = ei; ei_t.match_cur = ei_t.match_lo; ei_t.idx = 0; cur_eis.push(ei_t); } } } else { eof_eis.push(ei); } } else { match ei.top_elts.get_tt(idx) { /* need to descend into sequence */ TtSequence(sp, seq) => { if seq.op == ast::ZeroOrMore { let mut new_ei = ei.clone(); new_ei.match_cur += seq.num_captures; new_ei.idx += 1; //we specifically matched zero repeats. for idx in ei.match_cur..ei.match_cur + seq.num_captures { (&mut new_ei.matches[idx]).push(Rc::new(MatchedSeq(vec![], sp))); } cur_eis.push(new_ei); } let matches: Vec<_> = (0..ei.matches.len()) .map(|_| Vec::new()).collect(); let ei_t = ei; cur_eis.push(Box::new(MatcherPos { stack: vec![], sep: seq.separator.clone(), idx: 0, matches: matches, match_lo: ei_t.match_cur, match_cur: ei_t.match_cur, match_hi: ei_t.match_cur + seq.num_captures, up: Some(ei_t), sp_lo: sp.lo, top_elts: Tt(TtSequence(sp, seq)), })); } TtToken(_, MatchNt(..)) => { // Built-in nonterminals never start with these tokens, // so we can eliminate them from consideration. match tok { token::CloseDelim(_) => {}, _ => bb_eis.push(ei), } } TtToken(sp, SubstNt(..)) => { return Error(sp, "Cannot transcribe in macro LHS".to_string()) } seq @ TtDelimited(..) | seq @ TtToken(_, DocComment(..)) => { let lower_elts = mem::replace(&mut ei.top_elts, Tt(seq)); let idx = ei.idx; ei.stack.push(MatcherTtFrame { elts: lower_elts, idx: idx, }); ei.idx = 0; cur_eis.push(ei); } TtToken(_, ref t) => { let mut ei_t = ei.clone(); if token_name_eq(t,&tok) { ei_t.idx += 1; next_eis.push(ei_t); } } } } } /* error messages here could be improved with links to orig. rules */ if token_name_eq(&tok, &token::Eof) { if eof_eis.len() == 1 { let mut v = Vec::new(); for dv in &mut (&mut eof_eis[0]).matches { v.push(dv.pop().unwrap()); } return Success(nameize(sess, ms, &v[..])); } else if eof_eis.len() > 1 { return Error(sp, "ambiguity: multiple successful parses".to_string()); } else { return Failure(sp, "unexpected end of macro invocation".to_string()); } } else { if (!bb_eis.is_empty() && !next_eis.is_empty()) || bb_eis.len() > 1 { let nts = bb_eis.iter().map(|ei| { match ei.top_elts.get_tt(ei.idx) { TtToken(_, MatchNt(bind, name, _, _)) => { format!("{} ('{}')", name, bind) } _ => panic!() } }).collect::<Vec<String>>().join(" or "); return Error(sp, format!( "local ambiguity: multiple parsing options: \ built-in NTs {} or {} other options.", nts, next_eis.len()).to_string()); } else if bb_eis.is_empty() && next_eis.is_empty() { return Failure(sp, format!("no rules expected the token `{}`", pprust::token_to_string(&tok)).to_string()); } else if !next_eis.is_empty() { /* Now process the next token */ while !next_eis.is_empty() { cur_eis.push(next_eis.pop().unwrap()); } rdr.next_token(); } else /* bb_eis.len() == 1 */ { let mut rust_parser = Parser::new(sess, cfg.clone(), Box::new(rdr.clone())); let mut ei = bb_eis.pop().unwrap(); match ei.top_elts.get_tt(ei.idx) { TtToken(span, MatchNt(_, ident, _, _)) => { let match_cur = ei.match_cur; (&mut ei.matches[match_cur]).push(Rc::new(MatchedNonterminal( parse_nt(&mut rust_parser, span, &ident.name.as_str())))); ei.idx += 1; ei.match_cur += 1; } _ => panic!() } cur_eis.push(ei); for _ in 0..rust_parser.tokens_consumed { let _ = rdr.next_token(); } } } assert!(!cur_eis.is_empty()); } } pub fn parse_nt(p: &mut Parser, sp: Span, name: &str) -> Nonterminal { match name { "tt" => { p.quote_depth += 1; //but in theory, non-quoted tts might be useful let res = token::NtTT(P(panictry!(p.parse_token_tree()))); p.quote_depth -= 1; return res; } _ => {} } // check at the beginning and the parser checks after each bump panictry!(p.check_unknown_macro_variable()); match name { "item" => match p.parse_item() { Some(i) => token::NtItem(i), None => panic!(p.fatal("expected an item keyword")) }, "block" => token::NtBlock(panictry!(p.parse_block())), "stmt" => match p.parse_stmt() { Some(s) => token::NtStmt(s), None => panic!(p.fatal("expected a statement")) }, "pat" => token::NtPat(p.parse_pat()), "expr" => token::NtExpr(p.parse_expr()), "ty" => token::NtTy(p.parse_ty()), // this could be handled like a token, since it is one "ident" => match p.token { token::Ident(sn,b) => { panictry!(p.bump()); token::NtIdent(Box::new(sn),b) } _ => { let token_str = pprust::token_to_string(&p.token); panic!(p.fatal(&format!("expected ident, found {}", &token_str[..]))) } }, "path" => { token::NtPath
(Box::new(panictry!(p.parse_path(LifetimeAndTypesWithoutColons)))) } "meta" => token:
conditional_block
macro_parser.rs
count_names(&ms[..]); let matches: Vec<_> = (0..match_idx_hi).map(|_| Vec::new()).collect(); Box::new(MatcherPos { stack: vec![], top_elts: TtSeq(ms), sep: sep, idx: 0, up: None, matches: matches, match_lo: 0, match_cur: 0, match_hi: match_idx_hi, sp_lo: lo }) } /// NamedMatch is a pattern-match result for a single token::MATCH_NONTERMINAL: /// so it is associated with a single ident in a parse, and all /// `MatchedNonterminal`s in the NamedMatch have the same nonterminal type /// (expr, item, etc). Each leaf in a single NamedMatch corresponds to a /// single token::MATCH_NONTERMINAL in the TokenTree that produced it. /// /// The in-memory structure of a particular NamedMatch represents the match /// that occurred when a particular subset of a matcher was applied to a /// particular token tree. /// /// The width of each MatchedSeq in the NamedMatch, and the identity of the /// `MatchedNonterminal`s, will depend on the token tree it was applied to: /// each MatchedSeq corresponds to a single TTSeq in the originating /// token tree. The depth of the NamedMatch structure will therefore depend /// only on the nesting depth of `ast::TTSeq`s in the originating /// token tree it was derived from. pub enum NamedMatch { MatchedSeq(Vec<Rc<NamedMatch>>, codemap::Span), MatchedNonterminal(Nonterminal) } pub fn nameize(p_s: &ParseSess, ms: &[TokenTree], res: &[Rc<NamedMatch>]) -> HashMap<Ident, Rc<NamedMatch>> { fn n_rec(p_s: &ParseSess, m: &TokenTree, res: &[Rc<NamedMatch>], ret_val: &mut HashMap<Ident, Rc<NamedMatch>>, idx: &mut usize) { match m { &TtSequence(_, ref seq) => { for next_m in &seq.tts { n_rec(p_s, next_m, res, ret_val, idx) } } &TtDelimited(_, ref delim) => { for next_m in &delim.tts { n_rec(p_s, next_m, res, ret_val, idx) } } &TtToken(sp, MatchNt(bind_name, _, _, _)) => { match ret_val.entry(bind_name) { Vacant(spot) => { spot.insert(res[*idx].clone()); *idx += 1; } Occupied(..) => { panic!(p_s.span_diagnostic .span_fatal(sp, &format!("duplicated bind name: {}", bind_name))) } } } &TtToken(_, SubstNt(..)) => panic!("Cannot fill in a NT"), &TtToken(_, _) => (), } } let mut ret_val = HashMap::new(); let mut idx = 0; for m in ms { n_rec(p_s, m, res, &mut ret_val, &mut idx) } ret_val } pub enum ParseResult<T> { Success(T), Failure(codemap::Span, String), Error(codemap::Span, String) } pub type NamedParseResult = ParseResult<HashMap<Ident, Rc<NamedMatch>>>; pub type PositionalParseResult = ParseResult<Vec<Rc<NamedMatch>>>; pub fn parse_or_else(sess: &ParseSess, cfg: ast::CrateConfig, rdr: TtReader, ms: Vec<TokenTree> ) -> HashMap<Ident, Rc<NamedMatch>> { match parse(sess, cfg, rdr, &ms[..]) { Success(m) => m, Failure(sp, str) => { panic!(sess.span_diagnostic.span_fatal(sp, &str[..])) } Error(sp, str) => { panic!(sess.span_diagnostic.span_fatal(sp, &str[..])) } } } /// Perform a token equality check, ignoring syntax context (that is, an /// unhygienic comparison) pub fn token_name_eq(t1 : &Token, t2 : &Token) -> bool { match (t1,t2) { (&token::Ident(id1,_),&token::Ident(id2,_)) | (&token::Lifetime(id1),&token::Lifetime(id2)) => id1.name == id2.name, _ => *t1 == *t2 } } pub fn parse(sess: &ParseSess, cfg: ast::CrateConfig, mut rdr: TtReader, ms: &[TokenTree]) -> NamedParseResult { let mut cur_eis = Vec::new(); cur_eis.push(initial_matcher_pos(Rc::new(ms.iter() .cloned() .collect()), None, rdr.peek().sp.lo)); loop { let mut bb_eis = Vec::new(); // black-box parsed by parser.rs let mut next_eis = Vec::new(); // or proceed normally let mut eof_eis = Vec::new(); let TokenAndSpan { tok, sp } = rdr.peek(); /* we append new items to this while we go */ loop { let mut ei = match cur_eis.pop() { None => break, /* for each Earley Item */ Some(ei) => ei, }; // When unzipped trees end, remove them while ei.idx >= ei.top_elts.len() { match ei.stack.pop() { Some(MatcherTtFrame { elts, idx }) => { ei.top_elts = elts; ei.idx = idx + 1; } None => break } } let idx = ei.idx; let len = ei.top_elts.len(); /* at end of sequence */ if idx >= len { // can't move out of `match`es, so: if ei.up.is_some() { // hack: a matcher sequence is repeating iff it has a // parent (the top level is just a container) // disregard separator, try to go up // (remove this condition to make trailing seps ok) if idx == len { // pop from the matcher position let mut new_pos = ei.up.clone().unwrap(); // update matches (the MBE "parse tree") by appending // each tree as a subtree. // I bet this is a perf problem: we're preemptively // doing a lot of array work that will get thrown away // most of the time. // Only touch the binders we have actually bound for idx in ei.match_lo..ei.match_hi { let sub = (ei.matches[idx]).clone(); (&mut new_pos.matches[idx]) .push(Rc::new(MatchedSeq(sub, mk_sp(ei.sp_lo, sp.hi)))); } new_pos.match_cur = ei.match_hi; new_pos.idx += 1; cur_eis.push(new_pos); } // can we go around again? // the *_t vars are workarounds for the lack of unary move match ei.sep { Some(ref t) if idx == len => { // we need a separator // i'm conflicted about whether this should be hygienic.... // though in this case, if the separators are never legal // idents, it shouldn't matter. if token_name_eq(&tok, t) { //pass the separator let mut ei_t = ei.clone(); // ei_t.match_cur = ei_t.match_lo; ei_t.idx += 1; next_eis.push(ei_t); } } _ => { // we don't need a separator let mut ei_t = ei; ei_t.match_cur = ei_t.match_lo; ei_t.idx = 0; cur_eis.push(ei_t); } } } else { eof_eis.push(ei); } } else { match ei.top_elts.get_tt(idx) { /* need to descend into sequence */ TtSequence(sp, seq) => { if seq.op == ast::ZeroOrMore { let mut new_ei = ei.clone(); new_ei.match_cur += seq.num_captures; new_ei.idx += 1; //we specifically matched zero repeats. for idx in ei.match_cur..ei.match_cur + seq.num_captures { (&mut new_ei.matches[idx]).push(Rc::new(MatchedSeq(vec![], sp))); } cur_eis.push(new_ei); } let matches: Vec<_> = (0..ei.matches.len()) .map(|_| Vec::new()).collect(); let ei_t = ei; cur_eis.push(Box::new(MatcherPos { stack: vec![], sep: seq.separator.clone(), idx: 0, matches: matches, match_lo: ei_t.match_cur, match_cur: ei_t.match_cur, match_hi: ei_t.match_cur + seq.num_captures, up: Some(ei_t), sp_lo: sp.lo, top_elts: Tt(TtSequence(sp, seq)), })); }
random_line_split
flo2_pid.py
#!/usr/bin/env python #-------------------------------------------------------------------------- # flo2_pid.py # Rick Kauffman a.k.a. Chewie # # Hewlett Packard Company Revision: 1.0 # ~~~~~~~~~ WookieWare ~~~~~~~~~~~~~ # Change history....09/03/2014 # # ##-------------------------------------------------------------------------- # Initial release - Pulls VARS from webform. # build a database of all dpids not in glarn # Calls glarn chooser deletes dpids # # #------Might not need this but please they are handy------------------------ # # Do the imports!!!! #----------------------If you dont have it use "apt-get install (name)" import sys import subprocess import cgi import cgitb; cgitb.enable() import hpsdnclient as hp import sqlite3 import requests from requests.auth import HTTPDigestAuth import xml.etree.ElementTree as xml # import pdb; pdb.set_trace() #------------------------------------------------------------------------- # Get the field VARS from the calling HTML form #-------------------------------------------------------------------------
imc_user = form.getvalue('imc_user') imc_passw = form.getvalue('imc_passw') pid_list = form.getvalue('list_o_pids') imc = form.getvalue('imc') if pid_list == None: print "Content-type:text/html\r\n\r\n" print "<!DOCTYPE html>" print "<html>" print "<head>" print "<title> Wookieware.com</title>" print "<link rel=\"stylesheet\" type\"text/css\" href=\"../../css/corex.css\"/>" print "<script src=\"http://ajax.googleapis.com/ajax/libs/jquery/1.7.2/jquery.min.js\"></script>" print "</head>" print "<body>" print "<h1> <img src=\"../../images/glarn.png\" width=\"50\" height=\"50\">glarn: The dpid database</h1>" print "<HR> " print "<h1> No items selected</h1>" print "<FORM method='post' ACTION=\"./pid_main.py\">" print "<h3> List is empty </h3>" print "<p> Click button below to go back to the system chooser</p>" print "<hr>" print "<input type=\"submit\" style=\"font-face: 'Arial'; font-size: larger; color: black; background-color: #0066FF; border: 3pt ridge lightgrey\" value=\" Main Menu\">" print "<input type=\"hidden\" name=\"server\" value=%s>" % (server) print "<input type=\"hidden\" name=\"user\" value=%s>" % (user) print "<input type=\"hidden\" name=\"passw\" value=%s>" % (passw) print "<input type=\"hidden\" name=\"imc_server\" value=%s>" % (imc_server) print "<input type=\"hidden\" name=\"imc_user\" value=%s>" % (imc_user) print "<input type=\"hidden\" name=\"imc_passw\" value=%s>" % (imc_passw) print "<input type=\"hidden\" name=\"imc\" value=%s>" % (imc) print "<p>For more information on how to use this application <a href=\"/faq.html\">User Guide</a></p>" print "<center><font face=\"Arial\" size=\"1\">SDN Solutions From WookieWare 2014</font></center>" print "</body>" print "</html>" #glarn.close() sys.exit() x = len(pid_list) # Keep track of how many items we need to process # Check to see if anything was chozen. If x is zero goto Nothing Selected page and exit j = 0 #Create authorization Token for the SDN controller auth = hp.XAuthToken(user=user,password=passw,server=server) api=hp.Api(controller=server,auth=auth) #-------------------------------------------------------------------------- # dpid factory: Break up dpis and match to vendor to determin MAC address #--------------------------------------------------------------------------- print "Content-type:text/html\r\n\r\n" print "<!DOCTYPE html>" print "<html>" print "<head>" print "<title> Wookieware.com</title>" print "<link rel=\"stylesheet\" type\"text/css\" href=\"../../css/corex.css\"/>" print "<script src=\"http://ajax.googleapis.com/ajax/libs/jquery/1.7.2/jquery.min.js\"></script>" print "</head>" print "<body>" print "<h1> <img src=\"../../images/glarn.png\" width=\"50\" height=\"50\">glarn: The dpid database</h1>" print "<HR> " print "<h3> Dpid flows display</h3>" print "<p>List of current flows by dpid" print "<FORM method='post' ACTION=\"./pid_main.py\">" # Delete records in database if x == 23:#only one entry (dpids are 23 chars long) try: flows = api.get_flows(pid_list) print "<h1>Flows for dpid %s:</h1>" % (pid_list) print "<table border=\"1\" cellpadding=\"3\" class=\"TFtable\">" print "<tr> <td> Host MAC address </td> <td> Destination MAC address </td> <td> Output Port </td> </tr>" for f in flows: eth_src = f.match.eth_src eth_dst = f.match.eth_dst action = f.actions.output print "<tr> <td> %s </td> <td> %s </td> <td> %s </td> </tr>" % (eth_src, eth_dst, action) print "</table>" except: print "<h1>Error getting dpid information %s</h1>" % (pid_list) elif j == 0: for i in pid_list: flows = api.get_flows(i) print "<h1>Flows for dpid %s:</h1>" % (i) print "<table border=\"1\" cellpadding=\"3\" class=\"TFtable\">" print "<tr> <td> Host MAC address </td> <td> Destination MAC address </td> <td> Output Port </td> </tr>" for f in flows: eth_src = f.match.eth_src eth_dst = f.match.eth_dst action = f.actions.output print "<tr> <td> %s </td> <td> %s </td> <td> %s </td> </tr>" % (eth_src, eth_dst, action) print "</table>" #-------------------------------------------------------------------------- # Finish manual or go home #--------------------------------------------------------------------------- #print "Content-type:text/html\r\n\r\n" #print "<!DOCTYPE html>" #print "<html>" #print "<head>" #print "<title> Wookieware.com</title>" #print "<link rel=\"stylesheet\" type\"text/css\" href=\"../../css/corex.css\"/>" #print "<script src=\"http://ajax.googleapis.com/ajax/libs/jquery/1.7.2/jquery.min.js\"></script>" #print "</head>" print "<HR>" print "<input type=\"submit\" style=\"font-face: 'Arial'; font-size: larger; color: black; background-color: #0066FF; border: 3pt ridge lightgrey\" value=\" Main Menu\">" print "<input type=\"hidden\" name=\"server\" value=%s>" % (server) print "<input type=\"hidden\" name=\"user\" value=%s>" % (user) print "<input type=\"hidden\" name=\"passw\" value=%s>" % (passw) print "<input type=\"hidden\" name=\"imc_server\" value=%s>" % (imc_server) print "<input type=\"hidden\" name=\"imc_user\" value=%s>" % (imc_user) print "<input type=\"hidden\" name=\"imc_passw\" value=%s>" % (imc_passw) print "<input type=\"hidden\" name=\"imc\" value=%s>" % (imc) print "</form>" print "<footer>" print "<p>For more information on how to use this application <a href=\"/faq.html\">User Guide</a></p>" print "<a href=\"/index.html\">BACK</a>" print "<center><font face=\"Arial\" size=\"1\">SDN Solutions From WookieWare 2014</font></center>" print "</footer>" print "</body>" print "</html>" sys.exit()
form = cgi.FieldStorage() server = form.getvalue('server') user = form.getvalue('user') passw = form.getvalue('passw') imc_server = form.getvalue('imc_server')
random_line_split
flo2_pid.py
#!/usr/bin/env python #-------------------------------------------------------------------------- # flo2_pid.py # Rick Kauffman a.k.a. Chewie # # Hewlett Packard Company Revision: 1.0 # ~~~~~~~~~ WookieWare ~~~~~~~~~~~~~ # Change history....09/03/2014 # # ##-------------------------------------------------------------------------- # Initial release - Pulls VARS from webform. # build a database of all dpids not in glarn # Calls glarn chooser deletes dpids # # #------Might not need this but please they are handy------------------------ # # Do the imports!!!! #----------------------If you dont have it use "apt-get install (name)" import sys import subprocess import cgi import cgitb; cgitb.enable() import hpsdnclient as hp import sqlite3 import requests from requests.auth import HTTPDigestAuth import xml.etree.ElementTree as xml # import pdb; pdb.set_trace() #------------------------------------------------------------------------- # Get the field VARS from the calling HTML form #------------------------------------------------------------------------- form = cgi.FieldStorage() server = form.getvalue('server') user = form.getvalue('user') passw = form.getvalue('passw') imc_server = form.getvalue('imc_server') imc_user = form.getvalue('imc_user') imc_passw = form.getvalue('imc_passw') pid_list = form.getvalue('list_o_pids') imc = form.getvalue('imc') if pid_list == None: print "Content-type:text/html\r\n\r\n" print "<!DOCTYPE html>" print "<html>" print "<head>" print "<title> Wookieware.com</title>" print "<link rel=\"stylesheet\" type\"text/css\" href=\"../../css/corex.css\"/>" print "<script src=\"http://ajax.googleapis.com/ajax/libs/jquery/1.7.2/jquery.min.js\"></script>" print "</head>" print "<body>" print "<h1> <img src=\"../../images/glarn.png\" width=\"50\" height=\"50\">glarn: The dpid database</h1>" print "<HR> " print "<h1> No items selected</h1>" print "<FORM method='post' ACTION=\"./pid_main.py\">" print "<h3> List is empty </h3>" print "<p> Click button below to go back to the system chooser</p>" print "<hr>" print "<input type=\"submit\" style=\"font-face: 'Arial'; font-size: larger; color: black; background-color: #0066FF; border: 3pt ridge lightgrey\" value=\" Main Menu\">" print "<input type=\"hidden\" name=\"server\" value=%s>" % (server) print "<input type=\"hidden\" name=\"user\" value=%s>" % (user) print "<input type=\"hidden\" name=\"passw\" value=%s>" % (passw) print "<input type=\"hidden\" name=\"imc_server\" value=%s>" % (imc_server) print "<input type=\"hidden\" name=\"imc_user\" value=%s>" % (imc_user) print "<input type=\"hidden\" name=\"imc_passw\" value=%s>" % (imc_passw) print "<input type=\"hidden\" name=\"imc\" value=%s>" % (imc) print "<p>For more information on how to use this application <a href=\"/faq.html\">User Guide</a></p>" print "<center><font face=\"Arial\" size=\"1\">SDN Solutions From WookieWare 2014</font></center>" print "</body>" print "</html>" #glarn.close() sys.exit() x = len(pid_list) # Keep track of how many items we need to process # Check to see if anything was chozen. If x is zero goto Nothing Selected page and exit j = 0 #Create authorization Token for the SDN controller auth = hp.XAuthToken(user=user,password=passw,server=server) api=hp.Api(controller=server,auth=auth) #-------------------------------------------------------------------------- # dpid factory: Break up dpis and match to vendor to determin MAC address #--------------------------------------------------------------------------- print "Content-type:text/html\r\n\r\n" print "<!DOCTYPE html>" print "<html>" print "<head>" print "<title> Wookieware.com</title>" print "<link rel=\"stylesheet\" type\"text/css\" href=\"../../css/corex.css\"/>" print "<script src=\"http://ajax.googleapis.com/ajax/libs/jquery/1.7.2/jquery.min.js\"></script>" print "</head>" print "<body>" print "<h1> <img src=\"../../images/glarn.png\" width=\"50\" height=\"50\">glarn: The dpid database</h1>" print "<HR> " print "<h3> Dpid flows display</h3>" print "<p>List of current flows by dpid" print "<FORM method='post' ACTION=\"./pid_main.py\">" # Delete records in database if x == 23:#only one entry (dpids are 23 chars long) try: flows = api.get_flows(pid_list) print "<h1>Flows for dpid %s:</h1>" % (pid_list) print "<table border=\"1\" cellpadding=\"3\" class=\"TFtable\">" print "<tr> <td> Host MAC address </td> <td> Destination MAC address </td> <td> Output Port </td> </tr>" for f in flows: eth_src = f.match.eth_src eth_dst = f.match.eth_dst action = f.actions.output print "<tr> <td> %s </td> <td> %s </td> <td> %s </td> </tr>" % (eth_src, eth_dst, action) print "</table>" except: print "<h1>Error getting dpid information %s</h1>" % (pid_list) elif j == 0: for i in pid_list: flows = api.get_flows(i) print "<h1>Flows for dpid %s:</h1>" % (i) print "<table border=\"1\" cellpadding=\"3\" class=\"TFtable\">" print "<tr> <td> Host MAC address </td> <td> Destination MAC address </td> <td> Output Port </td> </tr>" for f in flows:
print "</table>" #-------------------------------------------------------------------------- # Finish manual or go home #--------------------------------------------------------------------------- #print "Content-type:text/html\r\n\r\n" #print "<!DOCTYPE html>" #print "<html>" #print "<head>" #print "<title> Wookieware.com</title>" #print "<link rel=\"stylesheet\" type\"text/css\" href=\"../../css/corex.css\"/>" #print "<script src=\"http://ajax.googleapis.com/ajax/libs/jquery/1.7.2/jquery.min.js\"></script>" #print "</head>" print "<HR>" print "<input type=\"submit\" style=\"font-face: 'Arial'; font-size: larger; color: black; background-color: #0066FF; border: 3pt ridge lightgrey\" value=\" Main Menu\">" print "<input type=\"hidden\" name=\"server\" value=%s>" % (server) print "<input type=\"hidden\" name=\"user\" value=%s>" % (user) print "<input type=\"hidden\" name=\"passw\" value=%s>" % (passw) print "<input type=\"hidden\" name=\"imc_server\" value=%s>" % (imc_server) print "<input type=\"hidden\" name=\"imc_user\" value=%s>" % (imc_user) print "<input type=\"hidden\" name=\"imc_passw\" value=%s>" % (imc_passw) print "<input type=\"hidden\" name=\"imc\" value=%s>" % (imc) print "</form>" print "<footer>" print "<p>For more information on how to use this application <a href=\"/faq.html\">User Guide</a></p>" print "<a href=\"/index.html\">BACK</a>" print "<center><font face=\"Arial\" size=\"1\">SDN Solutions From WookieWare 2014</font></center>" print "</footer>" print "</body>" print "</html>" sys.exit()
eth_src = f.match.eth_src eth_dst = f.match.eth_dst action = f.actions.output print "<tr> <td> %s </td> <td> %s </td> <td> %s </td> </tr>" % (eth_src, eth_dst, action)
conditional_block
stamplay-js-sdk-tests.ts
Stamplay.init('sample'); const userFn = Stamplay.User(); const user = new userFn.Model(); const colTags = Stamplay.Cobject('tag'); const tags = new colTags.Collection(); // Signing up const registrationData = { email : 'user@provider.com', password: 'mySecret' }; user.signup(registrationData).then(() => { user.set('phoneNumber', '020 123 4567'); return user.save(); }).then(() => { const number = user.get('phoneNumber'); console.log(number); // number value is 020 123 4567 }); // Action const colFoo = Stamplay.Cobject('foo');
() => { // success callback }, (err: any) => { // error callback } );
const fooMod = new colFoo.Model(); fooMod.fetch(5) .then(() => fooMod.upVote()) .then(
random_line_split
commands.rs
use std::process::{Command, Child, ExitStatus, Output, Stdio}; use std::io::{Read, Write, Error as IOError}; use std::collections::BTreeSet; use branches::Branches; use error::Error; use options::Options; pub fn spawn_piped(args: &[&str]) -> Child { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::piped()) .stdout(Stdio::piped()) .stderr(Stdio::piped()) .spawn() .unwrap_or_else(|e| panic!("Error with child process: {}", e)) } pub fn run_command_with_no_output(args: &[&str]) { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::null()) .stdout(Stdio::null()) .stderr(Stdio::null()) .output() .unwrap_or_else(|e| panic!("Error with command: {}", e)); } pub fn output(args: &[&str]) -> String { let result = run_command(args); String::from_utf8(result.stdout).unwrap().trim().to_owned() } pub fn run_command(args: &[&str]) -> Output { run_command_with_result(args).unwrap_or_else(|e| panic!("Error with command: {}", e)) } pub fn run_command_with_result(args: &[&str]) -> Result<Output, IOError> { Command::new(&args[0]) .args(&args[1..]) .output() } pub fn run_command_with_status(args: &[&str]) -> Result<ExitStatus, IOError> { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::null()) .stdout(Stdio::null()) .stderr(Stdio::null()) .status() } pub fn validate_git_installation() -> Result<(), Error> { match Command::new("git").output() { Ok(_) => Ok(()), Err(_) => Err(Error::GitInstallationError), } } pub fn delete_local_branches(branches: &Branches) -> String { let xargs = spawn_piped(&["xargs", "git", "branch", "-D"]); { xargs.stdin.unwrap().write_all(branches.string.as_bytes()).unwrap() } let mut branches_delete_result = String::new(); xargs.stdout.unwrap().read_to_string(&mut branches_delete_result).unwrap(); branches_delete_result } pub fn delete_remote_branches(branches: &Branches, options: &Options) -> String { let xargs = spawn_piped(&["xargs", "git", "push", &options.remote, "--delete"]); let remote_branches_cmd = run_command(&["git", "branch", "-r"]); let s = String::from_utf8(remote_branches_cmd.stdout).unwrap(); let all_remote_branches = s.split('\n').collect::<Vec<&str>>(); let origin_for_trim = &format!("{}/", &options.remote)[..]; let b_tree_remotes = all_remote_branches.iter() .map(|b| b.trim().trim_start_matches(origin_for_trim).to_owned()) .collect::<BTreeSet<String>>();
} let intersection: Vec<_> = b_tree_remotes.intersection(&b_tree_branches).cloned().collect(); { xargs.stdin.unwrap().write_all(intersection.join("\n").as_bytes()).unwrap() } let mut stderr = String::new(); xargs.stderr.unwrap().read_to_string(&mut stderr).unwrap(); // Everything is written to stderr, so we need to process that let split = stderr.split('\n'); let vec: Vec<&str> = split.collect(); let mut output = vec![]; for s in vec { if s.contains("error: unable to delete '") { let branch = s.trim_start_matches("error: unable to delete '") .trim_end_matches("': remote ref does not exist"); output.push(branch.to_owned() + " was already deleted in the remote."); } else if s.contains(" - [deleted]") { output.push(s.to_owned()); } } output.join("\n") } #[cfg(test)] mod test { use super::spawn_piped; use std::io::{Read, Write}; #[test] fn test_spawn_piped() { let echo = spawn_piped(&["grep", "foo"]); { echo.stdin.unwrap().write_all("foo\nbar\nbaz".as_bytes()).unwrap() } let mut stdout = String::new(); echo.stdout.unwrap().read_to_string(&mut stdout).unwrap(); assert_eq!(stdout, "foo\n"); } }
let mut b_tree_branches = BTreeSet::new(); for branch in branches.vec.clone() { b_tree_branches.insert(branch);
random_line_split
commands.rs
use std::process::{Command, Child, ExitStatus, Output, Stdio}; use std::io::{Read, Write, Error as IOError}; use std::collections::BTreeSet; use branches::Branches; use error::Error; use options::Options; pub fn spawn_piped(args: &[&str]) -> Child { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::piped()) .stdout(Stdio::piped()) .stderr(Stdio::piped()) .spawn() .unwrap_or_else(|e| panic!("Error with child process: {}", e)) } pub fn run_command_with_no_output(args: &[&str]) { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::null()) .stdout(Stdio::null()) .stderr(Stdio::null()) .output() .unwrap_or_else(|e| panic!("Error with command: {}", e)); } pub fn output(args: &[&str]) -> String { let result = run_command(args); String::from_utf8(result.stdout).unwrap().trim().to_owned() } pub fn run_command(args: &[&str]) -> Output { run_command_with_result(args).unwrap_or_else(|e| panic!("Error with command: {}", e)) } pub fn run_command_with_result(args: &[&str]) -> Result<Output, IOError> { Command::new(&args[0]) .args(&args[1..]) .output() } pub fn run_command_with_status(args: &[&str]) -> Result<ExitStatus, IOError> { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::null()) .stdout(Stdio::null()) .stderr(Stdio::null()) .status() } pub fn validate_git_installation() -> Result<(), Error> { match Command::new("git").output() { Ok(_) => Ok(()), Err(_) => Err(Error::GitInstallationError), } } pub fn delete_local_branches(branches: &Branches) -> String { let xargs = spawn_piped(&["xargs", "git", "branch", "-D"]); { xargs.stdin.unwrap().write_all(branches.string.as_bytes()).unwrap() } let mut branches_delete_result = String::new(); xargs.stdout.unwrap().read_to_string(&mut branches_delete_result).unwrap(); branches_delete_result } pub fn delete_remote_branches(branches: &Branches, options: &Options) -> String { let xargs = spawn_piped(&["xargs", "git", "push", &options.remote, "--delete"]); let remote_branches_cmd = run_command(&["git", "branch", "-r"]); let s = String::from_utf8(remote_branches_cmd.stdout).unwrap(); let all_remote_branches = s.split('\n').collect::<Vec<&str>>(); let origin_for_trim = &format!("{}/", &options.remote)[..]; let b_tree_remotes = all_remote_branches.iter() .map(|b| b.trim().trim_start_matches(origin_for_trim).to_owned()) .collect::<BTreeSet<String>>(); let mut b_tree_branches = BTreeSet::new(); for branch in branches.vec.clone() { b_tree_branches.insert(branch); } let intersection: Vec<_> = b_tree_remotes.intersection(&b_tree_branches).cloned().collect(); { xargs.stdin.unwrap().write_all(intersection.join("\n").as_bytes()).unwrap() } let mut stderr = String::new(); xargs.stderr.unwrap().read_to_string(&mut stderr).unwrap(); // Everything is written to stderr, so we need to process that let split = stderr.split('\n'); let vec: Vec<&str> = split.collect(); let mut output = vec![]; for s in vec { if s.contains("error: unable to delete '") { let branch = s.trim_start_matches("error: unable to delete '") .trim_end_matches("': remote ref does not exist"); output.push(branch.to_owned() + " was already deleted in the remote."); } else if s.contains(" - [deleted]")
} output.join("\n") } #[cfg(test)] mod test { use super::spawn_piped; use std::io::{Read, Write}; #[test] fn test_spawn_piped() { let echo = spawn_piped(&["grep", "foo"]); { echo.stdin.unwrap().write_all("foo\nbar\nbaz".as_bytes()).unwrap() } let mut stdout = String::new(); echo.stdout.unwrap().read_to_string(&mut stdout).unwrap(); assert_eq!(stdout, "foo\n"); } }
{ output.push(s.to_owned()); }
conditional_block
commands.rs
use std::process::{Command, Child, ExitStatus, Output, Stdio}; use std::io::{Read, Write, Error as IOError}; use std::collections::BTreeSet; use branches::Branches; use error::Error; use options::Options; pub fn spawn_piped(args: &[&str]) -> Child { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::piped()) .stdout(Stdio::piped()) .stderr(Stdio::piped()) .spawn() .unwrap_or_else(|e| panic!("Error with child process: {}", e)) } pub fn run_command_with_no_output(args: &[&str]) { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::null()) .stdout(Stdio::null()) .stderr(Stdio::null()) .output() .unwrap_or_else(|e| panic!("Error with command: {}", e)); } pub fn output(args: &[&str]) -> String { let result = run_command(args); String::from_utf8(result.stdout).unwrap().trim().to_owned() } pub fn run_command(args: &[&str]) -> Output { run_command_with_result(args).unwrap_or_else(|e| panic!("Error with command: {}", e)) } pub fn run_command_with_result(args: &[&str]) -> Result<Output, IOError> { Command::new(&args[0]) .args(&args[1..]) .output() } pub fn run_command_with_status(args: &[&str]) -> Result<ExitStatus, IOError> { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::null()) .stdout(Stdio::null()) .stderr(Stdio::null()) .status() } pub fn validate_git_installation() -> Result<(), Error> { match Command::new("git").output() { Ok(_) => Ok(()), Err(_) => Err(Error::GitInstallationError), } } pub fn delete_local_branches(branches: &Branches) -> String { let xargs = spawn_piped(&["xargs", "git", "branch", "-D"]); { xargs.stdin.unwrap().write_all(branches.string.as_bytes()).unwrap() } let mut branches_delete_result = String::new(); xargs.stdout.unwrap().read_to_string(&mut branches_delete_result).unwrap(); branches_delete_result } pub fn delete_remote_branches(branches: &Branches, options: &Options) -> String { let xargs = spawn_piped(&["xargs", "git", "push", &options.remote, "--delete"]); let remote_branches_cmd = run_command(&["git", "branch", "-r"]); let s = String::from_utf8(remote_branches_cmd.stdout).unwrap(); let all_remote_branches = s.split('\n').collect::<Vec<&str>>(); let origin_for_trim = &format!("{}/", &options.remote)[..]; let b_tree_remotes = all_remote_branches.iter() .map(|b| b.trim().trim_start_matches(origin_for_trim).to_owned()) .collect::<BTreeSet<String>>(); let mut b_tree_branches = BTreeSet::new(); for branch in branches.vec.clone() { b_tree_branches.insert(branch); } let intersection: Vec<_> = b_tree_remotes.intersection(&b_tree_branches).cloned().collect(); { xargs.stdin.unwrap().write_all(intersection.join("\n").as_bytes()).unwrap() } let mut stderr = String::new(); xargs.stderr.unwrap().read_to_string(&mut stderr).unwrap(); // Everything is written to stderr, so we need to process that let split = stderr.split('\n'); let vec: Vec<&str> = split.collect(); let mut output = vec![]; for s in vec { if s.contains("error: unable to delete '") { let branch = s.trim_start_matches("error: unable to delete '") .trim_end_matches("': remote ref does not exist"); output.push(branch.to_owned() + " was already deleted in the remote."); } else if s.contains(" - [deleted]") { output.push(s.to_owned()); } } output.join("\n") } #[cfg(test)] mod test { use super::spawn_piped; use std::io::{Read, Write}; #[test] fn test_spawn_piped()
}
{ let echo = spawn_piped(&["grep", "foo"]); { echo.stdin.unwrap().write_all("foo\nbar\nbaz".as_bytes()).unwrap() } let mut stdout = String::new(); echo.stdout.unwrap().read_to_string(&mut stdout).unwrap(); assert_eq!(stdout, "foo\n"); }
identifier_body
commands.rs
use std::process::{Command, Child, ExitStatus, Output, Stdio}; use std::io::{Read, Write, Error as IOError}; use std::collections::BTreeSet; use branches::Branches; use error::Error; use options::Options; pub fn
(args: &[&str]) -> Child { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::piped()) .stdout(Stdio::piped()) .stderr(Stdio::piped()) .spawn() .unwrap_or_else(|e| panic!("Error with child process: {}", e)) } pub fn run_command_with_no_output(args: &[&str]) { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::null()) .stdout(Stdio::null()) .stderr(Stdio::null()) .output() .unwrap_or_else(|e| panic!("Error with command: {}", e)); } pub fn output(args: &[&str]) -> String { let result = run_command(args); String::from_utf8(result.stdout).unwrap().trim().to_owned() } pub fn run_command(args: &[&str]) -> Output { run_command_with_result(args).unwrap_or_else(|e| panic!("Error with command: {}", e)) } pub fn run_command_with_result(args: &[&str]) -> Result<Output, IOError> { Command::new(&args[0]) .args(&args[1..]) .output() } pub fn run_command_with_status(args: &[&str]) -> Result<ExitStatus, IOError> { Command::new(&args[0]) .args(&args[1..]) .stdin(Stdio::null()) .stdout(Stdio::null()) .stderr(Stdio::null()) .status() } pub fn validate_git_installation() -> Result<(), Error> { match Command::new("git").output() { Ok(_) => Ok(()), Err(_) => Err(Error::GitInstallationError), } } pub fn delete_local_branches(branches: &Branches) -> String { let xargs = spawn_piped(&["xargs", "git", "branch", "-D"]); { xargs.stdin.unwrap().write_all(branches.string.as_bytes()).unwrap() } let mut branches_delete_result = String::new(); xargs.stdout.unwrap().read_to_string(&mut branches_delete_result).unwrap(); branches_delete_result } pub fn delete_remote_branches(branches: &Branches, options: &Options) -> String { let xargs = spawn_piped(&["xargs", "git", "push", &options.remote, "--delete"]); let remote_branches_cmd = run_command(&["git", "branch", "-r"]); let s = String::from_utf8(remote_branches_cmd.stdout).unwrap(); let all_remote_branches = s.split('\n').collect::<Vec<&str>>(); let origin_for_trim = &format!("{}/", &options.remote)[..]; let b_tree_remotes = all_remote_branches.iter() .map(|b| b.trim().trim_start_matches(origin_for_trim).to_owned()) .collect::<BTreeSet<String>>(); let mut b_tree_branches = BTreeSet::new(); for branch in branches.vec.clone() { b_tree_branches.insert(branch); } let intersection: Vec<_> = b_tree_remotes.intersection(&b_tree_branches).cloned().collect(); { xargs.stdin.unwrap().write_all(intersection.join("\n").as_bytes()).unwrap() } let mut stderr = String::new(); xargs.stderr.unwrap().read_to_string(&mut stderr).unwrap(); // Everything is written to stderr, so we need to process that let split = stderr.split('\n'); let vec: Vec<&str> = split.collect(); let mut output = vec![]; for s in vec { if s.contains("error: unable to delete '") { let branch = s.trim_start_matches("error: unable to delete '") .trim_end_matches("': remote ref does not exist"); output.push(branch.to_owned() + " was already deleted in the remote."); } else if s.contains(" - [deleted]") { output.push(s.to_owned()); } } output.join("\n") } #[cfg(test)] mod test { use super::spawn_piped; use std::io::{Read, Write}; #[test] fn test_spawn_piped() { let echo = spawn_piped(&["grep", "foo"]); { echo.stdin.unwrap().write_all("foo\nbar\nbaz".as_bytes()).unwrap() } let mut stdout = String::new(); echo.stdout.unwrap().read_to_string(&mut stdout).unwrap(); assert_eq!(stdout, "foo\n"); } }
spawn_piped
identifier_name
kanban-value-attribute-config.component.ts
/* * Lumeer: Modern Data Definition and Processing Platform * * Copyright (C) since 2017 Lumeer.io, s.r.o. and/or its affiliates. * * This program is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program. If not, see <http://www.gnu.org/licenses/>. */ import {ChangeDetectionStrategy, Component, EventEmitter, Input, Output} from '@angular/core'; import {SelectItemModel} from '../../../../../../shared/select/select-item/select-item.model'; import {DataAggregationType} from '../../../../../../shared/utils/data/data-aggregation'; import {I18n} from '@ngx-translate/i18n-polyfill'; import {KanbanAttribute, KanbanValueAttribute, KanbanValueType} from '../../../../../../core/store/kanbans/kanban'; @Component({ selector: 'kanban-value-attribute-config', templateUrl: './kanban-value-attribute-config.component.html', changeDetection: ChangeDetectionStrategy.OnPush, }) export class KanbanValueAttributeConfigComponent { @Input() public kanbanAttribute: KanbanValueAttribute; @Input() public availableAttributes: SelectItemModel[]; @Output() public attributeSelect = new EventEmitter<KanbanValueAttribute>(); @Output() public attributeChange = new EventEmitter<KanbanValueAttribute>(); @Output() public attributeRemove = new EventEmitter(); public readonly buttonClasses = 'flex-grow-1 text-truncate'; public readonly aggregationPlaceholder: string; public readonly aggregations = Object.values(DataAggregationType); public readonly valueTypes = Object.values(KanbanValueType); public readonly valueType = KanbanValueType; constructor(private i18n: I18n) { this.aggregationPlaceholder = i18n({id: 'aggregation', value: 'Aggregation'}); } public onAggregationSelect(aggregation: DataAggregationType) { const newAttribute = {...this.kanbanAttribute, aggregation}; this.attributeChange.emit(newAttribute); } public onAttributeSelected(attribute: KanbanAttribute) { const valueAttribute: KanbanValueAttribute = { ...attribute, aggregation: DataAggregationType.Sum, valueType: KanbanValueType.Default, }; this.attributeSelect.emit(valueAttribute); } public onAttributeRemoved() { this.attributeRemove.emit(); }
}
public onValueTypeSelected(valueType: KanbanValueType) { const valueAttribute: KanbanValueAttribute = {...this.kanbanAttribute, valueType}; this.attributeChange.emit(valueAttribute); }
random_line_split
kanban-value-attribute-config.component.ts
/* * Lumeer: Modern Data Definition and Processing Platform * * Copyright (C) since 2017 Lumeer.io, s.r.o. and/or its affiliates. * * This program is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program. If not, see <http://www.gnu.org/licenses/>. */ import {ChangeDetectionStrategy, Component, EventEmitter, Input, Output} from '@angular/core'; import {SelectItemModel} from '../../../../../../shared/select/select-item/select-item.model'; import {DataAggregationType} from '../../../../../../shared/utils/data/data-aggregation'; import {I18n} from '@ngx-translate/i18n-polyfill'; import {KanbanAttribute, KanbanValueAttribute, KanbanValueType} from '../../../../../../core/store/kanbans/kanban'; @Component({ selector: 'kanban-value-attribute-config', templateUrl: './kanban-value-attribute-config.component.html', changeDetection: ChangeDetectionStrategy.OnPush, }) export class KanbanValueAttributeConfigComponent { @Input() public kanbanAttribute: KanbanValueAttribute; @Input() public availableAttributes: SelectItemModel[]; @Output() public attributeSelect = new EventEmitter<KanbanValueAttribute>(); @Output() public attributeChange = new EventEmitter<KanbanValueAttribute>(); @Output() public attributeRemove = new EventEmitter(); public readonly buttonClasses = 'flex-grow-1 text-truncate'; public readonly aggregationPlaceholder: string; public readonly aggregations = Object.values(DataAggregationType); public readonly valueTypes = Object.values(KanbanValueType); public readonly valueType = KanbanValueType;
(private i18n: I18n) { this.aggregationPlaceholder = i18n({id: 'aggregation', value: 'Aggregation'}); } public onAggregationSelect(aggregation: DataAggregationType) { const newAttribute = {...this.kanbanAttribute, aggregation}; this.attributeChange.emit(newAttribute); } public onAttributeSelected(attribute: KanbanAttribute) { const valueAttribute: KanbanValueAttribute = { ...attribute, aggregation: DataAggregationType.Sum, valueType: KanbanValueType.Default, }; this.attributeSelect.emit(valueAttribute); } public onAttributeRemoved() { this.attributeRemove.emit(); } public onValueTypeSelected(valueType: KanbanValueType) { const valueAttribute: KanbanValueAttribute = {...this.kanbanAttribute, valueType}; this.attributeChange.emit(valueAttribute); } }
constructor
identifier_name
kanban-value-attribute-config.component.ts
/* * Lumeer: Modern Data Definition and Processing Platform * * Copyright (C) since 2017 Lumeer.io, s.r.o. and/or its affiliates. * * This program is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program. If not, see <http://www.gnu.org/licenses/>. */ import {ChangeDetectionStrategy, Component, EventEmitter, Input, Output} from '@angular/core'; import {SelectItemModel} from '../../../../../../shared/select/select-item/select-item.model'; import {DataAggregationType} from '../../../../../../shared/utils/data/data-aggregation'; import {I18n} from '@ngx-translate/i18n-polyfill'; import {KanbanAttribute, KanbanValueAttribute, KanbanValueType} from '../../../../../../core/store/kanbans/kanban'; @Component({ selector: 'kanban-value-attribute-config', templateUrl: './kanban-value-attribute-config.component.html', changeDetection: ChangeDetectionStrategy.OnPush, }) export class KanbanValueAttributeConfigComponent { @Input() public kanbanAttribute: KanbanValueAttribute; @Input() public availableAttributes: SelectItemModel[]; @Output() public attributeSelect = new EventEmitter<KanbanValueAttribute>(); @Output() public attributeChange = new EventEmitter<KanbanValueAttribute>(); @Output() public attributeRemove = new EventEmitter(); public readonly buttonClasses = 'flex-grow-1 text-truncate'; public readonly aggregationPlaceholder: string; public readonly aggregations = Object.values(DataAggregationType); public readonly valueTypes = Object.values(KanbanValueType); public readonly valueType = KanbanValueType; constructor(private i18n: I18n) { this.aggregationPlaceholder = i18n({id: 'aggregation', value: 'Aggregation'}); } public onAggregationSelect(aggregation: DataAggregationType) { const newAttribute = {...this.kanbanAttribute, aggregation}; this.attributeChange.emit(newAttribute); } public onAttributeSelected(attribute: KanbanAttribute) { const valueAttribute: KanbanValueAttribute = { ...attribute, aggregation: DataAggregationType.Sum, valueType: KanbanValueType.Default, }; this.attributeSelect.emit(valueAttribute); } public onAttributeRemoved() { this.attributeRemove.emit(); } public onValueTypeSelected(valueType: KanbanValueType)
}
{ const valueAttribute: KanbanValueAttribute = {...this.kanbanAttribute, valueType}; this.attributeChange.emit(valueAttribute); }
identifier_body
javax.swing.plaf.multi.MultiLabelUI.d.ts
declare namespace javax { namespace swing { namespace plaf { namespace multi { class
extends javax.swing.plaf.LabelUI { protected uis: java.util.Vector<javax.swing.plaf.ComponentUI> public constructor() public getUIs(): javax.swing.plaf.ComponentUI[] public contains(arg0: javax.swing.JComponent, arg1: number | java.lang.Integer, arg2: number | java.lang.Integer): boolean public update(arg0: java.awt.Graphics, arg1: javax.swing.JComponent): void public static createUI(arg0: javax.swing.JComponent): javax.swing.plaf.ComponentUI public installUI(arg0: javax.swing.JComponent): void public uninstallUI(arg0: javax.swing.JComponent): void public paint(arg0: java.awt.Graphics, arg1: javax.swing.JComponent): void public getPreferredSize(arg0: javax.swing.JComponent): java.awt.Dimension public getMinimumSize(arg0: javax.swing.JComponent): java.awt.Dimension public getMaximumSize(arg0: javax.swing.JComponent): java.awt.Dimension public getAccessibleChildrenCount(arg0: javax.swing.JComponent): number public getAccessibleChild(arg0: javax.swing.JComponent, arg1: number | java.lang.Integer): javax.accessibility.Accessible } } } } }
MultiLabelUI
identifier_name
javax.swing.plaf.multi.MultiLabelUI.d.ts
declare namespace javax { namespace swing { namespace plaf { namespace multi { class MultiLabelUI extends javax.swing.plaf.LabelUI { protected uis: java.util.Vector<javax.swing.plaf.ComponentUI> public constructor() public getUIs(): javax.swing.plaf.ComponentUI[] public contains(arg0: javax.swing.JComponent, arg1: number | java.lang.Integer, arg2: number | java.lang.Integer): boolean public update(arg0: java.awt.Graphics, arg1: javax.swing.JComponent): void public static createUI(arg0: javax.swing.JComponent): javax.swing.plaf.ComponentUI public installUI(arg0: javax.swing.JComponent): void public uninstallUI(arg0: javax.swing.JComponent): void public paint(arg0: java.awt.Graphics, arg1: javax.swing.JComponent): void public getPreferredSize(arg0: javax.swing.JComponent): java.awt.Dimension public getMinimumSize(arg0: javax.swing.JComponent): java.awt.Dimension public getMaximumSize(arg0: javax.swing.JComponent): java.awt.Dimension public getAccessibleChildrenCount(arg0: javax.swing.JComponent): number
} }
public getAccessibleChild(arg0: javax.swing.JComponent, arg1: number | java.lang.Integer): javax.accessibility.Accessible } } }
random_line_split
snapshot.js
'use strict'; var fs = require('fs'), _ = require('lodash'), colors = require('cli-color'), utils = require('./utils.js'); function doSnapShot(roadMapPath, urlPrefix) { var target, payload, response, url; var roadMap = utils.getRoadMapFromPath(roadMapPath); console.log('Processing road map file "' + roadMapPath + '":'); var bootInfo = utils.getBootInfo(); var souvenirPath = utils.getSouvenirPathForRoadMapPath(roadMapPath); var fixTestcases = utils.getTestcases(); if (_.isEmpty(fixTestcases)) { utils.mkEmptyDirSync(souvenirPath); } var no = 0, skipped = 0, failed = 0, fail = [], bad = []; for (target in roadMap) { no++; if (!_.isEmpty(fixTestcases) && !_.contains(fixTestcases, no)) { skipped++; continue; } console.log(''); console.log( 'Request #' + no + ' =================='); payload = roadMap[target]; url = urlPrefix + target; response = utils.getHttpResponse(url, payload, bootInfo.getHeaders()); if (_.isNull(response))
var targetFilePath = utils.getSouvenirPathForTarget(souvenirPath, target); fs.writeFileSync(targetFilePath, JSON.stringify(response)); let code = response.statusCode; if ((code >= 400) && (code < 500)) { code = colors.magenta(code); bad.push(no); } else if (code >= 500) { code = colors.red(code); failed++; fail.push(no); } else { code = colors.green(code); } console.log(' --> ' + code + ': Stored ' + response.body.length + ' bytes as souvenir'); } console.log(''); console.log('=========================='); let status = []; let succ = no - skipped - failed - bad.length; if (skipped > 0) { status.push(colors.blue(skipped + ' skipped')); } if (bad.length > 0) { status.push(colors.magenta(bad.length + ' bad')); } if (failed > 0) { status.push(colors.red(failed + ' failed')) } if (succ > 0) { status.push(colors.green(succ + ' successful')) } console.log(' Summary: ' + no + ' requests found [ ' + status.join(', ') + ' ]'); if (fail.length > 0) { console.log(colors.red(' Failed requests: ' + fail.join(','))); } if (bad.length > 0) { console.log(colors.magenta(' Bad requests: ' + bad.join(','))); } } var args = utils.getArguments(); if (args.hasOwnProperty('args') || args.length === 1) { var roadMapPath = utils.getRoadMapPath(); var baseUrl = utils.getBaseUrlFromArguments(); doSnapShot(roadMapPath, baseUrl); } else { args.printHelp(); }
{ failed++; fail.push(no); console.log('ERROR! Request timed out!'); continue; }
conditional_block
snapshot.js
'use strict'; var fs = require('fs'), _ = require('lodash'), colors = require('cli-color'), utils = require('./utils.js'); function
(roadMapPath, urlPrefix) { var target, payload, response, url; var roadMap = utils.getRoadMapFromPath(roadMapPath); console.log('Processing road map file "' + roadMapPath + '":'); var bootInfo = utils.getBootInfo(); var souvenirPath = utils.getSouvenirPathForRoadMapPath(roadMapPath); var fixTestcases = utils.getTestcases(); if (_.isEmpty(fixTestcases)) { utils.mkEmptyDirSync(souvenirPath); } var no = 0, skipped = 0, failed = 0, fail = [], bad = []; for (target in roadMap) { no++; if (!_.isEmpty(fixTestcases) && !_.contains(fixTestcases, no)) { skipped++; continue; } console.log(''); console.log( 'Request #' + no + ' =================='); payload = roadMap[target]; url = urlPrefix + target; response = utils.getHttpResponse(url, payload, bootInfo.getHeaders()); if (_.isNull(response)) { failed++; fail.push(no); console.log('ERROR! Request timed out!'); continue; } var targetFilePath = utils.getSouvenirPathForTarget(souvenirPath, target); fs.writeFileSync(targetFilePath, JSON.stringify(response)); let code = response.statusCode; if ((code >= 400) && (code < 500)) { code = colors.magenta(code); bad.push(no); } else if (code >= 500) { code = colors.red(code); failed++; fail.push(no); } else { code = colors.green(code); } console.log(' --> ' + code + ': Stored ' + response.body.length + ' bytes as souvenir'); } console.log(''); console.log('=========================='); let status = []; let succ = no - skipped - failed - bad.length; if (skipped > 0) { status.push(colors.blue(skipped + ' skipped')); } if (bad.length > 0) { status.push(colors.magenta(bad.length + ' bad')); } if (failed > 0) { status.push(colors.red(failed + ' failed')) } if (succ > 0) { status.push(colors.green(succ + ' successful')) } console.log(' Summary: ' + no + ' requests found [ ' + status.join(', ') + ' ]'); if (fail.length > 0) { console.log(colors.red(' Failed requests: ' + fail.join(','))); } if (bad.length > 0) { console.log(colors.magenta(' Bad requests: ' + bad.join(','))); } } var args = utils.getArguments(); if (args.hasOwnProperty('args') || args.length === 1) { var roadMapPath = utils.getRoadMapPath(); var baseUrl = utils.getBaseUrlFromArguments(); doSnapShot(roadMapPath, baseUrl); } else { args.printHelp(); }
doSnapShot
identifier_name
snapshot.js
'use strict'; var fs = require('fs'), _ = require('lodash'), colors = require('cli-color'), utils = require('./utils.js'); function doSnapShot(roadMapPath, urlPrefix)
fail = [], bad = []; for (target in roadMap) { no++; if (!_.isEmpty(fixTestcases) && !_.contains(fixTestcases, no)) { skipped++; continue; } console.log(''); console.log( 'Request #' + no + ' =================='); payload = roadMap[target]; url = urlPrefix + target; response = utils.getHttpResponse(url, payload, bootInfo.getHeaders()); if (_.isNull(response)) { failed++; fail.push(no); console.log('ERROR! Request timed out!'); continue; } var targetFilePath = utils.getSouvenirPathForTarget(souvenirPath, target); fs.writeFileSync(targetFilePath, JSON.stringify(response)); let code = response.statusCode; if ((code >= 400) && (code < 500)) { code = colors.magenta(code); bad.push(no); } else if (code >= 500) { code = colors.red(code); failed++; fail.push(no); } else { code = colors.green(code); } console.log(' --> ' + code + ': Stored ' + response.body.length + ' bytes as souvenir'); } console.log(''); console.log('=========================='); let status = []; let succ = no - skipped - failed - bad.length; if (skipped > 0) { status.push(colors.blue(skipped + ' skipped')); } if (bad.length > 0) { status.push(colors.magenta(bad.length + ' bad')); } if (failed > 0) { status.push(colors.red(failed + ' failed')) } if (succ > 0) { status.push(colors.green(succ + ' successful')) } console.log(' Summary: ' + no + ' requests found [ ' + status.join(', ') + ' ]'); if (fail.length > 0) { console.log(colors.red(' Failed requests: ' + fail.join(','))); } if (bad.length > 0) { console.log(colors.magenta(' Bad requests: ' + bad.join(','))); } } var args = utils.getArguments(); if (args.hasOwnProperty('args') || args.length === 1) { var roadMapPath = utils.getRoadMapPath(); var baseUrl = utils.getBaseUrlFromArguments(); doSnapShot(roadMapPath, baseUrl); } else { args.printHelp(); }
{ var target, payload, response, url; var roadMap = utils.getRoadMapFromPath(roadMapPath); console.log('Processing road map file "' + roadMapPath + '":'); var bootInfo = utils.getBootInfo(); var souvenirPath = utils.getSouvenirPathForRoadMapPath(roadMapPath); var fixTestcases = utils.getTestcases(); if (_.isEmpty(fixTestcases)) { utils.mkEmptyDirSync(souvenirPath); } var no = 0, skipped = 0, failed = 0,
identifier_body
snapshot.js
'use strict'; var fs = require('fs'), _ = require('lodash'), colors = require('cli-color'), utils = require('./utils.js'); function doSnapShot(roadMapPath, urlPrefix) { var target, payload, response, url; var roadMap = utils.getRoadMapFromPath(roadMapPath); console.log('Processing road map file "' + roadMapPath + '":'); var bootInfo = utils.getBootInfo(); var souvenirPath = utils.getSouvenirPathForRoadMapPath(roadMapPath); var fixTestcases = utils.getTestcases(); if (_.isEmpty(fixTestcases)) { utils.mkEmptyDirSync(souvenirPath); } var no = 0, skipped = 0, failed = 0, fail = [], bad = []; for (target in roadMap) { no++; if (!_.isEmpty(fixTestcases) && !_.contains(fixTestcases, no)) { skipped++; continue; } console.log(''); console.log( 'Request #' + no + ' =================='); payload = roadMap[target]; url = urlPrefix + target; response = utils.getHttpResponse(url, payload, bootInfo.getHeaders()); if (_.isNull(response)) { failed++; fail.push(no); console.log('ERROR! Request timed out!'); continue; } var targetFilePath = utils.getSouvenirPathForTarget(souvenirPath, target); fs.writeFileSync(targetFilePath, JSON.stringify(response)); let code = response.statusCode; if ((code >= 400) && (code < 500)) { code = colors.magenta(code); bad.push(no); } else if (code >= 500) { code = colors.red(code); failed++; fail.push(no); } else { code = colors.green(code); } console.log(' --> ' + code + ': Stored ' + response.body.length + ' bytes as souvenir'); } console.log(''); console.log('=========================='); let status = []; let succ = no - skipped - failed - bad.length; if (skipped > 0) { status.push(colors.blue(skipped + ' skipped')); } if (bad.length > 0) { status.push(colors.magenta(bad.length + ' bad')); } if (failed > 0) { status.push(colors.red(failed + ' failed')) } if (succ > 0) { status.push(colors.green(succ + ' successful')) } console.log(' Summary: ' + no + ' requests found [ ' + status.join(', ') + ' ]'); if (fail.length > 0) { console.log(colors.red(' Failed requests: ' + fail.join(','))); } if (bad.length > 0) {
console.log(colors.magenta(' Bad requests: ' + bad.join(','))); } } var args = utils.getArguments(); if (args.hasOwnProperty('args') || args.length === 1) { var roadMapPath = utils.getRoadMapPath(); var baseUrl = utils.getBaseUrlFromArguments(); doSnapShot(roadMapPath, baseUrl); } else { args.printHelp(); }
random_line_split
Cleaner.ts
import { Arr } from '@ephox/katamari'; type Task = () => void; export interface Cleaner { readonly add: (task: Task) => void; readonly run: () => void; readonly wrap: <T extends any[], U>(fn: (...a: T) => U) => (...args: T) => U; } export const Cleaner = (): Cleaner => { let tasks: Task[] = []; const add = (task: Task) => { tasks.push(task); }; const run = () => { Arr.each(tasks, (task) => { try { task(); } catch (e) { // eslint-disable-next-line no-console console.log(e); } }); tasks = []; }; const wrap = <T extends any[], U> (fn: (...a: T) => U) => (...args: T): U => { run(); return fn.apply(null, args); }; return { add,
wrap }; };
run,
random_line_split
JsxProcessor.ts
any) => { if (!(value instanceof Array)) { throw new Error('Not an array: ' + toValueString(value)); } return value; } }); const MARKUP: {[el: string]: AttrConverters} = Object.freeze({ br: {}, b: COMMON_ATTR, span: COMMON_ATTR, big: COMMON_ATTR, i: COMMON_ATTR, small: COMMON_ATTR, strong: COMMON_ATTR, ins: COMMON_ATTR, del: COMMON_ATTR, a: Object.assign({ href: (value: any) => { if (typeof value !== 'string') { throw new Error('Not a string: ' + toValueString(value)); } return value; } }, COMMON_ATTR) }); export function createJsxProcessor() { return new JsxProcessor(); } export default class JsxProcessor { createElement(Type: ElementFn | string, attributes?: any, ...children: any[]) { if (!(Type instanceof Function) && typeof Type !== 'string') { throw new Error(`JSX: Unsupported type ${toValueString(Type)}`); } const typeName = Type instanceof Function ? Type.name : Type; if (attributes?.children && children && children.length) { throw new Error(`JSX: Children for type ${typeName} given twice.`); } // Children may be part of attributes or given as varargs or both. // For JSX factories/functional components they should always be part of attributes const rawChildren = children.length ? children : attributes?.children || []; const {finalChildren, additionalAttributes} = parseChildren(rawChildren, Type); const finalAttributes = {...attributes}; joinAttributes(finalAttributes, additionalAttributes, Type); if (finalChildren) { finalAttributes.children = finalChildren; } if (typeof Type === 'string') { return this.createIntrinsicElement(Type, finalAttributes); } else if (Type.prototype && Type.prototype[JSX.jsxFactory]) { return this.createCustomComponent(Type, finalAttributes); } else { return this.createFunctionalComponent(Type, finalAttributes); } } createCustomComponent(Type: ElementFn, attributes: any) { return Type.prototype[JSX.jsxFactory].call(this, Type, attributes); } createFunctionalComponent(Type: ElementFn, attributes: any) { try { const result = Type.call(this, attributes); Type[symbols.jsxType] = true; if (result instanceof Object) { result[symbols.jsxType] = Type; } return result; } catch (ex) { throw new Error(`JSX: "${ex.message}" ${getCurrentLine(ex)}`); } } createIntrinsicElement(el: string, attributes: any) { if (el in MARKUP) { const encoded: any = {}; Object.keys(attributes || {}).forEach(attribute => { const encoder = MARKUP[el][attribute]; if (!encoder) { if (attribute === 'children') { throw new Error(`Element "${el}" can not have children`);
} else { throw new Error(`Element "${el}" does not support attribute "${attribute}"`); } } try { encoded[attribute] = encoder(attributes[attribute]); } catch(ex) { throw new Error(`Element "${el}" attribute "${attribute}" can not bet set: ${ex.message}`); } }); const text = joinTextContent(encoded.children, true); const tagOpen = [el].concat(Object.keys(encoded || {}).filter(attr => attr !== 'children').map( attribute => `${attribute}='${encoded[attribute]}'` )).join(' '); if (text) { return `<${tagOpen}>${text}</${el}>`; } return `<${tagOpen}/>`; } throw new Error(`JSX: Unsupported type ${el}`); } createNativeObject(Type: any, attributes: any) { if (attributes && 'children' in attributes) { throw new Error(`JSX: ${Type.name} can not have children`); } const {data, ...properties} = attributesWithoutListener(attributes || {}); const result = new Type(properties); registerListenerAttributes(result, attributes); if (data) { result.data = data; } return result; } getChildren(attributes: any) { if (!attributes || !('children' in attributes)) { return null; } return flattenChildren(attributes.children); } withoutChildren(attributes: any) { return omit(attributes, ['children']); } withContentText(attributes: any, content: any[], property: string, markupEnabled: boolean) { if (attributes && attributes[property] && content && content.length) { throw new Error(`JSX: ${property} given twice`); } const text = attributes && attributes[property] ? attributes[property].toString() : joinTextContent(content || [], markupEnabled); return Object.assign(attributes || {}, text ? {[property]: text} : {}); } withContentChildren(attributes: any, content: any[], property: string) { if (attributes && attributes[property] && content && content.length) { throw new Error(`JSX: ${property} given twice`); } const children = attributes && attributes[property] ? attributes[property] : (content || []); return Object.assign(attributes || {}, children ? {[property]: children} : {}); } withShorthands( attributes: object, shorthandsMapping: {[attr: string]: string}, merge: ((value1: any, value2: string) => any) ): object { const shorthandsKeys = Object.keys(shorthandsMapping); const shorthands = shorthandsKeys.filter(value => value in attributes); if (!shorthands.length) { return attributes; } const attrCopy: any = omit(attributes, shorthandsKeys); shorthands.forEach(shorthand => { const prop = shorthandsMapping[shorthand]; if (prop in attrCopy) { attrCopy[prop] = merge(attrCopy[prop], shorthand); } else { attrCopy[prop] = shorthand; } }); return attrCopy; } makeFactories(dic: {[key: string]: ElementFn}) { const result: {[key: string]: ElementFn} = {}; Object.keys(dic).forEach(key => { result[key] = this.makeFactory(dic[key]) as ElementFn; }); return result; } makeFactory(constructor: ElementFn): ElementFn { if (arguments.length !== 1) { throw new Error(`Expected exactly one argument, got ${arguments.length}`); } checkType(constructor, Function, 'first parameter'); if (!constructor.prototype || !constructor.prototype[JSX.jsxFactory]) { throw new Error(`Function ${constructor.name} is not a valid constructor`); } if (constructor[symbols.originalComponent]) { return this.makeFactory(constructor[symbols.originalComponent] as ElementFn); } return createFactoryProxy(this, constructor); } } function createFactoryProxy(processor: JsxProcessor, constructor: ElementFn): ElementFn { const handler: ProxyHandler<ElementFn> = { apply(target, _thisArg, args) { const [attributes, functionalComponent] = args; if (args.length > 1) { if (!(functionalComponent instanceof Function)) { throw new TypeError('Second parameter must be a function'); } if (functionalComponent.prototype && functionalComponent.prototype[JSX.jsxFactory]) { throw new TypeError('Second parameter must be a factory'); } } const result = processor.createElement(proxy, attributes); if (args.length > 1 && result instanceof Object) { functionalComponent[JSX.jsxType] = true; result[JSX.jsxType] = functionalComponent; } return result; }, get(target, property, receiver) { if (receiver === proxy) { if (property === symbols.originalComponent) { return constructor; } if (property === symbols.proxyHandler) { return handler; } } return Reflect.get(target, property, receiver); } }; /** @type {Factory} */ const proxy = new Proxy(constructor, handler); return proxy; } /** * Converts any value to a flat array. */ export function flattenChildren(children: unknown) { if (children instanceof Array) { let result: any[] = []; for (const child of children) { if (child && child.toArray) { result = result.concat(flattenChildren(child.toArray())); } else if (child instanceof Array) { result = result.concat(flattenChildren(child)); } else { result.push(child); } } return result; } return [children]; } export function joinTextContent(textArray: string[], markupEnabled: boolean) { if (!textArray) { return null; } if (markupEnabled) { return textArray .map(str => str + '') .join('') .replace(/\s+/g, ' ') .replace(/\s*<br\s*\/>\s*/g, '<br/>'); } return textArray.join(''); } export const JSX = { processor: null as JsxProcessor | null, jsxFactory: symbols.jsxFactory, jsxType: symbols.jsxType, install(jsxProcessor: JsxProcessor) { this.processor = jsxProcessor; }, createElement() { return this.processor!.createElement.apply(this
random_line_split
JsxProcessor.ts
any) => { if (!(value instanceof Array)) { throw new Error('Not an array: ' + toValueString(value)); } return value; } }); const MARKUP: {[el: string]: AttrConverters} = Object.freeze({ br: {}, b: COMMON_ATTR, span: COMMON_ATTR, big: COMMON_ATTR, i: COMMON_ATTR, small: COMMON_ATTR, strong: COMMON_ATTR, ins: COMMON_ATTR, del: COMMON_ATTR, a: Object.assign({ href: (value: any) => { if (typeof value !== 'string') { throw new Error('Not a string: ' + toValueString(value)); } return value; } }, COMMON_ATTR) }); export function createJsxProcessor()
export default class JsxProcessor { createElement(Type: ElementFn | string, attributes?: any, ...children: any[]) { if (!(Type instanceof Function) && typeof Type !== 'string') { throw new Error(`JSX: Unsupported type ${toValueString(Type)}`); } const typeName = Type instanceof Function ? Type.name : Type; if (attributes?.children && children && children.length) { throw new Error(`JSX: Children for type ${typeName} given twice.`); } // Children may be part of attributes or given as varargs or both. // For JSX factories/functional components they should always be part of attributes const rawChildren = children.length ? children : attributes?.children || []; const {finalChildren, additionalAttributes} = parseChildren(rawChildren, Type); const finalAttributes = {...attributes}; joinAttributes(finalAttributes, additionalAttributes, Type); if (finalChildren) { finalAttributes.children = finalChildren; } if (typeof Type === 'string') { return this.createIntrinsicElement(Type, finalAttributes); } else if (Type.prototype && Type.prototype[JSX.jsxFactory]) { return this.createCustomComponent(Type, finalAttributes); } else { return this.createFunctionalComponent(Type, finalAttributes); } } createCustomComponent(Type: ElementFn, attributes: any) { return Type.prototype[JSX.jsxFactory].call(this, Type, attributes); } createFunctionalComponent(Type: ElementFn, attributes: any) { try { const result = Type.call(this, attributes); Type[symbols.jsxType] = true; if (result instanceof Object) { result[symbols.jsxType] = Type; } return result; } catch (ex) { throw new Error(`JSX: "${ex.message}" ${getCurrentLine(ex)}`); } } createIntrinsicElement(el: string, attributes: any) { if (el in MARKUP) { const encoded: any = {}; Object.keys(attributes || {}).forEach(attribute => { const encoder = MARKUP[el][attribute]; if (!encoder) { if (attribute === 'children') { throw new Error(`Element "${el}" can not have children`); } else { throw new Error(`Element "${el}" does not support attribute "${attribute}"`); } } try { encoded[attribute] = encoder(attributes[attribute]); } catch(ex) { throw new Error(`Element "${el}" attribute "${attribute}" can not bet set: ${ex.message}`); } }); const text = joinTextContent(encoded.children, true); const tagOpen = [el].concat(Object.keys(encoded || {}).filter(attr => attr !== 'children').map( attribute => `${attribute}='${encoded[attribute]}'` )).join(' '); if (text) { return `<${tagOpen}>${text}</${el}>`; } return `<${tagOpen}/>`; } throw new Error(`JSX: Unsupported type ${el}`); } createNativeObject(Type: any, attributes: any) { if (attributes && 'children' in attributes) { throw new Error(`JSX: ${Type.name} can not have children`); } const {data, ...properties} = attributesWithoutListener(attributes || {}); const result = new Type(properties); registerListenerAttributes(result, attributes); if (data) { result.data = data; } return result; } getChildren(attributes: any) { if (!attributes || !('children' in attributes)) { return null; } return flattenChildren(attributes.children); } withoutChildren(attributes: any) { return omit(attributes, ['children']); } withContentText(attributes: any, content: any[], property: string, markupEnabled: boolean) { if (attributes && attributes[property] && content && content.length) { throw new Error(`JSX: ${property} given twice`); } const text = attributes && attributes[property] ? attributes[property].toString() : joinTextContent(content || [], markupEnabled); return Object.assign(attributes || {}, text ? {[property]: text} : {}); } withContentChildren(attributes: any, content: any[], property: string) { if (attributes && attributes[property] && content && content.length) { throw new Error(`JSX: ${property} given twice`); } const children = attributes && attributes[property] ? attributes[property] : (content || []); return Object.assign(attributes || {}, children ? {[property]: children} : {}); } withShorthands( attributes: object, shorthandsMapping: {[attr: string]: string}, merge: ((value1: any, value2: string) => any) ): object { const shorthandsKeys = Object.keys(shorthandsMapping); const shorthands = shorthandsKeys.filter(value => value in attributes); if (!shorthands.length) { return attributes; } const attrCopy: any = omit(attributes, shorthandsKeys); shorthands.forEach(shorthand => { const prop = shorthandsMapping[shorthand]; if (prop in attrCopy) { attrCopy[prop] = merge(attrCopy[prop], shorthand); } else { attrCopy[prop] = shorthand; } }); return attrCopy; } makeFactories(dic: {[key: string]: ElementFn}) { const result: {[key: string]: ElementFn} = {}; Object.keys(dic).forEach(key => { result[key] = this.makeFactory(dic[key]) as ElementFn; }); return result; } makeFactory(constructor: ElementFn): ElementFn { if (arguments.length !== 1) { throw new Error(`Expected exactly one argument, got ${arguments.length}`); } checkType(constructor, Function, 'first parameter'); if (!constructor.prototype || !constructor.prototype[JSX.jsxFactory]) { throw new Error(`Function ${constructor.name} is not a valid constructor`); } if (constructor[symbols.originalComponent]) { return this.makeFactory(constructor[symbols.originalComponent] as ElementFn); } return createFactoryProxy(this, constructor); } } function createFactoryProxy(processor: JsxProcessor, constructor: ElementFn): ElementFn { const handler: ProxyHandler<ElementFn> = { apply(target, _thisArg, args) { const [attributes, functionalComponent] = args; if (args.length > 1) { if (!(functionalComponent instanceof Function)) { throw new TypeError('Second parameter must be a function'); } if (functionalComponent.prototype && functionalComponent.prototype[JSX.jsxFactory]) { throw new TypeError('Second parameter must be a factory'); } } const result = processor.createElement(proxy, attributes); if (args.length > 1 && result instanceof Object) { functionalComponent[JSX.jsxType] = true; result[JSX.jsxType] = functionalComponent; } return result; }, get(target, property, receiver) { if (receiver === proxy) { if (property === symbols.originalComponent) { return constructor; } if (property === symbols.proxyHandler) { return handler; } } return Reflect.get(target, property, receiver); } }; /** @type {Factory} */ const proxy = new Proxy(constructor, handler); return proxy; } /** * Converts any value to a flat array. */ export function flattenChildren(children: unknown) { if (children instanceof Array) { let result: any[] = []; for (const child of children) { if (child && child.toArray) { result = result.concat(flattenChildren(child.toArray())); } else if (child instanceof Array) { result = result.concat(flattenChildren(child)); } else { result.push(child); } } return result; } return [children]; } export function joinTextContent(textArray: string[], markupEnabled: boolean) { if (!textArray) { return null; } if (markupEnabled) { return textArray .map(str => str + '') .join('') .replace(/\s+/g, ' ') .replace(/\s*<br\s*\/>\s*/g, '<br/>'); } return textArray.join(''); } export const JSX = { processor: null as JsxProcessor | null, jsxFactory: symbols.jsxFactory, jsxType: symbols.jsxType, install(jsxProcessor: JsxProcessor) { this.processor = jsxProcessor; }, createElement() { return this.processor!.createElement.apply
{ return new JsxProcessor(); }
identifier_body
JsxProcessor.ts
any) => { if (!(value instanceof Array)) { throw new Error('Not an array: ' + toValueString(value)); } return value; } }); const MARKUP: {[el: string]: AttrConverters} = Object.freeze({ br: {}, b: COMMON_ATTR, span: COMMON_ATTR, big: COMMON_ATTR, i: COMMON_ATTR, small: COMMON_ATTR, strong: COMMON_ATTR, ins: COMMON_ATTR, del: COMMON_ATTR, a: Object.assign({ href: (value: any) => { if (typeof value !== 'string') { throw new Error('Not a string: ' + toValueString(value)); } return value; } }, COMMON_ATTR) }); export function createJsxProcessor() { return new JsxProcessor(); } export default class JsxProcessor { createElement(Type: ElementFn | string, attributes?: any, ...children: any[]) { if (!(Type instanceof Function) && typeof Type !== 'string') { throw new Error(`JSX: Unsupported type ${toValueString(Type)}`); } const typeName = Type instanceof Function ? Type.name : Type; if (attributes?.children && children && children.length) { throw new Error(`JSX: Children for type ${typeName} given twice.`); } // Children may be part of attributes or given as varargs or both. // For JSX factories/functional components they should always be part of attributes const rawChildren = children.length ? children : attributes?.children || []; const {finalChildren, additionalAttributes} = parseChildren(rawChildren, Type); const finalAttributes = {...attributes}; joinAttributes(finalAttributes, additionalAttributes, Type); if (finalChildren) { finalAttributes.children = finalChildren; } if (typeof Type === 'string') { return this.createIntrinsicElement(Type, finalAttributes); } else if (Type.prototype && Type.prototype[JSX.jsxFactory])
else { return this.createFunctionalComponent(Type, finalAttributes); } } createCustomComponent(Type: ElementFn, attributes: any) { return Type.prototype[JSX.jsxFactory].call(this, Type, attributes); } createFunctionalComponent(Type: ElementFn, attributes: any) { try { const result = Type.call(this, attributes); Type[symbols.jsxType] = true; if (result instanceof Object) { result[symbols.jsxType] = Type; } return result; } catch (ex) { throw new Error(`JSX: "${ex.message}" ${getCurrentLine(ex)}`); } } createIntrinsicElement(el: string, attributes: any) { if (el in MARKUP) { const encoded: any = {}; Object.keys(attributes || {}).forEach(attribute => { const encoder = MARKUP[el][attribute]; if (!encoder) { if (attribute === 'children') { throw new Error(`Element "${el}" can not have children`); } else { throw new Error(`Element "${el}" does not support attribute "${attribute}"`); } } try { encoded[attribute] = encoder(attributes[attribute]); } catch(ex) { throw new Error(`Element "${el}" attribute "${attribute}" can not bet set: ${ex.message}`); } }); const text = joinTextContent(encoded.children, true); const tagOpen = [el].concat(Object.keys(encoded || {}).filter(attr => attr !== 'children').map( attribute => `${attribute}='${encoded[attribute]}'` )).join(' '); if (text) { return `<${tagOpen}>${text}</${el}>`; } return `<${tagOpen}/>`; } throw new Error(`JSX: Unsupported type ${el}`); } createNativeObject(Type: any, attributes: any) { if (attributes && 'children' in attributes) { throw new Error(`JSX: ${Type.name} can not have children`); } const {data, ...properties} = attributesWithoutListener(attributes || {}); const result = new Type(properties); registerListenerAttributes(result, attributes); if (data) { result.data = data; } return result; } getChildren(attributes: any) { if (!attributes || !('children' in attributes)) { return null; } return flattenChildren(attributes.children); } withoutChildren(attributes: any) { return omit(attributes, ['children']); } withContentText(attributes: any, content: any[], property: string, markupEnabled: boolean) { if (attributes && attributes[property] && content && content.length) { throw new Error(`JSX: ${property} given twice`); } const text = attributes && attributes[property] ? attributes[property].toString() : joinTextContent(content || [], markupEnabled); return Object.assign(attributes || {}, text ? {[property]: text} : {}); } withContentChildren(attributes: any, content: any[], property: string) { if (attributes && attributes[property] && content && content.length) { throw new Error(`JSX: ${property} given twice`); } const children = attributes && attributes[property] ? attributes[property] : (content || []); return Object.assign(attributes || {}, children ? {[property]: children} : {}); } withShorthands( attributes: object, shorthandsMapping: {[attr: string]: string}, merge: ((value1: any, value2: string) => any) ): object { const shorthandsKeys = Object.keys(shorthandsMapping); const shorthands = shorthandsKeys.filter(value => value in attributes); if (!shorthands.length) { return attributes; } const attrCopy: any = omit(attributes, shorthandsKeys); shorthands.forEach(shorthand => { const prop = shorthandsMapping[shorthand]; if (prop in attrCopy) { attrCopy[prop] = merge(attrCopy[prop], shorthand); } else { attrCopy[prop] = shorthand; } }); return attrCopy; } makeFactories(dic: {[key: string]: ElementFn}) { const result: {[key: string]: ElementFn} = {}; Object.keys(dic).forEach(key => { result[key] = this.makeFactory(dic[key]) as ElementFn; }); return result; } makeFactory(constructor: ElementFn): ElementFn { if (arguments.length !== 1) { throw new Error(`Expected exactly one argument, got ${arguments.length}`); } checkType(constructor, Function, 'first parameter'); if (!constructor.prototype || !constructor.prototype[JSX.jsxFactory]) { throw new Error(`Function ${constructor.name} is not a valid constructor`); } if (constructor[symbols.originalComponent]) { return this.makeFactory(constructor[symbols.originalComponent] as ElementFn); } return createFactoryProxy(this, constructor); } } function createFactoryProxy(processor: JsxProcessor, constructor: ElementFn): ElementFn { const handler: ProxyHandler<ElementFn> = { apply(target, _thisArg, args) { const [attributes, functionalComponent] = args; if (args.length > 1) { if (!(functionalComponent instanceof Function)) { throw new TypeError('Second parameter must be a function'); } if (functionalComponent.prototype && functionalComponent.prototype[JSX.jsxFactory]) { throw new TypeError('Second parameter must be a factory'); } } const result = processor.createElement(proxy, attributes); if (args.length > 1 && result instanceof Object) { functionalComponent[JSX.jsxType] = true; result[JSX.jsxType] = functionalComponent; } return result; }, get(target, property, receiver) { if (receiver === proxy) { if (property === symbols.originalComponent) { return constructor; } if (property === symbols.proxyHandler) { return handler; } } return Reflect.get(target, property, receiver); } }; /** @type {Factory} */ const proxy = new Proxy(constructor, handler); return proxy; } /** * Converts any value to a flat array. */ export function flattenChildren(children: unknown) { if (children instanceof Array) { let result: any[] = []; for (const child of children) { if (child && child.toArray) { result = result.concat(flattenChildren(child.toArray())); } else if (child instanceof Array) { result = result.concat(flattenChildren(child)); } else { result.push(child); } } return result; } return [children]; } export function joinTextContent(textArray: string[], markupEnabled: boolean) { if (!textArray) { return null; } if (markupEnabled) { return textArray .map(str => str + '') .join('') .replace(/\s+/g, ' ') .replace(/\s*<br\s*\/>\s*/g, '<br/>'); } return textArray.join(''); } export const JSX = { processor: null as JsxProcessor | null, jsxFactory: symbols.jsxFactory, jsxType: symbols.jsxType, install(jsxProcessor: JsxProcessor) { this.processor = jsxProcessor; }, createElement() { return this.processor!.createElement.apply(this
{ return this.createCustomComponent(Type, finalAttributes); }
conditional_block
JsxProcessor.ts
: any) => { if (!(value instanceof Array)) { throw new Error('Not an array: ' + toValueString(value)); } return value; } }); const MARKUP: {[el: string]: AttrConverters} = Object.freeze({ br: {}, b: COMMON_ATTR, span: COMMON_ATTR, big: COMMON_ATTR, i: COMMON_ATTR, small: COMMON_ATTR, strong: COMMON_ATTR, ins: COMMON_ATTR, del: COMMON_ATTR, a: Object.assign({ href: (value: any) => { if (typeof value !== 'string') { throw new Error('Not a string: ' + toValueString(value)); } return value; } }, COMMON_ATTR) }); export function createJsxProcessor() { return new JsxProcessor(); } export default class JsxProcessor { createElement(Type: ElementFn | string, attributes?: any, ...children: any[]) { if (!(Type instanceof Function) && typeof Type !== 'string') { throw new Error(`JSX: Unsupported type ${toValueString(Type)}`); } const typeName = Type instanceof Function ? Type.name : Type; if (attributes?.children && children && children.length) { throw new Error(`JSX: Children for type ${typeName} given twice.`); } // Children may be part of attributes or given as varargs or both. // For JSX factories/functional components they should always be part of attributes const rawChildren = children.length ? children : attributes?.children || []; const {finalChildren, additionalAttributes} = parseChildren(rawChildren, Type); const finalAttributes = {...attributes}; joinAttributes(finalAttributes, additionalAttributes, Type); if (finalChildren) { finalAttributes.children = finalChildren; } if (typeof Type === 'string') { return this.createIntrinsicElement(Type, finalAttributes); } else if (Type.prototype && Type.prototype[JSX.jsxFactory]) { return this.createCustomComponent(Type, finalAttributes); } else { return this.createFunctionalComponent(Type, finalAttributes); } }
(Type: ElementFn, attributes: any) { return Type.prototype[JSX.jsxFactory].call(this, Type, attributes); } createFunctionalComponent(Type: ElementFn, attributes: any) { try { const result = Type.call(this, attributes); Type[symbols.jsxType] = true; if (result instanceof Object) { result[symbols.jsxType] = Type; } return result; } catch (ex) { throw new Error(`JSX: "${ex.message}" ${getCurrentLine(ex)}`); } } createIntrinsicElement(el: string, attributes: any) { if (el in MARKUP) { const encoded: any = {}; Object.keys(attributes || {}).forEach(attribute => { const encoder = MARKUP[el][attribute]; if (!encoder) { if (attribute === 'children') { throw new Error(`Element "${el}" can not have children`); } else { throw new Error(`Element "${el}" does not support attribute "${attribute}"`); } } try { encoded[attribute] = encoder(attributes[attribute]); } catch(ex) { throw new Error(`Element "${el}" attribute "${attribute}" can not bet set: ${ex.message}`); } }); const text = joinTextContent(encoded.children, true); const tagOpen = [el].concat(Object.keys(encoded || {}).filter(attr => attr !== 'children').map( attribute => `${attribute}='${encoded[attribute]}'` )).join(' '); if (text) { return `<${tagOpen}>${text}</${el}>`; } return `<${tagOpen}/>`; } throw new Error(`JSX: Unsupported type ${el}`); } createNativeObject(Type: any, attributes: any) { if (attributes && 'children' in attributes) { throw new Error(`JSX: ${Type.name} can not have children`); } const {data, ...properties} = attributesWithoutListener(attributes || {}); const result = new Type(properties); registerListenerAttributes(result, attributes); if (data) { result.data = data; } return result; } getChildren(attributes: any) { if (!attributes || !('children' in attributes)) { return null; } return flattenChildren(attributes.children); } withoutChildren(attributes: any) { return omit(attributes, ['children']); } withContentText(attributes: any, content: any[], property: string, markupEnabled: boolean) { if (attributes && attributes[property] && content && content.length) { throw new Error(`JSX: ${property} given twice`); } const text = attributes && attributes[property] ? attributes[property].toString() : joinTextContent(content || [], markupEnabled); return Object.assign(attributes || {}, text ? {[property]: text} : {}); } withContentChildren(attributes: any, content: any[], property: string) { if (attributes && attributes[property] && content && content.length) { throw new Error(`JSX: ${property} given twice`); } const children = attributes && attributes[property] ? attributes[property] : (content || []); return Object.assign(attributes || {}, children ? {[property]: children} : {}); } withShorthands( attributes: object, shorthandsMapping: {[attr: string]: string}, merge: ((value1: any, value2: string) => any) ): object { const shorthandsKeys = Object.keys(shorthandsMapping); const shorthands = shorthandsKeys.filter(value => value in attributes); if (!shorthands.length) { return attributes; } const attrCopy: any = omit(attributes, shorthandsKeys); shorthands.forEach(shorthand => { const prop = shorthandsMapping[shorthand]; if (prop in attrCopy) { attrCopy[prop] = merge(attrCopy[prop], shorthand); } else { attrCopy[prop] = shorthand; } }); return attrCopy; } makeFactories(dic: {[key: string]: ElementFn}) { const result: {[key: string]: ElementFn} = {}; Object.keys(dic).forEach(key => { result[key] = this.makeFactory(dic[key]) as ElementFn; }); return result; } makeFactory(constructor: ElementFn): ElementFn { if (arguments.length !== 1) { throw new Error(`Expected exactly one argument, got ${arguments.length}`); } checkType(constructor, Function, 'first parameter'); if (!constructor.prototype || !constructor.prototype[JSX.jsxFactory]) { throw new Error(`Function ${constructor.name} is not a valid constructor`); } if (constructor[symbols.originalComponent]) { return this.makeFactory(constructor[symbols.originalComponent] as ElementFn); } return createFactoryProxy(this, constructor); } } function createFactoryProxy(processor: JsxProcessor, constructor: ElementFn): ElementFn { const handler: ProxyHandler<ElementFn> = { apply(target, _thisArg, args) { const [attributes, functionalComponent] = args; if (args.length > 1) { if (!(functionalComponent instanceof Function)) { throw new TypeError('Second parameter must be a function'); } if (functionalComponent.prototype && functionalComponent.prototype[JSX.jsxFactory]) { throw new TypeError('Second parameter must be a factory'); } } const result = processor.createElement(proxy, attributes); if (args.length > 1 && result instanceof Object) { functionalComponent[JSX.jsxType] = true; result[JSX.jsxType] = functionalComponent; } return result; }, get(target, property, receiver) { if (receiver === proxy) { if (property === symbols.originalComponent) { return constructor; } if (property === symbols.proxyHandler) { return handler; } } return Reflect.get(target, property, receiver); } }; /** @type {Factory} */ const proxy = new Proxy(constructor, handler); return proxy; } /** * Converts any value to a flat array. */ export function flattenChildren(children: unknown) { if (children instanceof Array) { let result: any[] = []; for (const child of children) { if (child && child.toArray) { result = result.concat(flattenChildren(child.toArray())); } else if (child instanceof Array) { result = result.concat(flattenChildren(child)); } else { result.push(child); } } return result; } return [children]; } export function joinTextContent(textArray: string[], markupEnabled: boolean) { if (!textArray) { return null; } if (markupEnabled) { return textArray .map(str => str + '') .join('') .replace(/\s+/g, ' ') .replace(/\s*<br\s*\/>\s*/g, '<br/>'); } return textArray.join(''); } export const JSX = { processor: null as JsxProcessor | null, jsxFactory: symbols.jsxFactory, jsxType: symbols.jsxType, install(jsxProcessor: JsxProcessor) { this.processor = jsxProcessor; }, createElement() { return this.processor!.createElement.apply(this
createCustomComponent
identifier_name
java.io.PipedInputStream.d.ts
declare namespace java { namespace io { class PipedInputStream extends java.io.InputStream { closedByWriter: boolean closedByReader: boolean connected: boolean readSide: java.lang.Thread writeSide: java.lang.Thread protected static readonly PIPE_SIZE: int protected buffer: byte[] protected in: int protected out: int static readonly $assertionsDisabled: boolean public constructor(arg0: java.io.PipedOutputStream) public constructor(arg0: java.io.PipedOutputStream, arg1: number | java.lang.Integer) public constructor() public constructor(arg0: number | java.lang.Integer) public connect(arg0: java.io.PipedOutputStream): void protected receive(arg0: number | java.lang.Integer): void receive(arg0: number[] | java.lang.Byte[], arg1: number | java.lang.Integer, arg2: number | java.lang.Integer): void receivedLast(): void public read(): number public read(arg0: number[] | java.lang.Byte[], arg1: number | java.lang.Integer, arg2: number | java.lang.Integer): number public available(): number
} } }
public close(): void
random_line_split
java.io.PipedInputStream.d.ts
declare namespace java { namespace io { class
extends java.io.InputStream { closedByWriter: boolean closedByReader: boolean connected: boolean readSide: java.lang.Thread writeSide: java.lang.Thread protected static readonly PIPE_SIZE: int protected buffer: byte[] protected in: int protected out: int static readonly $assertionsDisabled: boolean public constructor(arg0: java.io.PipedOutputStream) public constructor(arg0: java.io.PipedOutputStream, arg1: number | java.lang.Integer) public constructor() public constructor(arg0: number | java.lang.Integer) public connect(arg0: java.io.PipedOutputStream): void protected receive(arg0: number | java.lang.Integer): void receive(arg0: number[] | java.lang.Byte[], arg1: number | java.lang.Integer, arg2: number | java.lang.Integer): void receivedLast(): void public read(): number public read(arg0: number[] | java.lang.Byte[], arg1: number | java.lang.Integer, arg2: number | java.lang.Integer): number public available(): number public close(): void } } }
PipedInputStream
identifier_name
controllers.py
from application import CONFIG, app from .models import * from flask import current_app, session from flask.ext.login import login_user, logout_user, current_user from flask.ext.principal import Principal, Identity, AnonymousIdentity, identity_changed, identity_loaded, RoleNeed import bcrypt import re import sendgrid import time from itsdangerous import URLSafeTimedSerializer AuthenticationError = Exception("AuthenticationError", "Invalid credentials.") UserExistsError = Exception("UserExistsError", "Email already exists in database.") UserDoesNotExistError = Exception("UserDoesNotExistError", "Account with given email does not exist.") login_manager = LoginManager() login_manager.init_app(app) principals = Principal(app) sg = sendgrid.SendGridClient(CONFIG["SENDGRID_API_KEY"]) ts = URLSafeTimedSerializer(CONFIG["SECRET_KEY"]) @login_manager.user_loader def load_user(user_id): user_entries = StaffUserEntry.objects(id = user_id) if user_entries.count() != 1: return None currUser = user_entries[0] user = User(currUser.id, currUser.email, currUser.firstname, currUser.lastname, currUser.roles) return user @identity_loaded.connect_via(app) def on_identity_loaded(sender, identity): identity.user = current_user if hasattr(current_user, 'roles'): for role in current_user.roles: identity.provides.add(RoleNeed(role)) def get_user(email): entries = StaffUserEntry.objects(email = email) if entries.count() == 1: return entries[0] return None def verify_user(email, password): currUser = get_user(email) if currUser is None: return None hashed = currUser.hashed if bcrypt.hashpw(password.encode("utf-8"), hashed.encode("utf-8")) == hashed.encode("utf-8"): return load_user(currUser.id) else: return None def login(email): user = load_user(get_user(email).id) if user != None: login_user(user) identity_changed.send(current_app._get_current_object(), identity = Identity(user.uid)) else: raise UserDoesNotExistError def logout(): logout_user() for key in ('identity.name', 'identity.auth_type'): session.pop(key, None) identity_changed.send(current_app._get_current_object(), identity = AnonymousIdentity()) def tokenize_email(email): return ts.dumps(email, salt = CONFIG["EMAIL_TOKENIZER_SALT"]) def detokenize_email(token): return ts.loads(token, salt = CONFIG["EMAIL_TOKENIZER_SALT"], max_age = 86400) def send_recovery_email(email): user = get_user(email) if user is None: raise UserDoesNotExistError token = tokenize_email(email) message = sendgrid.Mail() message.add_to(email) message.set_from("noreply@hackbca.com") message.set_subject("hackBCA III - Account Recovery") message.set_html("<p></p>") message.add_filter("templates", "enable", "1") message.add_filter("templates", "template_id", CONFIG["SENDGRID_ACCOUNT_RECOVERY_TEMPLATE"]) message.add_substitution("prefix", "staff") message.add_substitution("token", token) status, msg = sg.send(message) def change_name(email, firstname, lastname): account = get_user(email) if account is None: raise UserDoesNotExistError account.firstname = firstname
account.lastname = lastname account.save() login(email) #To update navbar def change_password(email, password): account = get_user(email) if account is None: raise UserDoesNotExistError hashed = str(bcrypt.hashpw(password.encode("utf-8"), bcrypt.gensalt()))[2:-1] account.hashed = hashed account.save() def get_user_attr(email, attr): user = get_user(email) if user is None: raise UserDoesNotExistError return getattr(user, attr) def set_user_attr(email, attr, value): user = get_user(email) if user is None: raise UserDoesNotExistError setattr(user, attr, value) user.save()
random_line_split
controllers.py
from application import CONFIG, app from .models import * from flask import current_app, session from flask.ext.login import login_user, logout_user, current_user from flask.ext.principal import Principal, Identity, AnonymousIdentity, identity_changed, identity_loaded, RoleNeed import bcrypt import re import sendgrid import time from itsdangerous import URLSafeTimedSerializer AuthenticationError = Exception("AuthenticationError", "Invalid credentials.") UserExistsError = Exception("UserExistsError", "Email already exists in database.") UserDoesNotExistError = Exception("UserDoesNotExistError", "Account with given email does not exist.") login_manager = LoginManager() login_manager.init_app(app) principals = Principal(app) sg = sendgrid.SendGridClient(CONFIG["SENDGRID_API_KEY"]) ts = URLSafeTimedSerializer(CONFIG["SECRET_KEY"]) @login_manager.user_loader def load_user(user_id): user_entries = StaffUserEntry.objects(id = user_id) if user_entries.count() != 1: return None currUser = user_entries[0] user = User(currUser.id, currUser.email, currUser.firstname, currUser.lastname, currUser.roles) return user @identity_loaded.connect_via(app) def
(sender, identity): identity.user = current_user if hasattr(current_user, 'roles'): for role in current_user.roles: identity.provides.add(RoleNeed(role)) def get_user(email): entries = StaffUserEntry.objects(email = email) if entries.count() == 1: return entries[0] return None def verify_user(email, password): currUser = get_user(email) if currUser is None: return None hashed = currUser.hashed if bcrypt.hashpw(password.encode("utf-8"), hashed.encode("utf-8")) == hashed.encode("utf-8"): return load_user(currUser.id) else: return None def login(email): user = load_user(get_user(email).id) if user != None: login_user(user) identity_changed.send(current_app._get_current_object(), identity = Identity(user.uid)) else: raise UserDoesNotExistError def logout(): logout_user() for key in ('identity.name', 'identity.auth_type'): session.pop(key, None) identity_changed.send(current_app._get_current_object(), identity = AnonymousIdentity()) def tokenize_email(email): return ts.dumps(email, salt = CONFIG["EMAIL_TOKENIZER_SALT"]) def detokenize_email(token): return ts.loads(token, salt = CONFIG["EMAIL_TOKENIZER_SALT"], max_age = 86400) def send_recovery_email(email): user = get_user(email) if user is None: raise UserDoesNotExistError token = tokenize_email(email) message = sendgrid.Mail() message.add_to(email) message.set_from("noreply@hackbca.com") message.set_subject("hackBCA III - Account Recovery") message.set_html("<p></p>") message.add_filter("templates", "enable", "1") message.add_filter("templates", "template_id", CONFIG["SENDGRID_ACCOUNT_RECOVERY_TEMPLATE"]) message.add_substitution("prefix", "staff") message.add_substitution("token", token) status, msg = sg.send(message) def change_name(email, firstname, lastname): account = get_user(email) if account is None: raise UserDoesNotExistError account.firstname = firstname account.lastname = lastname account.save() login(email) #To update navbar def change_password(email, password): account = get_user(email) if account is None: raise UserDoesNotExistError hashed = str(bcrypt.hashpw(password.encode("utf-8"), bcrypt.gensalt()))[2:-1] account.hashed = hashed account.save() def get_user_attr(email, attr): user = get_user(email) if user is None: raise UserDoesNotExistError return getattr(user, attr) def set_user_attr(email, attr, value): user = get_user(email) if user is None: raise UserDoesNotExistError setattr(user, attr, value) user.save()
on_identity_loaded
identifier_name
controllers.py
from application import CONFIG, app from .models import * from flask import current_app, session from flask.ext.login import login_user, logout_user, current_user from flask.ext.principal import Principal, Identity, AnonymousIdentity, identity_changed, identity_loaded, RoleNeed import bcrypt import re import sendgrid import time from itsdangerous import URLSafeTimedSerializer AuthenticationError = Exception("AuthenticationError", "Invalid credentials.") UserExistsError = Exception("UserExistsError", "Email already exists in database.") UserDoesNotExistError = Exception("UserDoesNotExistError", "Account with given email does not exist.") login_manager = LoginManager() login_manager.init_app(app) principals = Principal(app) sg = sendgrid.SendGridClient(CONFIG["SENDGRID_API_KEY"]) ts = URLSafeTimedSerializer(CONFIG["SECRET_KEY"]) @login_manager.user_loader def load_user(user_id): user_entries = StaffUserEntry.objects(id = user_id) if user_entries.count() != 1: return None currUser = user_entries[0] user = User(currUser.id, currUser.email, currUser.firstname, currUser.lastname, currUser.roles) return user @identity_loaded.connect_via(app) def on_identity_loaded(sender, identity): identity.user = current_user if hasattr(current_user, 'roles'): for role in current_user.roles: identity.provides.add(RoleNeed(role)) def get_user(email): entries = StaffUserEntry.objects(email = email) if entries.count() == 1: return entries[0] return None def verify_user(email, password): currUser = get_user(email) if currUser is None: return None hashed = currUser.hashed if bcrypt.hashpw(password.encode("utf-8"), hashed.encode("utf-8")) == hashed.encode("utf-8"): return load_user(currUser.id) else: return None def login(email): user = load_user(get_user(email).id) if user != None:
else: raise UserDoesNotExistError def logout(): logout_user() for key in ('identity.name', 'identity.auth_type'): session.pop(key, None) identity_changed.send(current_app._get_current_object(), identity = AnonymousIdentity()) def tokenize_email(email): return ts.dumps(email, salt = CONFIG["EMAIL_TOKENIZER_SALT"]) def detokenize_email(token): return ts.loads(token, salt = CONFIG["EMAIL_TOKENIZER_SALT"], max_age = 86400) def send_recovery_email(email): user = get_user(email) if user is None: raise UserDoesNotExistError token = tokenize_email(email) message = sendgrid.Mail() message.add_to(email) message.set_from("noreply@hackbca.com") message.set_subject("hackBCA III - Account Recovery") message.set_html("<p></p>") message.add_filter("templates", "enable", "1") message.add_filter("templates", "template_id", CONFIG["SENDGRID_ACCOUNT_RECOVERY_TEMPLATE"]) message.add_substitution("prefix", "staff") message.add_substitution("token", token) status, msg = sg.send(message) def change_name(email, firstname, lastname): account = get_user(email) if account is None: raise UserDoesNotExistError account.firstname = firstname account.lastname = lastname account.save() login(email) #To update navbar def change_password(email, password): account = get_user(email) if account is None: raise UserDoesNotExistError hashed = str(bcrypt.hashpw(password.encode("utf-8"), bcrypt.gensalt()))[2:-1] account.hashed = hashed account.save() def get_user_attr(email, attr): user = get_user(email) if user is None: raise UserDoesNotExistError return getattr(user, attr) def set_user_attr(email, attr, value): user = get_user(email) if user is None: raise UserDoesNotExistError setattr(user, attr, value) user.save()
login_user(user) identity_changed.send(current_app._get_current_object(), identity = Identity(user.uid))
conditional_block
controllers.py
from application import CONFIG, app from .models import * from flask import current_app, session from flask.ext.login import login_user, logout_user, current_user from flask.ext.principal import Principal, Identity, AnonymousIdentity, identity_changed, identity_loaded, RoleNeed import bcrypt import re import sendgrid import time from itsdangerous import URLSafeTimedSerializer AuthenticationError = Exception("AuthenticationError", "Invalid credentials.") UserExistsError = Exception("UserExistsError", "Email already exists in database.") UserDoesNotExistError = Exception("UserDoesNotExistError", "Account with given email does not exist.") login_manager = LoginManager() login_manager.init_app(app) principals = Principal(app) sg = sendgrid.SendGridClient(CONFIG["SENDGRID_API_KEY"]) ts = URLSafeTimedSerializer(CONFIG["SECRET_KEY"]) @login_manager.user_loader def load_user(user_id): user_entries = StaffUserEntry.objects(id = user_id) if user_entries.count() != 1: return None currUser = user_entries[0] user = User(currUser.id, currUser.email, currUser.firstname, currUser.lastname, currUser.roles) return user @identity_loaded.connect_via(app) def on_identity_loaded(sender, identity): identity.user = current_user if hasattr(current_user, 'roles'): for role in current_user.roles: identity.provides.add(RoleNeed(role)) def get_user(email): entries = StaffUserEntry.objects(email = email) if entries.count() == 1: return entries[0] return None def verify_user(email, password):
def login(email): user = load_user(get_user(email).id) if user != None: login_user(user) identity_changed.send(current_app._get_current_object(), identity = Identity(user.uid)) else: raise UserDoesNotExistError def logout(): logout_user() for key in ('identity.name', 'identity.auth_type'): session.pop(key, None) identity_changed.send(current_app._get_current_object(), identity = AnonymousIdentity()) def tokenize_email(email): return ts.dumps(email, salt = CONFIG["EMAIL_TOKENIZER_SALT"]) def detokenize_email(token): return ts.loads(token, salt = CONFIG["EMAIL_TOKENIZER_SALT"], max_age = 86400) def send_recovery_email(email): user = get_user(email) if user is None: raise UserDoesNotExistError token = tokenize_email(email) message = sendgrid.Mail() message.add_to(email) message.set_from("noreply@hackbca.com") message.set_subject("hackBCA III - Account Recovery") message.set_html("<p></p>") message.add_filter("templates", "enable", "1") message.add_filter("templates", "template_id", CONFIG["SENDGRID_ACCOUNT_RECOVERY_TEMPLATE"]) message.add_substitution("prefix", "staff") message.add_substitution("token", token) status, msg = sg.send(message) def change_name(email, firstname, lastname): account = get_user(email) if account is None: raise UserDoesNotExistError account.firstname = firstname account.lastname = lastname account.save() login(email) #To update navbar def change_password(email, password): account = get_user(email) if account is None: raise UserDoesNotExistError hashed = str(bcrypt.hashpw(password.encode("utf-8"), bcrypt.gensalt()))[2:-1] account.hashed = hashed account.save() def get_user_attr(email, attr): user = get_user(email) if user is None: raise UserDoesNotExistError return getattr(user, attr) def set_user_attr(email, attr, value): user = get_user(email) if user is None: raise UserDoesNotExistError setattr(user, attr, value) user.save()
currUser = get_user(email) if currUser is None: return None hashed = currUser.hashed if bcrypt.hashpw(password.encode("utf-8"), hashed.encode("utf-8")) == hashed.encode("utf-8"): return load_user(currUser.id) else: return None
identifier_body
router.js
import Ember from 'ember'; import config from './config/environment'; var Router = Ember.Router.extend({ location: config.locationType }); Router.map(function() { // Auth this.route('activation', { path: 'activation/' }, function() { this.route('activate', { path: ':user_id/:token/' });
}); // Options this.route('options', { path: 'options/' }, function() { this.route('forum', { path: 'forum-options/' }); this.route('signature', { path: 'edit-signature/' }); this.route('username', { path: 'change-username/' }); this.route('password', { path: 'change-password/' }, function() { this.route('confirm', { path: ':token/' }); }); this.route('email', { path: 'change-email/' }, function() { this.route('confirm', { path: ':token/' }); }); }); // Legal this.route('terms-of-service', { path: 'terms-of-service/' }); this.route('privacy-policy', { path: 'privacy-policy/' }); // Error this.route('error-0', { path: 'error-0/' }); this.route('error-403', { path: 'error-403/:reason/' }); this.route('error-404', { path: 'error-404/' }); this.route('error-banned', { path: 'banned/:reason/' }); this.route('not-found', { path: '*path' }); }); export default Router;
}); this.route('forgotten-password', { path: 'forgotten-password/' }, function() { this.route('change-form', { path: ':user_id/:token/' });
random_line_split
validate-lab.ts
import isEmpty from 'lodash/isEmpty' import Lab from '../../shared/model/Lab' export class LabError extends Error { message: string result?: string patient?: string type?: string constructor(message: string, result: string, patient: string, type: string) { super(message) this.message = message this.result = result this.patient = patient this.type = type } }
const labError = {} as LabError if (!lab.patient) { labError.patient = 'labs.requests.error.patientRequired' } if (!lab.type) { labError.type = 'labs.requests.error.typeRequired' } if (!isEmpty(labError)) { labError.message = 'labs.requests.error.unableToRequest' } return labError } export function validateLabComplete(lab: Partial<Lab>): LabError { const labError = {} as LabError if (!lab.result) { labError.result = 'labs.requests.error.resultRequiredToComplete' labError.message = 'labs.requests.error.unableToComplete' } return labError }
export function validateLabRequest(lab: Partial<Lab>): LabError {
random_line_split
validate-lab.ts
import isEmpty from 'lodash/isEmpty' import Lab from '../../shared/model/Lab' export class LabError extends Error { message: string result?: string patient?: string type?: string
(message: string, result: string, patient: string, type: string) { super(message) this.message = message this.result = result this.patient = patient this.type = type } } export function validateLabRequest(lab: Partial<Lab>): LabError { const labError = {} as LabError if (!lab.patient) { labError.patient = 'labs.requests.error.patientRequired' } if (!lab.type) { labError.type = 'labs.requests.error.typeRequired' } if (!isEmpty(labError)) { labError.message = 'labs.requests.error.unableToRequest' } return labError } export function validateLabComplete(lab: Partial<Lab>): LabError { const labError = {} as LabError if (!lab.result) { labError.result = 'labs.requests.error.resultRequiredToComplete' labError.message = 'labs.requests.error.unableToComplete' } return labError }
constructor
identifier_name
validate-lab.ts
import isEmpty from 'lodash/isEmpty' import Lab from '../../shared/model/Lab' export class LabError extends Error { message: string result?: string patient?: string type?: string constructor(message: string, result: string, patient: string, type: string) { super(message) this.message = message this.result = result this.patient = patient this.type = type } } export function validateLabRequest(lab: Partial<Lab>): LabError { const labError = {} as LabError if (!lab.patient) { labError.patient = 'labs.requests.error.patientRequired' } if (!lab.type) { labError.type = 'labs.requests.error.typeRequired' } if (!isEmpty(labError)) { labError.message = 'labs.requests.error.unableToRequest' } return labError } export function validateLabComplete(lab: Partial<Lab>): LabError
{ const labError = {} as LabError if (!lab.result) { labError.result = 'labs.requests.error.resultRequiredToComplete' labError.message = 'labs.requests.error.unableToComplete' } return labError }
identifier_body
validate-lab.ts
import isEmpty from 'lodash/isEmpty' import Lab from '../../shared/model/Lab' export class LabError extends Error { message: string result?: string patient?: string type?: string constructor(message: string, result: string, patient: string, type: string) { super(message) this.message = message this.result = result this.patient = patient this.type = type } } export function validateLabRequest(lab: Partial<Lab>): LabError { const labError = {} as LabError if (!lab.patient)
if (!lab.type) { labError.type = 'labs.requests.error.typeRequired' } if (!isEmpty(labError)) { labError.message = 'labs.requests.error.unableToRequest' } return labError } export function validateLabComplete(lab: Partial<Lab>): LabError { const labError = {} as LabError if (!lab.result) { labError.result = 'labs.requests.error.resultRequiredToComplete' labError.message = 'labs.requests.error.unableToComplete' } return labError }
{ labError.patient = 'labs.requests.error.patientRequired' }
conditional_block
test.rs
::Session, crate: ast::Crate) -> ast::Crate { // We generate the test harness when building in the 'test' // configuration, either with the '--test' or '--cfg test' // command line options. let should_test = attr::contains_name(crate.config, "test"); if should_test { generate_test_harness(sess, crate) } else { strip_test_functions(crate) } } struct TestHarnessGenerator { cx: @mut TestCtxt, } impl fold::ast_fold for TestHarnessGenerator { fn fold_crate(&self, c: ast::Crate) -> ast::Crate { let folded = fold::noop_fold_crate(c, self); // Add a special __test module to the crate that will contain code // generated for the test harness ast::Crate { module: add_test_module(self.cx, &folded.module), .. folded } } fn fold_item(&self, i: @ast::item) -> SmallVector<@ast::item> { self.cx.path.push(i.ident); debug!("current path: {}", ast_util::path_name_i(self.cx.path.clone())); if is_test_fn(self.cx, i) || is_bench_fn(i) { match i.node { ast::item_fn(_, purity, _, _, _) if purity == ast::unsafe_fn =>
_ => { debug!("this is a test function"); let test = Test { span: i.span, path: self.cx.path.clone(), bench: is_bench_fn(i), ignore: is_ignored(self.cx, i), should_fail: should_fail(i) }; self.cx.testfns.push(test); // debug!("have {} test/bench functions", // cx.testfns.len()); } } } let res = fold::noop_fold_item(i, self); self.cx.path.pop(); res } fn fold_mod(&self, m: &ast::_mod) -> ast::_mod { // Remove any #[main] from the AST so it doesn't clash with // the one we're going to add. Only if compiling an executable. fn nomain(cx: @mut TestCtxt, item: @ast::item) -> @ast::item { if !*cx.sess.building_library { @ast::item { attrs: item.attrs.iter().filter_map(|attr| { if "main" != attr.name() { Some(*attr) } else { None } }).collect(), .. (*item).clone() } } else { item } } let mod_nomain = ast::_mod { view_items: m.view_items.clone(), items: m.items.iter().map(|i| nomain(self.cx, *i)).collect(), }; fold::noop_fold_mod(&mod_nomain, self) } } fn generate_test_harness(sess: session::Session, crate: ast::Crate) -> ast::Crate { let cx: @mut TestCtxt = @mut TestCtxt { sess: sess, ext_cx: ExtCtxt::new(sess.parse_sess, sess.opts.cfg.clone()), path: ~[], testfns: ~[], is_extra: is_extra(&crate), config: crate.config.clone(), }; let ext_cx = cx.ext_cx; ext_cx.bt_push(ExpnInfo { call_site: dummy_sp(), callee: NameAndSpan { name: @"test", format: MacroAttribute, span: None } }); let fold = TestHarnessGenerator { cx: cx }; let res = fold.fold_crate(crate); ext_cx.bt_pop(); return res; } fn strip_test_functions(crate: ast::Crate) -> ast::Crate { // When not compiling with --test we should not compile the // #[test] functions config::strip_items(crate, |attrs| { !attr::contains_name(attrs, "test") && !attr::contains_name(attrs, "bench") }) } fn is_test_fn(cx: @mut TestCtxt, i: @ast::item) -> bool { let has_test_attr = attr::contains_name(i.attrs, "test"); fn has_test_signature(i: @ast::item) -> bool { match &i.node { &ast::item_fn(ref decl, _, _, ref generics, _) => { let no_output = match decl.output.node { ast::ty_nil => true, _ => false }; decl.inputs.is_empty() && no_output && !generics.is_parameterized() } _ => false } } if has_test_attr && !has_test_signature(i) { let sess = cx.sess; sess.span_err( i.span, "functions used as tests must have signature fn() -> ()." ); } return has_test_attr && has_test_signature(i); } fn is_bench_fn(i: @ast::item) -> bool { let has_bench_attr = attr::contains_name(i.attrs, "bench"); fn has_test_signature(i: @ast::item) -> bool { match i.node { ast::item_fn(ref decl, _, _, ref generics, _) => { let input_cnt = decl.inputs.len(); let no_output = match decl.output.node { ast::ty_nil => true, _ => false }; let tparm_cnt = generics.ty_params.len(); // NB: inadequate check, but we're running // well before resolve, can't get too deep. input_cnt == 1u && no_output && tparm_cnt == 0u } _ => false } } return has_bench_attr && has_test_signature(i); } fn is_ignored(cx: @mut TestCtxt, i: @ast::item) -> bool { i.attrs.iter().any(|attr| { // check ignore(cfg(foo, bar)) "ignore" == attr.name() && match attr.meta_item_list() { Some(ref cfgs) => attr::test_cfg(cx.config, cfgs.iter().map(|x| *x)), None => true } }) } fn should_fail(i: @ast::item) -> bool { attr::contains_name(i.attrs, "should_fail") } fn add_test_module(cx: &TestCtxt, m: &ast::_mod) -> ast::_mod { let testmod = mk_test_module(cx); ast::_mod { items: vec::append_one(m.items.clone(), testmod), ..(*m).clone() } } /* We're going to be building a module that looks more or less like: mod __test { #[!resolve_unexported] extern mod extra (name = "extra", vers = "..."); fn main() { #[main]; extra::test::test_main_static(::os::args(), tests) } static tests : &'static [extra::test::TestDescAndFn] = &[ ... the list of tests in the crate ... ]; } */ fn mk_std(cx: &TestCtxt) -> ast::view_item { let id_extra = cx.sess.ident_of("extra"); let vi = if cx.is_extra { ast::view_item_use( ~[@nospan(ast::view_path_simple(id_extra, path_node(~[id_extra]), ast::DUMMY_NODE_ID))]) } else { let mi = attr::mk_name_value_item_str(@"vers", @"0.9-pre"); ast::view_item_extern_mod(id_extra, None, ~[mi], ast::DUMMY_NODE_ID) }; ast::view_item { node: vi, attrs: ~[], vis: ast::public, span: dummy_sp() } } fn mk_test_module(cx: &TestCtxt) -> @ast::item { // Link to extra let view_items = ~[mk_std(cx)]; // A constant vector of test descriptors. let tests = mk_tests(cx); // The synthesized main function which will call the console test runner // with our list of tests let mainfn = (quote_item!(cx.ext_cx, pub fn main() { #[main]; extra::test::test_main_static(::std::os::args(), TESTS); } )).unwrap(); let testmod = ast::_mod { view_items: view_items, items: ~[mainfn, tests], }; let item_ = ast::item_mod(testmod); // This attribute tells resolve to let us call unexported functions let resolve_unexported_attr = attr::mk_attr(attr::mk_word_item(@"!resolve_unexported")); let item = ast::item { ident: cx.sess.ident_of("__test"), attrs: ~[resolve_unexported_attr], id: ast::DUMMY_NODE_ID, node: item_, vis: ast::public, span: dummy_sp(), }; debug!("Synthetic test module:\n{}\n", pprust::item_to_str(&item, cx.sess.intr())); return @item; } fn nospan<T>(t: T) ->
{ let sess = self.cx.sess; sess.span_fatal(i.span, "unsafe functions cannot be used for \ tests"); }
conditional_block
test.rs
// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or // http://www.apache.org/licenses/LICENSE-2.0> or the MIT license // <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your // option. This file may not be copied, modified, or distributed // except according to those terms. // Code that generates a test runner to run all the tests in a crate use driver::session; use front::config; use std::vec; use syntax::ast_util::*; use syntax::attr::AttrMetaMethods; use syntax::attr; use syntax::codemap::{dummy_sp, Span, ExpnInfo, NameAndSpan, MacroAttribute}; use syntax::codemap; use syntax::ext::base::ExtCtxt; use syntax::fold::ast_fold; use syntax::fold; use syntax::opt_vec; use syntax::print::pprust; use syntax::{ast, ast_util}; use syntax::util::small_vector::SmallVector; struct Test { span: Span, path: ~[ast::Ident], bench: bool, ignore: bool, should_fail: bool } struct TestCtxt { sess: session::Session, path: ~[ast::Ident], ext_cx: @ExtCtxt, testfns: ~[Test], is_extra: bool, config: ast::CrateConfig, } // Traverse the crate, collecting all the test functions, eliding any // existing main functions, and synthesizing a main test harness pub fn modify_for_testing(sess: session::Session, crate: ast::Crate) -> ast::Crate { // We generate the test harness when building in the 'test' // configuration, either with the '--test' or '--cfg test' // command line options. let should_test = attr::contains_name(crate.config, "test"); if should_test { generate_test_harness(sess, crate) } else { strip_test_functions(crate) } } struct TestHarnessGenerator { cx: @mut TestCtxt, } impl fold::ast_fold for TestHarnessGenerator { fn fold_crate(&self, c: ast::Crate) -> ast::Crate { let folded = fold::noop_fold_crate(c, self); // Add a special __test module to the crate that will contain code // generated for the test harness ast::Crate { module: add_test_module(self.cx, &folded.module), .. folded } } fn fold_item(&self, i: @ast::item) -> SmallVector<@ast::item> { self.cx.path.push(i.ident); debug!("current path: {}", ast_util::path_name_i(self.cx.path.clone())); if is_test_fn(self.cx, i) || is_bench_fn(i) { match i.node { ast::item_fn(_, purity, _, _, _) if purity == ast::unsafe_fn => { let sess = self.cx.sess; sess.span_fatal(i.span, "unsafe functions cannot be used for \ tests"); } _ => { debug!("this is a test function"); let test = Test { span: i.span, path: self.cx.path.clone(), bench: is_bench_fn(i), ignore: is_ignored(self.cx, i), should_fail: should_fail(i) }; self.cx.testfns.push(test); // debug!("have {} test/bench functions", // cx.testfns.len()); } } } let res = fold::noop_fold_item(i, self); self.cx.path.pop(); res } fn fold_mod(&self, m: &ast::_mod) -> ast::_mod { // Remove any #[main] from the AST so it doesn't clash with // the one we're going to add. Only if compiling an executable. fn nomain(cx: @mut TestCtxt, item: @ast::item) -> @ast::item { if !*cx.sess.building_library { @ast::item { attrs: item.attrs.iter().filter_map(|attr| { if "main" != attr.name() { Some(*attr) } else { None } }).collect(), .. (*item).clone() } } else { item } } let mod_nomain = ast::_mod { view_items: m.view_items.clone(), items: m.items.iter().map(|i| nomain(self.cx, *i)).collect(), }; fold::noop_fold_mod(&mod_nomain, self) } } fn generate_test_harness(sess: session::Session, crate: ast::Crate) -> ast::Crate { let cx: @mut TestCtxt = @mut TestCtxt { sess: sess, ext_cx: ExtCtxt::new(sess.parse_sess, sess.opts.cfg.clone()), path: ~[], testfns: ~[], is_extra: is_extra(&crate), config: crate.config.clone(), }; let ext_cx = cx.ext_cx; ext_cx.bt_push(ExpnInfo { call_site: dummy_sp(), callee: NameAndSpan { name: @"test", format: MacroAttribute, span: None } }); let fold = TestHarnessGenerator { cx: cx }; let res = fold.fold_crate(crate); ext_cx.bt_pop(); return res; } fn strip_test_functions(crate: ast::Crate) -> ast::Crate { // When not compiling with --test we should not compile the // #[test] functions config::strip_items(crate, |attrs| { !attr::contains_name(attrs, "test") && !attr::contains_name(attrs, "bench") }) } fn is_test_fn(cx: @mut TestCtxt, i: @ast::item) -> bool { let has_test_attr = attr::contains_name(i.attrs, "test"); fn has_test_signature(i: @ast::item) -> bool { match &i.node { &ast::item_fn(ref decl, _, _, ref generics, _) => { let no_output = match decl.output.node { ast::ty_nil => true, _ => false }; decl.inputs.is_empty() && no_output && !generics.is_parameterized() } _ => false } } if has_test_attr && !has_test_signature(i) { let sess = cx.sess; sess.span_err( i.span, "functions used as tests must have signature fn() -> ()." ); } return has_test_attr && has_test_signature(i); } fn is_bench_fn(i: @ast::item) -> bool { let has_bench_attr = attr::contains_name(i.attrs, "bench"); fn has_test_signature(i: @ast::item) -> bool { match i.node { ast::item_fn(ref decl, _, _, ref generics, _) => { let input_cnt = decl.inputs.len(); let no_output = match decl.output.node { ast::ty_nil => true, _ => false }; let tparm_cnt = generics.ty_params.len(); // NB: inadequate check, but we're running // well before resolve, can't get too deep. input_cnt == 1u && no_output && tparm_cnt == 0u } _ => false } } return has_bench_attr && has_test_signature(i); } fn is_ignored(cx: @mut TestCtxt, i: @ast::item) -> bool { i.attrs.iter().any(|attr| { // check ignore(cfg(foo, bar)) "ignore" == attr.name() && match attr.meta_item_list() { Some(ref cfgs) => attr::test_cfg(cx.config, cfgs.iter().map(|x| *x)), None => true } }) } fn should_fail(i: @ast::item) -> bool { attr::contains_name(i.attrs, "should_fail") } fn add_test_module(cx: &TestCtxt, m: &ast::_mod) -> ast::_mod { let testmod = mk_test_module(cx); ast::_mod { items: vec::append_one(m.items.clone(), testmod), ..(*m).clone() } } /* We're going to be building a module that looks more or less like: mod __test { #[!resolve_unexported] extern mod extra (name = "extra", vers = "..."); fn main() { #[main]; extra::test::test_main_static(::os::args(), tests) } static tests : &'static [extra::test::TestDescAndFn] = &[ ... the list of tests in the crate ... ]; } */ fn mk_std(cx: &TestCtxt) -> ast::view_item { let id_extra = cx.sess.ident_of("extra"); let vi = if cx.is_extra { ast::view_item_use( ~[@nospan(ast::view_path_simple(id_extra, path_node(~[id_extra]), ast::DUMMY_NODE_ID))]) } else { let mi = attr::mk_name
// Copyright 2012 The Rust Project Developers. See the COPYRIGHT // file at the top-level directory of this distribution and at // http://rust-lang.org/COPYRIGHT. //
random_line_split
test.rs
::Session, crate: ast::Crate) -> ast::Crate { // We generate the test harness when building in the 'test' // configuration, either with the '--test' or '--cfg test' // command line options. let should_test = attr::contains_name(crate.config, "test"); if should_test { generate_test_harness(sess, crate) } else { strip_test_functions(crate) } } struct TestHarnessGenerator { cx: @mut TestCtxt, } impl fold::ast_fold for TestHarnessGenerator { fn fold_crate(&self, c: ast::Crate) -> ast::Crate { let folded = fold::noop_fold_crate(c, self); // Add a special __test module to the crate that will contain code // generated for the test harness ast::Crate { module: add_test_module(self.cx, &folded.module), .. folded } } fn fold_item(&self, i: @ast::item) -> SmallVector<@ast::item> { self.cx.path.push(i.ident); debug!("current path: {}", ast_util::path_name_i(self.cx.path.clone())); if is_test_fn(self.cx, i) || is_bench_fn(i) { match i.node { ast::item_fn(_, purity, _, _, _) if purity == ast::unsafe_fn => { let sess = self.cx.sess; sess.span_fatal(i.span, "unsafe functions cannot be used for \ tests"); } _ => { debug!("this is a test function"); let test = Test { span: i.span, path: self.cx.path.clone(), bench: is_bench_fn(i), ignore: is_ignored(self.cx, i), should_fail: should_fail(i) }; self.cx.testfns.push(test); // debug!("have {} test/bench functions", // cx.testfns.len()); } } } let res = fold::noop_fold_item(i, self); self.cx.path.pop(); res } fn fold_mod(&self, m: &ast::_mod) -> ast::_mod { // Remove any #[main] from the AST so it doesn't clash with // the one we're going to add. Only if compiling an executable. fn nomain(cx: @mut TestCtxt, item: @ast::item) -> @ast::item { if !*cx.sess.building_library { @ast::item { attrs: item.attrs.iter().filter_map(|attr| { if "main" != attr.name() { Some(*attr) } else { None } }).collect(), .. (*item).clone() } } else { item } } let mod_nomain = ast::_mod { view_items: m.view_items.clone(), items: m.items.iter().map(|i| nomain(self.cx, *i)).collect(), }; fold::noop_fold_mod(&mod_nomain, self) } } fn generate_test_harness(sess: session::Session, crate: ast::Crate) -> ast::Crate { let cx: @mut TestCtxt = @mut TestCtxt { sess: sess, ext_cx: ExtCtxt::new(sess.parse_sess, sess.opts.cfg.clone()), path: ~[], testfns: ~[], is_extra: is_extra(&crate), config: crate.config.clone(), }; let ext_cx = cx.ext_cx; ext_cx.bt_push(ExpnInfo { call_site: dummy_sp(), callee: NameAndSpan { name: @"test", format: MacroAttribute, span: None } }); let fold = TestHarnessGenerator { cx: cx }; let res = fold.fold_crate(crate); ext_cx.bt_pop(); return res; } fn strip_test_functions(crate: ast::Crate) -> ast::Crate { // When not compiling with --test we should not compile the // #[test] functions config::strip_items(crate, |attrs| { !attr::contains_name(attrs, "test") && !attr::contains_name(attrs, "bench") }) } fn is_test_fn(cx: @mut TestCtxt, i: @ast::item) -> bool { let has_test_attr = attr::contains_name(i.attrs, "test"); fn has_test_signature(i: @ast::item) -> bool { match &i.node { &ast::item_fn(ref decl, _, _, ref generics, _) => { let no_output = match decl.output.node { ast::ty_nil => true, _ => false }; decl.inputs.is_empty() && no_output && !generics.is_parameterized() } _ => false } } if has_test_attr && !has_test_signature(i) { let sess = cx.sess; sess.span_err( i.span, "functions used as tests must have signature fn() -> ()." ); } return has_test_attr && has_test_signature(i); } fn is_bench_fn(i: @ast::item) -> bool { let has_bench_attr = attr::contains_name(i.attrs, "bench"); fn has_test_signature(i: @ast::item) -> bool
return has_bench_attr && has_test_signature(i); } fn is_ignored(cx: @mut TestCtxt, i: @ast::item) -> bool { i.attrs.iter().any(|attr| { // check ignore(cfg(foo, bar)) "ignore" == attr.name() && match attr.meta_item_list() { Some(ref cfgs) => attr::test_cfg(cx.config, cfgs.iter().map(|x| *x)), None => true } }) } fn should_fail(i: @ast::item) -> bool { attr::contains_name(i.attrs, "should_fail") } fn add_test_module(cx: &TestCtxt, m: &ast::_mod) -> ast::_mod { let testmod = mk_test_module(cx); ast::_mod { items: vec::append_one(m.items.clone(), testmod), ..(*m).clone() } } /* We're going to be building a module that looks more or less like: mod __test { #[!resolve_unexported] extern mod extra (name = "extra", vers = "..."); fn main() { #[main]; extra::test::test_main_static(::os::args(), tests) } static tests : &'static [extra::test::TestDescAndFn] = &[ ... the list of tests in the crate ... ]; } */ fn mk_std(cx: &TestCtxt) -> ast::view_item { let id_extra = cx.sess.ident_of("extra"); let vi = if cx.is_extra { ast::view_item_use( ~[@nospan(ast::view_path_simple(id_extra, path_node(~[id_extra]), ast::DUMMY_NODE_ID))]) } else { let mi = attr::mk_name_value_item_str(@"vers", @"0.9-pre"); ast::view_item_extern_mod(id_extra, None, ~[mi], ast::DUMMY_NODE_ID) }; ast::view_item { node: vi, attrs: ~[], vis: ast::public, span: dummy_sp() } } fn mk_test_module(cx: &TestCtxt) -> @ast::item { // Link to extra let view_items = ~[mk_std(cx)]; // A constant vector of test descriptors. let tests = mk_tests(cx); // The synthesized main function which will call the console test runner // with our list of tests let mainfn = (quote_item!(cx.ext_cx, pub fn main() { #[main]; extra::test::test_main_static(::std::os::args(), TESTS); } )).unwrap(); let testmod = ast::_mod { view_items: view_items, items: ~[mainfn, tests], }; let item_ = ast::item_mod(testmod); // This attribute tells resolve to let us call unexported functions let resolve_unexported_attr = attr::mk_attr(attr::mk_word_item(@"!resolve_unexported")); let item = ast::item { ident: cx.sess.ident_of("__test"), attrs: ~[resolve_unexported_attr], id: ast::DUMMY_NODE_ID, node: item_, vis: ast::public, span: dummy_sp(), }; debug!("Synthetic test module:\n{}\n", pprust::item_to_str(&item, cx.sess.intr())); return @item; } fn nospan<T>(t: T) ->
{ match i.node { ast::item_fn(ref decl, _, _, ref generics, _) => { let input_cnt = decl.inputs.len(); let no_output = match decl.output.node { ast::ty_nil => true, _ => false }; let tparm_cnt = generics.ty_params.len(); // NB: inadequate check, but we're running // well before resolve, can't get too deep. input_cnt == 1u && no_output && tparm_cnt == 0u } _ => false } }
identifier_body
test.rs
'test' // configuration, either with the '--test' or '--cfg test' // command line options. let should_test = attr::contains_name(crate.config, "test"); if should_test { generate_test_harness(sess, crate) } else { strip_test_functions(crate) } } struct TestHarnessGenerator { cx: @mut TestCtxt, } impl fold::ast_fold for TestHarnessGenerator { fn fold_crate(&self, c: ast::Crate) -> ast::Crate { let folded = fold::noop_fold_crate(c, self); // Add a special __test module to the crate that will contain code // generated for the test harness ast::Crate { module: add_test_module(self.cx, &folded.module), .. folded } } fn fold_item(&self, i: @ast::item) -> SmallVector<@ast::item> { self.cx.path.push(i.ident); debug!("current path: {}", ast_util::path_name_i(self.cx.path.clone())); if is_test_fn(self.cx, i) || is_bench_fn(i) { match i.node { ast::item_fn(_, purity, _, _, _) if purity == ast::unsafe_fn => { let sess = self.cx.sess; sess.span_fatal(i.span, "unsafe functions cannot be used for \ tests"); } _ => { debug!("this is a test function"); let test = Test { span: i.span, path: self.cx.path.clone(), bench: is_bench_fn(i), ignore: is_ignored(self.cx, i), should_fail: should_fail(i) }; self.cx.testfns.push(test); // debug!("have {} test/bench functions", // cx.testfns.len()); } } } let res = fold::noop_fold_item(i, self); self.cx.path.pop(); res } fn fold_mod(&self, m: &ast::_mod) -> ast::_mod { // Remove any #[main] from the AST so it doesn't clash with // the one we're going to add. Only if compiling an executable. fn nomain(cx: @mut TestCtxt, item: @ast::item) -> @ast::item { if !*cx.sess.building_library { @ast::item { attrs: item.attrs.iter().filter_map(|attr| { if "main" != attr.name() { Some(*attr) } else { None } }).collect(), .. (*item).clone() } } else { item } } let mod_nomain = ast::_mod { view_items: m.view_items.clone(), items: m.items.iter().map(|i| nomain(self.cx, *i)).collect(), }; fold::noop_fold_mod(&mod_nomain, self) } } fn generate_test_harness(sess: session::Session, crate: ast::Crate) -> ast::Crate { let cx: @mut TestCtxt = @mut TestCtxt { sess: sess, ext_cx: ExtCtxt::new(sess.parse_sess, sess.opts.cfg.clone()), path: ~[], testfns: ~[], is_extra: is_extra(&crate), config: crate.config.clone(), }; let ext_cx = cx.ext_cx; ext_cx.bt_push(ExpnInfo { call_site: dummy_sp(), callee: NameAndSpan { name: @"test", format: MacroAttribute, span: None } }); let fold = TestHarnessGenerator { cx: cx }; let res = fold.fold_crate(crate); ext_cx.bt_pop(); return res; } fn strip_test_functions(crate: ast::Crate) -> ast::Crate { // When not compiling with --test we should not compile the // #[test] functions config::strip_items(crate, |attrs| { !attr::contains_name(attrs, "test") && !attr::contains_name(attrs, "bench") }) } fn is_test_fn(cx: @mut TestCtxt, i: @ast::item) -> bool { let has_test_attr = attr::contains_name(i.attrs, "test"); fn has_test_signature(i: @ast::item) -> bool { match &i.node { &ast::item_fn(ref decl, _, _, ref generics, _) => { let no_output = match decl.output.node { ast::ty_nil => true, _ => false }; decl.inputs.is_empty() && no_output && !generics.is_parameterized() } _ => false } } if has_test_attr && !has_test_signature(i) { let sess = cx.sess; sess.span_err( i.span, "functions used as tests must have signature fn() -> ()." ); } return has_test_attr && has_test_signature(i); } fn is_bench_fn(i: @ast::item) -> bool { let has_bench_attr = attr::contains_name(i.attrs, "bench"); fn has_test_signature(i: @ast::item) -> bool { match i.node { ast::item_fn(ref decl, _, _, ref generics, _) => { let input_cnt = decl.inputs.len(); let no_output = match decl.output.node { ast::ty_nil => true, _ => false }; let tparm_cnt = generics.ty_params.len(); // NB: inadequate check, but we're running // well before resolve, can't get too deep. input_cnt == 1u && no_output && tparm_cnt == 0u } _ => false } } return has_bench_attr && has_test_signature(i); } fn is_ignored(cx: @mut TestCtxt, i: @ast::item) -> bool { i.attrs.iter().any(|attr| { // check ignore(cfg(foo, bar)) "ignore" == attr.name() && match attr.meta_item_list() { Some(ref cfgs) => attr::test_cfg(cx.config, cfgs.iter().map(|x| *x)), None => true } }) } fn should_fail(i: @ast::item) -> bool { attr::contains_name(i.attrs, "should_fail") } fn add_test_module(cx: &TestCtxt, m: &ast::_mod) -> ast::_mod { let testmod = mk_test_module(cx); ast::_mod { items: vec::append_one(m.items.clone(), testmod), ..(*m).clone() } } /* We're going to be building a module that looks more or less like: mod __test { #[!resolve_unexported] extern mod extra (name = "extra", vers = "..."); fn main() { #[main]; extra::test::test_main_static(::os::args(), tests) } static tests : &'static [extra::test::TestDescAndFn] = &[ ... the list of tests in the crate ... ]; } */ fn mk_std(cx: &TestCtxt) -> ast::view_item { let id_extra = cx.sess.ident_of("extra"); let vi = if cx.is_extra { ast::view_item_use( ~[@nospan(ast::view_path_simple(id_extra, path_node(~[id_extra]), ast::DUMMY_NODE_ID))]) } else { let mi = attr::mk_name_value_item_str(@"vers", @"0.9-pre"); ast::view_item_extern_mod(id_extra, None, ~[mi], ast::DUMMY_NODE_ID) }; ast::view_item { node: vi, attrs: ~[], vis: ast::public, span: dummy_sp() } } fn mk_test_module(cx: &TestCtxt) -> @ast::item { // Link to extra let view_items = ~[mk_std(cx)]; // A constant vector of test descriptors. let tests = mk_tests(cx); // The synthesized main function which will call the console test runner // with our list of tests let mainfn = (quote_item!(cx.ext_cx, pub fn main() { #[main]; extra::test::test_main_static(::std::os::args(), TESTS); } )).unwrap(); let testmod = ast::_mod { view_items: view_items, items: ~[mainfn, tests], }; let item_ = ast::item_mod(testmod); // This attribute tells resolve to let us call unexported functions let resolve_unexported_attr = attr::mk_attr(attr::mk_word_item(@"!resolve_unexported")); let item = ast::item { ident: cx.sess.ident_of("__test"), attrs: ~[resolve_unexported_attr], id: ast::DUMMY_NODE_ID, node: item_, vis: ast::public, span: dummy_sp(), }; debug!("Synthetic test module:\n{}\n", pprust::item_to_str(&item, cx.sess.intr())); return @item; } fn nospan<T>(t: T) -> codemap::Spanned<T> { codemap::Spanned { node: t, span: dummy_sp() } } fn
path_node
identifier_name
roulette.component.js
"use strict"; var __decorate = (this && this.__decorate) || function (decorators, target, key, desc) { var c = arguments.length, r = c < 3 ? target : desc === null ? desc = Object.getOwnPropertyDescriptor(target, key) : desc, d; if (typeof Reflect === "object" && typeof Reflect.decorate === "function") r = Reflect.decorate(decorators, target, key, desc); else for (var i = decorators.length - 1; i >= 0; i--) if (d = decorators[i]) r = (c < 3 ? d(r) : c > 3 ? d(target, key, r) : d(target, key)) || r; return c > 3 && r && Object.defineProperty(target, key, r), r; }; var __metadata = (this && this.__metadata) || function (k, v) { if (typeof Reflect === "object" && typeof Reflect.metadata === "function") return Reflect.metadata(k, v); }; Object.defineProperty(exports, "__esModule", { value: true }); var core_1 = require("@angular/core"); var app_service_1 = require("./app.service"); require("./rxjs-extensions"); var RouletteComponent = (function () { function RouletteComponent(services) { this.services = services; this.a = 3000 / (Math.pow(20 * this.services.timba.players.length, 35)); this.totalRounds = 20 * this.services.timba.players.length; this.initialRounds = 10 * this.services.timba.players.length; this.accRounds = 15 * this.services.timba.players.length; } RouletteComponent.prototype.ngOnInit = function () { var _this = this; $("#welcome").css("opacity", "1"); setTimeout(function () { $("#welcome").css("opacity", "0"); setTimeout(function () { _this.addPlayerRoulette(0); _this.addPlayerRouletteFade(0); setTimeout(function () { _this.showAndHide("three"); setTimeout(function () { _this.showAndHide("two"); setTimeout(function () { _this.showAndHide("one"); setTimeout(function () { _this.rotate(_this.services.timba.winnerIndex); setTimeout(function () { _this.services.playing = false; _this.services.nav = 'winner'; }, 24000); }, 2000); }, 2000); }, 2000); }, 500 * _this.services.timba.players.length); }, 1000); }, 4000); }; RouletteComponent.prototype.showAndHide = function (n) { $("#" + n).css("opacity", "1"); setTimeout(function () { $("#" + n).css("opacity", "0"); }, 1000); }; RouletteComponent.prototype.addPlayerRoulette = function (i) { if (i < this.services.timba.players.length) { $("#roulette").append("<div id=\"roulette" + i + "\" class=\"roulette-cell\" style=\"transition:opacity 0.5s ease-in-out;opacity:0;transform: rotate(" + i * 360 / this.services.timba.players.length + "deg) translateX(200px);\">" + this.services.timba.players[i].email + "</div>"); this.addPlayerRoulette(++i); } }; RouletteComponent.prototype.addPlayerRouletteFade = function (i) { var _this = this; setTimeout(function () { if (i < _this.services.timba.players.length) { $("#roulette" + i).css("opacity", "1"); if (_this.services.timba.players[i].email == _this.services.user.email)
_this.addPlayerRouletteFade(++i); } }, 500); }; RouletteComponent.prototype.rotate = function (i) { $("#roulette").css("transition", "transform 20s cubic-bezier(0.2, 0, 0.000000000000000000000000000000000000000001, 1)"); $("#roulette").css("transform", "rotate(" + (4320 - Math.floor(i * 360 / this.services.timba.players.length)) + "deg)"); }; return RouletteComponent; }()); RouletteComponent = __decorate([ core_1.Component({ selector: 'roulette', templateUrl: 'app/roulette.component.html', styleUrls: ['app/roulette.component.css'] }), __metadata("design:paramtypes", [app_service_1.AppService]) ], RouletteComponent); exports.RouletteComponent = RouletteComponent; //# sourceMappingURL=roulette.component.js.map
{ $("#roulette" + i).css("text-shadow", "0 0 10px #fff"); $("#roulette" + i).css("font-weight", "bold"); }
conditional_block
roulette.component.js
"use strict"; var __decorate = (this && this.__decorate) || function (decorators, target, key, desc) { var c = arguments.length, r = c < 3 ? target : desc === null ? desc = Object.getOwnPropertyDescriptor(target, key) : desc, d; if (typeof Reflect === "object" && typeof Reflect.decorate === "function") r = Reflect.decorate(decorators, target, key, desc); else for (var i = decorators.length - 1; i >= 0; i--) if (d = decorators[i]) r = (c < 3 ? d(r) : c > 3 ? d(target, key, r) : d(target, key)) || r; return c > 3 && r && Object.defineProperty(target, key, r), r; }; var __metadata = (this && this.__metadata) || function (k, v) { if (typeof Reflect === "object" && typeof Reflect.metadata === "function") return Reflect.metadata(k, v); }; Object.defineProperty(exports, "__esModule", { value: true }); var core_1 = require("@angular/core"); var app_service_1 = require("./app.service"); require("./rxjs-extensions"); var RouletteComponent = (function () { function
(services) { this.services = services; this.a = 3000 / (Math.pow(20 * this.services.timba.players.length, 35)); this.totalRounds = 20 * this.services.timba.players.length; this.initialRounds = 10 * this.services.timba.players.length; this.accRounds = 15 * this.services.timba.players.length; } RouletteComponent.prototype.ngOnInit = function () { var _this = this; $("#welcome").css("opacity", "1"); setTimeout(function () { $("#welcome").css("opacity", "0"); setTimeout(function () { _this.addPlayerRoulette(0); _this.addPlayerRouletteFade(0); setTimeout(function () { _this.showAndHide("three"); setTimeout(function () { _this.showAndHide("two"); setTimeout(function () { _this.showAndHide("one"); setTimeout(function () { _this.rotate(_this.services.timba.winnerIndex); setTimeout(function () { _this.services.playing = false; _this.services.nav = 'winner'; }, 24000); }, 2000); }, 2000); }, 2000); }, 500 * _this.services.timba.players.length); }, 1000); }, 4000); }; RouletteComponent.prototype.showAndHide = function (n) { $("#" + n).css("opacity", "1"); setTimeout(function () { $("#" + n).css("opacity", "0"); }, 1000); }; RouletteComponent.prototype.addPlayerRoulette = function (i) { if (i < this.services.timba.players.length) { $("#roulette").append("<div id=\"roulette" + i + "\" class=\"roulette-cell\" style=\"transition:opacity 0.5s ease-in-out;opacity:0;transform: rotate(" + i * 360 / this.services.timba.players.length + "deg) translateX(200px);\">" + this.services.timba.players[i].email + "</div>"); this.addPlayerRoulette(++i); } }; RouletteComponent.prototype.addPlayerRouletteFade = function (i) { var _this = this; setTimeout(function () { if (i < _this.services.timba.players.length) { $("#roulette" + i).css("opacity", "1"); if (_this.services.timba.players[i].email == _this.services.user.email) { $("#roulette" + i).css("text-shadow", "0 0 10px #fff"); $("#roulette" + i).css("font-weight", "bold"); } _this.addPlayerRouletteFade(++i); } }, 500); }; RouletteComponent.prototype.rotate = function (i) { $("#roulette").css("transition", "transform 20s cubic-bezier(0.2, 0, 0.000000000000000000000000000000000000000001, 1)"); $("#roulette").css("transform", "rotate(" + (4320 - Math.floor(i * 360 / this.services.timba.players.length)) + "deg)"); }; return RouletteComponent; }()); RouletteComponent = __decorate([ core_1.Component({ selector: 'roulette', templateUrl: 'app/roulette.component.html', styleUrls: ['app/roulette.component.css'] }), __metadata("design:paramtypes", [app_service_1.AppService]) ], RouletteComponent); exports.RouletteComponent = RouletteComponent; //# sourceMappingURL=roulette.component.js.map
RouletteComponent
identifier_name