query stringlengths 5 1.23k | positive stringlengths 53 15.2k | id_ int64 0 252k | task_name stringlengths 87 242 | negative listlengths 20 553 |
|---|---|---|---|---|
Restoring cluster from a given hbase snapshot id | def restore_point ( cls , cluster_id_label , s3_location , backup_id , table_names , overwrite = True , automatic = True ) : conn = Qubole . agent ( version = Cluster . api_version ) parameters = { } parameters [ 's3_location' ] = s3_location parameters [ 'backup_id' ] = backup_id parameters [ 'table_names' ] = table_names parameters [ 'overwrite' ] = overwrite parameters [ 'automatic' ] = automatic return conn . post ( cls . element_path ( cluster_id_label ) + "/restore_point" , data = parameters ) | 251,100 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/cluster.py#L682-L693 | [
"def",
"set_until",
"(",
"self",
",",
"date",
")",
":",
"if",
"isinstance",
"(",
"date",
",",
"datetime",
".",
"date",
")",
"and",
"date",
"<=",
"datetime",
".",
"date",
".",
"today",
"(",
")",
":",
"self",
".",
"arguments",
".",
"update",
"(",
"{"... |
Update for snapshot schedule | def update_snapshot_schedule ( cls , cluster_id_label , s3_location = None , frequency_unit = None , frequency_num = None , status = None ) : conn = Qubole . agent ( version = Cluster . api_version ) data = { } if s3_location is not None : data [ "s3_location" ] = s3_location if frequency_unit is not None : data [ "frequency_unit" ] = frequency_unit if frequency_num is not None : data [ "frequency_num" ] = frequency_num if status is not None : data [ "status" ] = status return conn . put ( cls . element_path ( cluster_id_label ) + "/snapshot_schedule" , data ) | 251,101 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/cluster.py#L704-L719 | [
"def",
"tofile",
"(",
"self",
",",
"filename",
",",
"format",
"=",
"'ascii'",
")",
":",
"if",
"not",
"common",
".",
"is_string",
"(",
"filename",
")",
":",
"raise",
"TypeError",
"(",
"'argument filename must be string but got %s'",
"%",
"(",
"type",
"(",
"fi... |
Purchase options for spot instances . Valid only when slave_request_type is hybrid or spot . | def set_spot_instance_settings ( self , maximum_bid_price_percentage = None , timeout_for_request = None , maximum_spot_instance_percentage = None ) : self . hadoop_settings [ 'spot_instance_settings' ] = { 'maximum_bid_price_percentage' : maximum_bid_price_percentage , 'timeout_for_request' : timeout_for_request , 'maximum_spot_instance_percentage' : maximum_spot_instance_percentage } | 251,102 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/cluster.py#L883-L904 | [
"def",
"upload_cbn_dir",
"(",
"dir_path",
",",
"manager",
")",
":",
"t",
"=",
"time",
".",
"time",
"(",
")",
"for",
"jfg_path",
"in",
"os",
".",
"listdir",
"(",
"dir_path",
")",
":",
"if",
"not",
"jfg_path",
".",
"endswith",
"(",
"'.jgf'",
")",
":",
... |
Purchase options for stable spot instances . | def set_stable_spot_instance_settings ( self , maximum_bid_price_percentage = None , timeout_for_request = None , allow_fallback = True ) : self . hadoop_settings [ 'stable_spot_instance_settings' ] = { 'maximum_bid_price_percentage' : maximum_bid_price_percentage , 'timeout_for_request' : timeout_for_request , 'allow_fallback' : allow_fallback } | 251,103 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/cluster.py#L907-L926 | [
"def",
"delete_topic_groups",
"(",
"self",
",",
"group_id",
",",
"topic_id",
")",
":",
"path",
"=",
"{",
"}",
"data",
"=",
"{",
"}",
"params",
"=",
"{",
"}",
"# REQUIRED - PATH - group_id\r",
"\"\"\"ID\"\"\"",
"path",
"[",
"\"group_id\"",
"]",
"=",
"group_id... |
This method can be used to create the payload which is sent while creating or updating a cluster . | def minimal_payload ( self ) : payload_dict = self . __dict__ payload_dict . pop ( "api_version" , None ) return util . _make_minimal ( payload_dict ) | 251,104 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/cluster.py#L1307-L1314 | [
"async",
"def",
"forget_ticket",
"(",
"self",
",",
"request",
")",
":",
"session",
"=",
"await",
"get_session",
"(",
"request",
")",
"session",
".",
"pop",
"(",
"self",
".",
"cookie_name",
",",
"''",
")"
] |
Raise exceptions in response to any http errors | def _handle_error ( response ) : code = response . status_code if 200 <= code < 400 : return if code == 400 : sys . stderr . write ( response . text + "\n" ) raise BadRequest ( response ) elif code == 401 : sys . stderr . write ( response . text + "\n" ) raise UnauthorizedAccess ( response ) elif code == 403 : sys . stderr . write ( response . text + "\n" ) raise ForbiddenAccess ( response ) elif code == 404 : sys . stderr . write ( response . text + "\n" ) raise ResourceNotFound ( response ) elif code == 405 : sys . stderr . write ( response . text + "\n" ) raise MethodNotAllowed ( response ) elif code == 409 : sys . stderr . write ( response . text + "\n" ) raise ResourceConflict ( response ) elif code == 422 : sys . stderr . write ( response . text + "\n" ) raise ResourceInvalid ( response ) elif code in ( 449 , 502 , 503 , 504 ) : sys . stderr . write ( response . text + "\n" ) raise RetryWithDelay ( response ) elif 401 <= code < 500 : sys . stderr . write ( response . text + "\n" ) raise ClientError ( response ) elif 500 <= code < 600 : sys . stderr . write ( response . text + "\n" ) raise ServerError ( response ) else : raise ConnectionError ( response ) | 251,105 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/connection.py#L111-L165 | [
"def",
"extract_bus_routine",
"(",
"page",
")",
":",
"if",
"not",
"isinstance",
"(",
"page",
",",
"pq",
")",
":",
"page",
"=",
"pq",
"(",
"page",
")",
"stations",
"=",
"extract_stations",
"(",
"page",
")",
"return",
"{",
"# Routine name.",
"'name'",
":",... |
Create a new template . | def createTemplate ( data ) : conn = Qubole . agent ( ) return conn . post ( Template . rest_entity_path , data ) | 251,106 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/template.py#L162-L172 | [
"async",
"def",
"_wait_exponentially",
"(",
"self",
",",
"exception",
",",
"max_wait_time",
"=",
"300",
")",
":",
"wait_time",
"=",
"min",
"(",
"(",
"2",
"**",
"self",
".",
"_connection_attempts",
")",
"+",
"random",
".",
"random",
"(",
")",
",",
"max_wa... |
Edit an existing template . | def editTemplate ( id , data ) : conn = Qubole . agent ( ) return conn . put ( Template . element_path ( id ) , data ) | 251,107 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/template.py#L175-L186 | [
"def",
"configure_splitevaluator",
"(",
"self",
")",
":",
"if",
"self",
".",
"classification",
":",
"speval",
"=",
"javabridge",
".",
"make_instance",
"(",
"\"weka/experiment/ClassifierSplitEvaluator\"",
",",
"\"()V\"",
")",
"else",
":",
"speval",
"=",
"javabridge",... |
View an existing Template details . | def viewTemplate ( id ) : conn = Qubole . agent ( ) return conn . get ( Template . element_path ( id ) ) | 251,108 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/template.py#L204-L215 | [
"def",
"get",
"(",
"self",
")",
":",
"now",
"=",
"time",
".",
"time",
"(",
")",
"while",
"self",
".",
"_clients",
"and",
"self",
".",
"_clients",
"[",
"0",
"]",
"[",
"0",
"]",
"<",
"now",
":",
"_",
",",
"(",
"client",
",",
"last_wait",
")",
"... |
Submit an existing Template . | def submitTemplate ( id , data = { } ) : conn = Qubole . agent ( ) path = str ( id ) + "/run" return conn . post ( Template . element_path ( path ) , data ) | 251,109 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/template.py#L218-L230 | [
"def",
"configure_splitevaluator",
"(",
"self",
")",
":",
"if",
"self",
".",
"classification",
":",
"speval",
"=",
"javabridge",
".",
"make_instance",
"(",
"\"weka/experiment/ClassifierSplitEvaluator\"",
",",
"\"()V\"",
")",
"else",
":",
"speval",
"=",
"javabridge",... |
Run an existing Template and waits for the Result . Prints result to stdout . | def runTemplate ( id , data = { } ) : conn = Qubole . agent ( ) path = str ( id ) + "/run" res = conn . post ( Template . element_path ( path ) , data ) cmdType = res [ 'command_type' ] cmdId = res [ 'id' ] cmdClass = eval ( cmdType ) cmd = cmdClass . find ( cmdId ) while not Command . is_done ( cmd . status ) : time . sleep ( Qubole . poll_interval ) cmd = cmdClass . find ( cmd . id ) return Template . getResult ( cmdClass , cmd ) | 251,110 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/template.py#L233-L255 | [
"def",
"_normalize_options",
"(",
"options",
")",
":",
"normalized_options",
"=",
"{",
"}",
"for",
"key",
",",
"value",
"in",
"iteritems",
"(",
"options",
")",
":",
"optname",
"=",
"str",
"(",
"key",
")",
".",
"lower",
"(",
")",
"intname",
"=",
"INTERN... |
Fetch existing Templates details . | def listTemplates ( data = { } ) : conn = Qubole . agent ( ) url_path = Template . rest_entity_path page_attr = [ ] if "page" in data and data [ "page" ] is not None : page_attr . append ( "page=%s" % data [ "page" ] ) if "per_page" in data and data [ "per_page" ] is not None : page_attr . append ( "per_page=%s" % data [ "per_page" ] ) if page_attr : url_path = "%s?%s" % ( url_path , "&" . join ( page_attr ) ) return conn . get ( url_path ) | 251,111 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/template.py#L269-L288 | [
"def",
"get",
"(",
"self",
")",
":",
"now",
"=",
"time",
".",
"time",
"(",
")",
"while",
"self",
".",
"_clients",
"and",
"self",
".",
"_clients",
"[",
"0",
"]",
"[",
"0",
"]",
"<",
"now",
":",
"_",
",",
"(",
"client",
",",
"last_wait",
")",
"... |
Carefully setup a dict | def edit ( args ) : tap = DbTap . find ( args . id ) options = { } if not args . name is None : options [ "db_name" ] = args . name if args . host is not None : options [ "db_host" ] = args . host if args . user is not None : options [ "db_user" ] = args . user if args . password is not None : options [ "db_passwd" ] = args . password if args . type is not None : options [ "db_type" ] = args . type if args . location is not None : options [ "db_location" ] = args . location if args . port is not None : options [ "port" ] = args . port tap = tap . edit ( * * options ) return json . dumps ( tap . attributes , sort_keys = True , indent = 4 ) | 251,112 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/dbtaps.py#L135-L154 | [
"def",
"get_log_events",
"(",
"awsclient",
",",
"log_group_name",
",",
"log_stream_name",
",",
"start_ts",
"=",
"None",
")",
":",
"client_logs",
"=",
"awsclient",
".",
"get_client",
"(",
"'logs'",
")",
"request",
"=",
"{",
"'logGroupName'",
":",
"log_group_name"... |
Create a new app . | def create ( cls , name , config = None , kind = "spark" ) : conn = Qubole . agent ( ) return conn . post ( cls . rest_entity_path , data = { 'name' : name , 'config' : config , 'kind' : kind } ) | 251,113 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/app.py#L127-L140 | [
"def",
"user_deleted_from_site_event",
"(",
"event",
")",
":",
"userid",
"=",
"event",
".",
"principal",
"catalog",
"=",
"api",
".",
"portal",
".",
"get_tool",
"(",
"'portal_catalog'",
")",
"query",
"=",
"{",
"'object_provides'",
":",
"WORKSPACE_INTERFACE",
"}",... |
Set parameters governing interaction with QDS | def configure ( cls , api_token , api_url = "https://api.qubole.com/api/" , version = "v1.2" , poll_interval = 5 , skip_ssl_cert_check = False , cloud_name = "AWS" ) : cls . _auth = QuboleAuth ( api_token ) cls . api_token = api_token cls . version = version cls . baseurl = api_url if poll_interval < Qubole . MIN_POLL_INTERVAL : log . warn ( "Poll interval cannot be less than %s seconds. Setting it to %s seconds.\n" % ( Qubole . MIN_POLL_INTERVAL , Qubole . MIN_POLL_INTERVAL ) ) cls . poll_interval = Qubole . MIN_POLL_INTERVAL else : cls . poll_interval = poll_interval cls . skip_ssl_cert_check = skip_ssl_cert_check cls . cloud_name = cloud_name . lower ( ) cls . cached_agent = None | 251,114 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/qubole.py#L36-L63 | [
"def",
"after_app_websocket",
"(",
"self",
",",
"func",
":",
"Callable",
")",
"->",
"Callable",
":",
"self",
".",
"record_once",
"(",
"lambda",
"state",
":",
"state",
".",
"app",
".",
"after_websocket",
"(",
"func",
")",
")",
"return",
"func"
] |
Use this to return final minimal request from cluster_info cloud_config or engine_config objects Alternatively call util . _make_minimal if only one object needs to be implemented | def get_cluster_request_parameters ( cluster_info , cloud_config , engine_config ) : cluster_request = { } cloud_config = util . _make_minimal ( cloud_config . __dict__ ) if bool ( cloud_config ) : cluster_request [ 'cloud_config' ] = cloud_config engine_config = util . _make_minimal ( engine_config . __dict__ ) if bool ( engine_config ) : cluster_request [ 'engine_config' ] = engine_config cluster_request . update ( util . _make_minimal ( cluster_info . __dict__ ) ) return cluster_request | 251,115 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/clusterv2.py#L131-L145 | [
"def",
"close",
"(",
"self",
")",
":",
"windll",
".",
"kernel32",
".",
"CloseHandle",
"(",
"self",
".",
"conout_pipe",
")",
"windll",
".",
"kernel32",
".",
"CloseHandle",
"(",
"self",
".",
"conin_pipe",
")"
] |
Create a new cluster using information provided in cluster_info . | def create ( cls , cluster_info ) : conn = Qubole . agent ( version = "v2" ) return conn . post ( cls . rest_entity_path , data = cluster_info ) | 251,116 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/clusterv2.py#L713-L718 | [
"def",
"timeout",
"(",
"limit",
",",
"handler",
")",
":",
"def",
"wrapper",
"(",
"f",
")",
":",
"def",
"wrapped_f",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"old_handler",
"=",
"signal",
".",
"getsignal",
"(",
"signal",
".",
"SIGALRM",
... |
Downloads the contents of all objects in s3_path into fp | def _download_to_local ( boto_conn , s3_path , fp , num_result_dir , delim = None ) : #Progress bar to display download progress def _callback ( downloaded , total ) : '''
Call function for upload.
`downloaded`: File size already downloaded (int)
`total`: Total file size to be downloaded (int)
''' if ( total is 0 ) or ( downloaded == total ) : return progress = downloaded * 100 / total sys . stderr . write ( '\r[{0}] {1}%' . format ( '#' * progress , progress ) ) sys . stderr . flush ( ) m = _URI_RE . match ( s3_path ) bucket_name = m . group ( 1 ) bucket = boto_conn . get_bucket ( bucket_name ) retries = 6 if s3_path . endswith ( '/' ) is False : #It is a file key_name = m . group ( 2 ) key_instance = bucket . get_key ( key_name ) while key_instance is None and retries > 0 : retries = retries - 1 log . info ( "Results file is not available on s3. Retry: " + str ( 6 - retries ) ) time . sleep ( 10 ) key_instance = bucket . get_key ( key_name ) if key_instance is None : raise Exception ( "Results file not available on s3 yet. This can be because of s3 eventual consistency issues." ) log . info ( "Downloading file from %s" % s3_path ) if delim is None : try : key_instance . get_contents_to_file ( fp ) # cb=_callback except boto . exception . S3ResponseError as e : if ( e . status == 403 ) : # SDK-191, boto gives an error while fetching the objects using versions which happens by default # in the get_contents_to_file() api. So attempt one without specifying version. log . warn ( "Access denied while fetching the s3 object. Retrying without specifying the version...." ) key_instance . open ( ) fp . write ( key_instance . read ( ) ) key_instance . close ( ) else : raise else : # Get contents as string. Replace parameters and write to file. _read_iteratively ( key_instance , fp , delim = delim ) else : #It is a folder key_prefix = m . group ( 2 ) bucket_paths = bucket . list ( key_prefix ) for one_path in bucket_paths : name = one_path . name # Eliminate _tmp_ files which ends with $folder$ if name . endswith ( '$folder$' ) : continue log . info ( "Downloading file from %s" % name ) if delim is None : one_path . get_contents_to_file ( fp ) # cb=_callback else : _read_iteratively ( one_path , fp , delim = delim ) | 251,117 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/commands.py#L1415-L1489 | [
"def",
"unregister",
"(",
"self",
",",
"device",
",",
"callback",
")",
":",
"if",
"not",
"device",
":",
"logger",
".",
"error",
"(",
"\"Received an invalid device: %r\"",
",",
"device",
")",
"return",
"logger",
".",
"debug",
"(",
"\"Removing subscription for {}\... |
Cancels command denoted by this id | def cancel_id ( cls , id ) : conn = Qubole . agent ( ) data = { "status" : "kill" } return conn . put ( cls . element_path ( id ) , data ) | 251,118 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/commands.py#L181-L190 | [
"def",
"ensure_compatible_admin",
"(",
"view",
")",
":",
"def",
"wrapper",
"(",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"user_roles",
"=",
"request",
".",
"user",
".",
"user_data",
".",
"get",
"(",
"'roles'",
",",
"[",
"]",
")... |
Fetches log for the command represented by this id | def get_log_id ( cls , id ) : conn = Qubole . agent ( ) r = conn . get_raw ( cls . element_path ( id ) + "/logs" ) return r . text | 251,119 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/commands.py#L199-L208 | [
"def",
"imsave",
"(",
"filename",
",",
"data",
",",
"maxval",
"=",
"None",
",",
"pam",
"=",
"False",
")",
":",
"try",
":",
"netpbm",
"=",
"NetpbmFile",
"(",
"data",
",",
"maxval",
"=",
"maxval",
")",
"netpbm",
".",
"write",
"(",
"filename",
",",
"p... |
Fetches log for the command represented by this object | def get_log ( self ) : log_path = self . meta_data [ 'logs_resource' ] conn = Qubole . agent ( ) r = conn . get_raw ( log_path ) return r . text | 251,120 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/commands.py#L210-L220 | [
"def",
"rejection_sample",
"(",
"n_samples",
",",
"pool_size",
",",
"rng_state",
")",
":",
"result",
"=",
"np",
".",
"empty",
"(",
"n_samples",
",",
"dtype",
"=",
"np",
".",
"int64",
")",
"for",
"i",
"in",
"range",
"(",
"n_samples",
")",
":",
"reject_s... |
Fetches the result for the command represented by this object | def get_results ( self , fp = sys . stdout , inline = True , delim = None , fetch = True , qlog = None , arguments = [ ] ) : result_path = self . meta_data [ 'results_resource' ] conn = Qubole . agent ( ) include_header = "false" if len ( arguments ) == 1 : include_header = arguments . pop ( 0 ) if include_header not in ( 'true' , 'false' ) : raise ParseError ( "incude_header can be either true or false" ) r = conn . get ( result_path , { 'inline' : inline , 'include_headers' : include_header } ) if r . get ( 'inline' ) : raw_results = r [ 'results' ] encoded_results = raw_results . encode ( 'utf8' ) if sys . version_info < ( 3 , 0 , 0 ) : fp . write ( encoded_results ) else : import io if isinstance ( fp , io . TextIOBase ) : if hasattr ( fp , 'buffer' ) : fp . buffer . write ( encoded_results ) else : fp . write ( raw_results ) elif isinstance ( fp , io . BufferedIOBase ) or isinstance ( fp , io . RawIOBase ) : fp . write ( encoded_results ) else : # Can this happen? Don't know what's the right thing to do in this case. pass else : if fetch : storage_credentials = conn . get ( Account . credentials_rest_entity_path ) if storage_credentials [ 'region_endpoint' ] is not None : boto_conn = boto . connect_s3 ( aws_access_key_id = storage_credentials [ 'storage_access_key' ] , aws_secret_access_key = storage_credentials [ 'storage_secret_key' ] , security_token = storage_credentials [ 'session_token' ] , host = storage_credentials [ 'region_endpoint' ] ) else : boto_conn = boto . connect_s3 ( aws_access_key_id = storage_credentials [ 'storage_access_key' ] , aws_secret_access_key = storage_credentials [ 'storage_secret_key' ] , security_token = storage_credentials [ 'session_token' ] ) log . info ( "Starting download from result locations: [%s]" % "," . join ( r [ 'result_location' ] ) ) #fetch latest value of num_result_dir num_result_dir = Command . find ( self . id ) . num_result_dir # If column/header names are not able to fetch then use include header as true if include_header . lower ( ) == "true" and qlog is not None : write_headers ( qlog , fp ) for s3_path in r [ 'result_location' ] : # In Python 3, # If the delim is None, fp should be in binary mode because # boto expects it to be. # If the delim is not None, then both text and binary modes # work. _download_to_local ( boto_conn , s3_path , fp , num_result_dir , delim = delim ) else : fp . write ( "," . join ( r [ 'result_location' ] ) ) | 251,121 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/commands.py#L257-L333 | [
"def",
"ensure_compatible_admin",
"(",
"view",
")",
":",
"def",
"wrapper",
"(",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"user_roles",
"=",
"request",
".",
"user",
".",
"user_data",
".",
"get",
"(",
"'roles'",
",",
"[",
"]",
")... |
Convert singular word to its plural form . | def pluralize ( singular ) : if singular in UNCOUNTABLES : return singular for i in IRREGULAR : if i [ 0 ] == singular : return i [ 1 ] for i in PLURALIZE_PATTERNS : if re . search ( i [ 0 ] , singular ) : return re . sub ( i [ 0 ] , i [ 1 ] , singular ) | 251,122 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/util.py#L85-L101 | [
"def",
"shutdown",
"(",
"self",
")",
":",
"self",
".",
"stop_all_periodic_tasks",
"(",
")",
"for",
"channel",
"in",
"self",
".",
"_bcm_sockets",
":",
"log",
".",
"debug",
"(",
"\"Closing bcm socket for channel {}\"",
".",
"format",
"(",
"channel",
")",
")",
... |
Convert plural word to its singular form . | def singularize ( plural ) : if plural in UNCOUNTABLES : return plural for i in IRREGULAR : if i [ 1 ] == plural : return i [ 0 ] for i in SINGULARIZE_PATTERNS : if re . search ( i [ 0 ] , plural ) : return re . sub ( i [ 0 ] , i [ 1 ] , plural ) return plural | 251,123 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/util.py#L104-L120 | [
"def",
"data_storage_dir",
"(",
"self",
")",
":",
"if",
"self",
".",
"_data_storage_dir",
"is",
"None",
":",
"self",
".",
"_data_storage_dir",
"=",
"tempfile",
".",
"mkdtemp",
"(",
"prefix",
"=",
"\"openfisca_\"",
")",
"log",
".",
"warn",
"(",
"(",
"\"Inte... |
Convert a word from lower_with_underscores to CamelCase . | def camelize ( word ) : return '' . join ( w [ 0 ] . upper ( ) + w [ 1 : ] for w in re . sub ( '[^A-Z^a-z^0-9^:]+' , ' ' , word ) . split ( ' ' ) ) | 251,124 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/util.py#L123-L132 | [
"def",
"save",
"(",
"self",
")",
":",
"self",
".",
"session",
".",
"add",
"(",
"self",
")",
"self",
".",
"session",
".",
"flush",
"(",
")",
"return",
"self"
] |
This function removes all the keys whose value is either None or an empty dictionary . | def _make_minimal ( dictionary ) : new_dict = { } for key , value in dictionary . items ( ) : if value is not None : if isinstance ( value , dict ) : new_value = _make_minimal ( value ) if new_value : new_dict [ key ] = new_value else : new_dict [ key ] = value return new_dict | 251,125 | https://github.com/qubole/qds-sdk-py/blob/77210fb64e5a7d567aedeea3b742a1d872fd0e5e/qds_sdk/util.py#L146-L160 | [
"def",
"prepare",
"(",
"self",
")",
":",
"request_time",
"=",
"1000.0",
"*",
"self",
".",
"request",
".",
"request_time",
"(",
")",
"access_log",
".",
"info",
"(",
"\"%d %s %.2fms\"",
",",
"self",
".",
"get_status",
"(",
")",
",",
"self",
".",
"_request_... |
Uploads a profiler report to IOpipe | def upload_profiler_report ( url , filename , config ) : try : logger . debug ( "Uploading profiler report to IOpipe" ) with open ( filename , "rb" ) as data : response = requests . put ( url , data = data , timeout = config [ "network_timeout" ] ) response . raise_for_status ( ) except Exception as e : logger . debug ( "Error while uploading profiler report: %s" , e ) if hasattr ( e , "response" ) : logger . debug ( e . response . content ) else : logger . debug ( "Profiler report uploaded successfully" ) finally : if os . path . isfile ( filename ) : os . remove ( filename ) | 251,126 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/contrib/profiler/request.py#L12-L33 | [
"def",
"delete",
"(",
"self",
",",
"bucket",
"=",
"None",
",",
"key",
"=",
"None",
",",
"version_id",
"=",
"None",
",",
"upload_id",
"=",
"None",
",",
"uploads",
"=",
"None",
")",
":",
"if",
"upload_id",
"is",
"not",
"None",
":",
"return",
"self",
... |
Mocks read_pid_stat as this is a Linux - specific operation . | def read_pid_stat ( pid ) : return { "utime" : random . randint ( 0 , 999999999 ) , "stime" : random . randint ( 0 , 999999999 ) , "cutime" : random . randint ( 0 , 999999999 ) , "cstime" : random . randint ( 0 , 999999999 ) , } | 251,127 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/mock_system.py#L40-L49 | [
"def",
"create_api_call",
"(",
"func",
",",
"settings",
")",
":",
"def",
"base_caller",
"(",
"api_call",
",",
"_",
",",
"*",
"args",
")",
":",
"\"\"\"Simply call api_call and ignore settings.\"\"\"",
"return",
"api_call",
"(",
"*",
"args",
")",
"def",
"inner",
... |
Mocks read_stat as this is a Linux - specific operation . | def read_stat ( ) : return [ { "times" : { "user" : random . randint ( 0 , 999999999 ) , "nice" : random . randint ( 0 , 999999999 ) , "sys" : random . randint ( 0 , 999999999 ) , "idle" : random . randint ( 0 , 999999999 ) , "irq" : random . randint ( 0 , 999999999 ) , } } ] | 251,128 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/mock_system.py#L63-L77 | [
"def",
"create_api_call",
"(",
"func",
",",
"settings",
")",
":",
"def",
"base_caller",
"(",
"api_call",
",",
"_",
",",
"*",
"args",
")",
":",
"\"\"\"Simply call api_call and ignore settings.\"\"\"",
"return",
"api_call",
"(",
"*",
"args",
")",
"def",
"inner",
... |
Loads plugins that match the Plugin interface and are instantiated . | def load_plugins ( self , plugins ) : def instantiate ( plugin ) : return plugin ( ) if inspect . isclass ( plugin ) else plugin loaded_plugins = [ ] plugins_seen = [ ] # Iterate over plugins in reverse to permit users to override default # plugin config for plugin in reversed ( plugins ) : if not is_plugin ( plugin ) or plugin . name in plugins_seen : continue # Build the plugins list in reverse to restore original order loaded_plugins . insert ( 0 , instantiate ( plugin ) ) plugins_seen . append ( plugin . name ) return loaded_plugins | 251,129 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/agent.py#L198-L220 | [
"def",
"remove_cookie",
"(",
"self",
",",
"cookie_name",
")",
":",
"if",
"self",
".",
"__ro_flag",
":",
"raise",
"RuntimeError",
"(",
"'Read-only cookie-jar changing attempt'",
")",
"if",
"cookie_name",
"in",
"self",
".",
"__cookies",
".",
"keys",
"(",
")",
":... |
Runs plugin hooks for each registered plugin . | def run_hooks ( self , name , event = None , context = None ) : hooks = { "pre:setup" : lambda p : p . pre_setup ( self ) , "post:setup" : lambda p : p . post_setup ( self ) , "pre:invoke" : lambda p : p . pre_invoke ( event , context ) , "post:invoke" : lambda p : p . post_invoke ( event , context ) , "pre:report" : lambda p : p . pre_report ( self . report ) , "post:report" : lambda p : p . post_report ( self . report ) , } if name in hooks : for p in self . plugins : if p . enabled : try : hooks [ name ] ( p ) except Exception as e : logger . error ( "IOpipe plugin %s hook raised error" % ( name , str ( e ) ) ) logger . exception ( e ) | 251,130 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/agent.py#L222-L244 | [
"def",
"ReleaseFileObject",
"(",
"self",
",",
"file_object",
")",
":",
"identifier",
",",
"cache_value",
"=",
"self",
".",
"_file_object_cache",
".",
"GetCacheValueByObject",
"(",
"file_object",
")",
"if",
"not",
"identifier",
":",
"raise",
"RuntimeError",
"(",
... |
Wait for all futures to complete . This should be done at the end of an an invocation . | def wait_for_futures ( self ) : [ future for future in futures . as_completed ( self . futures ) ] self . futures = [ ] | 251,131 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/agent.py#L260-L266 | [
"def",
"_local_update",
"(",
"self",
",",
"rdict",
")",
":",
"sanitized",
"=",
"self",
".",
"_check_keys",
"(",
"rdict",
")",
"temp_meta",
"=",
"self",
".",
"_meta_data",
"self",
".",
"__dict__",
"=",
"sanitized",
"self",
".",
"_meta_data",
"=",
"temp_meta... |
Checks to see if we re working with a valid lambda context object . | def validate_context ( self , context ) : return all ( [ hasattr ( context , attr ) for attr in [ "aws_request_id" , "function_name" , "function_version" , "get_remaining_time_in_millis" , "invoked_function_arn" , "log_group_name" , "log_stream_name" , "memory_limit_in_mb" , ] ] ) and callable ( context . get_remaining_time_in_millis ) | 251,132 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/agent.py#L268-L289 | [
"def",
"supplementary_material",
"(",
"soup",
")",
":",
"supplementary_material",
"=",
"[",
"]",
"supplementary_material_tags",
"=",
"raw_parser",
".",
"supplementary_material",
"(",
"soup",
")",
"position",
"=",
"1",
"for",
"tag",
"in",
"supplementary_material_tags",... |
Monkey patches requests Session class if available . Overloads the send method to add tracing and metrics collection . | def patch_session_send ( context , http_filter ) : if Session is None : return def send ( self , * args , * * kwargs ) : id = ensure_utf8 ( str ( uuid . uuid4 ( ) ) ) with context . iopipe . mark ( id ) : response = original_session_send ( self , * args , * * kwargs ) trace = context . iopipe . mark . measure ( id ) context . iopipe . mark . delete ( id ) collect_metrics_for_response ( response , context , trace , http_filter ) return response Session . send = send | 251,133 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/contrib/trace/auto_http.py#L63-L80 | [
"def",
"ParentAndBaseName",
"(",
"path",
")",
":",
"dirname",
",",
"basename",
"=",
"os",
".",
"path",
".",
"split",
"(",
"path",
")",
"dirname",
"=",
"dirname",
".",
"rstrip",
"(",
"os",
".",
"path",
".",
"sep",
")",
"if",
"os",
".",
"path",
".",
... |
Monkey patches botocore s vendored requests if available . Overloads the Session class send method to add tracing and metric collection . | def patch_botocore_session_send ( context , http_filter ) : if BotocoreSession is None : return def send ( self , * args , * * kwargs ) : id = str ( uuid . uuid4 ( ) ) with context . iopipe . mark ( id ) : response = original_botocore_session_send ( self , * args , * * kwargs ) trace = context . iopipe . mark . measure ( id ) context . iopipe . mark . delete ( id ) collect_metrics_for_response ( response , context , trace , http_filter ) return response BotocoreSession . send = send | 251,134 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/contrib/trace/auto_http.py#L83-L100 | [
"def",
"ParentAndBaseName",
"(",
"path",
")",
":",
"dirname",
",",
"basename",
"=",
"os",
".",
"path",
".",
"split",
"(",
"path",
")",
"dirname",
"=",
"dirname",
".",
"rstrip",
"(",
"os",
".",
"path",
".",
"sep",
")",
"if",
"os",
".",
"path",
".",
... |
Collects relevant metrics from a requests Response object and adds them to the IOpipe context . | def collect_metrics_for_response ( http_response , context , trace , http_filter ) : http_response = copy . deepcopy ( http_response ) if http_filter is not None and callable ( http_filter ) : http_response = http_filter ( http_response ) if http_response is False : return request = None if hasattr ( http_response , "request" ) : parsed_url = None if hasattr ( http_response . request , "url" ) : parsed_url = urlparse ( http_response . request . url ) request_headers = [ ] if hasattr ( http_response . request , "headers" ) : request_headers = [ { "key" : ensure_utf8 ( k ) , "string" : ensure_utf8 ( v ) } for k , v in http_response . request . headers . items ( ) if k . lower ( ) in INCLUDE_HEADERS ] request = Request ( hash = ensure_utf8 ( getattr ( parsed_url , "fragment" , None ) ) , headers = request_headers , hostname = ensure_utf8 ( getattr ( parsed_url , "hostname" , None ) ) , method = ensure_utf8 ( getattr ( http_response . request , "method" , None ) ) , path = ensure_utf8 ( getattr ( parsed_url , "path" , None ) ) , # TODO: Determine if this is redundant pathname = ensure_utf8 ( getattr ( parsed_url , "path" , None ) ) , port = ensure_utf8 ( getattr ( parsed_url , "port" , None ) ) , protocol = ensure_utf8 ( getattr ( parsed_url , "scheme" , None ) ) , query = ensure_utf8 ( getattr ( parsed_url , "query" , None ) ) , url = ensure_utf8 ( getattr ( http_response . request , "url" , None ) ) , ) response_headers = [ ] if hasattr ( http_response , "headers" ) : response_headers = [ { "key" : ensure_utf8 ( k ) , "string" : ensure_utf8 ( v ) } for k , v in http_response . headers . items ( ) if k . lower ( ) in INCLUDE_HEADERS ] response = Response ( headers = response_headers , statusCode = ensure_utf8 ( getattr ( http_response , "status_code" , None ) ) , statusMessage = None , ) context . iopipe . mark . http_trace ( trace , request , response ) | 251,135 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/contrib/trace/auto_http.py#L125-L178 | [
"def",
"_make_variant_locus_id",
"(",
"self",
",",
"gene_id",
",",
"disease_id",
")",
":",
"alt_locus_id",
"=",
"'_:'",
"+",
"re",
".",
"sub",
"(",
"r':'",
",",
"''",
",",
"gene_id",
")",
"+",
"'-'",
"+",
"re",
".",
"sub",
"(",
"r':'",
",",
"''",
"... |
Returns meta data about plugins . | def get_plugin_meta ( plugins ) : return [ { "name" : p . name , "version" : p . version , "homepage" : p . homepage , "enabled" : p . enabled , } for p in plugins if is_plugin ( p ) ] | 251,136 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/plugins.py#L4-L22 | [
"def",
"_wrap_client_error",
"(",
"e",
")",
":",
"error_code",
"=",
"e",
".",
"response",
"[",
"'Error'",
"]",
"[",
"'Code'",
"]",
"message",
"=",
"e",
".",
"response",
"[",
"'Error'",
"]",
"[",
"'Message'",
"]",
"if",
"error_code",
"==",
"'BadRequestExc... |
Returns true if the plugin implements the Plugin interface . | def is_plugin ( plugin ) : try : return isinstance ( plugin , Plugin ) or issubclass ( plugin , Plugin ) except TypeError : return False | 251,137 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/plugins.py#L25-L36 | [
"def",
"save_binary",
"(",
"self",
",",
"filename",
")",
":",
"_safe_call",
"(",
"_LIB",
".",
"LGBM_DatasetSaveBinary",
"(",
"self",
".",
"construct",
"(",
")",
".",
"handle",
",",
"c_str",
"(",
"filename",
")",
")",
")",
"return",
"self"
] |
Python 2 and 3 compatible way to do meta classes | def with_metaclass ( meta , * bases ) : class metaclass ( meta ) : def __new__ ( cls , name , this_bases , d ) : return meta ( name , bases , d ) return type . __new__ ( metaclass , "temporary_class" , ( ) , { } ) | 251,138 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/plugins.py#L39-L46 | [
"def",
"_sections_to_variance_sections",
"(",
"self",
",",
"sections_over_time",
")",
":",
"variance_sections",
"=",
"[",
"]",
"for",
"i",
"in",
"range",
"(",
"len",
"(",
"sections_over_time",
"[",
"0",
"]",
")",
")",
":",
"time_sections",
"=",
"[",
"section... |
Returns the contents of a AWS Lambda context . | def extract_context_data ( self ) : data = { } for k , v in { # camel case names in the report to align with AWS standards "functionName" : "function_name" , "functionVersion" : "function_version" , "memoryLimitInMB" : "memory_limit_in_mb" , "invokedFunctionArn" : "invoked_function_arn" , "awsRequestId" : "aws_request_id" , "logGroupName" : "log_group_name" , "logStreamName" : "log_stream_name" , } . items ( ) : if hasattr ( self . context , v ) : data [ k ] = getattr ( self . context , v ) if ( hasattr ( self . context , "invoked_function_arn" ) and "AWS_SAM_LOCAL" in os . environ ) : data [ "invokedFunctionArn" ] = ( "arn:aws:lambda:local:0:function:%s" % data . get ( "functionName" , "unknown" ) ) if hasattr ( self . context , "get_remaining_time_in_millis" ) and callable ( self . context . get_remaining_time_in_millis ) : data [ "getRemainingTimeInMillis" ] = self . context . get_remaining_time_in_millis ( ) data [ "traceId" ] = os . getenv ( "_X_AMZN_TRACE_ID" , "" ) return data | 251,139 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/report.py#L82-L117 | [
"def",
"_strip_ctype",
"(",
"name",
",",
"ctype",
",",
"protocol",
"=",
"2",
")",
":",
"# parse channel type from name (e.g. 'L1:GDS-CALIB_STRAIN,reduced')",
"try",
":",
"name",
",",
"ctypestr",
"=",
"name",
".",
"rsplit",
"(",
"','",
",",
"1",
")",
"except",
... |
Adds details of an error to the report . | def retain_error ( self , error , frame = None ) : if frame is None : stack = traceback . format_exc ( ) self . labels . add ( "@iopipe/error" ) else : stack = "\n" . join ( traceback . format_stack ( frame ) ) self . labels . add ( "@iopipe/timeout" ) details = { "name" : type ( error ) . __name__ , "message" : "{}" . format ( error ) , "stack" : stack , } self . report [ "errors" ] = details | 251,140 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/report.py#L119-L136 | [
"def",
"_spintaylor_aligned_prec_swapper",
"(",
"*",
"*",
"p",
")",
":",
"orig_approximant",
"=",
"p",
"[",
"'approximant'",
"]",
"if",
"p",
"[",
"'spin2x'",
"]",
"==",
"0",
"and",
"p",
"[",
"'spin2y'",
"]",
"==",
"0",
"and",
"p",
"[",
"'spin1x'",
"]",... |
Prepare the report to be sent to IOpipe . | def prepare ( self , error = None , frame = None ) : if error : self . retain_error ( error , frame ) self . report [ "environment" ] [ "host" ] [ "boot_id" ] = system . read_bootid ( ) # convert labels to list for sending self . report [ "labels" ] = list ( self . labels ) meminfo = system . read_meminfo ( ) self . report . update ( { "aws" : self . extract_context_data ( ) , "timestampEnd" : int ( time . time ( ) * 1000 ) , } ) self . report [ "environment" ] [ "os" ] . update ( { "cpus" : system . read_stat ( ) , "freemem" : meminfo [ "MemFree" ] , "hostname" : system . read_hostname ( ) , "totalmem" : meminfo [ "MemTotal" ] , "usedmem" : meminfo [ "MemTotal" ] - meminfo [ "MemFree" ] , } ) self . report [ "environment" ] [ "os" ] [ "linux" ] [ "pid" ] = { "self" : { "stat" : system . read_pid_stat ( "self" ) , "stat_start" : self . stat_start , "status" : system . read_pid_status ( "self" ) , } } self . report [ "disk" ] = system . read_disk ( ) self . report [ "duration" ] = int ( ( monotonic ( ) - self . start_time ) * 1e9 ) | 251,141 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/report.py#L138-L182 | [
"def",
"copy",
"(",
"chain",
":",
"MiningChain",
")",
"->",
"MiningChain",
":",
"if",
"not",
"isinstance",
"(",
"chain",
",",
"MiningChain",
")",
":",
"raise",
"ValidationError",
"(",
"\"`at_block_number` may only be used with 'MiningChain\"",
")",
"base_db",
"=",
... |
Sends the report to IOpipe . | def send ( self ) : if self . sent is True : return self . sent = True logger . debug ( "Sending report to IOpipe:" ) logger . debug ( json . dumps ( self . report , indent = 2 , sort_keys = True ) ) self . client . submit_future ( send_report , copy . deepcopy ( self . report ) , self . config ) | 251,142 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/report.py#L184-L195 | [
"def",
"delete",
"(",
"self",
",",
"msg",
",",
"claim_id",
"=",
"None",
")",
":",
"msg_id",
"=",
"utils",
".",
"get_id",
"(",
"msg",
")",
"if",
"claim_id",
":",
"uri",
"=",
"\"/%s/%s?claim_id=%s\"",
"%",
"(",
"self",
".",
"uri_base",
",",
"msg_id",
"... |
Sends the report to IOpipe s collector . | def send_report ( report , config ) : headers = { "Authorization" : "Bearer {}" . format ( config [ "token" ] ) } url = "https://{host}{path}" . format ( * * config ) try : response = session . post ( url , json = report , headers = headers , timeout = config [ "network_timeout" ] ) response . raise_for_status ( ) except Exception as e : logger . debug ( "Error sending report to IOpipe: %s" % e ) else : logger . debug ( "Report sent to IOpipe successfully" ) | 251,143 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/send_report.py#L12-L30 | [
"def",
"cluster_types",
"(",
"types",
",",
"max_clust",
"=",
"12",
")",
":",
"if",
"len",
"(",
"types",
")",
"<",
"max_clust",
":",
"max_clust",
"=",
"len",
"(",
"types",
")",
"# Do actual clustering",
"cluster_dict",
"=",
"do_clustering",
"(",
"types",
",... |
Uploads log data to IOpipe . | def upload_log_data ( url , stream_or_file , config ) : try : logger . debug ( "Uploading log data to IOpipe" ) if isinstance ( stream_or_file , StringIO ) : stream_or_file . seek ( 0 ) response = requests . put ( url , data = stream_or_file , timeout = config [ "network_timeout" ] ) else : with open ( stream_or_file , "rb" ) as data : response = requests . put ( url , data = data , timeout = config [ "network_timeout" ] ) response . raise_for_status ( ) except Exception as e : logger . debug ( "Error while uploading log data: %s" , e ) logger . exception ( e ) if hasattr ( e , "response" ) and hasattr ( e . response , "content" ) : logger . debug ( e . response . content ) else : logger . debug ( "Log data uploaded successfully" ) finally : if isinstance ( stream_or_file , str ) and os . path . exists ( stream_or_file ) : os . remove ( stream_or_file ) | 251,144 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/contrib/logger/request.py#L14-L44 | [
"def",
"get_response_content_type",
"(",
"self",
")",
":",
"if",
"self",
".",
"_best_response_match",
"is",
"None",
":",
"settings",
"=",
"get_settings",
"(",
"self",
".",
"application",
",",
"force_instance",
"=",
"True",
")",
"acceptable",
"=",
"headers",
".... |
Returns the IOpipe signer hostname for a region | def get_signer_hostname ( ) : region = os . getenv ( "AWS_REGION" , "" ) region = region if region and region in SUPPORTED_REGIONS else "us-west-2" return "signer.{region}.iopipe.com" . format ( region = region ) | 251,145 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/signer.py#L15-L24 | [
"def",
"subdivide",
"(",
"self",
",",
"N",
"=",
"1",
",",
"method",
"=",
"0",
")",
":",
"triangles",
"=",
"vtk",
".",
"vtkTriangleFilter",
"(",
")",
"triangles",
".",
"SetInputData",
"(",
"self",
".",
"polydata",
"(",
")",
")",
"triangles",
".",
"Upd... |
Returns a signed request URL from IOpipe | def get_signed_request ( config , context , extension ) : url = "https://{hostname}/" . format ( hostname = get_signer_hostname ( ) ) try : logger . debug ( "Requesting signed request URL from %s" , url ) response = requests . post ( url , json = { "arn" : context . invoked_function_arn , "requestId" : context . aws_request_id , "timestamp" : int ( time . time ( ) * 1000 ) , "extension" : extension , } , headers = { "Authorization" : config [ "token" ] } , timeout = config [ "network_timeout" ] , ) response . raise_for_status ( ) except Exception as e : logger . debug ( "Error requesting signed request URL: %s" , e ) if hasattr ( e , "response" ) : logger . debug ( e . response . content ) else : response = response . json ( ) logger . debug ( "Signed request URL received for %s" , response [ "url" ] ) return response | 251,146 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/signer.py#L27-L60 | [
"def",
"validate_categories_equal_entities",
"(",
"categories",
",",
"entities",
",",
"message",
"=",
"None",
",",
"exception",
"=",
"MatrixError",
")",
":",
"nb_categories",
"=",
"len",
"(",
"categories",
")",
"nb_entities",
"=",
"len",
"(",
"entities",
")",
... |
Sample pure Lambda function | def handler ( event , context ) : try : ip = requests . get ( "http://checkip.amazonaws.com/" ) except requests . RequestException as e : # Send some context about this error to Lambda Logs print ( e ) raise e return { "statusCode" : 200 , "body" : json . dumps ( { "message" : "hello world" , "location" : ip . text . replace ( "\n" , "" ) } ) , } | 251,147 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/acceptance/serverless-layers/handler.py#L6-L87 | [
"def",
"get_url_repair_patterns",
"(",
")",
":",
"file_types_list",
"=",
"[",
"ur'h\\s*t\\s*m'",
",",
"# htm",
"ur'h\\s*t\\s*m\\s*l'",
",",
"# html",
"ur't\\s*x\\s*t'",
"# txt",
"ur'p\\s*h\\s*p'",
"# php",
"ur'a\\s*s\\s*p\\s*'",
"# asp",
"ur'j\\s*s\\s*p'",
",",
"# jsp",
... |
Returns system memory usage information . | def read_meminfo ( ) : data = { } with open ( "/proc/meminfo" , "rb" ) as meminfo_file : for row in meminfo_file : fields = row . split ( ) # Example content: # MemTotal: 3801016 kB # MemFree: 1840972 kB # MemAvailable: 3287752 kB # HugePages_Total: 0 data [ fields [ 0 ] . decode ( "ascii" ) [ : - 1 ] ] = int ( fields [ 1 ] ) * 1024 return data | 251,148 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/system.py#L50-L67 | [
"def",
"build_news",
"(",
"ctx",
",",
"draft",
"=",
"False",
",",
"yes",
"=",
"False",
")",
":",
"report",
".",
"info",
"(",
"ctx",
",",
"\"docs.build-news\"",
",",
"\"building changelog from news fragments\"",
")",
"build_command",
"=",
"f\"towncrier --version {c... |
Returns system process stat information . | def read_pid_stat ( pid = "self" ) : with open ( "/proc/%s/stat" % ( pid , ) , "rb" ) as f : stat = f . readline ( ) . split ( ) return { "utime" : int ( stat [ 13 ] ) , "stime" : int ( stat [ 14 ] ) , "cutime" : int ( stat [ 15 ] ) , "cstime" : int ( stat [ 16 ] ) , } | 251,149 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/system.py#L70-L85 | [
"def",
"create",
"(",
"self",
",",
"image",
",",
"geometry",
",",
"options",
")",
":",
"image",
"=",
"self",
".",
"cropbox",
"(",
"image",
",",
"geometry",
",",
"options",
")",
"image",
"=",
"self",
".",
"orientation",
"(",
"image",
",",
"geometry",
... |
Returns the system process sstatus . | def read_pid_status ( pid = "self" ) : data = { } with open ( "/proc/%s/status" % ( pid , ) , "rb" ) as status_file : for row in status_file : fields = row . split ( ) if fields and fields [ 0 ] in [ b"VmRSS:" , b"Threads:" , b"FDSize:" ] : try : data [ fields [ 0 ] . decode ( "ascii" ) [ : - 1 ] ] = int ( fields [ 1 ] ) except ValueError : data [ fields [ 0 ] . decode ( "ascii" ) [ : - 1 ] ] = fields [ 1 ] . decode ( "ascii" ) return data | 251,150 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/system.py#L88-L105 | [
"def",
"delete",
"(",
"request",
",",
"obj_id",
"=",
"None",
")",
":",
"data",
"=",
"request",
".",
"DELETE",
"or",
"json",
".",
"loads",
"(",
"request",
".",
"body",
")",
"guids",
"=",
"data",
".",
"get",
"(",
"'guids'",
")",
".",
"split",
"(",
... |
Returns the system stat information . | def read_stat ( ) : data = [ ] with open ( "/proc/stat" , "rb" ) as stat_file : for line in stat_file : cpu_stat = line . split ( ) if cpu_stat [ 0 ] [ : 3 ] != b"cpu" : break # First cpu line is aggregation of following lines, skip it if len ( cpu_stat [ 0 ] ) == 3 : continue data . append ( { "times" : { "user" : int ( cpu_stat [ 1 ] ) , "nice" : int ( cpu_stat [ 2 ] ) , "sys" : int ( cpu_stat [ 3 ] ) , "idle" : int ( cpu_stat [ 4 ] ) , "irq" : int ( cpu_stat [ 6 ] ) , } } ) return data | 251,151 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/system.py#L108-L135 | [
"def",
"remove_binary_support",
"(",
"self",
",",
"api_id",
",",
"cors",
"=",
"False",
")",
":",
"response",
"=",
"self",
".",
"apigateway_client",
".",
"get_rest_api",
"(",
"restApiId",
"=",
"api_id",
")",
"if",
"\"binaryMediaTypes\"",
"in",
"response",
"and"... |
Returns IOpipe configuration options setting defaults as necessary . | def set_config ( * * config ) : config . setdefault ( "debug" , bool ( strtobool ( os . getenv ( "IOPIPE_DEBUG" , "false" ) ) ) ) config . setdefault ( "enabled" , bool ( strtobool ( os . getenv ( "IOPIPE_ENABLED" , "true" ) ) ) ) config . setdefault ( "host" , get_hostname ( ) ) config . setdefault ( "install_method" , os . getenv ( "IOPIPE_INSTALL_METHOD" , "manual" ) ) config . setdefault ( "network_timeout" , os . getenv ( "IOPIPE_NETWORK_TIMEOUT" , 5000 ) ) config . setdefault ( "path" , get_collector_path ( ) ) config . setdefault ( "plugins" , [ ] ) config . setdefault ( "sync_http" , False ) config . setdefault ( "timeout_window" , os . getenv ( "IOPIPE_TIMEOUT_WINDOW" , 500 ) ) config . setdefault ( "token" , os . getenv ( "IOPIPE_TOKEN" ) or os . getenv ( "IOPIPE_CLIENTID" ) or "" ) if "client_id" in config : config [ "token" ] = config . pop ( "client_id" ) if "url" in config : url = config . pop ( "url" ) config [ "host" ] = get_hostname ( url ) config [ "path" ] = get_collector_path ( url ) if "." in str ( config [ "network_timeout" ] ) : warnings . warn ( "IOpipe's 'network_timeout' is now in milliseconds, expressed as an integer" ) try : config [ "debug" ] = bool ( config [ "debug" ] ) except ValueError : config [ "debug" ] = False try : config [ "network_timeout" ] = int ( config [ "network_timeout" ] ) / 1000.0 except ValueError : config [ "network_timeout" ] = 5.0 if "." in str ( config [ "timeout_window" ] ) : warnings . warn ( "IOpipe's 'timeout_window' is now in milliseconds, expressed as an integer" ) try : config [ "timeout_window" ] = int ( config [ "timeout_window" ] ) / 1000.0 except ValueError : config [ "timeout_window" ] = 0.5 return config | 251,152 | https://github.com/iopipe/iopipe-python/blob/4eb653977341bc67f8b1b87aedb3aaaefc25af61/iopipe/config.py#L8-L59 | [
"def",
"write_bits",
"(",
"self",
",",
"*",
"args",
")",
":",
"# Would be nice to make this a bit smarter",
"if",
"len",
"(",
"args",
")",
">",
"8",
":",
"raise",
"ValueError",
"(",
"\"Can only write 8 bits at a time\"",
")",
"self",
".",
"_output_buffer",
".",
... |
Encodes Unicode strings to byte strings if necessary . | def b ( s ) : return s if isinstance ( s , bytes ) else s . encode ( locale . getpreferredencoding ( ) ) | 251,153 | https://github.com/pief/python-netsnmpagent/blob/b1aad1c7f034509c40d9ab17d59be32e809bd31d/netsnmpapi.py#L16-L19 | [
"def",
"stop_experiment",
"(",
"args",
")",
":",
"experiment_id_list",
"=",
"parse_ids",
"(",
"args",
")",
"if",
"experiment_id_list",
":",
"experiment_config",
"=",
"Experiments",
"(",
")",
"experiment_dict",
"=",
"experiment_config",
".",
"get_all_experiments",
"(... |
Writes a formatted log message with a timestamp to stdout . | def LogMsg ( msg ) : global headerlogged if headerlogged == 0 : print ( "{0:<8} {1:<90} {2}" . format ( "Time" , "MainThread" , "UpdateSNMPObjsThread" ) ) print ( "{0:-^120}" . format ( "-" ) ) headerlogged = 1 threadname = threading . currentThread ( ) . name funcname = sys . _getframe ( 1 ) . f_code . co_name if funcname == "<module>" : funcname = "Main code path" elif funcname == "LogNetSnmpMsg" : funcname = "net-snmp code" else : funcname = "{0}()" . format ( funcname ) if threadname == "MainThread" : logmsg = "{0} {1:<112.112}" . format ( time . strftime ( "%T" , time . localtime ( time . time ( ) ) ) , "{0}: {1}" . format ( funcname , msg ) ) else : logmsg = "{0} {1:>112.112}" . format ( time . strftime ( "%T" , time . localtime ( time . time ( ) ) ) , "{0}: {1}" . format ( funcname , msg ) ) print ( logmsg ) | 251,154 | https://github.com/pief/python-netsnmpagent/blob/b1aad1c7f034509c40d9ab17d59be32e809bd31d/examples/threading_agent.py#L70-L104 | [
"def",
"create_ambiente",
"(",
"self",
")",
":",
"return",
"Ambiente",
"(",
"self",
".",
"networkapi_url",
",",
"self",
".",
"user",
",",
"self",
".",
"password",
",",
"self",
".",
"user_ldap",
")"
] |
Function that does the actual data update . | def UpdateSNMPObjs ( ) : global threadingString LogMsg ( "Beginning data update." ) data = "" # Obtain the data by calling an external command. We don't use # subprocess.check_output() here for compatibility with Python versions # older than 2.7. LogMsg ( "Calling external command \"sleep 5; date\"." ) proc = subprocess . Popen ( "sleep 5; date" , shell = True , env = { "LANG" : "C" } , stdout = subprocess . PIPE , stderr = subprocess . STDOUT ) output = proc . communicate ( ) [ 0 ] . splitlines ( ) [ 0 ] rc = proc . poll ( ) if rc != 0 : LogMsg ( "An error occured executing the command: {0}" . format ( output ) ) return msg = "Updating \"threadingString\" object with data \"{0}\"." LogMsg ( msg . format ( output ) ) threadingString . update ( output ) LogMsg ( "Data update done, exiting thread." ) | 251,155 | https://github.com/pief/python-netsnmpagent/blob/b1aad1c7f034509c40d9ab17d59be32e809bd31d/examples/threading_agent.py#L131-L157 | [
"def",
"_get_partition_info",
"(",
"storage_system",
",",
"device_path",
")",
":",
"try",
":",
"partition_infos",
"=",
"storage_system",
".",
"RetrieveDiskPartitionInfo",
"(",
"devicePath",
"=",
"[",
"device_path",
"]",
")",
"except",
"vim",
".",
"fault",
".",
"... |
Returns a dictionary with the currently registered SNMP objects . | def getRegistered ( self , context = "" ) : myobjs = { } try : # Python 2.x objs_iterator = self . _objs [ context ] . iteritems ( ) except AttributeError : # Python 3.x objs_iterator = self . _objs [ context ] . items ( ) for oidstr , snmpobj in objs_iterator : myobjs [ oidstr ] = { "type" : type ( snmpobj ) . __name__ , "value" : snmpobj . value ( ) } return dict ( myobjs ) | 251,156 | https://github.com/pief/python-netsnmpagent/blob/b1aad1c7f034509c40d9ab17d59be32e809bd31d/netsnmpagent.py#L696-L713 | [
"def",
"delete_video",
"(",
"video_id",
",",
"cascade",
"=",
"False",
",",
"delete_shares",
"=",
"False",
",",
"_connection",
"=",
"None",
")",
":",
"c",
"=",
"_connection",
"if",
"not",
"c",
":",
"c",
"=",
"connection",
".",
"APIConnection",
"(",
")",
... |
Starts the agent . Among other things this means connecting to the master agent if configured that way . | def start ( self ) : if self . _status != netsnmpAgentStatus . CONNECTED and self . _status != netsnmpAgentStatus . RECONNECTING : self . _status = netsnmpAgentStatus . FIRSTCONNECT libnsa . init_snmp ( b ( self . AgentName ) ) if self . _status == netsnmpAgentStatus . CONNECTFAILED : msg = "Error connecting to snmpd instance at \"{0}\" -- " "incorrect \"MasterSocket\" or snmpd not running?" msg = msg . format ( self . MasterSocket ) raise netsnmpAgentException ( msg ) | 251,157 | https://github.com/pief/python-netsnmpagent/blob/b1aad1c7f034509c40d9ab17d59be32e809bd31d/netsnmpagent.py#L715-L726 | [
"def",
"win32_refresh_window",
"(",
"cls",
")",
":",
"# Get console handle",
"handle",
"=",
"windll",
".",
"kernel32",
".",
"GetConsoleWindow",
"(",
")",
"RDW_INVALIDATE",
"=",
"0x0001",
"windll",
".",
"user32",
".",
"RedrawWindow",
"(",
"handle",
",",
"None",
... |
Adjust self . _trim_top according to self . _scroll_action | def _adjust_trim_top ( self , canv , size ) : action = self . _scroll_action self . _scroll_action = None maxcol , maxrow = size trim_top = self . _trim_top canv_rows = canv . rows ( ) if trim_top < 0 : # Negative trim_top values use bottom of canvas as reference trim_top = canv_rows - maxrow + trim_top + 1 if canv_rows <= maxrow : self . _trim_top = 0 # Reset scroll position return def ensure_bounds ( new_trim_top ) : return max ( 0 , min ( canv_rows - maxrow , new_trim_top ) ) if action == SCROLL_LINE_UP : self . _trim_top = ensure_bounds ( trim_top - 1 ) elif action == SCROLL_LINE_DOWN : self . _trim_top = ensure_bounds ( trim_top + 1 ) elif action == SCROLL_PAGE_UP : self . _trim_top = ensure_bounds ( trim_top - maxrow + 1 ) elif action == SCROLL_PAGE_DOWN : self . _trim_top = ensure_bounds ( trim_top + maxrow - 1 ) elif action == SCROLL_TO_TOP : self . _trim_top = 0 elif action == SCROLL_TO_END : self . _trim_top = canv_rows - maxrow else : self . _trim_top = ensure_bounds ( trim_top ) # If the cursor was moved by the most recent keypress, adjust trim_top # so that the new cursor position is within the displayed canvas part. # But don't do this if the cursor is at the top/bottom edge so we can still scroll out if self . _old_cursor_coords is not None and self . _old_cursor_coords != canv . cursor : self . _old_cursor_coords = None curscol , cursrow = canv . cursor if cursrow < self . _trim_top : self . _trim_top = cursrow elif cursrow >= self . _trim_top + maxrow : self . _trim_top = max ( 0 , cursrow - maxrow + 1 ) | 251,158 | https://github.com/tonycpsu/panwid/blob/e83a1f612cf5c53de88a7180c1b84b3b7b85460a/panwid/scroll.py#L177-L224 | [
"def",
"complete_experiment",
"(",
"self",
",",
"status",
")",
":",
"self",
".",
"log",
"(",
"\"Bot player completing experiment. Status: {}\"",
".",
"format",
"(",
"status",
")",
")",
"while",
"True",
":",
"url",
"=",
"\"{host}/{status}?participant_id={participant_id... |
Return the number of rows for size | def rows_max ( self , size = None , focus = False ) : if size is not None : ow = self . _original_widget ow_size = self . _get_original_widget_size ( size ) sizing = ow . sizing ( ) if FIXED in sizing : self . _rows_max_cached = ow . pack ( ow_size , focus ) [ 1 ] elif FLOW in sizing : self . _rows_max_cached = ow . rows ( ow_size , focus ) else : raise RuntimeError ( 'Not a flow/box widget: %r' % self . _original_widget ) return self . _rows_max_cached | 251,159 | https://github.com/tonycpsu/panwid/blob/e83a1f612cf5c53de88a7180c1b84b3b7b85460a/panwid/scroll.py#L257-L272 | [
"def",
"update_experiment",
"(",
")",
":",
"experiment_config",
"=",
"Experiments",
"(",
")",
"experiment_dict",
"=",
"experiment_config",
".",
"get_all_experiments",
"(",
")",
"if",
"not",
"experiment_dict",
":",
"return",
"None",
"for",
"key",
"in",
"experiment_... |
Nearest original_widget that is compatible with the scrolling API | def scrolling_base_widget ( self ) : def orig_iter ( w ) : while hasattr ( w , 'original_widget' ) : w = w . original_widget yield w yield w def is_scrolling_widget ( w ) : return hasattr ( w , 'get_scrollpos' ) and hasattr ( w , 'rows_max' ) for w in orig_iter ( self ) : if is_scrolling_widget ( w ) : return w raise ValueError ( 'Not compatible to be wrapped by ScrollBar: %r' % w ) | 251,160 | https://github.com/tonycpsu/panwid/blob/e83a1f612cf5c53de88a7180c1b84b3b7b85460a/panwid/scroll.py#L389-L403 | [
"def",
"load",
"(",
"self",
",",
"path",
")",
":",
"(",
"folder",
",",
"filename",
")",
"=",
"os",
".",
"path",
".",
"split",
"(",
"path",
")",
"(",
"name",
",",
"extension",
")",
"=",
"os",
".",
"path",
".",
"splitext",
"(",
"filename",
")",
"... |
Execute the specified coroutine and return its result . Issue a cancellation request after seconds have elapsed . When a timeout occurs no exception is raised . Instead timeout_result is returned . | def ignore_after ( seconds , coro = None , * args , timeout_result = None ) : if coro : return _ignore_after_func ( seconds , False , coro , args , timeout_result ) return TimeoutAfter ( seconds , ignore = True ) | 251,161 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/curio.py#L392-L411 | [
"def",
"gen_etag",
"(",
")",
":",
"my_salt",
"=",
"gen_uuid",
"(",
")",
"if",
"six",
".",
"PY2",
":",
"my_salt",
"=",
"my_salt",
".",
"decode",
"(",
"'utf-8'",
")",
"elif",
"six",
".",
"PY3",
":",
"my_salt",
"=",
"my_salt",
".",
"encode",
"(",
"'ut... |
Add an already existing task to the task group . | def _add_task ( self , task ) : if hasattr ( task , '_task_group' ) : raise RuntimeError ( 'task is already part of a group' ) if self . _closed : raise RuntimeError ( 'task group is closed' ) task . _task_group = self if task . done ( ) : self . _done . append ( task ) else : self . _pending . add ( task ) task . add_done_callback ( self . _on_done ) | 251,162 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/curio.py#L118-L129 | [
"def",
"get_vault_connection",
"(",
")",
":",
"def",
"_use_local_config",
"(",
")",
":",
"log",
".",
"debug",
"(",
"'Using Vault connection details from local config'",
")",
"try",
":",
"if",
"__opts__",
"[",
"'vault'",
"]",
"[",
"'auth'",
"]",
"[",
"'method'",
... |
Returns the next completed task . Returns None if no more tasks remain . A TaskGroup may also be used as an asynchronous iterator . | async def next_done ( self ) : if not self . _done and self . _pending : self . _done_event . clear ( ) await self . _done_event . wait ( ) if self . _done : return self . _done . popleft ( ) return None | 251,163 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/curio.py#L155-L164 | [
"def",
"set_connection_type",
"(",
"self",
",",
"connection_type",
")",
":",
"result",
"=",
"self",
".",
"library",
".",
"Cli_SetConnectionType",
"(",
"self",
".",
"pointer",
",",
"c_uint16",
"(",
"connection_type",
")",
")",
"if",
"result",
"!=",
"0",
":",
... |
Wait for tasks in the group to terminate according to the wait policy for the group . | async def join ( self ) : def errored ( task ) : return not task . cancelled ( ) and task . exception ( ) try : if self . _wait in ( all , object ) : while True : task = await self . next_done ( ) if task is None : return if errored ( task ) : break if self . _wait is object : if task . cancelled ( ) or task . result ( ) is not None : return else : # any task = await self . next_done ( ) if task is None or not errored ( task ) : return finally : await self . cancel_remaining ( ) if errored ( task ) : raise task . exception ( ) | 251,164 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/curio.py#L176-L211 | [
"def",
"createExternalTable",
"(",
"self",
",",
"tableName",
",",
"path",
"=",
"None",
",",
"source",
"=",
"None",
",",
"schema",
"=",
"None",
",",
"*",
"*",
"options",
")",
":",
"return",
"self",
".",
"sparkSession",
".",
"catalog",
".",
"createExternal... |
Cancel all remaining tasks . | async def cancel_remaining ( self ) : self . _closed = True task_list = list ( self . _pending ) for task in task_list : task . cancel ( ) for task in task_list : with suppress ( CancelledError ) : await task | 251,165 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/curio.py#L213-L221 | [
"def",
"set_editor",
"(",
"self",
",",
"editor",
")",
":",
"if",
"self",
".",
"_editor",
"is",
"not",
"None",
":",
"try",
":",
"self",
".",
"_editor",
".",
"offset_calculator",
".",
"pic_infos_available",
".",
"disconnect",
"(",
"self",
".",
"_update",
"... |
Connect to the proxy and perform a handshake requesting a connection . | async def _connect_one ( self , remote_address ) : loop = asyncio . get_event_loop ( ) for info in await loop . getaddrinfo ( str ( self . address . host ) , self . address . port , type = socket . SOCK_STREAM ) : # This object has state so is only good for one connection client = self . protocol ( remote_address , self . auth ) sock = socket . socket ( family = info [ 0 ] ) try : # A non-blocking socket is required by loop socket methods sock . setblocking ( False ) await loop . sock_connect ( sock , info [ 4 ] ) await self . _handshake ( client , sock , loop ) self . peername = sock . getpeername ( ) return sock except ( OSError , SOCKSProtocolError ) as e : exception = e # Don't close the socket because of an asyncio bug # see https://github.com/kyuupichan/aiorpcX/issues/8 return exception | 251,166 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/socks.py#L300-L323 | [
"def",
"get_exif_data",
"(",
"self",
",",
"image",
")",
":",
"exif_data",
"=",
"{",
"}",
"info",
"=",
"image",
".",
"_getexif",
"(",
")",
"if",
"info",
":",
"for",
"tag",
",",
"value",
"in",
"info",
".",
"items",
"(",
")",
":",
"decoded",
"=",
"T... |
Connect to the proxy and perform a handshake requesting a connection to each address in addresses . | async def _connect ( self , remote_addresses ) : assert remote_addresses exceptions = [ ] for remote_address in remote_addresses : sock = await self . _connect_one ( remote_address ) if isinstance ( sock , socket . socket ) : return sock , remote_address exceptions . append ( sock ) strings = set ( f'{exc!r}' for exc in exceptions ) raise ( exceptions [ 0 ] if len ( strings ) == 1 else OSError ( f'multiple exceptions: {", ".join(strings)}' ) ) | 251,167 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/socks.py#L325-L342 | [
"def",
"get_cumulative_metadata",
"(",
"self",
")",
":",
"# Implemented from template for osid.resource.ResourceForm.get_group_metadata_template",
"metadata",
"=",
"dict",
"(",
"self",
".",
"_mdata",
"[",
"'cumulative'",
"]",
")",
"metadata",
".",
"update",
"(",
"{",
"'... |
Return True if it appears we can connect to a SOCKS proxy otherwise False . | async def _detect_proxy ( self ) : if self . protocol is SOCKS4a : remote_address = NetAddress ( 'www.apple.com' , 80 ) else : remote_address = NetAddress ( '8.8.8.8' , 53 ) sock = await self . _connect_one ( remote_address ) if isinstance ( sock , socket . socket ) : sock . close ( ) return True # SOCKSFailure indicates something failed, but that we are likely talking to a # proxy return isinstance ( sock , SOCKSFailure ) | 251,168 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/socks.py#L344-L360 | [
"def",
"combine_bam",
"(",
"in_files",
",",
"out_file",
",",
"config",
")",
":",
"runner",
"=",
"broad",
".",
"runner_from_path",
"(",
"\"picard\"",
",",
"config",
")",
"runner",
".",
"run_fn",
"(",
"\"picard_merge\"",
",",
"in_files",
",",
"out_file",
")",
... |
Try to detect a SOCKS proxy on a host on one of the ports . | async def auto_detect_at_host ( cls , host , ports , auth ) : for port in ports : proxy = await cls . auto_detect_at_address ( NetAddress ( host , port ) , auth ) if proxy : return proxy return None | 251,169 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/socks.py#L380-L393 | [
"def",
"run",
"(",
"self",
",",
"conf",
",",
"arg",
",",
"err",
")",
":",
"weld_context_new",
"=",
"weld",
".",
"weld_context_new",
"weld_context_new",
".",
"argtypes",
"=",
"[",
"c_weld_conf",
"]",
"weld_context_new",
".",
"restype",
"=",
"c_weld_context",
... |
Initiate a connection . | async def create_connection ( self ) : connector = self . proxy or self . loop return await connector . create_connection ( self . session_factory , self . host , self . port , * * self . kwargs ) | 251,170 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/session.py#L73-L77 | [
"def",
"Run",
"(",
"self",
",",
"args",
")",
":",
"with",
"vfs",
".",
"VFSOpen",
"(",
"args",
".",
"pathspec",
",",
"progress_callback",
"=",
"self",
".",
"Progress",
")",
"as",
"file_obj",
":",
"fingerprinter",
"=",
"Fingerprinter",
"(",
"self",
".",
... |
Called by asyncio when a message comes in . | def data_received ( self , framed_message ) : if self . verbosity >= 4 : self . logger . debug ( f'Received framed message {framed_message}' ) self . recv_size += len ( framed_message ) self . bump_cost ( len ( framed_message ) * self . bw_cost_per_byte ) self . framer . received_bytes ( framed_message ) | 251,171 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/session.py#L224-L230 | [
"def",
"resolve_expression",
"(",
"self",
",",
"query",
"=",
"None",
",",
"allow_joins",
"=",
"True",
",",
"reuse",
"=",
"None",
",",
"summarize",
"=",
"False",
",",
"for_save",
"=",
"False",
")",
":",
"c",
"=",
"self",
".",
"copy",
"(",
")",
"c",
... |
Transport calls when the send buffer is full . | def pause_writing ( self ) : if not self . is_closing ( ) : self . _can_send . clear ( ) self . transport . pause_reading ( ) | 251,172 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/session.py#L232-L236 | [
"def",
"_set_sequence",
"(",
"cls",
",",
"val",
")",
":",
"if",
"db",
".",
"engine",
".",
"dialect",
".",
"name",
"==",
"'postgresql'",
":",
"# pragma: no cover",
"db",
".",
"session",
".",
"execute",
"(",
"\"SELECT setval(pg_get_serial_sequence(\"",
"\"'{0}', '... |
Transport calls when the send buffer has room . | def resume_writing ( self ) : if not self . _can_send . is_set ( ) : self . _can_send . set ( ) self . transport . resume_reading ( ) | 251,173 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/session.py#L238-L242 | [
"def",
"list_categories",
"(",
"self",
",",
"package_keyname",
",",
"*",
"*",
"kwargs",
")",
":",
"get_kwargs",
"=",
"{",
"}",
"get_kwargs",
"[",
"'mask'",
"]",
"=",
"kwargs",
".",
"get",
"(",
"'mask'",
",",
"CATEGORY_MASK",
")",
"if",
"'filter'",
"in",
... |
Called by asyncio when a connection is established . | def connection_made ( self , transport ) : self . transport = transport # If the Socks proxy was used then _proxy and _remote_address are already set if self . _proxy is None : # This would throw if called on a closed SSL transport. Fixed in asyncio in # Python 3.6.1 and 3.5.4 peername = transport . get_extra_info ( 'peername' ) self . _remote_address = NetAddress ( peername [ 0 ] , peername [ 1 ] ) self . _task = spawn_sync ( self . _process_messages ( ) , loop = self . loop ) | 251,174 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/session.py#L244-L255 | [
"def",
"removeApplicationManifest",
"(",
"self",
",",
"pchApplicationManifestFullPath",
")",
":",
"fn",
"=",
"self",
".",
"function_table",
".",
"removeApplicationManifest",
"result",
"=",
"fn",
"(",
"pchApplicationManifestFullPath",
")",
"return",
"result"
] |
Called by asyncio when the connection closes . | def connection_lost ( self , exc ) : # Work around uvloop bug; see https://github.com/MagicStack/uvloop/issues/246 if self . transport : self . transport = None self . closed_event . set ( ) # Release waiting tasks self . _can_send . set ( ) # Cancelling directly leads to self-cancellation problems for member # functions await-ing self.close() self . loop . call_soon ( self . _task . cancel ) | 251,175 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/session.py#L257-L269 | [
"def",
"do_form_dash_instance",
"(",
"self",
",",
"replacements",
"=",
"None",
",",
"specific_identifier",
"=",
"None",
",",
"cache_id",
"=",
"None",
")",
":",
"ndid",
",",
"base_pathname",
"=",
"self",
".",
"get_base_pathname",
"(",
"specific_identifier",
",",
... |
Call to recalculate sleeps and concurrency for the session . Called automatically if cost has drifted significantly . Otherwise can be called at regular intervals if desired . | def recalc_concurrency ( self ) : # Refund resource usage proportionally to elapsed time; the bump passed is negative now = time . time ( ) self . cost = max ( 0 , self . cost - ( now - self . _cost_time ) * self . cost_decay_per_sec ) self . _cost_time = now self . _cost_last = self . cost # Setting cost_hard_limit <= 0 means to not limit concurrency value = self . _incoming_concurrency . max_concurrent cost_soft_range = self . cost_hard_limit - self . cost_soft_limit if cost_soft_range <= 0 : return cost = self . cost + self . extra_cost ( ) self . _cost_fraction = max ( 0.0 , ( cost - self . cost_soft_limit ) / cost_soft_range ) target = max ( 0 , ceil ( ( 1.0 - self . _cost_fraction ) * self . initial_concurrent ) ) if abs ( target - value ) > 1 : self . logger . info ( f'changing task concurrency from {value} to {target}' ) self . _incoming_concurrency . set_target ( target ) | 251,176 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/session.py#L282-L305 | [
"def",
"revert_metadata",
"(",
"files",
")",
":",
"for",
"file_path",
"in",
"files",
":",
"tags",
"=",
"EasyMP3",
"(",
"file_path",
")",
"tags",
".",
"delete",
"(",
")",
"tags",
".",
"save",
"(",
")"
] |
Close the connection and return when closed . | async def close ( self , * , force_after = 30 ) : if self . transport : self . transport . close ( ) try : async with timeout_after ( force_after ) : await self . closed_event . wait ( ) except TaskTimeout : self . abort ( ) await self . closed_event . wait ( ) | 251,177 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/session.py#L342-L351 | [
"def",
"setOverlayTransformTrackedDeviceRelative",
"(",
"self",
",",
"ulOverlayHandle",
",",
"unTrackedDevice",
")",
":",
"fn",
"=",
"self",
".",
"function_table",
".",
"setOverlayTransformTrackedDeviceRelative",
"pmatTrackedDeviceToOverlayTransform",
"=",
"HmdMatrix34_t",
"(... |
Send an RPC request over the network . | async def send_request ( self , method , args = ( ) ) : message , event = self . connection . send_request ( Request ( method , args ) ) return await self . _send_concurrent ( message , event , 1 ) | 251,178 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/session.py#L631-L634 | [
"def",
"create_meta_main",
"(",
"create_path",
",",
"config",
",",
"role",
",",
"categories",
")",
":",
"meta_file",
"=",
"c",
".",
"DEFAULT_META_FILE",
".",
"replace",
"(",
"\"%author_name\"",
",",
"config",
"[",
"\"author_name\"",
"]",
")",
"meta_file",
"=",... |
Send an RPC notification over the network . | async def send_notification ( self , method , args = ( ) ) : message = self . connection . send_notification ( Notification ( method , args ) ) await self . _send_message ( message ) | 251,179 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/session.py#L636-L639 | [
"def",
"compose",
"(",
"self",
")",
":",
"rr",
"=",
"self",
".",
"__reagents",
"+",
"self",
".",
"__reactants",
"if",
"rr",
":",
"if",
"not",
"all",
"(",
"isinstance",
"(",
"x",
",",
"(",
"MoleculeContainer",
",",
"CGRContainer",
")",
")",
"for",
"x"... |
Close the listening socket . This does not close any ServerSession objects created to handle incoming connections . | async def close ( self ) : if self . server : self . server . close ( ) await self . server . wait_closed ( ) self . server = None | 251,180 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/session.py#L674-L681 | [
"def",
"add_firmware_image",
"(",
"self",
",",
"name",
",",
"datafile",
",",
"*",
"*",
"kwargs",
")",
":",
"kwargs",
".",
"update",
"(",
"{",
"'name'",
":",
"name",
"}",
")",
"firmware_image",
"=",
"FirmwareImage",
".",
"_create_request_map",
"(",
"kwargs"... |
Returns a Python object or a ProtocolError . | def _message_to_payload ( cls , message ) : try : return json . loads ( message . decode ( ) ) except UnicodeDecodeError : message = 'messages must be encoded in UTF-8' except json . JSONDecodeError : message = 'invalid JSON' raise cls . _error ( cls . PARSE_ERROR , message , True , None ) | 251,181 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/jsonrpc.py#L223-L231 | [
"def",
"merge_entities",
"(",
"self",
",",
"from_entity_ids",
",",
"to_entity_id",
",",
"force",
"=",
"False",
",",
"mount_point",
"=",
"DEFAULT_MOUNT_POINT",
")",
":",
"params",
"=",
"{",
"'from_entity_ids'",
":",
"from_entity_ids",
",",
"'to_entity_id'",
":",
... |
Convert a request Batch to a message . | def batch_message ( cls , batch , request_ids ) : assert isinstance ( batch , Batch ) if not cls . allow_batches : raise ProtocolError . invalid_request ( 'protocol does not permit batches' ) id_iter = iter ( request_ids ) rm = cls . request_message nm = cls . notification_message parts = ( rm ( request , next ( id_iter ) ) if isinstance ( request , Request ) else nm ( request ) for request in batch ) return cls . batch_message_from_parts ( parts ) | 251,182 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/jsonrpc.py#L305-L316 | [
"def",
"__isOpenThreadWpanRunning",
"(",
"self",
")",
":",
"print",
"'call __isOpenThreadWpanRunning'",
"if",
"self",
".",
"__stripValue",
"(",
"self",
".",
"__sendCommand",
"(",
"WPANCTL_CMD",
"+",
"'getprop -v NCP:State'",
")",
"[",
"0",
"]",
")",
"==",
"'associ... |
Convert messages one per batch item into a batch message . At least one message must be passed . | def batch_message_from_parts ( cls , messages ) : # Comma-separate the messages and wrap the lot in square brackets middle = b', ' . join ( messages ) if not middle : raise ProtocolError . empty_batch ( ) return b'' . join ( [ b'[' , middle , b']' ] ) | 251,183 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/jsonrpc.py#L319-L327 | [
"def",
"_connect_hive",
"(",
"self",
",",
"hive",
")",
":",
"try",
":",
"handle",
"=",
"self",
".",
"_remote_hives",
"[",
"hive",
"]",
"except",
"KeyError",
":",
"handle",
"=",
"win32",
".",
"RegConnectRegistry",
"(",
"self",
".",
"_machine",
",",
"hive"... |
Encode a Python object as JSON and convert it to bytes . | def encode_payload ( cls , payload ) : try : return json . dumps ( payload ) . encode ( ) except TypeError : msg = f'JSON payload encoding error: {payload}' raise ProtocolError ( cls . INTERNAL_ERROR , msg ) from None | 251,184 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/jsonrpc.py#L330-L336 | [
"async",
"def",
"set_max_relative_mod",
"(",
"self",
",",
"max_mod",
",",
"timeout",
"=",
"OTGW_DEFAULT_TIMEOUT",
")",
":",
"if",
"isinstance",
"(",
"max_mod",
",",
"int",
")",
"and",
"not",
"0",
"<=",
"max_mod",
"<=",
"100",
":",
"return",
"None",
"cmd",
... |
Attempt to detect the protocol from the message . | def detect_protocol ( cls , message ) : main = cls . _message_to_payload ( message ) def protocol_for_payload ( payload ) : if not isinstance ( payload , dict ) : return JSONRPCLoose # Will error # Obey an explicit "jsonrpc" version = payload . get ( 'jsonrpc' ) if version == '2.0' : return JSONRPCv2 if version == '1.0' : return JSONRPCv1 # Now to decide between JSONRPCLoose and JSONRPCv1 if possible if 'result' in payload and 'error' in payload : return JSONRPCv1 return JSONRPCLoose if isinstance ( main , list ) : parts = set ( protocol_for_payload ( payload ) for payload in main ) # If all same protocol, return it if len ( parts ) == 1 : return parts . pop ( ) # If strict protocol detected, return it, preferring JSONRPCv2. # This means a batch of JSONRPCv1 will fail for protocol in ( JSONRPCv2 , JSONRPCv1 ) : if protocol in parts : return protocol # Will error if no parts return JSONRPCLoose return protocol_for_payload ( main ) | 251,185 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/jsonrpc.py#L544-L576 | [
"def",
"blit",
"(",
"self",
",",
"console",
":",
"tcod",
".",
"console",
".",
"Console",
",",
"x",
":",
"float",
",",
"y",
":",
"float",
",",
"bg_blend",
":",
"int",
",",
"scale_x",
":",
"float",
",",
"scale_y",
":",
"float",
",",
"angle",
":",
"... |
Call with an unframed message received from the network . | def receive_message ( self , message ) : if self . _protocol is JSONRPCAutoDetect : self . _protocol = JSONRPCAutoDetect . detect_protocol ( message ) try : item , request_id = self . _protocol . message_to_item ( message ) except ProtocolError as e : if e . response_msg_id is not id : return self . _receive_response ( e , e . response_msg_id ) raise if isinstance ( item , Request ) : item . send_result = partial ( self . _send_result , request_id ) return [ item ] if isinstance ( item , Notification ) : return [ item ] if isinstance ( item , Response ) : return self . _receive_response ( item . result , request_id ) assert isinstance ( item , list ) if all ( isinstance ( payload , dict ) and ( 'result' in payload or 'error' in payload ) for payload in item ) : return self . _receive_response_batch ( item ) else : return self . _receive_request_batch ( item ) | 251,186 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/jsonrpc.py#L706-L737 | [
"def",
"create_meta_main",
"(",
"create_path",
",",
"config",
",",
"role",
",",
"categories",
")",
":",
"meta_file",
"=",
"c",
".",
"DEFAULT_META_FILE",
".",
"replace",
"(",
"\"%author_name\"",
",",
"config",
"[",
"\"author_name\"",
"]",
")",
"meta_file",
"=",... |
Cancel all pending requests . | def cancel_pending_requests ( self ) : exception = CancelledError ( ) for _request , event in self . _requests . values ( ) : event . result = exception event . set ( ) self . _requests . clear ( ) | 251,187 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/jsonrpc.py#L739-L745 | [
"def",
"customer_discount_webhook_handler",
"(",
"event",
")",
":",
"crud_type",
"=",
"CrudType",
".",
"determine",
"(",
"event",
"=",
"event",
")",
"discount_data",
"=",
"event",
".",
"data",
".",
"get",
"(",
"\"object\"",
",",
"{",
"}",
")",
"coupon_data",... |
Return True if hostname is valid otherwise False . | def is_valid_hostname ( hostname ) : if not isinstance ( hostname , str ) : raise TypeError ( 'hostname must be a string' ) # strip exactly one dot from the right, if present if hostname and hostname [ - 1 ] == "." : hostname = hostname [ : - 1 ] if not hostname or len ( hostname ) > 253 : return False labels = hostname . split ( '.' ) # the TLD must be not all-numeric if re . match ( NUMERIC_REGEX , labels [ - 1 ] ) : return False return all ( LABEL_REGEX . match ( label ) for label in labels ) | 251,188 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/util.py#L47-L60 | [
"def",
"wrap2spy",
"(",
"self",
")",
":",
"if",
"self",
".",
"args_type",
"==",
"\"MODULE_FUNCTION\"",
":",
"self",
".",
"orig_func",
"=",
"deepcopy",
"(",
"getattr",
"(",
"self",
".",
"obj",
",",
"self",
".",
"prop",
")",
")",
"setattr",
"(",
"self",
... |
Host is an IPv4Address IPv6Address or a string . | def classify_host ( host ) : if isinstance ( host , ( IPv4Address , IPv6Address ) ) : return host if is_valid_hostname ( host ) : return host return ip_address ( host ) | 251,189 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/util.py#L63-L77 | [
"def",
"increment_lessons",
"(",
"self",
",",
"measure_vals",
",",
"reward_buff_sizes",
"=",
"None",
")",
":",
"ret",
"=",
"{",
"}",
"if",
"reward_buff_sizes",
":",
"for",
"brain_name",
",",
"buff_size",
"in",
"reward_buff_sizes",
".",
"items",
"(",
")",
":"... |
Validate port and return it as an integer . | def validate_port ( port ) : if not isinstance ( port , ( str , int ) ) : raise TypeError ( f'port must be an integer or string: {port}' ) if isinstance ( port , str ) and port . isdigit ( ) : port = int ( port ) if isinstance ( port , int ) and 0 < port <= 65535 : return port raise ValueError ( f'invalid port: {port}' ) | 251,190 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/util.py#L80-L90 | [
"def",
"Run",
"(",
"self",
",",
"arg",
")",
":",
"if",
"arg",
"is",
"None",
":",
"arg",
"=",
"rdf_client_action",
".",
"GetClientStatsRequest",
"(",
")",
"proc",
"=",
"psutil",
".",
"Process",
"(",
"os",
".",
"getpid",
"(",
")",
")",
"meminfo",
"=",
... |
Validate a protocol a string and return it . | def validate_protocol ( protocol ) : if not re . match ( PROTOCOL_REGEX , protocol ) : raise ValueError ( f'invalid protocol: {protocol}' ) return protocol . lower ( ) | 251,191 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/util.py#L93-L97 | [
"def",
"get_listing",
"(",
"self",
")",
":",
"if",
"not",
"hasattr",
"(",
"self",
",",
"'listing'",
")",
":",
"allEvents",
"=",
"self",
".",
"get_allEvents",
"(",
")",
"openEvents",
"=",
"allEvents",
".",
"filter",
"(",
"registrationOpen",
"=",
"True",
"... |
inspect . iscoroutinefunction that looks through partials . | def is_async_call ( func ) : while isinstance ( func , partial ) : func = func . func return inspect . iscoroutinefunction ( func ) | 251,192 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/util.py#L261-L265 | [
"def",
"filter_model_items",
"(",
"index_instance",
",",
"model_items",
",",
"model_name",
",",
"start_date",
",",
"end_date",
")",
":",
"if",
"index_instance",
".",
"updated_field",
"is",
"None",
":",
"logger",
".",
"warning",
"(",
"\"No updated date field found fo... |
Construct a Service from a string . | def from_string ( cls , string , * , default_func = None ) : if not isinstance ( string , str ) : raise TypeError ( f'service must be a string: {string}' ) parts = string . split ( '://' , 1 ) if len ( parts ) == 2 : protocol , address = parts else : item , = parts protocol = None if default_func : if default_func ( item , ServicePart . HOST ) and default_func ( item , ServicePart . PORT ) : protocol , address = item , '' else : protocol , address = default_func ( None , ServicePart . PROTOCOL ) , item if not protocol : raise ValueError ( f'invalid service string: {string}' ) if default_func : default_func = partial ( default_func , protocol . lower ( ) ) address = NetAddress . from_string ( address , default_func = default_func ) return cls ( protocol , address ) | 251,193 | https://github.com/kyuupichan/aiorpcX/blob/707c989ed1c67ac9a40cd20b0161b1ce1f4d7db0/aiorpcx/util.py#L218-L244 | [
"def",
"_tupleCompare",
"(",
"tuple1",
",",
"ineq",
",",
"tuple2",
",",
"eq",
"=",
"lambda",
"a",
",",
"b",
":",
"(",
"a",
"==",
"b",
")",
",",
"ander",
"=",
"AND",
",",
"orer",
"=",
"OR",
")",
":",
"orholder",
"=",
"[",
"]",
"for",
"limit",
... |
The XML file seems to have mixed - encoding ; we scrub out the control characters from the file for processing . | def scrub ( self ) : LOG . info ( "Scrubbing out the nasty characters that break our parser." ) myfile = '/' . join ( ( self . rawdir , self . files [ 'data' ] [ 'file' ] ) ) tmpfile = '/' . join ( ( self . rawdir , self . files [ 'data' ] [ 'file' ] + '.tmp.gz' ) ) tmp = gzip . open ( tmpfile , 'wb' ) du = DipperUtil ( ) with gzip . open ( myfile , 'rb' ) as fh : filereader = io . TextIOWrapper ( fh , newline = "" ) for line in filereader : line = du . remove_control_characters ( line ) + '\n' tmp . write ( line . encode ( 'utf-8' ) ) tmp . close ( ) # TEC I do not like this at all. original data must be preserved as is. # also may be heavy handed as chars which do not break the parser # are stripped as well (i.e. tabs and newlines) # move the temp file LOG . info ( "Replacing the original data with the scrubbed file." ) shutil . move ( tmpfile , myfile ) return | 251,194 | https://github.com/monarch-initiative/dipper/blob/24cc80db355bbe15776edc5c7b41e0886959ba41/dipper/sources/OMIA.py#L202-L234 | [
"def",
"_getNearestMappingIndexList",
"(",
"fromValList",
",",
"toValList",
")",
":",
"indexList",
"=",
"[",
"]",
"for",
"fromTimestamp",
"in",
"fromValList",
":",
"smallestDiff",
"=",
"_getSmallestDifference",
"(",
"toValList",
",",
"fromTimestamp",
")",
"i",
"="... |
Loop through the xml file and process the article - breed article - phene breed - phene phene - gene associations and the external links to LIDA . | def process_associations ( self , limit ) : myfile = '/' . join ( ( self . rawdir , self . files [ 'data' ] [ 'file' ] ) ) f = gzip . open ( myfile , 'rb' ) filereader = io . TextIOWrapper ( f , newline = "" ) filereader . readline ( ) # remove the xml declaration line for event , elem in ET . iterparse ( filereader ) : # iterparse is not deprecated self . process_xml_table ( elem , 'Article_Breed' , self . _process_article_breed_row , limit ) self . process_xml_table ( elem , 'Article_Phene' , self . _process_article_phene_row , limit ) self . process_xml_table ( elem , 'Breed_Phene' , self . _process_breed_phene_row , limit ) self . process_xml_table ( elem , 'Lida_Links' , self . _process_lida_links_row , limit ) self . process_xml_table ( elem , 'Phene_Gene' , self . _process_phene_gene_row , limit ) self . process_xml_table ( elem , 'Group_MPO' , self . _process_group_mpo_row , limit ) f . close ( ) return | 251,195 | https://github.com/monarch-initiative/dipper/blob/24cc80db355bbe15776edc5c7b41e0886959ba41/dipper/sources/OMIA.py#L362-L390 | [
"def",
"set_units",
"(",
"self",
",",
"unit",
")",
":",
"self",
".",
"_units",
"=",
"validate_type",
"(",
"unit",
",",
"type",
"(",
"None",
")",
",",
"*",
"six",
".",
"string_types",
")"
] |
Linking articles to species - specific phenes . | def _process_article_phene_row ( self , row ) : # article_id, phene_id, added_by # look up the article in the hashmap phenotype_id = self . id_hash [ 'phene' ] . get ( row [ 'phene_id' ] ) article_id = self . id_hash [ 'article' ] . get ( row [ 'article_id' ] ) omia_id = self . _get_omia_id_from_phene_id ( phenotype_id ) if self . test_mode or omia_id not in self . test_ids [ 'disease' ] or phenotype_id is None or article_id is None : return # make a triple, where the article is about the phenotype self . graph . addTriple ( article_id , self . globaltt [ 'is_about' ] , phenotype_id ) return | 251,196 | https://github.com/monarch-initiative/dipper/blob/24cc80db355bbe15776edc5c7b41e0886959ba41/dipper/sources/OMIA.py#L645-L667 | [
"def",
"cleanup",
"(",
"connector_manager",
",",
"red_data",
",",
"tmp_dir",
")",
":",
"for",
"key",
",",
"arg",
"in",
"red_data",
"[",
"'inputs'",
"]",
".",
"items",
"(",
")",
":",
"val",
"=",
"arg",
"if",
"isinstance",
"(",
"arg",
",",
"list",
")",... |
doubt this should be kept | def filter_keep_phenotype_entry_ids ( self , entry ) : omim_id = str ( entry [ 'mimNumber' ] ) otype = self . globaltt [ 'obsolete' ] if omim_id in self . omim_type : otype = self . omim_type [ omim_id ] if otype == self . globaltt [ 'obsolete' ] and omim_id in self . omim_replaced : omim_id = self . omim_replaced [ omim_id ] otype = self . omim_type [ omim_id ] # else: # removed or multiple if otype not in ( self . globaltt [ 'Phenotype' ] , self . globaltt [ 'has_affected_feature' ] ) : omim_id = None return omim_id | 251,197 | https://github.com/monarch-initiative/dipper/blob/24cc80db355bbe15776edc5c7b41e0886959ba41/dipper/sources/OMIA.py#L825-L841 | [
"def",
"api_call",
"(",
"self",
",",
"api_method",
":",
"str",
",",
"*",
",",
"http_verb",
":",
"str",
"=",
"\"POST\"",
",",
"files",
":",
"dict",
"=",
"None",
",",
"data",
":",
"dict",
"=",
"None",
",",
"params",
":",
"dict",
"=",
"None",
",",
"... |
Decorates the three given strings as a line of ntriples | def make_spo ( sub , prd , obj ) : # To establish string as a curie and expand, # we use a global curie_map(.yaml) # sub are allways uri (unless a bnode) # prd are allways uri (unless prd is 'a') # should fail loudly if curie does not exist if prd == 'a' : prd = 'rdf:type' try : ( subcuri , subid ) = re . split ( r':' , sub ) except Exception : LOG . error ( "not a Subject Curie '%s'" , sub ) raise ValueError try : ( prdcuri , prdid ) = re . split ( r':' , prd ) except Exception : LOG . error ( "not a Predicate Curie '%s'" , prd ) raise ValueError objt = '' # object is a curie or bnode or literal [string|number] objcuri = None match = re . match ( CURIERE , obj ) if match is not None : try : ( objcuri , objid ) = re . split ( r':' , obj ) except ValueError : match = None if match is not None and objcuri in CURIEMAP : objt = CURIEMAP [ objcuri ] + objid . strip ( ) # allow unexpanded bnodes in object if objcuri != '_' or CURIEMAP [ objcuri ] != '_:b' : objt = '<' + objt + '>' elif obj . isnumeric ( ) : objt = '"' + obj + '"' else : # Literals may not contain the characters ", LF, CR '\' # except in their escaped forms. internal quotes as well. obj = obj . strip ( '"' ) . replace ( '\\' , '\\\\' ) . replace ( '"' , '\'' ) obj = obj . replace ( '\n' , '\\n' ) . replace ( '\r' , '\\r' ) objt = '"' + obj + '"' # allow unexpanded bnodes in subject if subcuri is not None and subcuri in CURIEMAP and prdcuri is not None and prdcuri in CURIEMAP : subjt = CURIEMAP [ subcuri ] + subid . strip ( ) if subcuri != '_' or CURIEMAP [ subcuri ] != '_:b' : subjt = '<' + subjt + '>' return subjt + ' <' + CURIEMAP [ prdcuri ] + prdid . strip ( ) + '> ' + objt + ' .' else : LOG . error ( 'Cant work with: <%s> %s , <%s> %s, %s' , subcuri , subid , prdcuri , prdid , objt ) return None | 251,198 | https://github.com/monarch-initiative/dipper/blob/24cc80db355bbe15776edc5c7b41e0886959ba41/dipper/sources/ClinVarXML_alpha.py#L183-L244 | [
"def",
"cudaMemcpy_htod",
"(",
"dst",
",",
"src",
",",
"count",
")",
":",
"status",
"=",
"_libcudart",
".",
"cudaMemcpy",
"(",
"dst",
",",
"src",
",",
"ctypes",
".",
"c_size_t",
"(",
"count",
")",
",",
"cudaMemcpyHostToDevice",
")",
"cudaCheckStatus",
"(",... |
write triples to a buffer incase we decide to drop them | def write_spo ( sub , prd , obj ) : rcvtriples . append ( make_spo ( sub , prd , obj ) ) | 251,199 | https://github.com/monarch-initiative/dipper/blob/24cc80db355bbe15776edc5c7b41e0886959ba41/dipper/sources/ClinVarXML_alpha.py#L247-L251 | [
"def",
"populateFromRow",
"(",
"self",
",",
"readGroupSetRecord",
")",
":",
"self",
".",
"_dataUrl",
"=",
"readGroupSetRecord",
".",
"dataurl",
"self",
".",
"_indexFile",
"=",
"readGroupSetRecord",
".",
"indexfile",
"self",
".",
"_programs",
"=",
"[",
"]",
"fo... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.