idx int64 0 251k | question stringlengths 53 3.53k | target stringlengths 5 1.23k | len_question int64 20 893 | len_target int64 3 238 |
|---|---|---|---|---|
239,000 | def motion_detection_sensitivity ( self ) : if not self . triggers : return None for trigger in self . triggers : if trigger . get ( "type" ) != "pirMotionActive" : continue sensitivity = trigger . get ( "sensitivity" ) if sensitivity : return sensitivity . get ( "default" ) return None | Sensitivity level of Camera motion detection . | 69 | 8 |
239,001 | def audio_detection_sensitivity ( self ) : if not self . triggers : return None for trigger in self . triggers : if trigger . get ( "type" ) != "audioAmplitude" : continue sensitivity = trigger . get ( "sensitivity" ) if sensitivity : return sensitivity . get ( "default" ) return None | Sensitivity level of Camera audio detection . | 70 | 8 |
239,002 | def live_streaming ( self ) : url = STREAM_ENDPOINT # override params params = STREAMING_BODY params [ 'from' ] = "{0}_web" . format ( self . user_id ) params [ 'to' ] = self . device_id params [ 'resource' ] = "cameras/{0}" . format ( self . device_id ) params [ 'transId' ] = "web!{0}" . format ( self . xcloud_id ) # override headers headers = { 'xCloudId' : self . xcloud_id } _LOGGER . debug ( "Streaming device %s" , self . name ) _LOGGER . debug ( "Device params %s" , params ) _LOGGER . debug ( "Device headers %s" , headers ) ret = self . _session . query ( url , method = 'POST' , extra_params = params , extra_headers = headers ) _LOGGER . debug ( "Streaming results %s" , ret ) if ret . get ( 'success' ) : return ret . get ( 'data' ) . get ( 'url' ) return ret . get ( 'data' ) | Return live streaming generator . | 258 | 5 |
239,003 | def schedule_snapshot ( self ) : # Notes: # - Snapshots are not immediate. # - Snapshots will be cached for predefined amount # of time. # - Snapshots are not balanced. To get a better # image, it must be taken from the stream, a few # seconds after stream start. url = SNAPSHOTS_ENDPOINT params = SNAPSHOTS_BODY params [ 'from' ] = "{0}_web" . format ( self . user_id ) params [ 'to' ] = self . device_id params [ 'resource' ] = "cameras/{0}" . format ( self . device_id ) params [ 'transId' ] = "web!{0}" . format ( self . xcloud_id ) # override headers headers = { 'xCloudId' : self . xcloud_id } _LOGGER . debug ( "Snapshot device %s" , self . name ) _LOGGER . debug ( "Device params %s" , params ) _LOGGER . debug ( "Device headers %s" , headers ) ret = self . _session . query ( url , method = 'POST' , extra_params = params , extra_headers = headers ) _LOGGER . debug ( "Snapshot results %s" , ret ) return ret is not None and ret . get ( 'success' ) | Trigger snapshot to be uploaded to AWS . Return success state . | 292 | 12 |
239,004 | def thread_function ( self ) : self . __subscribed = True url = SUBSCRIBE_ENDPOINT + "?token=" + self . _session_token data = self . _session . query ( url , method = 'GET' , raw = True , stream = True ) if not data or not data . ok : _LOGGER . debug ( "Did not receive a valid response. Aborting.." ) return None self . __sseclient = sseclient . SSEClient ( data ) try : for event in ( self . __sseclient ) . events ( ) : if not self . __subscribed : break data = json . loads ( event . data ) if data . get ( 'status' ) == "connected" : _LOGGER . debug ( "Successfully subscribed this base station" ) elif data . get ( 'action' ) : action = data . get ( 'action' ) resource = data . get ( 'resource' ) if action == "logout" : _LOGGER . debug ( "Logged out by some other entity" ) self . __subscribed = False break elif action == "is" and "subscriptions/" not in resource : self . __events . append ( data ) self . __event_handle . set ( ) except TypeError as error : _LOGGER . debug ( "Got unexpected error: %s" , error ) return None return True | Thread function . | 303 | 3 |
239,005 | def _get_event_stream ( self ) : self . __event_handle = threading . Event ( ) event_thread = threading . Thread ( target = self . thread_function ) event_thread . start ( ) | Spawn a thread and monitor the Arlo Event Stream . | 48 | 11 |
239,006 | def _unsubscribe_myself ( self ) : url = UNSUBSCRIBE_ENDPOINT return self . _session . query ( url , method = 'GET' , raw = True , stream = False ) | Unsubscribe this base station for all events . | 49 | 10 |
239,007 | def _close_event_stream ( self ) : self . __subscribed = False del self . __events [ : ] self . __event_handle . clear ( ) | Stop the Event stream thread . | 36 | 6 |
239,008 | def publish_and_get_event ( self , resource ) : l_subscribed = False this_event = None if not self . __subscribed : self . _get_event_stream ( ) self . _subscribe_myself ( ) l_subscribed = True status = self . publish ( action = 'get' , resource = resource , mode = None , publish_response = False ) if status == 'success' : i = 0 while not this_event and i < 2 : self . __event_handle . wait ( 5.0 ) self . __event_handle . clear ( ) _LOGGER . debug ( "Instance %s resource: %s" , str ( i ) , resource ) for event in self . __events : if event [ 'resource' ] == resource : this_event = event self . __events . remove ( event ) break i = i + 1 if l_subscribed : self . _unsubscribe_myself ( ) self . _close_event_stream ( ) l_subscribed = False return this_event | Publish and get the event from base station . | 226 | 10 |
239,009 | def publish ( self , action = 'get' , resource = None , camera_id = None , mode = None , publish_response = None , properties = None ) : url = NOTIFY_ENDPOINT . format ( self . device_id ) body = ACTION_BODY . copy ( ) if properties is None : properties = { } if resource : body [ 'resource' ] = resource if action == 'get' : body [ 'properties' ] = None else : # consider moving this logic up a layer if resource == 'schedule' : properties . update ( { 'active' : True } ) elif resource == 'subscribe' : body [ 'resource' ] = "subscriptions/" + "{0}_web" . format ( self . user_id ) dev = [ ] dev . append ( self . device_id ) properties . update ( { 'devices' : dev } ) elif resource == 'modes' : available_modes = self . available_modes_with_ids properties . update ( { 'active' : available_modes . get ( mode ) } ) elif resource == 'privacy' : properties . update ( { 'privacyActive' : not mode } ) body [ 'resource' ] = "cameras/{0}" . format ( camera_id ) body [ 'action' ] = action body [ 'properties' ] = properties body [ 'publishResponse' ] = publish_response body [ 'from' ] = "{0}_web" . format ( self . user_id ) body [ 'to' ] = self . device_id body [ 'transId' ] = "web!{0}" . format ( self . xcloud_id ) _LOGGER . debug ( "Action body: %s" , body ) ret = self . _session . query ( url , method = 'POST' , extra_params = body , extra_headers = { "xCloudId" : self . xcloud_id } ) if ret and ret . get ( 'success' ) : return 'success' return None | Run action . | 445 | 3 |
239,010 | def refresh_rate ( self , value ) : if isinstance ( value , ( int , float ) ) : self . _refresh_rate = value | Override the refresh_rate attribute . | 32 | 7 |
239,011 | def available_modes ( self ) : if not self . _available_modes : modes = self . available_modes_with_ids if not modes : return None self . _available_modes = list ( modes . keys ( ) ) return self . _available_modes | Return list of available mode names . | 61 | 7 |
239,012 | def available_modes_with_ids ( self ) : if not self . _available_mode_ids : all_modes = FIXED_MODES . copy ( ) self . _available_mode_ids = all_modes modes = self . get_available_modes ( ) try : if modes : # pylint: disable=consider-using-dict-comprehension simple_modes = dict ( [ ( m . get ( "type" , m . get ( "name" ) ) , m . get ( "id" ) ) for m in modes ] ) all_modes . update ( simple_modes ) self . _available_mode_ids = all_modes except TypeError : _LOGGER . debug ( "Did not receive a valid response. Passing.." ) return self . _available_mode_ids | Return list of objects containing available mode name and id . | 183 | 11 |
239,013 | def mode ( self ) : if self . is_in_schedule_mode : return "schedule" resource = "modes" mode_event = self . publish_and_get_event ( resource ) if mode_event : properties = mode_event . get ( 'properties' ) active_mode = properties . get ( 'active' ) modes = properties . get ( 'modes' ) if not modes : return None for mode in modes : if mode . get ( 'id' ) == active_mode : return mode . get ( 'type' ) if mode . get ( 'type' ) is not None else mode . get ( 'name' ) return None | Return current mode key . | 142 | 5 |
239,014 | def is_in_schedule_mode ( self ) : resource = "schedule" mode_event = self . publish_and_get_event ( resource ) if mode_event and mode_event . get ( "resource" , None ) == "schedule" : properties = mode_event . get ( 'properties' ) return properties . get ( "active" , False ) return False | Returns True if base_station is currently on a scheduled mode . | 83 | 13 |
239,015 | def get_available_modes ( self ) : resource = "modes" resource_event = self . publish_and_get_event ( resource ) if resource_event : properties = resource_event . get ( "properties" ) return properties . get ( "modes" ) return None | Return a list of available mode objects for an Arlo user . | 62 | 13 |
239,016 | def get_cameras_properties ( self ) : resource = "cameras" resource_event = self . publish_and_get_event ( resource ) if resource_event : self . _last_refresh = int ( time . time ( ) ) self . _camera_properties = resource_event . get ( 'properties' ) | Return camera properties . | 73 | 4 |
239,017 | def get_cameras_battery_level ( self ) : battery_levels = { } if not self . camera_properties : return None for camera in self . camera_properties : serialnum = camera . get ( 'serialNumber' ) cam_battery = camera . get ( 'batteryLevel' ) battery_levels [ serialnum ] = cam_battery return battery_levels | Return a list of battery levels of all cameras . | 83 | 10 |
239,018 | def get_cameras_signal_strength ( self ) : signal_strength = { } if not self . camera_properties : return None for camera in self . camera_properties : serialnum = camera . get ( 'serialNumber' ) cam_strength = camera . get ( 'signalStrength' ) signal_strength [ serialnum ] = cam_strength return signal_strength | Return a list of signal strength of all cameras . | 81 | 10 |
239,019 | def get_camera_extended_properties ( self ) : resource = 'cameras/{}' . format ( self . device_id ) resource_event = self . publish_and_get_event ( resource ) if resource_event is None : return None self . _camera_extended_properties = resource_event . get ( 'properties' ) return self . _camera_extended_properties | Return camera extended properties . | 87 | 5 |
239,020 | def get_speaker_muted ( self ) : if not self . camera_extended_properties : return None speaker = self . camera_extended_properties . get ( 'speaker' ) if not speaker : return None return speaker . get ( 'mute' ) | Return whether or not the speaker is muted . | 59 | 9 |
239,021 | def get_speaker_volume ( self ) : if not self . camera_extended_properties : return None speaker = self . camera_extended_properties . get ( 'speaker' ) if not speaker : return None return speaker . get ( 'volume' ) | Return the volume setting of the speaker . | 57 | 8 |
239,022 | def properties ( self ) : resource = "basestation" basestn_event = self . publish_and_get_event ( resource ) if basestn_event : return basestn_event . get ( 'properties' ) return None | Return the base station info . | 52 | 6 |
239,023 | def get_cameras_rules ( self ) : resource = "rules" rules_event = self . publish_and_get_event ( resource ) if rules_event : return rules_event . get ( 'properties' ) return None | Return the camera rules . | 51 | 5 |
239,024 | def get_cameras_schedule ( self ) : resource = "schedule" schedule_event = self . publish_and_get_event ( resource ) if schedule_event : return schedule_event . get ( 'properties' ) return None | Return the schedule set for cameras . | 53 | 7 |
239,025 | def get_ambient_sensor_data ( self ) : resource = 'cameras/{}/ambientSensors/history' . format ( self . device_id ) history_event = self . publish_and_get_event ( resource ) if history_event is None : return None properties = history_event . get ( 'properties' ) self . _ambient_sensor_data = ArloBaseStation . _decode_sensor_data ( properties ) return self . _ambient_sensor_data | Refresh ambient sensor history | 116 | 5 |
239,026 | def _decode_sensor_data ( properties ) : b64_input = "" for s in properties . get ( 'payload' ) : # pylint: disable=consider-using-join b64_input += s decoded = base64 . b64decode ( b64_input ) data = zlib . decompress ( decoded ) points = [ ] i = 0 while i < len ( data ) : points . append ( { 'timestamp' : int ( 1e3 * ArloBaseStation . _parse_statistic ( data [ i : ( i + 4 ) ] , 0 ) ) , 'temperature' : ArloBaseStation . _parse_statistic ( data [ ( i + 8 ) : ( i + 10 ) ] , 1 ) , 'humidity' : ArloBaseStation . _parse_statistic ( data [ ( i + 14 ) : ( i + 16 ) ] , 1 ) , 'airQuality' : ArloBaseStation . _parse_statistic ( data [ ( i + 20 ) : ( i + 22 ) ] , 1 ) } ) i += 22 return points | Decode decompress and parse the data from the history API | 242 | 12 |
239,027 | def _parse_statistic ( data , scale ) : i = 0 for byte in bytearray ( data ) : i = ( i << 8 ) + byte if i == 32768 : return None if scale == 0 : return i return float ( i ) / ( scale * 10 ) | Parse binary statistics returned from the history API | 61 | 9 |
239,028 | def play_track ( self , track_id = DEFAULT_TRACK_ID , position = 0 ) : self . publish ( action = 'playTrack' , resource = 'audioPlayback/player' , publish_response = False , properties = { 'trackId' : track_id , 'position' : position } ) | Plays a track at the given position . | 70 | 9 |
239,029 | def set_shuffle_on ( self ) : self . publish ( action = 'set' , resource = 'audioPlayback/config' , publish_response = False , properties = { 'config' : { 'shuffleActive' : True } } ) | Sets playback to shuffle . | 55 | 6 |
239,030 | def set_shuffle_off ( self ) : self . publish ( action = 'set' , resource = 'audioPlayback/config' , publish_response = False , properties = { 'config' : { 'shuffleActive' : False } } ) | Sets playback to sequential . | 55 | 6 |
239,031 | def set_night_light_on ( self ) : self . publish ( action = 'set' , resource = 'cameras/{}' . format ( self . device_id ) , publish_response = False , properties = { 'nightLight' : { 'enabled' : True } } ) | Turns on the night light . | 65 | 7 |
239,032 | def set_night_light_off ( self ) : self . publish ( action = 'set' , resource = 'cameras/{}' . format ( self . device_id ) , publish_response = False , properties = { 'nightLight' : { 'enabled' : False } } ) | Turns off the night light . | 65 | 7 |
239,033 | def mode ( self , mode ) : modes = self . available_modes if ( not modes ) or ( mode not in modes ) : return self . publish ( action = 'set' , resource = 'modes' if mode != 'schedule' else 'schedule' , mode = mode , publish_response = True ) self . update ( ) | Set Arlo camera mode . | 74 | 6 |
239,034 | def unindent ( self , lines ) : indent = min ( len ( self . re . match ( r'^ *' , line ) . group ( ) ) for line in lines ) return [ line [ indent : ] . rstrip ( ) for line in lines ] | Removes any indentation that is common to all of the given lines . | 57 | 15 |
239,035 | def get_call_exprs ( self , line ) : line = line . lstrip ( ) try : tree = self . ast . parse ( line ) except SyntaxError : return None for node in self . ast . walk ( tree ) : if isinstance ( node , self . ast . Call ) : offsets = [ ] for arg in node . args : # In Python 3.4 the col_offset is calculated wrong. See # https://bugs.python.org/issue21295 if isinstance ( arg , self . ast . Attribute ) and ( ( 3 , 4 , 0 ) <= self . sys . version_info <= ( 3 , 4 , 3 ) ) : offsets . append ( arg . col_offset - len ( arg . value . id ) - 1 ) else : offsets . append ( arg . col_offset ) if node . keywords : line = line [ : node . keywords [ 0 ] . value . col_offset ] line = self . re . sub ( r'\w+\s*=\s*$' , '' , line ) else : line = self . re . sub ( r'\s*\)\s*$' , '' , line ) offsets . append ( len ( line ) ) args = [ ] for i in range ( len ( node . args ) ) : args . append ( line [ offsets [ i ] : offsets [ i + 1 ] ] . rstrip ( ', ' ) ) return args | Gets the argument expressions from the source of a function call . | 306 | 13 |
239,036 | def show ( self , func_name , values , labels = None ) : s = self . Stanza ( self . indent ) if func_name == '<module>' and self . in_console : func_name = '<console>' s . add ( [ func_name + ': ' ] ) reprs = map ( self . safe_repr , values ) if labels : sep = '' for label , repr in zip ( labels , reprs ) : s . add ( [ label + '=' , self . CYAN , repr , self . NORMAL ] , sep ) sep = ', ' else : sep = '' for repr in reprs : s . add ( [ self . CYAN , repr , self . NORMAL ] , sep ) sep = ', ' self . writer . write ( s . chunks ) | Prints out nice representations of the given values . | 174 | 10 |
239,037 | def trace ( self , func ) : def wrapper ( * args , * * kwargs ) : # Print out the call to the function with its arguments. s = self . Stanza ( self . indent ) s . add ( [ self . GREEN , func . __name__ , self . NORMAL , '(' ] ) s . indent += 4 sep = '' for arg in args : s . add ( [ self . CYAN , self . safe_repr ( arg ) , self . NORMAL ] , sep ) sep = ', ' for name , value in sorted ( kwargs . items ( ) ) : s . add ( [ name + '=' , self . CYAN , self . safe_repr ( value ) , self . NORMAL ] , sep ) sep = ', ' s . add ( ')' , wrap = False ) self . writer . write ( s . chunks ) # Call the function. self . indent += 2 try : result = func ( * args , * * kwargs ) except : # Display an exception. self . indent -= 2 etype , evalue , etb = self . sys . exc_info ( ) info = self . inspect . getframeinfo ( etb . tb_next , context = 3 ) s = self . Stanza ( self . indent ) s . add ( [ self . RED , '!> ' , self . safe_repr ( evalue ) , self . NORMAL ] ) s . add ( [ 'at ' , info . filename , ':' , info . lineno ] , ' ' ) lines = self . unindent ( info . code_context ) firstlineno = info . lineno - info . index fmt = '%' + str ( len ( str ( firstlineno + len ( lines ) ) ) ) + 'd' for i , line in enumerate ( lines ) : s . newline ( ) s . add ( [ i == info . index and self . MAGENTA or '' , fmt % ( i + firstlineno ) , i == info . index and '> ' or ': ' , line , self . NORMAL ] ) self . writer . write ( s . chunks ) raise # Display the return value. self . indent -= 2 s = self . Stanza ( self . indent ) s . add ( [ self . GREEN , '-> ' , self . CYAN , self . safe_repr ( result ) , self . NORMAL ] ) self . writer . write ( s . chunks ) return result return self . functools . update_wrapper ( wrapper , func ) | Decorator to print out a function s arguments and return value . | 546 | 14 |
239,038 | def d ( self , depth = 1 ) : info = self . inspect . getframeinfo ( self . sys . _getframe ( 1 ) ) s = self . Stanza ( self . indent ) s . add ( [ info . function + ': ' ] ) s . add ( [ self . MAGENTA , 'Interactive console opened' , self . NORMAL ] ) self . writer . write ( s . chunks ) frame = self . sys . _getframe ( depth ) env = frame . f_globals . copy ( ) env . update ( frame . f_locals ) self . indent += 2 self . in_console = True self . code . interact ( 'Python console opened by q.d() in ' + info . function , local = env ) self . in_console = False self . indent -= 2 s = self . Stanza ( self . indent ) s . add ( [ info . function + ': ' ] ) s . add ( [ self . MAGENTA , 'Interactive console closed' , self . NORMAL ] ) self . writer . write ( s . chunks ) | Launches an interactive console at the point where it s called . | 235 | 13 |
239,039 | def swagger ( self ) : # prepare Swagger.host & Swagger.basePath if not self . __swagger : return None common_path = os . path . commonprefix ( list ( self . __swagger . paths ) ) # remove tailing slash, # because all paths in Paths Object would prefixed with slah. common_path = common_path [ : - 1 ] if common_path [ - 1 ] == '/' else common_path if len ( common_path ) > 0 : p = six . moves . urllib . parse . urlparse ( common_path ) self . __swagger . update_field ( 'host' , p . netloc ) new_common_path = six . moves . urllib . parse . urlunparse ( ( p . scheme , p . netloc , '' , '' , '' , '' ) ) new_path = { } for k in self . __swagger . paths . keys ( ) : new_path [ k [ len ( new_common_path ) : ] ] = self . __swagger . paths [ k ] self . __swagger . update_field ( 'paths' , new_path ) return self . __swagger | some preparation before returning Swagger object | 260 | 7 |
239,040 | def scope_compose ( scope , name , sep = private . SCOPE_SEPARATOR ) : if name == None : new_scope = scope else : new_scope = scope if scope else name if scope and name : new_scope = scope + sep + name return new_scope | compose a new scope | 61 | 5 |
239,041 | def nv_tuple_list_replace ( l , v ) : _found = False for i , x in enumerate ( l ) : if x [ 0 ] == v [ 0 ] : l [ i ] = v _found = True if not _found : l . append ( v ) | replace a tuple in a tuple list | 63 | 7 |
239,042 | def url_dirname ( url ) : p = six . moves . urllib . parse . urlparse ( url ) for e in [ private . FILE_EXT_JSON , private . FILE_EXT_YAML ] : if p . path . endswith ( e ) : return six . moves . urllib . parse . urlunparse ( p [ : 2 ] + ( os . path . dirname ( p . path ) , ) + p [ 3 : ] ) return url | Return the folder containing the . json file | 105 | 8 |
239,043 | def url_join ( url , path ) : p = six . moves . urllib . parse . urlparse ( url ) t = None if p . path and p . path [ - 1 ] == '/' : if path and path [ 0 ] == '/' : path = path [ 1 : ] t = '' . join ( [ p . path , path ] ) else : t = ( '' if path and path [ 0 ] == '/' else '/' ) . join ( [ p . path , path ] ) return six . moves . urllib . parse . urlunparse ( p [ : 2 ] + ( t , ) + # os.sep is different on windows, don't use it here. p [ 3 : ] ) | url version of os . path . join | 157 | 8 |
239,044 | def derelativise_url ( url ) : parsed = six . moves . urllib . parse . urlparse ( url ) newpath = [ ] for chunk in parsed . path [ 1 : ] . split ( '/' ) : if chunk == '.' : continue elif chunk == '..' : # parent dir. newpath = newpath [ : - 1 ] continue # TODO: Verify this behaviour. elif _fullmatch ( r'\.{3,}' , chunk ) is not None : # parent dir. newpath = newpath [ : - 1 ] continue newpath += [ chunk ] return six . moves . urllib . parse . urlunparse ( parsed [ : 2 ] + ( '/' + ( '/' . join ( newpath ) ) , ) + parsed [ 3 : ] ) | Normalizes URLs gets rid of .. and . | 175 | 9 |
239,045 | def get_swagger_version ( obj ) : if isinstance ( obj , dict ) : if 'swaggerVersion' in obj : return obj [ 'swaggerVersion' ] elif 'swagger' in obj : return obj [ 'swagger' ] return None else : # should be an instance of BaseObj return obj . swaggerVersion if hasattr ( obj , 'swaggerVersion' ) else obj . swagger | get swagger version from loaded json | 90 | 7 |
239,046 | def walk ( start , ofn , cyc = None ) : ctx , stk = { } , [ start ] cyc = [ ] if cyc == None else cyc while len ( stk ) : top = stk [ - 1 ] if top not in ctx : ctx . update ( { top : list ( ofn ( top ) ) } ) if len ( ctx [ top ] ) : n = ctx [ top ] [ 0 ] if n in stk : # cycles found, # normalize the representation of cycles, # start from the smallest vertex, ex. # 4 -> 5 -> 2 -> 7 -> 9 would produce # (2, 7, 9, 4, 5) nc = stk [ stk . index ( n ) : ] ni = nc . index ( min ( nc ) ) nc = nc [ ni : ] + nc [ : ni ] + [ min ( nc ) ] if nc not in cyc : cyc . append ( nc ) ctx [ top ] . pop ( 0 ) else : stk . append ( n ) else : ctx . pop ( top ) stk . pop ( ) if len ( stk ) : ctx [ stk [ - 1 ] ] . remove ( top ) return cyc | Non recursive DFS to detect cycles | 277 | 7 |
239,047 | def _validate_param ( self , path , obj , _ ) : errs = [ ] if obj . allowMultiple : if not obj . paramType in ( 'path' , 'query' , 'header' ) : errs . append ( 'allowMultiple should be applied on path, header, or query parameters' ) if obj . type == 'array' : errs . append ( 'array Type with allowMultiple is not supported.' ) if obj . paramType == 'body' and obj . name not in ( '' , 'body' ) : errs . append ( 'body parameter with invalid name: {0}' . format ( obj . name ) ) if obj . type == 'File' : if obj . paramType != 'form' : errs . append ( 'File parameter should be form type: {0}' . format ( obj . name ) ) if 'multipart/form-data' not in obj . _parent_ . consumes : errs . append ( 'File parameter should consume multipart/form-data: {0}' . format ( obj . name ) ) if obj . type == 'void' : errs . append ( 'void is only allowed in Operation object.' ) return path , obj . __class__ . __name__ , errs | validate option combination of Parameter object | 273 | 8 |
239,048 | def _validate_items ( self , path , obj , _ ) : errs = [ ] if obj . type == 'void' : errs . append ( 'void is only allowed in Operation object.' ) return path , obj . __class__ . __name__ , errs | validate option combination of Property object | 60 | 7 |
239,049 | def _validate_auth ( self , path , obj , _ ) : errs = [ ] if obj . type == 'apiKey' : if not obj . passAs : errs . append ( 'need "passAs" for apiKey' ) if not obj . keyname : errs . append ( 'need "keyname" for apiKey' ) elif obj . type == 'oauth2' : if not obj . grantTypes : errs . append ( 'need "grantTypes" for oauth2' ) return path , obj . __class__ . __name__ , errs | validate that apiKey and oauth2 requirements | 130 | 10 |
239,050 | def _validate_granttype ( self , path , obj , _ ) : errs = [ ] if not obj . implicit and not obj . authorization_code : errs . append ( 'Either implicit or authorization_code should be defined.' ) return path , obj . __class__ . __name__ , errs | make sure either implicit or authorization_code is defined | 68 | 10 |
239,051 | def _validate_auths ( self , path , obj , app ) : errs = [ ] for k , v in six . iteritems ( obj . authorizations or { } ) : if k not in app . raw . authorizations : errs . append ( 'auth {0} not found in resource list' . format ( k ) ) if app . raw . authorizations [ k ] . type in ( 'basicAuth' , 'apiKey' ) and v != [ ] : errs . append ( 'auth {0} should be an empty list' . format ( k ) ) return path , obj . __class__ . __name__ , errs | make sure that apiKey and basicAuth are empty list in Operation object . | 141 | 15 |
239,052 | def default_tree_traversal ( root , leaves ) : objs = [ ( '#' , root ) ] while len ( objs ) > 0 : path , obj = objs . pop ( ) # name of child are json-pointer encoded, we don't have # to encode it again. if obj . __class__ not in leaves : objs . extend ( map ( lambda i : ( path + '/' + i [ 0 ] , ) + ( i [ 1 ] , ) , six . iteritems ( obj . _children_ ) ) ) # the path we expose here follows JsonPointer described here # http://tools.ietf.org/html/draft-ietf-appsawg-json-pointer-07 yield path , obj | default tree traversal | 162 | 4 |
239,053 | def render_all ( self , op , exclude = [ ] , opt = None ) : opt = self . default ( ) if opt == None else opt if not isinstance ( op , Operation ) : raise ValueError ( 'Not a Operation: {0}' . format ( op ) ) if not isinstance ( opt , dict ) : raise ValueError ( 'Not a dict: {0}' . format ( opt ) ) template = opt [ 'parameter_template' ] max_p = opt [ 'max_parameter' ] out = { } for p in op . parameters : if p . name in exclude : continue if p . name in template : out . update ( { p . name : template [ p . name ] } ) continue if not max_p and not p . required : if random . randint ( 0 , 1 ) == 0 or opt [ 'minimal_parameter' ] : continue out . update ( { p . name : self . render ( p , opt = opt ) } ) return out | render a set of parameter for an Operation | 217 | 8 |
239,054 | def _prepare_forms ( self ) : content_type = 'application/x-www-form-urlencoded' if self . __op . consumes and content_type not in self . __op . consumes : raise errs . SchemaError ( 'content type {0} does not present in {1}' . format ( content_type , self . __op . consumes ) ) return content_type , six . moves . urllib . parse . urlencode ( self . __p [ 'formData' ] ) | private function to prepare content for paramType = form | 114 | 10 |
239,055 | def _prepare_body ( self ) : content_type = self . __consume if not content_type : content_type = self . __op . consumes [ 0 ] if self . __op . consumes else 'application/json' if self . __op . consumes and content_type not in self . __op . consumes : raise errs . SchemaError ( 'content type {0} does not present in {1}' . format ( content_type , self . __op . consumes ) ) # according to spec, payload should be one and only, # so we just return the first value in dict. for parameter in self . __op . parameters : parameter = final ( parameter ) if getattr ( parameter , 'in' ) == 'body' : schema = deref ( parameter . schema ) _type = schema . type _format = schema . format name = schema . name body = self . __p [ 'body' ] [ parameter . name ] return content_type , self . __op . _mime_codec . marshal ( content_type , body , _type = _type , _format = _format , name = name ) return None , None | private function to prepare content for paramType = body | 250 | 10 |
239,056 | def _prepare_files ( self , encoding ) : content_type = 'multipart/form-data' if self . __op . consumes and content_type not in self . __op . consumes : raise errs . SchemaError ( 'content type {0} does not present in {1}' . format ( content_type , self . __op . consumes ) ) boundary = uuid4 ( ) . hex content_type += '; boundary={0}' content_type = content_type . format ( boundary ) # init stream body = io . BytesIO ( ) w = codecs . getwriter ( encoding ) def append ( name , obj ) : body . write ( six . b ( '--{0}\r\n' . format ( boundary ) ) ) # header w ( body ) . write ( 'Content-Disposition: form-data; name="{0}"; filename="{1}"' . format ( name , obj . filename ) ) body . write ( six . b ( '\r\n' ) ) if 'Content-Type' in obj . header : w ( body ) . write ( 'Content-Type: {0}' . format ( obj . header [ 'Content-Type' ] ) ) body . write ( six . b ( '\r\n' ) ) if 'Content-Transfer-Encoding' in obj . header : w ( body ) . write ( 'Content-Transfer-Encoding: {0}' . format ( obj . header [ 'Content-Transfer-Encoding' ] ) ) body . write ( six . b ( '\r\n' ) ) body . write ( six . b ( '\r\n' ) ) # body if not obj . data : with open ( obj . filename , 'rb' ) as f : body . write ( f . read ( ) ) else : data = obj . data . read ( ) if isinstance ( data , six . text_type ) : w ( body ) . write ( data ) else : body . write ( data ) body . write ( six . b ( '\r\n' ) ) for k , v in self . __p [ 'formData' ] : body . write ( six . b ( '--{0}\r\n' . format ( boundary ) ) ) w ( body ) . write ( 'Content-Disposition: form-data; name="{0}"' . format ( k ) ) body . write ( six . b ( '\r\n' ) ) body . write ( six . b ( '\r\n' ) ) w ( body ) . write ( v ) body . write ( six . b ( '\r\n' ) ) # begin of file section for k , v in six . iteritems ( self . __p [ 'file' ] ) : if isinstance ( v , list ) : for vv in v : append ( k , vv ) else : append ( k , v ) # final boundary body . write ( six . b ( '--{0}--\r\n' . format ( boundary ) ) ) return content_type , body . getvalue ( ) | private function to prepare content for paramType = form with File | 677 | 12 |
239,057 | def prepare ( self , scheme = 'http' , handle_files = True , encoding = 'utf-8' ) : if isinstance ( scheme , list ) : if self . __scheme is None : scheme = scheme . pop ( ) else : if self . __scheme in scheme : scheme = self . __scheme else : raise Exception ( 'preferred scheme:{} is not supported by the client or spec:{}' . format ( self . __scheme , scheme ) ) elif not isinstance ( scheme , six . string_types ) : raise ValueError ( '"scheme" should be a list or string' ) # combine path parameters into path # TODO: 'dot' is allowed in swagger, but it won't work in python-format path_params = { } for k , v in six . iteritems ( self . __p [ 'path' ] ) : path_params [ k ] = six . moves . urllib . parse . quote_plus ( v ) self . __path = self . __path . format ( * * path_params ) # combine path parameters into url self . __url = '' . join ( [ scheme , ':' , self . __url . format ( * * path_params ) ] ) # header parameters self . __header . update ( self . __p [ 'header' ] ) # update data parameter content_type = None if self . __p [ 'file' ] : if handle_files : content_type , self . __data = self . _prepare_files ( encoding ) else : # client that can encode multipart/form-data, should # access form-data via data property and file from file # property. # only form data can be carried along with files, self . __data = self . __p [ 'formData' ] elif self . __p [ 'formData' ] : content_type , self . __data = self . _prepare_forms ( ) elif self . __p [ 'body' ] : content_type , self . __data = self . _prepare_body ( ) else : self . __data = None if content_type : self . __header . update ( { 'Content-Type' : content_type } ) accept = self . __produce if not accept and self . __op . produces : accept = self . __op . produces [ 0 ] if accept : if self . __op . produces and accept not in self . __op . produces : raise errs . SchemaError ( 'accept {0} does not present in {1}' . format ( accept , self . __op . produces ) ) self . __header . update ( { 'Accept' : accept } ) return self | make this request ready for Clients | 581 | 7 |
239,058 | def apply_with ( self , status = None , raw = None , header = None ) : if status != None : self . __status = status r = ( final ( self . __op . responses . get ( str ( self . __status ) , None ) ) or final ( self . __op . responses . get ( 'default' , None ) ) ) if header != None : if isinstance ( header , ( collections . Mapping , collections . MutableMapping ) ) : for k , v in six . iteritems ( header ) : self . _convert_header ( r , k , v ) else : for k , v in header : self . _convert_header ( r , k , v ) if raw != None : # update 'raw' self . __raw = raw if self . __status == None : raise Exception ( 'Update status code before assigning raw data' ) if r and r . schema and not self . __raw_body_only : # update data from Opeartion if succeed else from responseMessage.responseModel content_type = 'application/json' for k , v in six . iteritems ( self . header ) : if k . lower ( ) == 'content-type' : content_type = v [ 0 ] . lower ( ) break schema = deref ( r . schema ) _type = schema . type _format = schema . format name = schema . name data = self . __op . _mime_codec . unmarshal ( content_type , self . raw , _type = _type , _format = _format , name = name ) self . __data = r . schema . _prim_ ( data , self . __op . _prim_factory , ctx = dict ( read = True ) ) return self | update header status code raw datum ... etc | 379 | 9 |
239,059 | def parse ( self , obj = None ) : if obj == None : return if not isinstance ( obj , dict ) : raise ValueError ( 'invalid obj passed: ' + str ( type ( obj ) ) ) def _apply ( x , kk , ct , v ) : if key not in self . _obj : self . _obj [ kk ] = { } if ct == None else [ ] with x ( self . _obj , kk ) as ctx : ctx . parse ( obj = v ) def _apply_dict ( x , kk , ct , v , k ) : if k not in self . _obj [ kk ] : self . _obj [ kk ] [ k ] = { } if ct == ContainerType . dict_ else [ ] with x ( self . _obj [ kk ] , k ) as ctx : ctx . parse ( obj = v ) def _apply_dict_before_list ( kk , ct , v , k ) : self . _obj [ kk ] [ k ] = [ ] if hasattr ( self , '__swagger_child__' ) : # to nested objects for key , ( ct , ctx_kls ) in six . iteritems ( self . __swagger_child__ ) : items = obj . get ( key , None ) # create an empty child, even it's None in input. # this makes other logic easier. if ct == ContainerType . list_ : self . _obj [ key ] = [ ] elif ct : self . _obj [ key ] = { } if items == None : continue container_apply ( ct , items , functools . partial ( _apply , ctx_kls , key ) , functools . partial ( _apply_dict , ctx_kls , key ) , functools . partial ( _apply_dict_before_list , key ) ) # update _obj with obj if self . _obj != None : for key in ( set ( obj . keys ( ) ) - set ( self . _obj . keys ( ) ) ) : self . _obj [ key ] = obj [ key ] else : self . _obj = obj | major part do parsing . | 478 | 5 |
239,060 | def _assign_parent ( self , ctx ) : def _assign ( cls , _ , obj ) : if obj == None : return if cls . is_produced ( obj ) : if isinstance ( obj , BaseObj ) : obj . _parent__ = self else : raise ValueError ( 'Object is not instance of {0} but {1}' . format ( cls . __swagger_ref_object__ . __name__ , obj . __class__ . __name__ ) ) # set self as childrent's parent for name , ( ct , ctx ) in six . iteritems ( ctx . __swagger_child__ ) : obj = getattr ( self , name ) if obj == None : continue container_apply ( ct , obj , functools . partial ( _assign , ctx ) ) | parent assignment internal usage only | 183 | 5 |
239,061 | def get_private_name ( self , f ) : f = self . __swagger_rename__ [ f ] if f in self . __swagger_rename__ . keys ( ) else f return '_' + self . __class__ . __name__ + '__' + f | get private protected name of an attribute | 64 | 7 |
239,062 | def update_field ( self , f , obj ) : n = self . get_private_name ( f ) if not hasattr ( self , n ) : raise AttributeError ( '{0} is not in {1}' . format ( n , self . __class__ . __name__ ) ) setattr ( self , n , obj ) self . __origin_keys . add ( f ) | update a field | 86 | 3 |
239,063 | def resolve ( self , ts ) : if isinstance ( ts , six . string_types ) : ts = [ ts ] obj = self while len ( ts ) > 0 : t = ts . pop ( 0 ) if issubclass ( obj . __class__ , BaseObj ) : obj = getattr ( obj , t ) elif isinstance ( obj , list ) : obj = obj [ int ( t ) ] elif isinstance ( obj , dict ) : obj = obj [ t ] return obj | resolve a list of tokens to an child object | 106 | 10 |
239,064 | def compare ( self , other , base = None ) : if self . __class__ != other . __class__ : return False , '' names = self . _field_names_ def cmp_func ( name , s , o ) : # special case for string types if isinstance ( s , six . string_types ) and isinstance ( o , six . string_types ) : return s == o , name if s . __class__ != o . __class__ : return False , name if isinstance ( s , BaseObj ) : if not isinstance ( s , weakref . ProxyTypes ) : return s . compare ( o , name ) elif isinstance ( s , list ) : for i , v in zip ( range ( len ( s ) ) , s ) : same , n = cmp_func ( jp_compose ( str ( i ) , name ) , v , o [ i ] ) if not same : return same , n elif isinstance ( s , dict ) : # compare if any key diff diff = list ( set ( s . keys ( ) ) - set ( o . keys ( ) ) ) if diff : return False , jp_compose ( str ( diff [ 0 ] ) , name ) diff = list ( set ( o . keys ( ) ) - set ( s . keys ( ) ) ) if diff : return False , jp_compose ( str ( diff [ 0 ] ) , name ) for k , v in six . iteritems ( s ) : same , n = cmp_func ( jp_compose ( k , name ) , v , o [ k ] ) if not same : return same , n else : return s == o , name return True , name for n in names : same , n = cmp_func ( jp_compose ( n , base ) , getattr ( self , n ) , getattr ( other , n ) ) if not same : return same , n return True , '' | comparison will return the first difference | 421 | 8 |
239,065 | def _field_names_ ( self ) : ret = [ ] for n in six . iterkeys ( self . __swagger_fields__ ) : new_n = self . __swagger_rename__ . get ( n , None ) ret . append ( new_n ) if new_n else ret . append ( n ) return ret | get list of field names defined in Swagger spec | 73 | 10 |
239,066 | def _children_ ( self ) : ret = { } names = self . _field_names_ def down ( name , obj ) : if isinstance ( obj , BaseObj ) : if not isinstance ( obj , weakref . ProxyTypes ) : ret [ name ] = obj elif isinstance ( obj , list ) : for i , v in zip ( range ( len ( obj ) ) , obj ) : down ( jp_compose ( str ( i ) , name ) , v ) elif isinstance ( obj , dict ) : for k , v in six . iteritems ( obj ) : down ( jp_compose ( k , name ) , v ) for n in names : down ( jp_compose ( n ) , getattr ( self , n ) ) return ret | get children objects | 169 | 3 |
239,067 | def load ( kls , url , getter = None , parser = None , url_load_hook = None , sep = consts . private . SCOPE_SEPARATOR , prim = None , mime_codec = None , resolver = None ) : logger . info ( 'load with [{0}]' . format ( url ) ) url = utils . normalize_url ( url ) app = kls ( url , url_load_hook = url_load_hook , sep = sep , prim = prim , mime_codec = mime_codec , resolver = resolver ) app . __raw , app . __version = app . load_obj ( url , getter = getter , parser = parser ) if app . __version not in [ '1.2' , '2.0' ] : raise NotImplementedError ( 'Unsupported Version: {0}' . format ( self . __version ) ) # update scheme if any p = six . moves . urllib . parse . urlparse ( url ) if p . scheme : app . schemes . append ( p . scheme ) return app | load json as a raw App | 245 | 6 |
239,068 | def prepare ( self , strict = True ) : self . __root = self . prepare_obj ( self . raw , self . __url ) self . validate ( strict = strict ) if hasattr ( self . __root , 'schemes' ) and self . __root . schemes : if len ( self . __root . schemes ) > 0 : self . __schemes = self . __root . schemes else : # extract schemes from the url to load spec self . __schemes = [ six . moves . urlparse ( self . __url ) . schemes ] s = Scanner ( self ) s . scan ( root = self . __root , route = [ Merge ( ) ] ) s . scan ( root = self . __root , route = [ PatchObject ( ) ] ) s . scan ( root = self . __root , route = [ Aggregate ( ) ] ) # reducer for Operation tr = TypeReduce ( self . __sep ) cy = CycleDetector ( ) s . scan ( root = self . __root , route = [ tr , cy ] ) # 'op' -- shortcut for Operation with tag and operaionId self . __op = utils . ScopeDict ( tr . op ) # 'm' -- shortcut for model in Swagger 1.2 if hasattr ( self . __root , 'definitions' ) and self . __root . definitions != None : self . __m = utils . ScopeDict ( self . __root . definitions ) else : self . __m = utils . ScopeDict ( { } ) # update scope-separater self . __m . sep = self . __sep self . __op . sep = self . __sep # cycle detection if len ( cy . cycles [ 'schema' ] ) > 0 and strict : raise errs . CycleDetectionError ( 'Cycles detected in Schema Object: {0}' . format ( cy . cycles [ 'schema' ] ) ) | preparation for loaded json | 421 | 6 |
239,069 | def create ( kls , url , strict = True ) : app = kls . load ( url ) app . prepare ( strict = strict ) return app | factory of App | 32 | 4 |
239,070 | def resolve ( self , jref , parser = None ) : logger . info ( 'resolving: [{0}]' . format ( jref ) ) if jref == None or len ( jref ) == 0 : raise ValueError ( 'Empty Path is not allowed' ) obj = None url , jp = utils . jr_split ( jref ) # check cacahed object against json reference by # comparing url first, and find those object prefixed with # the JSON pointer. o = self . __objs . get ( url , None ) if o : if isinstance ( o , BaseObj ) : obj = o . resolve ( utils . jp_split ( jp ) [ 1 : ] ) elif isinstance ( o , dict ) : for k , v in six . iteritems ( o ) : if jp . startswith ( k ) : obj = v . resolve ( utils . jp_split ( jp [ len ( k ) : ] ) [ 1 : ] ) break else : raise Exception ( 'Unknown Cached Object: {0}' . format ( str ( type ( o ) ) ) ) # this object is not found in cache if obj == None : if url : obj , _ = self . load_obj ( jref , parser = parser ) if obj : obj = self . prepare_obj ( obj , jref ) else : # a local reference, 'jref' is just a json-pointer if not jp . startswith ( '#' ) : raise ValueError ( 'Invalid Path, root element should be \'#\', but [{0}]' . format ( jref ) ) obj = self . root . resolve ( utils . jp_split ( jp ) [ 1 : ] ) # heading element is #, mapping to self.root if obj == None : raise ValueError ( 'Unable to resolve path, [{0}]' . format ( jref ) ) if isinstance ( obj , ( six . string_types , six . integer_types , list , dict ) ) : return obj return weakref . proxy ( obj ) | JSON reference resolver | 452 | 4 |
239,071 | def prepare_schemes ( self , req ) : # fix test bug when in python3 scheme, more details in commint msg ret = sorted ( self . __schemes__ & set ( req . schemes ) , reverse = True ) if len ( ret ) == 0 : raise ValueError ( 'No schemes available: {0}' . format ( req . schemes ) ) return ret | make sure this client support schemes required by current request | 82 | 10 |
239,072 | def compose_headers ( self , req , headers = None , opt = None , as_dict = False ) : if headers is None : return list ( req . header . items ( ) ) if not as_dict else req . header opt = opt or { } join_headers = opt . pop ( BaseClient . join_headers , None ) if as_dict and not join_headers : # pick the most efficient way for special case headers = dict ( headers ) if isinstance ( headers , list ) else headers headers . update ( req . header ) return headers # include Request.headers aggregated_headers = list ( req . header . items ( ) ) # include customized headers if isinstance ( headers , list ) : aggregated_headers . extend ( headers ) elif isinstance ( headers , dict ) : aggregated_headers . extend ( headers . items ( ) ) else : raise Exception ( 'unknown type as header: {}' . format ( str ( type ( headers ) ) ) ) if join_headers : joined = { } for h in aggregated_headers : key = h [ 0 ] if key in joined : joined [ key ] = ',' . join ( [ joined [ key ] , h [ 1 ] ] ) else : joined [ key ] = h [ 1 ] aggregated_headers = list ( joined . items ( ) ) return dict ( aggregated_headers ) if as_dict else aggregated_headers | a utility to compose headers from pyswagger . io . Request and customized headers | 298 | 17 |
239,073 | def request ( self , req_and_resp , opt = None , headers = None ) : req , resp = req_and_resp # dump info for debugging logger . info ( 'request.url: {0}' . format ( req . url ) ) logger . info ( 'request.header: {0}' . format ( req . header ) ) logger . info ( 'request.query: {0}' . format ( req . query ) ) logger . info ( 'request.file: {0}' . format ( req . files ) ) logger . info ( 'request.schemes: {0}' . format ( req . schemes ) ) # apply authorizations if self . __security : self . __security ( req ) return req , resp | preprocess before performing a request usually some patching . authorization also applied here . | 162 | 16 |
239,074 | def to_url ( self ) : if self . __collection_format == 'multi' : return [ str ( s ) for s in self ] else : return [ str ( self ) ] | special function for handling multi refer to Swagger 2 . 0 Parameter Object collectionFormat | 40 | 17 |
239,075 | def apply_with ( self , obj , val , ctx ) : for k , v in six . iteritems ( val ) : if k in obj . properties : pobj = obj . properties . get ( k ) if pobj . readOnly == True and ctx [ 'read' ] == False : raise Exception ( 'read-only property is set in write context.' ) self [ k ] = ctx [ 'factory' ] . produce ( pobj , v ) # TODO: patternProperties here # TODO: fix bug, everything would not go into additionalProperties, instead of recursive elif obj . additionalProperties == True : ctx [ 'addp' ] = True elif obj . additionalProperties not in ( None , False ) : ctx [ 'addp_schema' ] = obj in_obj = set ( six . iterkeys ( obj . properties ) ) in_self = set ( six . iterkeys ( self ) ) other_prop = in_obj - in_self for k in other_prop : p = obj . properties [ k ] if p . is_set ( "default" ) : self [ k ] = ctx [ 'factory' ] . produce ( p , p . default ) not_found = set ( obj . required ) - set ( six . iterkeys ( self ) ) if len ( not_found ) : raise ValueError ( 'Model missing required key(s): {0}' . format ( ', ' . join ( not_found ) ) ) # remove assigned properties to avoid duplicated # primitive creation _val = { } for k in set ( six . iterkeys ( val ) ) - in_obj : _val [ k ] = val [ k ] if obj . discriminator : self [ obj . discriminator ] = ctx [ 'name' ] return _val | recursively apply Schema object | 393 | 7 |
239,076 | def _op ( self , _ , obj , app ) : if obj . responses == None : return tmp = { } for k , v in six . iteritems ( obj . responses ) : if isinstance ( k , six . integer_types ) : tmp [ str ( k ) ] = v else : tmp [ k ] = v obj . update_field ( 'responses' , tmp ) | convert status code in Responses from int to string | 83 | 11 |
239,077 | def produce ( self , obj , val , ctx = None ) : val = obj . default if val == None else val if val == None : return None obj = deref ( obj ) ctx = { } if ctx == None else ctx if 'name' not in ctx and hasattr ( obj , 'name' ) : ctx [ 'name' ] = obj . name if 'guard' not in ctx : ctx [ 'guard' ] = CycleGuard ( ) if 'addp_schema' not in ctx : # Schema Object of additionalProperties ctx [ 'addp_schema' ] = None if 'addp' not in ctx : # additionalProperties ctx [ 'addp' ] = False if '2nd_pass' not in ctx : # 2nd pass processing function ctx [ '2nd_pass' ] = None if 'factory' not in ctx : # primitive factory ctx [ 'factory' ] = self if 'read' not in ctx : # default is in 'read' context ctx [ 'read' ] = True # cycle guard ctx [ 'guard' ] . update ( obj ) ret = None if obj . type : creater , _2nd = self . get ( _type = obj . type , _format = obj . format ) if not creater : raise ValueError ( 'Can\'t resolve type from:(' + str ( obj . type ) + ', ' + str ( obj . format ) + ')' ) ret = creater ( obj , val , ctx ) if _2nd : val = _2nd ( obj , ret , val , ctx ) ctx [ '2nd_pass' ] = _2nd elif len ( obj . properties ) or obj . additionalProperties : ret = Model ( ) val = ret . apply_with ( obj , val , ctx ) if isinstance ( ret , ( Date , Datetime , Byte , File ) ) : # it's meanless to handle allOf for these types. return ret def _apply ( o , r , v , c ) : if hasattr ( ret , 'apply_with' ) : v = r . apply_with ( o , v , c ) else : _2nd = c [ '2nd_pass' ] if _2nd == None : _ , _2nd = self . get ( _type = o . type , _format = o . format ) if _2nd : _2nd ( o , r , v , c ) # update it back to context c [ '2nd_pass' ] = _2nd return v # handle allOf for Schema Object allOf = getattr ( obj , 'allOf' , None ) if allOf : not_applied = [ ] for a in allOf : a = deref ( a ) if not ret : # try to find right type for this primitive. ret = self . produce ( a , val , ctx ) is_member = hasattr ( ret , 'apply_with' ) else : val = _apply ( a , ret , val , ctx ) if not ret : # if we still can't determine the type, # keep this Schema object for later use. not_applied . append ( a ) if ret : for a in not_applied : val = _apply ( a , ret , val , ctx ) if ret != None and hasattr ( ret , 'cleanup' ) : val = ret . cleanup ( val , ctx ) return ret | factory function to create primitives | 758 | 7 |
239,078 | def _check_elementary_axis_name ( name : str ) -> bool : if len ( name ) == 0 : return False if not 'a' <= name [ 0 ] <= 'z' : return False for letter in name : if ( not letter . isdigit ( ) ) and not ( 'a' <= letter <= 'z' ) : return False return True | Valid elementary axes contain only lower latin letters and digits and start with a letter . | 78 | 17 |
239,079 | def open ( self ) : self . _window = Window ( caption = self . caption , height = self . height , width = self . width , vsync = False , resizable = True , ) | Open the window . | 42 | 4 |
239,080 | def show ( self , frame ) : # check that the frame has the correct dimensions if len ( frame . shape ) != 3 : raise ValueError ( 'frame should have shape with only 3 dimensions' ) # open the window if it isn't open already if not self . is_open : self . open ( ) # prepare the window for the next frame self . _window . clear ( ) self . _window . switch_to ( ) self . _window . dispatch_events ( ) # create an image data object image = ImageData ( frame . shape [ 1 ] , frame . shape [ 0 ] , 'RGB' , frame . tobytes ( ) , pitch = frame . shape [ 1 ] * - 3 ) # send the image to the window image . blit ( 0 , 0 , width = self . _window . width , height = self . _window . height ) self . _window . flip ( ) | Show an array of pixels on the window . | 191 | 9 |
239,081 | def display_arr ( screen , arr , video_size , transpose ) : # take the transpose if necessary if transpose : pyg_img = pygame . surfarray . make_surface ( arr . swapaxes ( 0 , 1 ) ) else : pyg_img = arr # resize the image according to given image size pyg_img = pygame . transform . scale ( pyg_img , video_size ) # blit the image to the surface screen . blit ( pyg_img , ( 0 , 0 ) ) | Display an image to the pygame screen . | 116 | 9 |
239,082 | def play_human ( env ) : # play the game and catch a potential keyboard interrupt try : play ( env , fps = env . metadata [ 'video.frames_per_second' ] ) except KeyboardInterrupt : pass # reset and close the environment env . close ( ) | Play the environment using keyboard as a human . | 58 | 9 |
239,083 | def get_action_meanings ( self ) : actions = sorted ( self . _action_meanings . keys ( ) ) return [ self . _action_meanings [ action ] for action in actions ] | Return a list of actions meanings . | 44 | 7 |
239,084 | def prg_rom ( self ) : try : return self . raw_data [ self . prg_rom_start : self . prg_rom_stop ] except IndexError : raise ValueError ( 'failed to read PRG-ROM on ROM.' ) | Return the PRG ROM of the ROM file . | 56 | 10 |
239,085 | def chr_rom ( self ) : try : return self . raw_data [ self . chr_rom_start : self . chr_rom_stop ] except IndexError : raise ValueError ( 'failed to read CHR-ROM on ROM.' ) | Return the CHR ROM of the ROM file . | 55 | 9 |
239,086 | def _screen_buffer ( self ) : # get the address of the screen address = _LIB . Screen ( self . _env ) # create a buffer from the contents of the address location buffer_ = ctypes . cast ( address , ctypes . POINTER ( SCREEN_TENSOR ) ) . contents # create a NumPy array from the buffer screen = np . frombuffer ( buffer_ , dtype = 'uint8' ) # reshape the screen from a column vector to a tensor screen = screen . reshape ( SCREEN_SHAPE_32_BIT ) # flip the bytes if the machine is little-endian (which it likely is) if sys . byteorder == 'little' : # invert the little-endian BGRx channels to big-endian xRGB screen = screen [ : , : , : : - 1 ] # remove the 0th axis (padding from storing colors in 32 bit) return screen [ : , : , 1 : ] | Setup the screen buffer from the C ++ code . | 209 | 10 |
239,087 | def _ram_buffer ( self ) : # get the address of the RAM address = _LIB . Memory ( self . _env ) # create a buffer from the contents of the address location buffer_ = ctypes . cast ( address , ctypes . POINTER ( RAM_VECTOR ) ) . contents # create a NumPy array from the buffer return np . frombuffer ( buffer_ , dtype = 'uint8' ) | Setup the RAM buffer from the C ++ code . | 90 | 10 |
239,088 | def _controller_buffer ( self , port ) : # get the address of the controller address = _LIB . Controller ( self . _env , port ) # create a memory buffer using the ctypes pointer for this vector buffer_ = ctypes . cast ( address , ctypes . POINTER ( CONTROLLER_VECTOR ) ) . contents # create a NumPy buffer from the binary data and return it return np . frombuffer ( buffer_ , dtype = 'uint8' ) | Find the pointer to a controller and setup a NumPy buffer . | 103 | 13 |
239,089 | def _frame_advance ( self , action ) : # set the action on the controller self . controllers [ 0 ] [ : ] = action # perform a step on the emulator _LIB . Step ( self . _env ) | Advance a frame in the emulator with an action . | 47 | 11 |
239,090 | def reset ( self ) : # call the before reset callback self . _will_reset ( ) # reset the emulator if self . _has_backup : self . _restore ( ) else : _LIB . Reset ( self . _env ) # call the after reset callback self . _did_reset ( ) # set the done flag to false self . done = False # return the screen from the emulator return self . screen | Reset the state of the environment and returns an initial observation . | 89 | 13 |
239,091 | def step ( self , action ) : # if the environment is done, raise an error if self . done : raise ValueError ( 'cannot step in a done environment! call `reset`' ) # set the action on the controller self . controllers [ 0 ] [ : ] = action # pass the action to the emulator as an unsigned byte _LIB . Step ( self . _env ) # get the reward for this step reward = self . _get_reward ( ) # get the done flag for this step self . done = self . _get_done ( ) # get the info for this step info = self . _get_info ( ) # call the after step callback self . _did_step ( self . done ) # bound the reward in [min, max] if reward < self . reward_range [ 0 ] : reward = self . reward_range [ 0 ] elif reward > self . reward_range [ 1 ] : reward = self . reward_range [ 1 ] # return the screen from the emulator and other relevant data return self . screen , reward , self . done , info | Run one frame of the NES and return the relevant observation data . | 231 | 13 |
239,092 | def close ( self ) : # make sure the environment hasn't already been closed if self . _env is None : raise ValueError ( 'env has already been closed.' ) # purge the environment from C++ memory _LIB . Close ( self . _env ) # deallocate the object locally self . _env = None # if there is an image viewer open, delete it if self . viewer is not None : self . viewer . close ( ) | Close the environment . | 93 | 4 |
239,093 | def render ( self , mode = 'human' ) : if mode == 'human' : # if the viewer isn't setup, import it and create one if self . viewer is None : from . _image_viewer import ImageViewer # get the caption for the ImageViewer if self . spec is None : # if there is no spec, just use the .nes filename caption = self . _rom_path . split ( '/' ) [ - 1 ] else : # set the caption to the OpenAI Gym id caption = self . spec . id # create the ImageViewer to display frames self . viewer = ImageViewer ( caption = caption , height = SCREEN_HEIGHT , width = SCREEN_WIDTH , ) # show the screen on the image viewer self . viewer . show ( self . screen ) elif mode == 'rgb_array' : return self . screen else : # unpack the modes as comma delineated strings ('a', 'b', ...) render_modes = [ repr ( x ) for x in self . metadata [ 'render.modes' ] ] msg = 'valid render modes are: {}' . format ( ', ' . join ( render_modes ) ) raise NotImplementedError ( msg ) | Render the environment . | 268 | 4 |
239,094 | def _get_args ( ) : parser = argparse . ArgumentParser ( description = __doc__ ) # add the argument for the Super Mario Bros environment to run parser . add_argument ( '--rom' , '-r' , type = str , help = 'The path to the ROM to play.' , required = True , ) # add the argument for the mode of execution as either human or random parser . add_argument ( '--mode' , '-m' , type = str , default = 'human' , choices = [ 'human' , 'random' ] , help = 'The execution mode for the emulation.' , ) # add the argument for the number of steps to take in random mode parser . add_argument ( '--steps' , '-s' , type = int , default = 500 , help = 'The number of random steps to take.' , ) return parser . parse_args ( ) | Parse arguments from the command line and return them . | 194 | 11 |
239,095 | def main ( ) : # get arguments from the command line args = _get_args ( ) # create the environment env = NESEnv ( args . rom ) # play the environment with the given mode if args . mode == 'human' : play_human ( env ) else : play_random ( env , args . steps ) | The main entry point for the command line interface . | 69 | 10 |
239,096 | def play_random ( env , steps ) : try : done = True progress = tqdm ( range ( steps ) ) for _ in progress : if done : _ = env . reset ( ) action = env . action_space . sample ( ) _ , reward , done , info = env . step ( action ) progress . set_postfix ( reward = reward , info = info ) env . render ( ) except KeyboardInterrupt : pass # close the environment env . close ( ) | Play the environment making uniformly random decisions . | 101 | 8 |
239,097 | def markowitz_portfolio ( cov_mat , exp_rets , target_ret , allow_short = False , market_neutral = False ) : if not isinstance ( cov_mat , pd . DataFrame ) : raise ValueError ( "Covariance matrix is not a DataFrame" ) if not isinstance ( exp_rets , pd . Series ) : raise ValueError ( "Expected returns is not a Series" ) if not isinstance ( target_ret , float ) : raise ValueError ( "Target return is not a float" ) if not cov_mat . index . equals ( exp_rets . index ) : raise ValueError ( "Indices do not match" ) if market_neutral and not allow_short : warnings . warn ( "A market neutral portfolio implies shorting" ) allow_short = True n = len ( cov_mat ) P = opt . matrix ( cov_mat . values ) q = opt . matrix ( 0.0 , ( n , 1 ) ) # Constraints Gx <= h if not allow_short : # exp_rets*x >= target_ret and x >= 0 G = opt . matrix ( np . vstack ( ( - exp_rets . values , - np . identity ( n ) ) ) ) h = opt . matrix ( np . vstack ( ( - target_ret , + np . zeros ( ( n , 1 ) ) ) ) ) else : # exp_rets*x >= target_ret G = opt . matrix ( - exp_rets . values ) . T h = opt . matrix ( - target_ret ) # Constraints Ax = b # sum(x) = 1 A = opt . matrix ( 1.0 , ( 1 , n ) ) if not market_neutral : b = opt . matrix ( 1.0 ) else : b = opt . matrix ( 0.0 ) # Solve optsolvers . options [ 'show_progress' ] = False sol = optsolvers . qp ( P , q , G , h , A , b ) if sol [ 'status' ] != 'optimal' : warnings . warn ( "Convergence problem" ) # Put weights into a labeled series weights = pd . Series ( sol [ 'x' ] , index = cov_mat . index ) return weights | Computes a Markowitz portfolio . | 491 | 7 |
239,098 | def min_var_portfolio ( cov_mat , allow_short = False ) : if not isinstance ( cov_mat , pd . DataFrame ) : raise ValueError ( "Covariance matrix is not a DataFrame" ) n = len ( cov_mat ) P = opt . matrix ( cov_mat . values ) q = opt . matrix ( 0.0 , ( n , 1 ) ) # Constraints Gx <= h if not allow_short : # x >= 0 G = opt . matrix ( - np . identity ( n ) ) h = opt . matrix ( 0.0 , ( n , 1 ) ) else : G = None h = None # Constraints Ax = b # sum(x) = 1 A = opt . matrix ( 1.0 , ( 1 , n ) ) b = opt . matrix ( 1.0 ) # Solve optsolvers . options [ 'show_progress' ] = False sol = optsolvers . qp ( P , q , G , h , A , b ) if sol [ 'status' ] != 'optimal' : warnings . warn ( "Convergence problem" ) # Put weights into a labeled series weights = pd . Series ( sol [ 'x' ] , index = cov_mat . index ) return weights | Computes the minimum variance portfolio . | 277 | 7 |
239,099 | def print_portfolio_info ( returns , avg_rets , weights ) : ret = ( weights * avg_rets ) . sum ( ) std = ( weights * returns ) . sum ( 1 ) . std ( ) sharpe = ret / std print ( "Optimal weights:\n{}\n" . format ( weights ) ) print ( "Expected return: {}" . format ( ret ) ) print ( "Expected variance: {}" . format ( std ** 2 ) ) print ( "Expected Sharpe: {}" . format ( sharpe ) ) | Print information on expected portfolio performance . | 119 | 7 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.