idx int64 0 251k | question stringlengths 53 3.53k | target stringlengths 5 1.23k | len_question int64 20 893 | len_target int64 3 238 |
|---|---|---|---|---|
247,000 | def get_queryset ( self , * * kwargs ) -> QuerySet : assert self . queryset is not None , ( "'%s' should either include a `queryset` attribute, " "or override the `get_queryset()` method." % self . __class__ . __name__ ) queryset = self . queryset if isinstance ( queryset , QuerySet ) : # Ensure queryset is re-evaluated on each request. queryset = queryset . all ( ) return queryset | Get the list of items for this view . This must be an iterable and may be a queryset . Defaults to using self . queryset . | 121 | 33 |
247,001 | def get_serializer_class ( self , * * kwargs ) -> Type [ Serializer ] : assert self . serializer_class is not None , ( "'%s' should either include a `serializer_class` attribute, " "or override the `get_serializer_class()` method." % self . __class__ . __name__ ) return self . serializer_class | Return the class to use for the serializer . Defaults to using self . serializer_class . | 85 | 21 |
247,002 | def view_as_consumer ( wrapped_view : typing . Callable [ [ HttpRequest ] , HttpResponse ] , mapped_actions : typing . Optional [ typing . Dict [ str , str ] ] = None ) -> Type [ AsyncConsumer ] : if mapped_actions is None : mapped_actions = { 'create' : 'PUT' , 'update' : 'PATCH' , 'list' : 'GET' , 'retrieve' : 'GET' } class DjangoViewWrapper ( DjangoViewAsConsumer ) : view = wrapped_view actions = mapped_actions return DjangoViewWrapper | Wrap a django View so that it will be triggered by actions over this json websocket consumer . | 129 | 21 |
247,003 | async def check_permissions ( self , action : str , * * kwargs ) : for permission in await self . get_permissions ( action = action , * * kwargs ) : if not await ensure_async ( permission . has_permission ) ( scope = self . scope , consumer = self , action = action , * * kwargs ) : raise PermissionDenied ( ) | Check if the action should be permitted . Raises an appropriate exception if the request is not permitted . | 87 | 20 |
247,004 | async def handle_exception ( self , exc : Exception , action : str , request_id ) : if isinstance ( exc , APIException ) : await self . reply ( action = action , errors = self . _format_errors ( exc . detail ) , status = exc . status_code , request_id = request_id ) elif exc == Http404 or isinstance ( exc , Http404 ) : await self . reply ( action = action , errors = self . _format_errors ( 'Not found' ) , status = 404 , request_id = request_id ) else : raise exc | Handle any exception that occurs by sending an appropriate message | 130 | 10 |
247,005 | async def receive_json ( self , content : typing . Dict , * * kwargs ) : # TODO assert format, if does not match return message. request_id = content . pop ( 'request_id' ) action = content . pop ( 'action' ) await self . handle_action ( action , request_id = request_id , * * content ) | Called with decoded JSON content . | 81 | 8 |
247,006 | def action ( atomic = None , * * kwargs ) : def decorator ( func ) : if atomic is None : _atomic = getattr ( settings , 'ATOMIC_REQUESTS' , False ) else : _atomic = atomic func . action = True func . kwargs = kwargs if asyncio . iscoroutinefunction ( func ) : if _atomic : raise ValueError ( 'Only synchronous actions can be atomic' ) return func if _atomic : # wrap function in atomic wrapper func = transaction . atomic ( func ) @ wraps ( func ) async def async_f ( self : AsyncAPIConsumer , * args , * * _kwargs ) : result , status = await database_sync_to_async ( func ) ( self , * args , * * _kwargs ) return result , status async_f . action = True async_f . kwargs = kwargs async_f . __name__ = func . __name__ return async_f return decorator | Mark a method as an action . | 217 | 7 |
247,007 | def datetime_parser ( s ) : try : ts = arrow . get ( s ) # Convert UTC to local, result of get is UTC unless it specifies timezone, bonfire assumes # all time to be machine local if ts . tzinfo == arrow . get ( ) . tzinfo : ts = ts . replace ( tzinfo = 'local' ) except : c = pdt . Calendar ( ) result , what = c . parse ( s ) ts = None if what in ( 1 , 2 , 3 ) : ts = datetime . datetime ( * result [ : 6 ] ) ts = arrow . get ( ts ) ts = ts . replace ( tzinfo = 'local' ) return ts if ts is None : raise ValueError ( "Cannot parse timestamp '" + s + "'" ) return ts | Parse timestamp s in local time . First the arrow parser is used if it fails the parsedatetime parser is used . | 174 | 25 |
247,008 | def seek ( self , offset : int = 0 , * args , * * kwargs ) : return self . fp . seek ( offset , * args , * * kwargs ) | A shortcut to self . fp . seek . | 40 | 10 |
247,009 | def set_title ( self , title : str , url : str = None ) -> None : self . title = title self . url = url | Sets the title of the embed . | 30 | 8 |
247,010 | def set_timestamp ( self , time : Union [ str , datetime . datetime ] = None , now : bool = False ) -> None : if now : self . timestamp = str ( datetime . datetime . utcnow ( ) ) else : self . timestamp = str ( time ) | Sets the timestamp of the embed . | 63 | 8 |
247,011 | def add_field ( self , name : str , value : str , inline : bool = True ) -> None : field = { 'name' : name , 'value' : value , 'inline' : inline } self . fields . append ( field ) | Adds an embed field . | 53 | 5 |
247,012 | def set_author ( self , name : str , icon_url : str = None , url : str = None ) -> None : self . author = { 'name' : name , 'icon_url' : icon_url , 'url' : url } | Sets the author of the embed . | 55 | 8 |
247,013 | def set_footer ( self , text : str , icon_url : str = None ) -> None : self . footer = { 'text' : text , 'icon_url' : icon_url } | Sets the footer of the embed . | 45 | 9 |
247,014 | async def init ( app , loop ) : app . session = aiohttp . ClientSession ( loop = loop ) # to make web requests app . webhook = Webhook . Async ( webhook_url , session = app . session ) em = Embed ( color = 0x2ecc71 ) em . set_author ( '[INFO] Starting Worker' ) em . description = 'Host: {}' . format ( socket . gethostname ( ) ) await app . webhook . send ( embed = em ) | Sends a message to the webhook channel when server starts . | 111 | 13 |
247,015 | async def server_stop ( app , loop ) : em = Embed ( color = 0xe67e22 ) em . set_footer ( 'Host: {}' . format ( socket . gethostname ( ) ) ) em . description = '[INFO] Server Stopped' await app . webhook . send ( embed = em ) await app . session . close ( ) | Sends a message to the webhook channel when server stops . | 80 | 13 |
247,016 | def get_deprecated_msg ( self , wrapped , instance ) : if instance is None : if inspect . isclass ( wrapped ) : fmt = "Call to deprecated class {name}." else : fmt = "Call to deprecated function (or staticmethod) {name}." else : if inspect . isclass ( instance ) : fmt = "Call to deprecated class method {name}." else : fmt = "Call to deprecated method {name}." if self . reason : fmt += " ({reason})" if self . version : fmt += " -- Deprecated since version {version}." return fmt . format ( name = wrapped . __name__ , reason = self . reason or "" , version = self . version or "" ) | Get the deprecation warning message for the user . | 152 | 11 |
247,017 | def slack_user ( request , api_data ) : if request . user . is_anonymous : return request , api_data data = deepcopy ( api_data ) slacker , _ = SlackUser . objects . get_or_create ( slacker = request . user ) slacker . access_token = data . pop ( 'access_token' ) slacker . extras = data slacker . save ( ) messages . add_message ( request , messages . SUCCESS , 'Your account has been successfully updated with ' 'Slack. You can share your messages within your slack ' 'domain.' ) return request , api_data | Pipeline for backward compatibility prior to 1 . 0 . 0 version . In case if you re willing maintain slack_user table . | 135 | 27 |
247,018 | def read ( varin , fname = 'MS2_L10.mat.txt' ) : # # fname = 'MS09_L10.mat.txt' # # fname = 'MS09_L05.mat.txt' # has PAR # fname = 'MS2_L10.mat.txt' # empty PAR d = np . loadtxt ( fname , comments = '*' ) if fname == 'MS2_L10.mat.txt' : var = [ 'lat' , 'lon' , 'depth' , 'temp' , 'density' , 'sigma' , 'oxygen' , 'voltage 2' , 'voltage 3' , 'fluorescence-CDOM' , 'fluorescence-ECO' , 'turbidity' , 'pressure' , 'salinity' , 'RINKO temperature' , 'RINKO DO - CTD temp' , 'RINKO DO - RINKO temp' , 'bottom' , 'PAR' ] elif ( fname == 'MS09_L05.mat.txt' ) or ( fname == 'MS09_L10.mat.txt' ) or ( fname == 'MS08_L12.mat.txt' ) : var = [ 'lat' , 'lon' , 'depth' , 'temp' , 'density' , 'sigma' , 'oxygen' , 'voltage 2' , 'voltage 3' , 'voltage 4' , 'fluorescence-CDOM' , 'fluorescence-ECO' , 'turbidity' , 'pressure' , 'salinity' , 'RINKO temperature' , 'RINKO DO - CTD temp' , 'RINKO DO - RINKO temp' , 'bottom' , 'PAR' ] # return data for variable varin return d [ : , 0 ] , d [ : , 1 ] , d [ : , 2 ] , d [ : , var . index ( varin ) ] | Read in dataset for variable var | 435 | 6 |
247,019 | def show ( cmap , var , vmin = None , vmax = None ) : # get variable data lat , lon , z , data = read ( var ) fig = plt . figure ( figsize = ( 16 , 12 ) ) # Plot with grayscale ax = fig . add_subplot ( 3 , 1 , 1 ) map1 = ax . scatter ( lon , - z , c = data , cmap = 'gray' , s = 10 , linewidths = 0. , vmin = vmin , vmax = vmax ) plt . colorbar ( map1 , ax = ax ) # Plot with jet ax = fig . add_subplot ( 3 , 1 , 2 ) map1 = ax . scatter ( lon , - z , c = data , cmap = 'jet' , s = 10 , linewidths = 0. , vmin = vmin , vmax = vmax ) plt . colorbar ( map1 , ax = ax ) # Plot with cmap ax = fig . add_subplot ( 3 , 1 , 3 ) map1 = ax . scatter ( lon , - z , c = data , cmap = cmap , s = 10 , linewidths = 0. , vmin = vmin , vmax = vmax ) ax . set_xlabel ( 'Longitude [degrees]' ) ax . set_ylabel ( 'Depth [m]' ) plt . colorbar ( map1 , ax = ax ) plt . suptitle ( var ) | Show a colormap for a chosen input variable var side by side with black and white and jet colormaps . | 331 | 24 |
247,020 | def plot_data ( ) : var = [ 'temp' , 'oxygen' , 'salinity' , 'fluorescence-ECO' , 'density' , 'PAR' , 'turbidity' , 'fluorescence-CDOM' ] # colorbar limits for each property lims = np . array ( [ [ 26 , 33 ] , [ 0 , 10 ] , [ 0 , 36 ] , [ 0 , 6 ] , [ 1005 , 1025 ] , [ 0 , 0.6 ] , [ 0 , 2 ] , [ 0 , 9 ] ] ) # reasonable values # lims = np.array([[20,36], [26,33], [1.5,5.6], [0,4], [0,9], [0,1.5]]) # values to show colormaps for fname in fnames : fig , axes = plt . subplots ( nrows = 4 , ncols = 2 ) fig . set_size_inches ( 20 , 10 ) fig . subplots_adjust ( top = 0.95 , bottom = 0.01 , left = 0.2 , right = 0.99 , wspace = 0.0 , hspace = 0.07 ) i = 0 for ax , Var , cmap in zip ( axes . flat , var , cmaps ) : # loop through data to plot up # get variable data lat , lon , z , data = test . read ( Var , fname ) map1 = ax . scatter ( lat , - z , c = data , cmap = cmap , s = 10 , linewidths = 0. , vmin = lims [ i , 0 ] , vmax = lims [ i , 1 ] ) # no stupid offset y_formatter = mpl . ticker . ScalarFormatter ( useOffset = False ) ax . xaxis . set_major_formatter ( y_formatter ) if i == 6 : ax . set_xlabel ( 'Latitude [degrees]' ) ax . set_ylabel ( 'Depth [m]' ) else : ax . set_xticklabels ( [ ] ) ax . set_yticklabels ( [ ] ) ax . set_ylim ( - z . max ( ) , 0 ) ax . set_xlim ( lat . min ( ) , lat . max ( ) ) cb = plt . colorbar ( map1 , ax = ax , pad = 0.02 ) cb . set_label ( cmap . name + ' [' + '$' + cmap . units + '$]' ) i += 1 fig . savefig ( 'figures/' + fname . split ( '.' ) [ 0 ] + '.png' , bbox_inches = 'tight' ) | Plot sample data up with the fancy colormaps . | 597 | 11 |
247,021 | def plot_lightness ( saveplot = False ) : from colorspacious import cspace_converter dc = 1. x = np . linspace ( 0.0 , 1.0 , 256 ) locs = [ ] # locations for text labels fig = plt . figure ( figsize = ( 16 , 5 ) ) ax = fig . add_subplot ( 111 ) fig . subplots_adjust ( left = 0.03 , right = 0.97 ) ax . set_xlim ( - 0.1 , len ( cm . cmap_d ) / 2. + 0.1 ) ax . set_ylim ( 0 , 100 ) ax . set_xlabel ( 'Lightness for each colormap' , fontsize = 14 ) for j , cmapname in enumerate ( cm . cmapnames ) : if '_r' in cmapname : # skip reversed versions for plot continue cmap = cm . cmap_d [ cmapname ] # get the colormap instance rgb = cmap ( x ) [ np . newaxis , : , : 3 ] lab = cspace_converter ( "sRGB1" , "CAM02-UCS" ) ( rgb ) L = lab [ 0 , : , 0 ] if L [ - 1 ] > L [ 0 ] : ax . scatter ( x + j * dc , L , c = x , cmap = cmap , s = 200 , linewidths = 0. ) else : ax . scatter ( x + j * dc , L [ : : - 1 ] , c = x [ : : - 1 ] , cmap = cmap , s = 200 , linewidths = 0. ) locs . append ( x [ - 1 ] + j * dc ) # store locations for colormap labels # Set up labels for colormaps ax . xaxis . set_ticks_position ( 'top' ) ticker = mpl . ticker . FixedLocator ( locs ) ax . xaxis . set_major_locator ( ticker ) formatter = mpl . ticker . FixedFormatter ( [ cmapname for cmapname in cm . cmapnames ] ) ax . xaxis . set_major_formatter ( formatter ) labels = ax . get_xticklabels ( ) for label in labels : label . set_rotation ( 60 ) if saveplot : fig . savefig ( 'figures/lightness.png' , bbox_inches = 'tight' ) fig . savefig ( 'figures/lightness.pdf' , bbox_inches = 'tight' ) plt . show ( ) | Plot lightness of colormaps together . | 574 | 9 |
247,022 | def plot_gallery ( saveplot = False ) : from colorspacious import cspace_converter gradient = np . linspace ( 0 , 1 , 256 ) gradient = np . vstack ( ( gradient , gradient ) ) x = np . linspace ( 0.0 , 1.0 , 256 ) fig , axes = plt . subplots ( nrows = int ( len ( cm . cmap_d ) / 2 ) , ncols = 1 , figsize = ( 6 , 12 ) ) fig . subplots_adjust ( top = 0.99 , bottom = 0.01 , left = 0.2 , right = 0.99 , wspace = 0.05 ) for ax , cmapname in zip ( axes , cm . cmapnames ) : if '_r' in cmapname : # skip reversed versions for plot continue cmap = cm . cmap_d [ cmapname ] # get the colormap instance rgb = cmap ( x ) [ np . newaxis , : , : 3 ] # Find a good conversion to grayscale jch = cspace_converter ( "sRGB1" , "CAM02-UCS" ) ( rgb ) # Not sure why to use JCh instead so using this. L = jch [ 0 , : , 0 ] L = np . float32 ( np . vstack ( ( L , L , L ) ) ) ax . imshow ( gradient , aspect = 'auto' , cmap = cmap ) pos1 = ax . get_position ( ) # get the original position pos2 = [ pos1 . x0 , pos1 . y0 , pos1 . width , pos1 . height / 3.0 ] axbw = fig . add_axes ( pos2 ) # colorbar axes axbw . set_axis_off ( ) axbw . imshow ( L , aspect = 'auto' , cmap = cm . gray , vmin = 0 , vmax = 100. ) pos = list ( ax . get_position ( ) . bounds ) x_text = pos [ 0 ] - 0.01 y_text = pos [ 1 ] + pos [ 3 ] / 2. fig . text ( x_text , y_text , cmap . name , va = 'center' , ha = 'right' ) # Turn off *all* ticks & spines, not just the ones with colormaps. for ax in axes : ax . set_axis_off ( ) if saveplot : fig . savefig ( 'figures/gallery.pdf' , bbox_inches = 'tight' ) fig . savefig ( 'figures/gallery.png' , bbox_inches = 'tight' ) plt . show ( ) | Make plot of colormaps and labels like in the matplotlib gallery . | 593 | 16 |
247,023 | def wrap_viscm ( cmap , dpi = 100 , saveplot = False ) : from viscm import viscm viscm ( cmap ) fig = plt . gcf ( ) fig . set_size_inches ( 22 , 10 ) plt . show ( ) if saveplot : fig . savefig ( 'figures/eval_' + cmap . name + '.png' , bbox_inches = 'tight' , dpi = dpi ) fig . savefig ( 'figures/eval_' + cmap . name + '.pdf' , bbox_inches = 'tight' , dpi = dpi ) | Evaluate goodness of colormap using perceptual deltas . | 138 | 14 |
247,024 | def quick_plot ( cmap , fname = None , fig = None , ax = None , N = 10 ) : x = np . linspace ( 0 , 10 , N ) X , _ = np . meshgrid ( x , x ) if ax is None : fig = plt . figure ( ) ax = fig . add_subplot ( 111 ) mappable = ax . pcolor ( X , cmap = cmap ) ax . set_title ( cmap . name , fontsize = 14 ) ax . set_xticks ( [ ] ) ax . set_yticks ( [ ] ) plt . colorbar ( mappable ) plt . show ( ) if fname is not None : plt . savefig ( fname + '.png' , bbox_inches = 'tight' ) | Show quick test of a colormap . | 175 | 9 |
247,025 | def print_colormaps ( cmaps , N = 256 , returnrgb = True , savefiles = False ) : rgb = [ ] for cmap in cmaps : rgbtemp = cmap ( np . linspace ( 0 , 1 , N ) ) [ np . newaxis , : , : 3 ] [ 0 ] if savefiles : np . savetxt ( cmap . name + '-rgb.txt' , rgbtemp ) rgb . append ( rgbtemp ) if returnrgb : return rgb | Print colormaps in 256 RGB colors to text files . | 109 | 12 |
247,026 | def cmap ( rgbin , N = 256 ) : # rgb inputs here if not isinstance ( rgbin [ 0 ] , _string_types ) : # normalize to be out of 1 if out of 256 instead if rgbin . max ( ) > 1 : rgbin = rgbin / 256. cmap = mpl . colors . LinearSegmentedColormap . from_list ( 'mycmap' , rgbin , N = N ) return cmap | Input an array of rgb values to generate a colormap . | 99 | 13 |
247,027 | def lighten ( cmapin , alpha ) : # set the alpha value while retaining the number of rows in original cmap return cmap ( cmapin ( np . linspace ( 0 , 1 , cmapin . N ) , alpha ) ) | Lighten a colormap by adding alpha < 1 . | 54 | 12 |
247,028 | def crop_by_percent ( cmap , per , which = 'both' , N = None ) : if which == 'both' : # take percent off both ends of cmap vmin = - 100 vmax = 100 pivot = 0 dmax = per elif which == 'min' : # take percent off bottom of cmap vmax = 10 pivot = 5 vmin = ( 0 + per / 100 ) * 2 * pivot dmax = None elif which == 'max' : # take percent off top of cmap vmin = 0 pivot = 5 vmax = ( 1 - per / 100 ) * 2 * pivot dmax = None newcmap = crop ( cmap , vmin , vmax , pivot , dmax = dmax , N = N ) return newcmap | Crop end or ends of a colormap by per percent . | 168 | 14 |
247,029 | def _premium ( fn ) : @ _functools . wraps ( fn ) def _fn ( self , * args , * * kwargs ) : if self . _lite : raise RuntimeError ( 'Premium API not available in lite access.' ) return fn ( self , * args , * * kwargs ) return _fn | Premium decorator for APIs that require premium access level . | 72 | 11 |
247,030 | def make_retrieveParameters ( offset = 1 , count = 100 , name = 'RS' , sort = 'D' ) : return _OrderedDict ( [ ( 'firstRecord' , offset ) , ( 'count' , count ) , ( 'sortField' , _OrderedDict ( [ ( 'name' , name ) , ( 'sort' , sort ) ] ) ) ] ) | Create retrieve parameters dictionary to be used with APIs . | 86 | 10 |
247,031 | def connect ( self ) : if not self . _SID : self . _SID = self . _auth . service . authenticate ( ) print ( 'Authenticated (SID: %s)' % self . _SID ) self . _search . set_options ( headers = { 'Cookie' : 'SID="%s"' % self . _SID } ) self . _auth . options . headers . update ( { 'Cookie' : 'SID="%s"' % self . _SID } ) return self . _SID | Authenticate to WOS and set the SID cookie . | 120 | 12 |
247,032 | def close ( self ) : if self . _SID : self . _auth . service . closeSession ( ) self . _SID = None | The close operation loads the session if it is valid and then closes it and releases the session seat . All the session data are deleted and become invalid after the request is processed . The session ID can no longer be used in subsequent requests . | 31 | 47 |
247,033 | def search ( self , query , count = 5 , offset = 1 , editions = None , symbolicTimeSpan = None , timeSpan = None , retrieveParameters = None ) : return self . _search . service . search ( queryParameters = _OrderedDict ( [ ( 'databaseId' , 'WOS' ) , ( 'userQuery' , query ) , ( 'editions' , editions ) , ( 'symbolicTimeSpan' , symbolicTimeSpan ) , ( 'timeSpan' , timeSpan ) , ( 'queryLanguage' , 'en' ) ] ) , retrieveParameters = ( retrieveParameters or self . make_retrieveParameters ( offset , count ) ) ) | The search operation submits a search query to the specified database edition and retrieves data . This operation returns a query ID that can be used in subsequent operations to retrieve more records . | 150 | 36 |
247,034 | def citedReferences ( self , uid , count = 100 , offset = 1 , retrieveParameters = None ) : return self . _search . service . citedReferences ( databaseId = 'WOS' , uid = uid , queryLanguage = 'en' , retrieveParameters = ( retrieveParameters or self . make_retrieveParameters ( offset , count ) ) ) | The citedReferences operation returns references cited by an article identified by a unique identifier . You may specify only one identifier per request . | 76 | 25 |
247,035 | def citedReferencesRetrieve ( self , queryId , count = 100 , offset = 1 , retrieveParameters = None ) : return self . _search . service . citedReferencesRetrieve ( queryId = queryId , retrieveParameters = ( retrieveParameters or self . make_retrieveParameters ( offset , count ) ) ) | The citedReferencesRetrieve operation submits a query returned by a previous citedReferences operation . | 65 | 18 |
247,036 | def single ( wosclient , wos_query , xml_query = None , count = 5 , offset = 1 ) : result = wosclient . search ( wos_query , count , offset ) xml = _re . sub ( ' xmlns="[^"]+"' , '' , result . records , count = 1 ) . encode ( 'utf-8' ) if xml_query : xml = _ET . fromstring ( xml ) return [ el . text for el in xml . findall ( xml_query ) ] else : return _minidom . parseString ( xml ) . toprettyxml ( ) | Perform a single Web of Science query and then XML query the results . | 132 | 15 |
247,037 | def query ( wosclient , wos_query , xml_query = None , count = 5 , offset = 1 , limit = 100 ) : results = [ single ( wosclient , wos_query , xml_query , min ( limit , count - x + 1 ) , x ) for x in range ( offset , count + 1 , limit ) ] if xml_query : return [ el for res in results for el in res ] else : pattern = _re . compile ( r'^<\?xml.*?\n<records>\n|\n</records>$.*' ) return ( '<?xml version="1.0" ?>\n<records>' + '\n' . join ( pattern . sub ( '' , res ) for res in results ) + '</records>' ) | Query Web of Science and XML query results with multiple requests . | 179 | 12 |
247,038 | def doi_to_wos ( wosclient , doi ) : results = query ( wosclient , 'DO="%s"' % doi , './REC/UID' , count = 1 ) return results [ 0 ] . lstrip ( 'WOS:' ) if results else None | Convert DOI to WOS identifier . | 61 | 8 |
247,039 | def sql_fingerprint ( query , hide_columns = True ) : parsed_query = parse ( query ) [ 0 ] sql_recursively_simplify ( parsed_query , hide_columns = hide_columns ) return str ( parsed_query ) | Simplify a query taking away exact values and fields selected . | 58 | 13 |
247,040 | def match_keyword ( token , keywords ) : if not token : return False if not token . is_keyword : return False return token . value . upper ( ) in keywords | Checks if the given token represents one of the given keywords | 38 | 12 |
247,041 | def _is_group ( token ) : is_group = token . is_group if isinstance ( is_group , bool ) : return is_group else : return is_group ( ) | sqlparse 0 . 2 . 2 changed it from a callable to a bool property | 41 | 17 |
247,042 | def sorted_names ( names ) : names = list ( names ) have_default = False if 'default' in names : names . remove ( 'default' ) have_default = True sorted_names = sorted ( names ) if have_default : sorted_names = [ 'default' ] + sorted_names return sorted_names | Sort a list of names but keep the word default first if it s there . | 69 | 16 |
247,043 | def record_diff ( old , new ) : return '\n' . join ( difflib . ndiff ( [ '%s: %s' % ( k , v ) for op in old for k , v in op . items ( ) ] , [ '%s: %s' % ( k , v ) for op in new for k , v in op . items ( ) ] , ) ) | Generate a human - readable diff of two performance records . | 86 | 12 |
247,044 | def dequeue ( self , block = True ) : return self . queue . get ( block , self . queue_get_timeout ) | Dequeue a record and return item . | 28 | 8 |
247,045 | def start ( self ) : self . _thread = t = threading . Thread ( target = self . _monitor ) t . setDaemon ( True ) t . start ( ) | Start the listener . | 38 | 4 |
247,046 | def handle ( self , record ) : record = self . prepare ( record ) for handler in self . handlers : handler ( record ) | Handle an item . | 27 | 4 |
247,047 | def _monitor ( self ) : err_msg = ( "invalid internal state:" " _stop_nowait can not be set if _stop is not set" ) assert self . _stop . isSet ( ) or not self . _stop_nowait . isSet ( ) , err_msg q = self . queue has_task_done = hasattr ( q , 'task_done' ) while not self . _stop . isSet ( ) : try : record = self . dequeue ( True ) if record is self . _sentinel_item : break self . handle ( record ) if has_task_done : q . task_done ( ) except queue . Empty : pass # There might still be records in the queue, # handle then unless _stop_nowait is set. while not self . _stop_nowait . isSet ( ) : try : record = self . dequeue ( False ) if record is self . _sentinel_item : break self . handle ( record ) if has_task_done : q . task_done ( ) except queue . Empty : break | Monitor the queue for items and ask the handler to deal with them . | 233 | 14 |
247,048 | def stop ( self , nowait = False ) : self . _stop . set ( ) if nowait : self . _stop_nowait . set ( ) self . queue . put_nowait ( self . _sentinel_item ) if ( self . _thread . isAlive ( ) and self . _thread is not threading . currentThread ( ) ) : self . _thread . join ( ) self . _thread = None | Stop the listener . | 93 | 4 |
247,049 | def terminate ( self , nowait = False ) : logger . debug ( "Acquiring lock for service termination" ) with self . lock : logger . debug ( "Terminating service" ) if not self . listener : logger . warning ( "Service already stopped." ) return self . listener . stop ( nowait ) try : if not nowait : self . _post_log_batch ( ) except Exception : if self . error_handler : self . error_handler ( sys . exc_info ( ) ) else : raise finally : self . queue = None self . listener = None | Finalize and stop service | 122 | 5 |
247,050 | def process_log ( self , * * log_item ) : logger . debug ( "Processing log item: %s" , log_item ) self . log_batch . append ( log_item ) if len ( self . log_batch ) >= self . log_batch_size : self . _post_log_batch ( ) | Special handler for log messages . | 72 | 6 |
247,051 | def process_item ( self , item ) : logger . debug ( "Processing item: %s (queue size: %s)" , item , self . queue . qsize ( ) ) method , kwargs = item if method not in self . supported_methods : raise Error ( "Not expected service method: {}" . format ( method ) ) try : if method == "log" : self . process_log ( * * kwargs ) else : self . _post_log_batch ( ) getattr ( self . rp_client , method ) ( * * kwargs ) except Exception : if self . error_handler : self . error_handler ( sys . exc_info ( ) ) else : self . terminate ( nowait = True ) raise | Main item handler . | 163 | 4 |
247,052 | def log ( self , time , message , level = None , attachment = None ) : logger . debug ( "log queued" ) args = { "time" : time , "message" : message , "level" : level , "attachment" : attachment , } self . queue . put_nowait ( ( "log" , args ) ) | Logs a message with attachment . | 74 | 7 |
247,053 | def log_batch ( self , log_data ) : url = uri_join ( self . base_url , "log" ) attachments = [ ] for log_item in log_data : log_item [ "item_id" ] = self . stack [ - 1 ] attachment = log_item . get ( "attachment" , None ) if "attachment" in log_item : del log_item [ "attachment" ] if attachment : if not isinstance ( attachment , collections . Mapping ) : attachment = { "data" : attachment } name = attachment . get ( "name" , str ( uuid . uuid4 ( ) ) ) log_item [ "file" ] = { "name" : name } attachments . append ( ( "file" , ( name , attachment [ "data" ] , attachment . get ( "mime" , "application/octet-stream" ) ) ) ) files = [ ( "json_request_part" , ( None , json . dumps ( log_data ) , "application/json" ) ) ] files . extend ( attachments ) from reportportal_client import POST_LOGBATCH_RETRY_COUNT for i in range ( POST_LOGBATCH_RETRY_COUNT ) : try : r = self . session . post ( url = url , files = files , verify = self . verify_ssl ) except KeyError : if i < POST_LOGBATCH_RETRY_COUNT - 1 : continue else : raise break logger . debug ( "log_batch - Stack: %s" , self . stack ) logger . debug ( "log_batch response: %s" , r . text ) return _get_data ( r ) | Logs batch of messages with attachment . | 369 | 8 |
247,054 | def git_versions_from_keywords ( keywords , tag_prefix , verbose ) : if not keywords : raise NotThisMethod ( "no keywords at all, weird" ) date = keywords . get ( "date" ) if date is not None : # git-2.2.0 added "%cI", which expands to an ISO-8601 -compliant # datestamp. However we prefer "%ci" (which expands to an "ISO-8601 # -like" string, which we must then edit to make compliant), because # it's been around since git-1.5.3, and it's too difficult to # discover which version we're using, or to work around using an # older one. date = date . strip ( ) . replace ( " " , "T" , 1 ) . replace ( " " , "" , 1 ) refnames = keywords [ "refnames" ] . strip ( ) if refnames . startswith ( "$Format" ) : if verbose : print ( "keywords are unexpanded, not using" ) raise NotThisMethod ( "unexpanded keywords, not a git-archive tarball" ) refs = set ( [ r . strip ( ) for r in refnames . strip ( "()" ) . split ( "," ) ] ) # starting in git-1.8.3, tags are listed as "tag: foo-1.0" instead of # just "foo-1.0". If we see a "tag: " prefix, prefer those. TAG = "tag: " tags = set ( [ r [ len ( TAG ) : ] for r in refs if r . startswith ( TAG ) ] ) if not tags : # Either we're using git < 1.8.3, or there really are no tags. We use # a heuristic: assume all version tags have a digit. The old git %d # expansion behaves like git log --decorate=short and strips out the # refs/heads/ and refs/tags/ prefixes that would let us distinguish # between branches and tags. By ignoring refnames without digits, we # filter out many common branch names like "release" and # "stabilization", as well as "HEAD" and "master". tags = set ( [ r for r in refs if re . search ( r'\d' , r ) ] ) if verbose : print ( "discarding '%s', no digits" % "," . join ( refs - tags ) ) if verbose : print ( "likely tags: %s" % "," . join ( sorted ( tags ) ) ) for ref in sorted ( tags ) : # sorting will prefer e.g. "2.0" over "2.0rc1" if ref . startswith ( tag_prefix ) : r = ref [ len ( tag_prefix ) : ] if verbose : print ( "picking %s" % r ) return { "version" : r , "full-revisionid" : keywords [ "full" ] . strip ( ) , "dirty" : False , "error" : None , "date" : date , "branch" : None } # no suitable tags, so version is "0+unknown", but full hex is still there if verbose : print ( "no suitable tags, using unknown + full revision id" ) return { "version" : "0+unknown" , "full-revisionid" : keywords [ "full" ] . strip ( ) , "dirty" : False , "error" : "no suitable tags" , "date" : None , "branch" : None } | Get version information from git keywords . | 776 | 7 |
247,055 | def render_pep440_branch_based ( pieces ) : replacements = ( [ ' ' , '.' ] , [ '(' , '' ] , [ ')' , '' ] , [ '\\' , '.' ] , [ '/' , '.' ] ) branch_name = pieces . get ( 'branch' ) or '' if branch_name : for old , new in replacements : branch_name = branch_name . replace ( old , new ) else : branch_name = 'unknown_branch' if pieces [ "closest-tag" ] : rendered = pieces [ "closest-tag" ] if pieces [ "distance" ] or pieces [ "dirty" ] : rendered += '.dev0' + plus_or_dot ( pieces ) rendered += "%d.%s.g%s" % ( pieces [ "distance" ] , branch_name , pieces [ 'short' ] ) if pieces [ "dirty" ] : rendered += ".dirty" else : # exception #1 rendered = "0+untagged.%d.%s.g%s" % ( pieces [ "distance" ] , branch_name , pieces [ 'short' ] ) if pieces [ "dirty" ] : rendered += ".dirty" return rendered | Build up version string with post - release local version identifier . | 268 | 12 |
247,056 | def render ( pieces , style ) : if pieces [ "error" ] : return { "version" : "unknown" , "full-revisionid" : pieces . get ( "long" ) , "dirty" : None , "error" : pieces [ "error" ] , "date" : None } if not style or style == "default" : style = "pep440" # the default if style == "pep440" : rendered = render_pep440 ( pieces ) elif style == "pep440-pre" : rendered = render_pep440_pre ( pieces ) elif style == "pep440-post" : rendered = render_pep440_post ( pieces ) elif style == "pep440-old" : rendered = render_pep440_old ( pieces ) elif style == "pep440-branch-based" : rendered = render_pep440_branch_based ( pieces ) elif style == "git-describe" : rendered = render_git_describe ( pieces ) elif style == "git-describe-long" : rendered = render_git_describe_long ( pieces ) else : raise ValueError ( "unknown style '%s'" % style ) return { "version" : rendered , "full-revisionid" : pieces [ "long" ] , "dirty" : pieces [ "dirty" ] , "error" : None , "date" : pieces . get ( "date" ) } | Render the given version pieces into the requested style . | 324 | 10 |
247,057 | def do_setup ( ) : root = get_root ( ) try : cfg = get_config_from_root ( root ) except ( EnvironmentError , configparser . NoSectionError , configparser . NoOptionError ) as e : if isinstance ( e , ( EnvironmentError , configparser . NoSectionError ) ) : print ( "Adding sample versioneer config to setup.cfg" , file = sys . stderr ) with open ( os . path . join ( root , "setup.cfg" ) , "a" ) as f : f . write ( SAMPLE_CONFIG ) print ( CONFIG_ERROR , file = sys . stderr ) return 1 print ( " creating %s" % cfg . versionfile_source ) with open ( cfg . versionfile_source , "w" ) as f : LONG = LONG_VERSION_PY [ cfg . VCS ] f . write ( LONG % { "DOLLAR" : "$" , "STYLE" : cfg . style , "TAG_PREFIX" : cfg . tag_prefix , "PARENTDIR_PREFIX" : cfg . parentdir_prefix , "VERSIONFILE_SOURCE" : cfg . versionfile_source , } ) ipy = os . path . join ( os . path . dirname ( cfg . versionfile_source ) , "__init__.py" ) if os . path . exists ( ipy ) : try : with open ( ipy , "r" ) as f : old = f . read ( ) except EnvironmentError : old = "" if INIT_PY_SNIPPET_RE . search ( old ) is None : print ( " appending to %s" % ipy ) with open ( ipy , "a" ) as f : f . write ( INIT_PY_SNIPPET ) else : print ( " %s unmodified" % ipy ) else : print ( " %s doesn't exist, ok" % ipy ) ipy = None # Make sure both the top-level "versioneer.py" and versionfile_source # (PKG/_version.py, used by runtime code) are in MANIFEST.in, so # they'll be copied into source distributions. Pip won't be able to # install the package without this. manifest_in = os . path . join ( root , "MANIFEST.in" ) simple_includes = set ( ) try : with open ( manifest_in , "r" ) as f : for line in f : if line . startswith ( "include " ) : for include in line . split ( ) [ 1 : ] : simple_includes . add ( include ) except EnvironmentError : pass # That doesn't cover everything MANIFEST.in can do # (http://docs.python.org/2/distutils/sourcedist.html#commands), so # it might give some false negatives. Appending redundant 'include' # lines is safe, though. if "versioneer.py" not in simple_includes : print ( " appending 'versioneer.py' to MANIFEST.in" ) with open ( manifest_in , "a" ) as f : f . write ( "include versioneer.py\n" ) else : print ( " 'versioneer.py' already in MANIFEST.in" ) if cfg . versionfile_source not in simple_includes : print ( " appending versionfile_source ('%s') to MANIFEST.in" % cfg . versionfile_source ) with open ( manifest_in , "a" ) as f : f . write ( "include %s\n" % cfg . versionfile_source ) else : print ( " versionfile_source already in MANIFEST.in" ) # Make VCS-specific changes. For git, this means creating/changing # .gitattributes to mark _version.py for export-subst keyword # substitution. do_vcs_install ( manifest_in , cfg . versionfile_source , ipy ) return 0 | Do main VCS - independent setup function for installing Versioneer . | 885 | 13 |
247,058 | def scan_setup_py ( ) : found = set ( ) setters = False errors = 0 with open ( "setup.py" , "r" ) as f : for line in f . readlines ( ) : if "import versioneer" in line : found . add ( "import" ) if "versioneer.get_cmdclass(" in line : found . add ( "cmdclass" ) if "versioneer.get_version()" in line : found . add ( "get_version" ) if "versioneer.VCS" in line : setters = True if "versioneer.versionfile_source" in line : setters = True if len ( found ) != 3 : print ( "" ) print ( "Your setup.py appears to be missing some important items" ) print ( "(but I might be wrong). Please make sure it has something" ) print ( "roughly like the following:" ) print ( "" ) print ( " import versioneer" ) print ( " setup( version=versioneer.get_version()," ) print ( " cmdclass=versioneer.get_cmdclass(), ...)" ) print ( "" ) errors += 1 if setters : print ( "You should remove lines like 'versioneer.VCS = ' and" ) print ( "'versioneer.versionfile_source = ' . This configuration" ) print ( "now lives in setup.cfg, and should be removed from setup.py" ) print ( "" ) errors += 1 return errors | Validate the contents of setup . py against Versioneer s expectations . | 325 | 14 |
247,059 | def read ( fname ) : file_path = os . path . join ( SETUP_DIRNAME , fname ) with codecs . open ( file_path , encoding = 'utf-8' ) as rfh : return rfh . read ( ) | Read a file from the directory where setup . py resides | 57 | 11 |
247,060 | def func ( self , w , * args ) : x0 = args [ 0 ] x1 = args [ 1 ] n0 = x0 . shape [ 0 ] n1 = x1 . shape [ 0 ] # n -- number of pairs to evaluate n = max ( n0 , n1 ) * 10 idx0 = np . random . choice ( range ( n0 ) , size = n ) idx1 = np . random . choice ( range ( n1 ) , size = n ) # b -- bias for the input and h layers b0 = np . ones ( ( n0 , 1 ) ) b1 = np . ones ( ( n1 , 1 ) ) i1 = self . i + 1 h = self . h h1 = h + 1 # Predict for features -- cannot use predict_raw() because here # different weights can be used. if sparse . issparse ( x0 ) : p0 = np . hstack ( ( sigm ( sparse . hstack ( ( x0 , b0 ) ) . dot ( w [ : - h1 ] . reshape ( i1 , h ) ) ) , b0 ) ) . dot ( w [ - h1 : ] . reshape ( h1 , 1 ) ) p1 = np . hstack ( ( sigm ( sparse . hstack ( ( x1 , b1 ) ) . dot ( w [ : - h1 ] . reshape ( i1 , h ) ) ) , b1 ) ) . dot ( w [ - h1 : ] . reshape ( h1 , 1 ) ) else : p0 = np . hstack ( ( sigm ( np . hstack ( ( x0 , b0 ) ) . dot ( w [ : - h1 ] . reshape ( i1 , h ) ) ) , b0 ) ) . dot ( w [ - h1 : ] . reshape ( h1 , 1 ) ) p1 = np . hstack ( ( sigm ( np . hstack ( ( x1 , b1 ) ) . dot ( w [ : - h1 ] . reshape ( i1 , h ) ) ) , b1 ) ) . dot ( w [ - h1 : ] . reshape ( h1 , 1 ) ) p0 = p0 [ idx0 ] p1 = p1 [ idx1 ] # Return the cost that consists of the sum of squared error + # L2-regularization for weights between the input and h layers + # L2-regularization for weights between the h and output layers. #return .5 * (sum((1 - sigm(p1 - p0)) ** 2) + self.l1 * sum(w[:-h1] ** 2) + return .5 * ( sum ( ( 1 - p1 + p0 ) ** 2 ) / n + self . l1 * sum ( w [ : - h1 ] ** 2 ) / ( i1 * h ) + self . l2 * sum ( w [ - h1 : ] ** 2 ) / h1 ) | Return the costs of the neural network for predictions . | 649 | 10 |
247,061 | def fprime ( self , w , * args ) : x0 = args [ 0 ] x1 = args [ 1 ] n0 = x0 . shape [ 0 ] n1 = x1 . shape [ 0 ] # n -- number of pairs to evaluate n = max ( n0 , n1 ) * 10 idx0 = np . random . choice ( range ( n0 ) , size = n ) idx1 = np . random . choice ( range ( n1 ) , size = n ) # b -- bias for the input and h layers b = np . ones ( ( n , 1 ) ) i1 = self . i + 1 h = self . h h1 = h + 1 w2 = w [ - h1 : ] . reshape ( h1 , 1 ) w1 = w [ : - h1 ] . reshape ( i1 , h ) if sparse . issparse ( x0 ) : x0 = x0 . tocsr ( ) [ idx0 ] x1 = x1 . tocsr ( ) [ idx1 ] xb0 = sparse . hstack ( ( x0 , b ) ) xb1 = sparse . hstack ( ( x1 , b ) ) else : x0 = x0 [ idx0 ] x1 = x1 [ idx1 ] xb0 = np . hstack ( ( x0 , b ) ) xb1 = np . hstack ( ( x1 , b ) ) z0 = np . hstack ( ( sigm ( xb0 . dot ( w1 ) ) , b ) ) z1 = np . hstack ( ( sigm ( xb1 . dot ( w1 ) ) , b ) ) y0 = z0 . dot ( w2 ) y1 = z1 . dot ( w2 ) #e = 1 - sigm(y1 - y0) #dy = e * dsigm(y1 - y0) e = 1 - ( y1 - y0 ) dy = e / n # Calculate the derivative of the cost function w.r.t. F and w2 where: # F -- weights between the input and h layers # w2 -- weights between the h and output layers dw1 = - ( xb1 . T . dot ( dy . dot ( w2 [ : - 1 ] . reshape ( 1 , h ) ) * dsigm ( xb1 . dot ( w1 ) ) ) - xb0 . T . dot ( dy . dot ( w2 [ : - 1 ] . reshape ( 1 , h ) ) * dsigm ( xb0 . dot ( w1 ) ) ) ) . reshape ( i1 * h ) + self . l1 * w [ : - h1 ] / ( i1 * h ) dw2 = - ( z1 - z0 ) . T . dot ( dy ) . reshape ( h1 ) + self . l2 * w [ - h1 : ] / h1 return np . append ( dw1 , dw2 ) | Return the derivatives of the cost function for predictions . | 650 | 10 |
247,062 | def _transform_col ( self , x , col ) : return norm . ppf ( self . ecdfs [ col ] ( x ) * .998 + .001 ) | Normalize one numerical column . | 37 | 6 |
247,063 | def _get_label_encoder_and_max ( self , x ) : # NaN cannot be used as a key for dict. So replace it with a random integer. label_count = x . fillna ( NAN_INT ) . value_counts ( ) n_uniq = label_count . shape [ 0 ] label_count = label_count [ label_count >= self . min_obs ] n_uniq_new = label_count . shape [ 0 ] # If every label appears more than min_obs, new label starts from 0. # Otherwise, new label starts from 1 and 0 is used for all old labels # that appear less than min_obs. offset = 0 if n_uniq == n_uniq_new else 1 label_encoder = pd . Series ( np . arange ( n_uniq_new ) + offset , index = label_count . index ) max_label = label_encoder . max ( ) label_encoder = label_encoder . to_dict ( ) return label_encoder , max_label | Return a mapping from values and its maximum of a column to integer labels . | 233 | 15 |
247,064 | def _transform_col ( self , x , i ) : return x . fillna ( NAN_INT ) . map ( self . label_encoders [ i ] ) . fillna ( 0 ) | Encode one categorical column into labels . | 44 | 9 |
247,065 | def _transform_col ( self , x , i ) : labels = self . label_encoder . _transform_col ( x , i ) label_max = self . label_encoder . label_maxes [ i ] # build row and column index for non-zero values of a sparse matrix index = np . array ( range ( len ( labels ) ) ) i = index [ labels > 0 ] j = labels [ labels > 0 ] - 1 # column index starts from 0 if len ( i ) > 0 : return sparse . coo_matrix ( ( np . ones_like ( i ) , ( i , j ) ) , shape = ( x . shape [ 0 ] , label_max ) ) else : # if there is no non-zero value, return no matrix return None | Encode one categorical column into sparse matrix with one - hot - encoding . | 168 | 16 |
247,066 | def transform ( self , X ) : for i , col in enumerate ( X . columns ) : X_col = self . _transform_col ( X [ col ] , i ) if X_col is not None : if i == 0 : X_new = X_col else : X_new = sparse . hstack ( ( X_new , X_col ) ) logger . debug ( '{} --> {} features' . format ( col , self . label_encoder . label_maxes [ i ] ) ) return X_new | Encode categorical columns into sparse matrix with one - hot - encoding . | 116 | 15 |
247,067 | def predict ( self , x ) : if self . _is_leaf ( ) : d1 = self . predict_initialize [ 'count_dict' ] d2 = count_dict ( self . Y ) for key , value in d1 . iteritems ( ) : if key in d2 : d2 [ key ] += value else : d2 [ key ] = value return argmax ( d2 ) else : if self . criterion ( x ) : return self . right . predict ( x ) else : return self . left . predict ( x ) | Make prediction recursively . Use both the samples inside the current node and the statistics inherited from parent . | 117 | 21 |
247,068 | def netflix ( es , ps , e0 , l = .0001 ) : m = len ( es ) n = len ( ps [ 0 ] ) X = np . stack ( ps ) . T pTy = .5 * ( n * e0 ** 2 + ( X ** 2 ) . sum ( axis = 0 ) - n * np . array ( es ) ** 2 ) w = np . linalg . pinv ( X . T . dot ( X ) + l * n * np . eye ( m ) ) . dot ( pTy ) return X . dot ( w ) , w | Combine predictions with the optimal weights to minimize RMSE . | 126 | 12 |
247,069 | def save_data ( X , y , path ) : catalog = { '.csv' : save_csv , '.sps' : save_libsvm , '.h5' : save_hdf5 } ext = os . path . splitext ( path ) [ 1 ] func = catalog [ ext ] if y is None : y = np . zeros ( ( X . shape [ 0 ] , ) ) func ( X , y , path ) | Save data as a CSV LibSVM or HDF5 file based on the file extension . | 96 | 19 |
247,070 | def save_csv ( X , y , path ) : if sparse . issparse ( X ) : X = X . todense ( ) np . savetxt ( path , np . hstack ( ( y . reshape ( ( - 1 , 1 ) ) , X ) ) , delimiter = ',' ) | Save data as a CSV file . | 66 | 7 |
247,071 | def save_libsvm ( X , y , path ) : dump_svmlight_file ( X , y , path , zero_based = False ) | Save data as a LibSVM file . | 34 | 9 |
247,072 | def save_hdf5 ( X , y , path ) : with h5py . File ( path , 'w' ) as f : is_sparse = 1 if sparse . issparse ( X ) else 0 f [ 'issparse' ] = is_sparse f [ 'target' ] = y if is_sparse : if not sparse . isspmatrix_csr ( X ) : X = X . tocsr ( ) f [ 'shape' ] = np . array ( X . shape ) f [ 'data' ] = X . data f [ 'indices' ] = X . indices f [ 'indptr' ] = X . indptr else : f [ 'data' ] = X | Save data as a HDF5 file . | 154 | 9 |
247,073 | def load_data ( path , dense = False ) : catalog = { '.csv' : load_csv , '.sps' : load_svmlight_file , '.h5' : load_hdf5 } ext = os . path . splitext ( path ) [ 1 ] func = catalog [ ext ] X , y = func ( path ) if dense and sparse . issparse ( X ) : X = X . todense ( ) return X , y | Load data from a CSV LibSVM or HDF5 file based on the file extension . | 100 | 19 |
247,074 | def load_csv ( path ) : with open ( path ) as f : line = f . readline ( ) . strip ( ) X = np . loadtxt ( path , delimiter = ',' , skiprows = 0 if is_number ( line . split ( ',' ) [ 0 ] ) else 1 ) y = np . array ( X [ : , 0 ] ) . flatten ( ) X = X [ : , 1 : ] return X , y | Load data from a CSV file . | 97 | 7 |
247,075 | def load_hdf5 ( path ) : with h5py . File ( path , 'r' ) as f : is_sparse = f [ 'issparse' ] [ ... ] if is_sparse : shape = tuple ( f [ 'shape' ] [ ... ] ) data = f [ 'data' ] [ ... ] indices = f [ 'indices' ] [ ... ] indptr = f [ 'indptr' ] [ ... ] X = sparse . csr_matrix ( ( data , indices , indptr ) , shape = shape ) else : X = f [ 'data' ] [ ... ] y = f [ 'target' ] [ ... ] return X , y | Load data from a HDF5 file . | 148 | 9 |
247,076 | def read_sps ( path ) : for line in open ( path ) : # parse x xs = line . rstrip ( ) . split ( ' ' ) yield xs [ 1 : ] , int ( xs [ 0 ] ) | Read a LibSVM file line - by - line . | 51 | 12 |
247,077 | def gini ( y , p ) : # check and get number of samples assert y . shape == p . shape n_samples = y . shape [ 0 ] # sort rows on prediction column # (from largest to smallest) arr = np . array ( [ y , p ] ) . transpose ( ) true_order = arr [ arr [ : , 0 ] . argsort ( ) ] [ : : - 1 , 0 ] pred_order = arr [ arr [ : , 1 ] . argsort ( ) ] [ : : - 1 , 0 ] # get Lorenz curves l_true = np . cumsum ( true_order ) / np . sum ( true_order ) l_pred = np . cumsum ( pred_order ) / np . sum ( pred_order ) l_ones = np . linspace ( 1 / n_samples , 1 , n_samples ) # get Gini coefficients (area between curves) g_true = np . sum ( l_ones - l_true ) g_pred = np . sum ( l_ones - l_pred ) # normalize to true Gini coefficient return g_pred / g_true | Normalized Gini Coefficient . | 250 | 7 |
247,078 | def logloss ( y , p ) : p [ p < EPS ] = EPS p [ p > 1 - EPS ] = 1 - EPS return log_loss ( y , p ) | Bounded log loss error . | 38 | 6 |
247,079 | def convert ( input_file_name , * * kwargs ) : delimiter = kwargs [ "delimiter" ] or "," quotechar = kwargs [ "quotechar" ] or "|" if six . PY2 : delimiter = delimiter . encode ( "utf-8" ) quotechar = quotechar . encode ( "utf-8" ) # Read CSV and form a header and rows list with open ( input_file_name , "rb" ) as input_file : reader = csv . reader ( input_file , encoding = "utf-8" , delimiter = delimiter , quotechar = quotechar ) csv_headers = [ ] if not kwargs . get ( "no_header" ) : # Read header from first line csv_headers = next ( reader ) csv_rows = [ row for row in reader if row ] # Set default column name if header is not present if not csv_headers and len ( csv_rows ) > 0 : end = len ( csv_rows [ 0 ] ) + 1 csv_headers = [ "Column {}" . format ( n ) for n in range ( 1 , end ) ] # Render csv to HTML html = render_template ( csv_headers , csv_rows , * * kwargs ) # Freeze all JS files in template return freeze_js ( html ) | Convert CSV file to HTML table | 301 | 7 |
247,080 | def save ( file_name , content ) : with open ( file_name , "w" , encoding = "utf-8" ) as output_file : output_file . write ( content ) return output_file . name | Save content to a file | 48 | 5 |
247,081 | def serve ( content ) : temp_folder = tempfile . gettempdir ( ) temp_file_name = tempfile . gettempprefix ( ) + str ( uuid . uuid4 ( ) ) + ".html" # Generate a file path with a random name in temporary dir temp_file_path = os . path . join ( temp_folder , temp_file_name ) # save content to temp file save ( temp_file_path , content ) # Open templfile in a browser webbrowser . open ( "file://{}" . format ( temp_file_path ) ) # Block the thread while content is served try : while True : time . sleep ( 1 ) except KeyboardInterrupt : # cleanup the temp file os . remove ( temp_file_path ) | Write content to a temp file and serve it in browser | 168 | 11 |
247,082 | def render_template ( table_headers , table_items , * * options ) : caption = options . get ( "caption" ) or "Table" display_length = options . get ( "display_length" ) or - 1 height = options . get ( "height" ) or "70vh" default_length_menu = [ - 1 , 10 , 25 , 50 ] pagination = options . get ( "pagination" ) virtual_scroll_limit = options . get ( "virtual_scroll" ) # Change % to vh height = height . replace ( "%" , "vh" ) # Header columns columns = [ ] for header in table_headers : columns . append ( { "title" : header } ) # Data table options datatable_options = { "columns" : columns , "data" : table_items , "iDisplayLength" : display_length , "sScrollX" : "100%" , "sScrollXInner" : "100%" } # Enable virtual scroll for rows bigger than 1000 rows is_paging = pagination virtual_scroll = False scroll_y = height if virtual_scroll_limit : if virtual_scroll_limit != - 1 and len ( table_items ) > virtual_scroll_limit : virtual_scroll = True display_length = - 1 fmt = ( "\nVirtual scroll is enabled since number of rows exceeds {limit}." " You can set custom row limit by setting flag -vs, --virtual-scroll." " Virtual scroll can be disabled by setting the value to -1 and set it to 0 to always enable." ) logger . warn ( fmt . format ( limit = virtual_scroll_limit ) ) if not is_paging : fmt = "\nPagination can not be disabled in virtual scroll mode." logger . warn ( fmt ) is_paging = True if is_paging and not virtual_scroll : # Add display length to the default display length menu length_menu = [ ] if display_length != - 1 : length_menu = sorted ( default_length_menu + [ display_length ] ) else : length_menu = default_length_menu # Set label as "All" it display length is -1 length_menu_label = [ str ( "All" ) if i == - 1 else i for i in length_menu ] datatable_options [ "lengthMenu" ] = [ length_menu , length_menu_label ] datatable_options [ "iDisplayLength" ] = display_length if is_paging : datatable_options [ "paging" ] = True else : datatable_options [ "paging" ] = False if scroll_y : datatable_options [ "scrollY" ] = scroll_y if virtual_scroll : datatable_options [ "scroller" ] = True datatable_options [ "bPaginate" ] = False datatable_options [ "deferRender" ] = True datatable_options [ "bLengthChange" ] = False enable_export = options . get ( "export" ) if enable_export : if options [ "export_options" ] : allowed = list ( options [ "export_options" ] ) else : allowed = [ "copy" , "csv" , "json" , "print" ] datatable_options [ "dom" ] = "Bfrtip" datatable_options [ "buttons" ] = allowed datatable_options_json = json . dumps ( datatable_options , separators = ( "," , ":" ) ) return template . render ( title = caption or "Table" , caption = caption , datatable_options = datatable_options_json , virtual_scroll = virtual_scroll , enable_export = enable_export ) | Render Jinja2 template | 802 | 5 |
247,083 | def freeze_js ( html ) : matches = js_src_pattern . finditer ( html ) if not matches : return html # Reverse regex matches to replace match string with respective JS content for match in reversed ( tuple ( matches ) ) : # JS file name file_name = match . group ( 1 ) file_path = os . path . join ( js_files_path , file_name ) with open ( file_path , "r" , encoding = "utf-8" ) as f : file_content = f . read ( ) # Replace matched string with inline JS fmt = '<script type="text/javascript">{}</script>' js_content = fmt . format ( file_content ) html = html [ : match . start ( ) ] + js_content + html [ match . end ( ) : ] return html | Freeze all JS assets to the rendered html itself . | 178 | 11 |
247,084 | def cli ( * args , * * kwargs ) : # Convert CSV file content = convert . convert ( kwargs [ "input_file" ] , * * kwargs ) # Serve the temporary file in browser. if kwargs [ "serve" ] : convert . serve ( content ) # Write to output file elif kwargs [ "output_file" ] : # Check if file can be overwrite if ( not kwargs [ "overwrite" ] and not prompt_overwrite ( kwargs [ "output_file" ] ) ) : raise click . Abort ( ) convert . save ( kwargs [ "output_file" ] , content ) click . secho ( "File converted successfully: {}" . format ( kwargs [ "output_file" ] ) , fg = "green" ) else : # If its not server and output file is missing then raise error raise click . BadOptionUsage ( "Missing argument \"output_file\"." ) | CSVtoTable commandline utility . | 213 | 8 |
247,085 | def activate_retry ( request , activation_key , template_name = 'userena/activate_retry_success.html' , extra_context = None ) : if not userena_settings . USERENA_ACTIVATION_RETRY : return redirect ( reverse ( 'userena_activate' , args = ( activation_key , ) ) ) try : if UserenaSignup . objects . check_expired_activation ( activation_key ) : new_key = UserenaSignup . objects . reissue_activation ( activation_key ) if new_key : if not extra_context : extra_context = dict ( ) return ExtraContextTemplateView . as_view ( template_name = template_name , extra_context = extra_context ) ( request ) else : return redirect ( reverse ( 'userena_activate' , args = ( activation_key , ) ) ) else : return redirect ( reverse ( 'userena_activate' , args = ( activation_key , ) ) ) except UserenaSignup . DoesNotExist : return redirect ( reverse ( 'userena_activate' , args = ( activation_key , ) ) ) | Reissue a new activation_key for the user with the expired activation_key . | 256 | 17 |
247,086 | def disabled_account ( request , username , template_name , extra_context = None ) : user = get_object_or_404 ( get_user_model ( ) , username__iexact = username ) if user . is_active : raise Http404 if not extra_context : extra_context = dict ( ) extra_context [ 'viewed_user' ] = user extra_context [ 'profile' ] = get_user_profile ( user = user ) return ExtraContextTemplateView . as_view ( template_name = template_name , extra_context = extra_context ) ( request ) | Checks if the account is disabled if so returns the disabled account template . | 131 | 15 |
247,087 | def profile_list ( request , page = 1 , template_name = 'userena/profile_list.html' , paginate_by = 50 , extra_context = None , * * kwargs ) : # pragma: no cover warnings . warn ( "views.profile_list is deprecated. Use ProfileListView instead" , DeprecationWarning , stacklevel = 2 ) try : page = int ( request . GET . get ( 'page' , None ) ) except ( TypeError , ValueError ) : page = page if userena_settings . USERENA_DISABLE_PROFILE_LIST and not request . user . is_staff : raise Http404 profile_model = get_profile_model ( ) queryset = profile_model . objects . get_visible_profiles ( request . user ) if not extra_context : extra_context = dict ( ) return ProfileListView . as_view ( queryset = queryset , paginate_by = paginate_by , page = page , template_name = template_name , extra_context = extra_context , * * kwargs ) ( request ) | Returns a list of all profiles that are public . | 247 | 10 |
247,088 | def get_or_create ( self , um_from_user , um_to_user , message ) : created = False try : contact = self . get ( Q ( um_from_user = um_from_user , um_to_user = um_to_user ) | Q ( um_from_user = um_to_user , um_to_user = um_from_user ) ) except self . model . DoesNotExist : created = True contact = self . create ( um_from_user = um_from_user , um_to_user = um_to_user , latest_message = message ) return ( contact , created ) | Get or create a Contact | 144 | 5 |
247,089 | def update_contact ( self , um_from_user , um_to_user , message ) : contact , created = self . get_or_create ( um_from_user , um_to_user , message ) # If the contact already existed, update the message if not created : contact . latest_message = message contact . save ( ) return contact | Get or update a contacts information | 76 | 6 |
247,090 | def get_contacts_for ( self , user ) : contacts = self . filter ( Q ( um_from_user = user ) | Q ( um_to_user = user ) ) return contacts | Returns the contacts for this user . | 43 | 7 |
247,091 | def send_message ( self , sender , um_to_user_list , body ) : msg = self . model ( sender = sender , body = body ) msg . save ( ) # Save the recipients msg . save_recipients ( um_to_user_list ) msg . update_contacts ( um_to_user_list ) signals . email_sent . send ( sender = None , msg = msg ) return msg | Send a message from a user to a user . | 92 | 10 |
247,092 | def get_conversation_between ( self , um_from_user , um_to_user ) : messages = self . filter ( Q ( sender = um_from_user , recipients = um_to_user , sender_deleted_at__isnull = True ) | Q ( sender = um_to_user , recipients = um_from_user , messagerecipient__deleted_at__isnull = True ) ) return messages | Returns a conversation between two users | 96 | 6 |
247,093 | def count_unread_messages_for ( self , user ) : unread_total = self . filter ( user = user , read_at__isnull = True , deleted_at__isnull = True ) . count ( ) return unread_total | Returns the amount of unread messages for this user | 56 | 10 |
247,094 | def count_unread_messages_between ( self , um_to_user , um_from_user ) : unread_total = self . filter ( message__sender = um_from_user , user = um_to_user , read_at__isnull = True , deleted_at__isnull = True ) . count ( ) return unread_total | Returns the amount of unread messages between two users | 81 | 10 |
247,095 | def reissue_activation ( self , activation_key ) : try : userena = self . get ( activation_key = activation_key ) except self . model . DoesNotExist : return False try : salt , new_activation_key = generate_sha1 ( userena . user . username ) userena . activation_key = new_activation_key userena . save ( using = self . _db ) userena . user . date_joined = get_datetime_now ( ) userena . user . save ( using = self . _db ) userena . send_activation_email ( ) return True except Exception : return False | Creates a new activation_key resetting activation timeframe when users let the previous key expire . | 143 | 19 |
247,096 | def check_expired_activation ( self , activation_key ) : if SHA1_RE . search ( activation_key ) : userena = self . get ( activation_key = activation_key ) return userena . activation_key_expired ( ) raise self . model . DoesNotExist | Check if activation_key is still valid . | 66 | 9 |
247,097 | def check_permissions ( self ) : # Variable to supply some feedback changed_permissions = [ ] changed_users = [ ] warnings = [ ] # Check that all the permissions are available. for model , perms in ASSIGNED_PERMISSIONS . items ( ) : if model == 'profile' : model_obj = get_profile_model ( ) else : model_obj = get_user_model ( ) model_content_type = ContentType . objects . get_for_model ( model_obj ) for perm in perms : try : Permission . objects . get ( codename = perm [ 0 ] , content_type = model_content_type ) except Permission . DoesNotExist : changed_permissions . append ( perm [ 1 ] ) Permission . objects . create ( name = perm [ 1 ] , codename = perm [ 0 ] , content_type = model_content_type ) # it is safe to rely on settings.ANONYMOUS_USER_NAME since it is a # requirement of django-guardian for user in get_user_model ( ) . objects . exclude ( username = settings . ANONYMOUS_USER_NAME ) : try : user_profile = get_user_profile ( user = user ) except ObjectDoesNotExist : warnings . append ( _ ( "No profile found for %(username)s" ) % { 'username' : user . username } ) else : all_permissions = get_perms ( user , user_profile ) + get_perms ( user , user ) for model , perms in ASSIGNED_PERMISSIONS . items ( ) : if model == 'profile' : perm_object = get_user_profile ( user = user ) else : perm_object = user for perm in perms : if perm [ 0 ] not in all_permissions : assign_perm ( perm [ 0 ] , user , perm_object ) changed_users . append ( user ) return ( changed_permissions , changed_users , warnings ) | Checks that all permissions are set correctly for the users . | 437 | 12 |
247,098 | def get_unread_message_count_for ( parser , token ) : try : tag_name , arg = token . contents . split ( None , 1 ) except ValueError : raise template . TemplateSyntaxError ( "%s tag requires arguments" % token . contents . split ( ) [ 0 ] ) m = re . search ( r'(.*?) as (\w+)' , arg ) if not m : raise template . TemplateSyntaxError ( "%s tag had invalid arguments" % tag_name ) user , var_name = m . groups ( ) return MessageCount ( user , var_name ) | Returns the unread message count for a user . | 130 | 10 |
247,099 | def get_unread_message_count_between ( parser , token ) : try : tag_name , arg = token . contents . split ( None , 1 ) except ValueError : raise template . TemplateSyntaxError ( "%s tag requires arguments" % token . contents . split ( ) [ 0 ] ) m = re . search ( r'(.*?) and (.*?) as (\w+)' , arg ) if not m : raise template . TemplateSyntaxError ( "%s tag had invalid arguments" % tag_name ) um_from_user , um_to_user , var_name = m . groups ( ) return MessageCount ( um_from_user , var_name , um_to_user ) | Returns the unread message count between two users . | 154 | 10 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.