idx
int64
0
252k
question
stringlengths
48
5.28k
target
stringlengths
5
1.23k
3,800
def make_readable_path ( path ) : home = os . path . expanduser ( "~" ) if path . startswith ( home ) : path = "~" + path [ len ( home ) : ] return path
Make a path more readable
3,801
def shortlex ( start , other , excludestart = False ) : if not excludestart : yield start queue = collections . deque ( [ ( start , other ) ] ) while queue : current , other = queue . popleft ( ) while other : first , other = other [ 0 ] , other [ 1 : ] result = current | first yield result if other : queue . append ( ...
Yield all unions of start with other in shortlex order .
3,802
def reverse_shortlex ( end , other , excludeend = False ) : if not excludeend : yield end queue = collections . deque ( [ ( end , other ) ] ) while queue : current , other = queue . popleft ( ) while other : first , other = other [ 0 ] , other [ 1 : ] result = current & first yield result if other : queue . append ( ( ...
Yield all intersections of end with other in reverse shortlex order .
3,803
def generate_filename ( self , instance , filename ) : from unidecode import unidecode return super ( ) . generate_filename ( instance , unidecode ( force_text ( filename ) ) )
removes UTF chars from filename
3,804
def next ( self ) : if self . blocking >= 0 : res = self . redis . blpop ( [ self . name ] , timeout = self . blocking ) if res : res = res [ 1 ] else : res = self . redis . lpop ( self . name ) value = self . deserialize ( res ) logger . debug ( 'Popped from "%s": %s' , self . name , repr ( value ) ) return value
Retrieve the next item in the queue .
3,805
def send ( self , * args ) : if None in args : raise TypeError ( 'None is not a valid queue item.' ) serialized_values = [ self . serialize ( value ) for value in args ] logger . debug ( 'Sending to "%s": %s' , self . name , serialized_values ) return self . redis . rpush ( self . name , * serialized_values )
Send a value to this LIFO Queue .
3,806
def clear ( self ) : logger . debug ( 'Clearing queue: "%s"' , self . name ) return self . redis . delete ( self . name )
Clear any existing values from this queue .
3,807
def u2i ( uint32 ) : mask = ( 2 ** 32 ) - 1 if uint32 & ( 1 << 31 ) : v = uint32 | ~ mask else : v = uint32 & mask return v
Converts a 32 bit unsigned number to signed .
3,808
def _u2i ( uint32 ) : v = u2i ( uint32 ) if v < 0 : if exceptions : raise ApigpioError ( error_text ( v ) ) return v
Converts a 32 bit unsigned number to signed . If the number is negative it indicates an error . On error a pigpio exception will be raised if exceptions is True .
3,809
def append ( self , cb ) : self . callbacks . append ( cb . callb ) self . monitor = self . monitor | cb . callb . bit yield from self . pi . _pigpio_aio_command ( _PI_CMD_NB , self . handle , self . monitor )
Adds a callback .
3,810
def remove ( self , cb ) : if cb in self . callbacks : self . callbacks . remove ( cb ) new_monitor = 0 for c in self . callbacks : new_monitor |= c . bit if new_monitor != self . monitor : self . monitor = new_monitor yield from self . pi . _pigpio_aio_command ( _PI_CMD_NB , self . handle , self . monitor )
Removes a callback .
3,811
def _pigpio_aio_command ( self , cmd , p1 , p2 , ) : with ( yield from self . _lock ) : data = struct . pack ( 'IIII' , cmd , p1 , p2 , 0 ) self . _loop . sock_sendall ( self . s , data ) response = yield from self . _loop . sock_recv ( self . s , 16 ) _ , res = struct . unpack ( '12sI' , response ) return res
Runs a pigpio socket command .
3,812
def _pigpio_aio_command_ext ( self , cmd , p1 , p2 , p3 , extents , rl = True ) : with ( yield from self . _lock ) : ext = bytearray ( struct . pack ( 'IIII' , cmd , p1 , p2 , p3 ) ) for x in extents : if isinstance ( x , str ) : ext . extend ( _b ( x ) ) else : ext . extend ( x ) self . _loop . sock_sendall ( self . s...
Runs an extended pigpio socket command .
3,813
def store_script ( self , script ) : if len ( script ) : res = yield from self . _pigpio_aio_command_ext ( _PI_CMD_PROC , 0 , 0 , len ( script ) , [ script ] ) return _u2i ( res ) else : return 0
Store a script for later execution .
3,814
def run_script ( self , script_id , params = None ) : if params is not None : ext = bytearray ( ) for p in params : ext . extend ( struct . pack ( "I" , p ) ) nump = len ( params ) extents = [ ext ] else : nump = 0 extents = [ ] res = yield from self . _pigpio_aio_command_ext ( _PI_CMD_PROCR , script_id , 0 , nump * 4 ...
Runs a stored script .
3,815
def script_status ( self , script_id ) : res = yield from self . _pigpio_aio_command ( _PI_CMD_PROCP , script_id , 0 ) bytes = u2i ( res ) if bytes > 0 : data = yield from self . _loop . sock_recv ( self . s , bytes ) while len ( data ) < bytes : b = yield from self . _loop . sock_recv ( self . s , bytes - len ( data )...
Returns the run status of a stored script as well as the current values of parameters 0 to 9 .
3,816
def stop_script ( self , script_id ) : res = yield from self . _pigpio_aio_command ( _PI_CMD_PROCS , script_id , 0 ) return _u2i ( res )
Stops a running script .
3,817
def delete_script ( self , script_id ) : res = yield from self . _pigpio_aio_command ( _PI_CMD_PROCD , script_id , 0 ) return _u2i ( res )
Deletes a stored script .
3,818
def clear_bank_1 ( self , bits ) : res = yield from self . _pigpio_aio_command ( _PI_CMD_BC1 , bits , 0 ) return _u2i ( res )
Clears gpios 0 - 31 if the corresponding bit in bits is set .
3,819
def set_bank_1 ( self , bits ) : res = yield from self . _pigpio_aio_command ( _PI_CMD_BS1 , bits , 0 ) return _u2i ( res )
Sets gpios 0 - 31 if the corresponding bit in bits is set .
3,820
def set_mode ( self , gpio , mode ) : res = yield from self . _pigpio_aio_command ( _PI_CMD_MODES , gpio , mode ) return _u2i ( res )
Sets the gpio mode .
3,821
def get_mode ( self , gpio ) : res = yield from self . _pigpio_aio_command ( _PI_CMD_MODEG , gpio , 0 ) return _u2i ( res )
Returns the gpio mode .
3,822
def write ( self , gpio , level ) : res = yield from self . _pigpio_aio_command ( _PI_CMD_WRITE , gpio , level ) return _u2i ( res )
Sets the gpio level .
3,823
def camelcase2list ( s , lower = False ) : s = re . findall ( r'([A-Z][a-z0-9]+)' , s ) return [ w . lower ( ) for w in s ] if lower else s
Converts a camelcase string to a list .
3,824
def get_route_param_names ( endpoint ) : try : g = current_app . url_map . iter_rules ( endpoint ) return next ( g ) . arguments except KeyError : return { }
Returns parameter names from the route .
3,825
def update_meta_info ( self ) : result = super ( BaseStructuredCalibration , self ) . update_meta_info ( ) result [ 'instrument' ] = self . instrument result [ 'uuid' ] = self . uuid result [ 'tags' ] = self . tags result [ 'type' ] = self . name ( ) minfo = self . meta_info try : result [ 'mode' ] = minfo [ 'mode_name...
Extract metadata from myself
3,826
def gnuplot ( script_name , args_dict = { } , data = [ ] , silent = True ) : gnuplot_command = 'gnuplot' if data : assert 'data' not in args_dict , 'Can\'t use \'data\' variable twice.' data_temp = _GnuplotDataTemp ( * data ) args_dict [ 'data' ] = data_temp . name if args_dict : gnuplot_command += ' -e "' for arg in a...
Call a Gnuplot script passing it arguments and datasets .
3,827
def gnuplot_2d ( x , y , filename , title = '' , x_label = '' , y_label = '' ) : _ , ext = os . path . splitext ( filename ) if ext != '.png' : filename += '.png' gnuplot_cmds = scr = _GnuplotScriptTemp ( gnuplot_cmds ) data = _GnuplotDataTemp ( x , y ) args_dict = { 'filename' : filename , 'filename_data' : data . nam...
Function to produce a general 2D plot .
3,828
def gnuplot_3d_matrix ( z_matrix , filename , title = '' , x_label = '' , y_label = '' ) : _ , ext = os . path . splitext ( filename ) if ext != '.png' : filename += '.png' gnuplot_cmds = scr = _GnuplotScriptTemp ( gnuplot_cmds ) data = _GnuplotDataZMatrixTemp ( z_matrix ) args_dict = { 'filename' : filename , 'filenam...
Function to produce a general 3D plot from a 2D matrix .
3,829
def read ( self , skip = [ ] , goto_metal = None , goto_reaction = None ) : if len ( skip ) > 0 : for skip_f in skip : self . omit_folders . append ( skip_f ) if os . path . isfile ( self . data_base + '/publication.txt' ) : self . user_base_level -= 1 self . stdout . write ( '---------------------- \n' ) self . stdout...
Get reactions from folders .
3,830
def slice_create ( center , block , start = 0 , stop = None ) : do = coor_to_pix_1d ( center - block ) up = coor_to_pix_1d ( center + block ) l = max ( start , do ) if stop is not None : h = min ( up + 1 , stop ) else : h = up + 1 return slice ( l , h , 1 )
Return an slice with a symmetric region around center .
3,831
def image_box ( center , shape , box ) : return tuple ( slice_create ( c , b , stop = s ) for c , s , b in zip ( center , shape , box ) )
Create a region of size box around a center in a image of shape .
3,832
def expand_region ( tuple_of_s , a , b , start = 0 , stop = None ) : return tuple ( expand_slice ( s , a , b , start = start , stop = stop ) for s in tuple_of_s )
Apply expend_slice on a tuple of slices
3,833
def adapt_obsres ( self , obsres ) : _logger . debug ( 'adapt observation result for work dir' ) for f in obsres . images : f . filename = os . path . basename ( f . filename ) return obsres
Adapt obsres after file copy
3,834
def store ( self , completed_task , resultsdir ) : with working_directory ( resultsdir ) : _logger . info ( 'storing result' ) return completed_task . store ( self )
Store the values of the completed task .
3,835
def validate ( self , obj ) : if not isinstance ( obj , self . internal_type ) : raise ValidationError ( obj , self . internal_type ) return True
Validate convertibility to internal representation
3,836
def delete_user ( self , user ) : assert self . user == 'catroot' or self . user == 'postgres' assert not user == 'public' con = self . connection or self . _connect ( ) cur = con . cursor ( ) cur . execute ( 'DROP SCHEMA {user} CASCADE;' . format ( user = user ) ) cur . execute ( 'REVOKE USAGE ON SCHEMA public FROM {u...
Delete user and all data
3,837
def truncate_schema ( self ) : assert self . server == 'localhost' con = self . connection or self . _connect ( ) self . _initialize ( con ) cur = con . cursor ( ) cur . execute ( 'DELETE FROM publication;' ) cur . execute ( 'TRUNCATE systems CASCADE;' ) con . commit ( ) con . close ( ) return
Will delete all data in schema . Only for test use!
3,838
def write ( self , * args , ** kwargs ) : return self . stream . write ( ending = "" , * args , ** kwargs )
Call the stream s write method without linebreaks at line endings .
3,839
def _update ( self ) : if self . strains . size and self . strains . size == self . values . size : x = np . log ( self . strains ) y = self . values if x . size < 4 : self . _interpolater = interp1d ( x , y , 'linear' , bounds_error = False , fill_value = ( y [ 0 ] , y [ - 1 ] ) ) else : self . _interpolater = interp1...
Initialize the 1D interpolation .
3,840
def is_nonlinear ( self ) : return any ( isinstance ( p , NonlinearProperty ) for p in [ self . mod_reduc , self . damping ] )
If nonlinear properties are specified .
3,841
def relative_error ( self ) : if self . previous is not None : try : err = 100. * np . max ( ( self . previous - self . value ) / self . value ) except ZeroDivisionError : err = np . inf else : err = 0 return err
The relative error in percent between the two iterations .
3,842
def duplicate ( cls , other ) : return cls ( other . soil_type , other . thickness , other . shear_vel )
Create a copy of the layer .
3,843
def damping ( self ) : try : value = self . _damping . value except AttributeError : value = self . _damping return value
Strain - compatible damping .
3,844
def auto_discretize ( self , max_freq = 50. , wave_frac = 0.2 ) : layers = [ ] for l in self : if l . soil_type . is_nonlinear : opt_thickness = l . shear_vel / max_freq * wave_frac count = np . ceil ( l . thickness / opt_thickness ) . astype ( int ) thickness = l . thickness / count for _ in range ( count ) : layers ....
Subdivide the layers to capture strain variation .
3,845
def location ( self , wave_field , depth = None , index = None ) : if not isinstance ( wave_field , WaveField ) : wave_field = WaveField [ wave_field ] if index is None and depth is not None : for i , layer in enumerate ( self [ : - 1 ] ) : if layer . depth <= depth < layer . depth_base : depth_within = depth - layer ....
Create a Location for a specific depth .
3,846
def time_average_vel ( self , depth ) : depths = [ l . depth for l in self ] travel_times = [ 0 ] + [ l . travel_time for l in self [ : - 1 ] ] if depths [ - 1 ] < depth : depths . append ( depth ) travel_times . append ( ( depth - self [ - 1 ] . depth ) / self [ - 1 ] . shear_vel ) total_travel_times = np . cumsum ( t...
Calculate the time - average velocity .
3,847
def simplified_rayliegh_vel ( self ) : thicks = np . array ( [ l . thickness for l in self ] ) depths_mid = np . array ( [ l . depth_mid for l in self ] ) shear_vels = np . array ( [ l . shear_vel for l in self ] ) mode_incr = depths_mid * thicks / shear_vels ** 2 shape = np . r_ [ np . cumsum ( mode_incr [ : : - 1 ] )...
Simplified Rayliegh velocity of the site .
3,848
def iter_thickness ( self , depth_total ) : total = 0 depth_prev = 0 while depth_prev < depth_total : total += np . random . exponential ( 1.0 ) depth = np . power ( ( self . c_2 * total ) / self . c_3 + total / self . c_3 + np . power ( self . c_1 , self . c_2 + 1 ) , 1 / ( self . c_2 + 1 ) ) - self . c_1 thickness = ...
Iterate over the varied thicknesses .
3,849
def _calc_covar_matrix ( self , profile ) : corr = self . _calc_corr ( profile ) std = self . _calc_ln_std ( profile ) std *= randnorm . scale var = std ** 2 covar = corr * std [ : - 1 ] * std [ 1 : ] mat = diags ( [ covar , var , covar ] , [ - 1 , 0 , 1 ] ) . toarray ( ) return mat
Calculate the covariance matrix .
3,850
def _calc_corr ( self , profile ) : depth = np . array ( [ l . depth_mid for l in profile [ : - 1 ] ] ) thick = np . diff ( depth ) depth = depth [ 1 : ] corr_depth = ( self . rho_200 * np . power ( ( depth + self . rho_0 ) / ( 200 + self . rho_0 ) , self . b ) ) corr_depth [ depth > 200 ] = self . rho_200 corr_thick =...
Compute the adjacent - layer correlations
3,851
def generic_model ( cls , site_class , ** kwds ) : p = dict ( cls . PARAMS [ site_class ] ) p . update ( kwds ) return cls ( ** p )
Use generic model parameters based on site class .
3,852
def calc_std_damping ( damping ) : damping = np . asarray ( damping ) . astype ( float ) std = ( np . exp ( - 5 ) + np . exp ( - 0.25 ) * np . sqrt ( 100 * damping ) ) / 100. return std
Calculate the standard deviation as a function of damping in decimal .
3,853
def _append_to ( self , field , element ) : if element not in EMPTIES : self . obj . setdefault ( field , [ ] ) self . obj . get ( field ) . append ( element )
Append the element to the field of the record .
3,854
def add_name_variant ( self , name ) : self . _ensure_field ( 'name' , { } ) self . obj [ 'name' ] . setdefault ( 'name_variants' , [ ] ) . append ( name )
Add name variant .
3,855
def add_native_name ( self , name ) : self . _ensure_field ( 'name' , { } ) self . obj [ 'name' ] . setdefault ( 'native_names' , [ ] ) . append ( name )
Add native name .
3,856
def add_previous_name ( self , name ) : self . _ensure_field ( 'name' , { } ) self . obj [ 'name' ] . setdefault ( 'previous_names' , [ ] ) . append ( name )
Add previous name .
3,857
def add_email_address ( self , email , hidden = None ) : existing_emails = get_value ( self . obj , 'email_addresses' , [ ] ) found_email = next ( ( existing_email for existing_email in existing_emails if existing_email . get ( 'value' ) == email ) , None ) if found_email is None : new_email = { 'value' : email } if hi...
Add email address .
3,858
def add_url ( self , url , description = None ) : url = { 'value' : url , } if description : url [ 'description' ] = description self . _append_to ( 'urls' , url )
Add a personal website .
3,859
def add_project ( self , name , record = None , start_date = None , end_date = None , curated = False , current = False ) : new_experiment = { } new_experiment [ 'name' ] = name if start_date : new_experiment [ 'start_date' ] = normalize_date ( start_date ) if end_date : new_experiment [ 'end_date' ] = normalize_date (...
Add an experiment that the person worked on .
3,860
def add_advisor ( self , name , ids = None , degree_type = None , record = None , curated = False ) : new_advisor = { } new_advisor [ 'name' ] = normalize_name ( name ) if ids : new_advisor [ 'ids' ] = force_list ( ids ) if degree_type : new_advisor [ 'degree_type' ] = degree_type if record : new_advisor [ 'record' ] =...
Add an advisor .
3,861
def add_private_note ( self , note , source = None ) : note = { 'value' : note , } if source : note [ 'source' ] = source self . _append_to ( '_private_notes' , note )
Add a private note .
3,862
def _compute_value ( power , wg ) : if power not in wg : p1 , p2 = power if p1 == 0 : yy = wg [ ( 0 , - 1 ) ] wg [ power ] = numpy . power ( yy , p2 / 2 ) . sum ( ) / len ( yy ) else : xx = wg [ ( - 1 , 0 ) ] wg [ power ] = numpy . power ( xx , p1 / 2 ) . sum ( ) / len ( xx ) return wg [ power ]
Return the weight corresponding to single power .
3,863
def _compute_weight ( powers , wg ) : pow1 = ( powers [ 0 ] , 0 ) pow2 = ( 0 , powers [ 1 ] ) cal1 = _compute_value ( pow1 , wg ) cal2 = _compute_value ( pow2 , wg ) return cal1 * cal2
Return the weight corresponding to given powers .
3,864
def imsurfit ( data , order , output_fit = False ) : c0 = complex ( 0 , data . shape [ 0 ] ) c1 = complex ( 0 , data . shape [ 1 ] ) xx , yy = numpy . ogrid [ - 1 : 1 : c0 , - 1 : 1 : c1 ] ncoeff = ( order + 1 ) * ( order + 2 ) // 2 powerlist = list ( _powers ( order ) ) bb = numpy . zeros ( ncoeff ) for idx , powers i...
Fit a bidimensional polynomial to an image .
3,865
def _yql_query ( yql ) : url = _YAHOO_BASE_URL . format ( urlencode ( { 'q' : yql } ) ) _LOGGER . debug ( "Send request to url: %s" , url ) try : request = urlopen ( url ) rawData = request . read ( ) data = json . loads ( rawData . decode ( "utf-8" ) ) _LOGGER . debug ( "Query data from yahoo: %s" , str ( data ) ) ret...
Fetch data from Yahoo! Return a dict if successfull or None .
3,866
def get_woeid ( lat , lon ) : yql = _YQL_WOEID . format ( lat , lon ) tmpData = _yql_query ( yql ) if tmpData is None : _LOGGER . error ( "No woid is received!" ) return None return tmpData . get ( "place" , { } ) . get ( "woeid" , None )
Ask Yahoo! who is the woeid from GPS position .
3,867
def updateWeather ( self ) : yql = _YQL_WEATHER . format ( self . _woeid , self . _unit ) tmpData = _yql_query ( yql ) if tmpData is not None and "channel" in tmpData : self . _data = tmpData [ "channel" ] return True _LOGGER . error ( "Fetch no weather data Yahoo!" ) self . _data = { } return False
Fetch weather data from Yahoo! True if success .
3,868
def get_energies ( atoms_list ) : if len ( atoms_list ) == 1 : return atoms_list [ 0 ] . get_potential_energy ( ) elif len ( atoms_list ) > 1 : energies = [ ] for atoms in atoms_list : energies . append ( atoms . get_potential_energy ( ) ) return energies
Potential energy for a list of atoms objects
3,869
def check_in_ase ( atoms , ase_db , energy = None ) : db_ase = ase . db . connect ( ase_db ) if energy is None : energy = atoms . get_potential_energy ( ) formula = get_chemical_formula ( atoms ) rows = db_ase . select ( energy = energy ) n = 0 ids = [ ] for row in rows : if formula == row . formula : n += 1 ids . appe...
Check if entry is allready in ASE db
3,870
def task ( name , deps = None , fn = None ) : if callable ( deps ) : fn = deps deps = None if not deps and not fn : logger . log ( logger . red ( "The task '%s' is empty" % name ) ) else : tasks [ name ] = [ fn , deps ]
Define a new task .
3,871
def sum_of_gaussian_factory ( N ) : name = "SumNGauss%d" % N attr = { } for i in range ( N ) : key = "amplitude_%d" % i attr [ key ] = Parameter ( key ) key = "center_%d" % i attr [ key ] = Parameter ( key ) key = "stddev_%d" % i attr [ key ] = Parameter ( key ) attr [ 'background' ] = Parameter ( 'background' , defaul...
Return a model of the sum of N Gaussians and a constant background .
3,872
def Debounce ( threshold = 100 ) : threshold *= 1000 max_tick = 0xFFFFFFFF class _decorated ( object ) : def __init__ ( self , pigpio_cb ) : self . _fn = pigpio_cb self . last = 0 self . is_method = False def __call__ ( self , * args , ** kwargs ) : if self . is_method : tick = args [ 3 ] else : tick = args [ 2 ] if se...
Simple debouncing decorator for apigpio callbacks .
3,873
def verify_refresh_request ( request ) : jwtauth_settings = request . app . settings . jwtauth . __dict__ . copy ( ) identity_policy = JWTIdentityPolicy ( ** jwtauth_settings ) return identity_policy . verify_refresh ( request )
Wrapper around JWTIdentityPolicy . verify_refresh which verify if the request to refresh the token is valid . If valid it returns the userid which can be used to create to create an updated identity with remember_identity . Otherwise it raises an exception based on InvalidTokenError .
3,874
def detrend ( arr , x = None , deg = 5 , tol = 1e-3 , maxloop = 10 ) : xx = numpy . arange ( len ( arr ) ) if x is None else x base = arr . copy ( ) trend = base pol = numpy . ones ( ( deg + 1 , ) ) for _ in range ( maxloop ) : pol_new = numpy . polyfit ( xx , base , deg ) pol_norm = numpy . linalg . norm ( pol ) diff_...
Compute a baseline trend of a signal
3,875
def calc_osc_accels ( self , osc_freqs , osc_damping = 0.05 , tf = None ) : if tf is None : tf = np . ones_like ( self . freqs ) else : tf = np . asarray ( tf ) . astype ( complex ) resp = np . array ( [ self . calc_peak ( tf * self . _calc_sdof_tf ( of , osc_damping ) ) for of in osc_freqs ] ) return resp
Compute the pseudo - acceleration spectral response of an oscillator with a specific frequency and damping .
3,876
def _calc_fourier_spectrum ( self , fa_length = None ) : if fa_length is None : n = 1 while n < self . accels . size : n <<= 1 else : n = fa_length self . _fourier_amps = np . fft . rfft ( self . _accels , n ) freq_step = 1. / ( 2 * self . _time_step * ( n / 2 ) ) self . _freqs = freq_step * np . arange ( 1 + n / 2 )
Compute the Fourier Amplitude Spectrum of the time series .
3,877
def _calc_sdof_tf ( self , osc_freq , damping = 0.05 ) : return ( - osc_freq ** 2. / ( np . square ( self . freqs ) - np . square ( osc_freq ) - 2.j * damping * osc_freq * self . freqs ) )
Compute the transfer function for a single - degree - of - freedom oscillator .
3,878
def load_at2_file ( cls , filename ) : with open ( filename ) as fp : next ( fp ) description = next ( fp ) . strip ( ) next ( fp ) parts = next ( fp ) . split ( ) time_step = float ( parts [ 1 ] ) accels = [ float ( p ) for l in fp for p in l . split ( ) ] return cls ( filename , description , time_step , accels )
Read an AT2 formatted time series .
3,879
def load_smc_file ( cls , filename ) : from . tools import parse_fixed_width with open ( filename ) as fp : lines = list ( fp ) lines_str = [ lines . pop ( 0 ) for _ in range ( 11 ) ] if lines_str [ 0 ] . strip ( ) != '2 CORRECTED ACCELEROGRAM' : raise RuntimeWarning ( "Loading uncorrected SMC file." ) m = re . search ...
Read an SMC formatted time series .
3,880
def get_info ( self ) : reconstructed = self . is_reconstructed ( ) site , site_type = self . get_site ( ) return reconstructed , site , site_type
Return surface reconstruction as well as primary and secondary adsorption site labels
3,881
def check_dissociated ( self , cutoff = 1.2 ) : dissociated = False if not len ( self . B ) > self . nslab + 1 : return dissociated adsatoms = [ atom for atom in self . B [ self . nslab : ] ] ads0 , ads1 = set ( atom . symbol for atom in adsatoms ) bond_dist = get_ads_dist ( self . B , ads0 , ads1 ) Cradii = [ cradii [...
Check if adsorbate dissociates
3,882
def is_reconstructed ( self , xy_cutoff = 0.3 , z_cutoff = 0.4 ) : assert self . A , 'Initial slab geometry needed to classify reconstruction' A = self . A [ : - 1 ] . copy ( ) B = self . B [ : - 1 ] . copy ( ) x_indices = np . argsort ( A . positions [ : , 0 ] ) A = A [ x_indices ] B = B [ x_indices ] a = A . position...
Compare initial and final slab configuration to determine if slab reconstructs during relaxation
3,883
def get_under_bridge ( self ) : C0 = self . B [ - 1 : ] * ( 3 , 3 , 1 ) ads_pos = C0 . positions [ 4 ] C = self . get_subsurface_layer ( ) * ( 3 , 3 , 1 ) dis = self . B . cell [ 0 ] [ 0 ] * 2 ret = None for ele in C : new_dis = np . linalg . norm ( ads_pos - ele . position ) if new_dis < dis : dis = new_dis ret = ele ...
Return element closest to the adsorbate in the subsurface layer
3,884
def get_under_hollow ( self ) : C0 = self . B [ - 1 : ] * ( 3 , 3 , 1 ) ads_pos = C0 . positions [ 4 ] C = self . get_subsurface_layer ( ) * ( 3 , 3 , 1 ) ret = 'FCC' if np . any ( [ np . linalg . norm ( ads_pos [ : 2 ] - ele . position [ : 2 ] ) < 0.5 * cradii [ ele . number ] for ele in C ] ) : ret = 'HCP' return ret
Return HCP if an atom is present below the adsorbate in the subsurface layer and FCC if not
3,885
def fmap ( order , aij , bij , x , y ) : u = np . zeros_like ( x ) v = np . zeros_like ( y ) k = 0 for i in range ( order + 1 ) : for j in range ( i + 1 ) : u += aij [ k ] * ( x ** ( i - j ) ) * ( y ** j ) v += bij [ k ] * ( x ** ( i - j ) ) * ( y ** j ) k += 1 return u , v
Evaluate the 2D polynomial transformation .
3,886
def ncoef_fmap ( order ) : ncoef = 0 for i in range ( order + 1 ) : for j in range ( i + 1 ) : ncoef += 1 return ncoef
Expected number of coefficients in a 2D transformation of a given order .
3,887
def order_fmap ( ncoef ) : loop = True order = 1 while loop : loop = not ( ncoef == ncoef_fmap ( order ) ) if loop : order += 1 if order > NMAX_ORDER : print ( 'No. of coefficients: ' , ncoef ) raise ValueError ( "order > " + str ( NMAX_ORDER ) + " not implemented" ) return order
Compute order corresponding to a given number of coefficients .
3,888
def get_template_name ( self ) : if self . template_name is None : name = camelcase2list ( self . __class__ . __name__ ) name = '{}/{}.html' . format ( name . pop ( 0 ) , '_' . join ( name ) ) self . template_name = name . lower ( ) return self . template_name
Returns the name of the template .
3,889
def append_file ( self , file ) : self . files . append ( file ) if self . transformer : future = asyncio . ensure_future ( self . transformer . transform ( file ) ) future . add_done_callback ( self . handle_transform )
Append a new file in the stream .
3,890
def flush_if_ended ( self ) : if self . ended and self . next and len ( self . files ) == self . transformed : future = asyncio . ensure_future ( self . transformer . flush ( ) ) future . add_done_callback ( lambda x : self . next . end_of_stream ( ) )
Call flush function if all files have been transformed .
3,891
def handle_transform ( self , task ) : self . transformed += 1 file = task . result ( ) if file : self . next . append_file ( file ) self . flush_if_ended ( )
Handle a transform callback .
3,892
def pipe ( self , transformer ) : if self . next : return stream = Stream ( ) self . next = stream stream . prev = self self . transformer = transformer transformer . stream = self transformer . piped ( ) for file in self . files : future = asyncio . ensure_future ( self . transformer . transform ( file ) ) future . ad...
Pipe this stream to another .
3,893
def pipes ( stream , * transformers ) : for transformer in transformers : stream = stream . pipe ( transformer ) return stream
Pipe several transformers end to end .
3,894
def convert ( self , val ) : pre = super ( Parameter , self ) . convert ( val ) if self . custom_validator is not None : post = self . custom_validator ( pre ) else : post = pre return post
Convert input values to type values .
3,895
def validate ( self , val ) : if self . validation : self . type . validate ( val ) if self . custom_validator is not None : self . custom_validator ( val ) return True
Validate values according to the requirement
3,896
def route ( obj , rule , * args , ** kwargs ) : def decorator ( cls ) : endpoint = kwargs . get ( 'endpoint' , camel_to_snake ( cls . __name__ ) ) kwargs [ 'view_func' ] = cls . as_view ( endpoint ) obj . add_url_rule ( rule , * args , ** kwargs ) return cls return decorator
Decorator for the View classes .
3,897
def fit ( self , target_type , target , adjust_thickness = False , adjust_site_atten = False , adjust_source_vel = False ) : density = self . profile . density nl = len ( density ) slowness = self . profile . slowness thickness = self . profile . thickness site_atten = self . _site_atten initial = slowness bounds = 1 /...
Fit to a target crustal amplification or site term .
3,898
def wave_at_location ( self , l ) : cterm = 1j * self . _wave_nums [ l . index ] * l . depth_within if l . wave_field == WaveField . within : return ( self . _waves_a [ l . index ] * np . exp ( cterm ) + self . _waves_b [ l . index ] * np . exp ( - cterm ) ) elif l . wave_field == WaveField . outcrop : return 2 * self ...
Compute the wave field at specific location .
3,899
def calc_accel_tf ( self , lin , lout ) : tf = self . wave_at_location ( lout ) / self . wave_at_location ( lin ) return tf
Compute the acceleration transfer function .