_id
stringlengths
2
7
title
stringlengths
1
88
partition
stringclasses
3 values
text
stringlengths
31
13.1k
language
stringclasses
1 value
meta_information
dict
q273700
Kernel.iteration_length
test
def iteration_length(self, dimension=None): """ Return the number of global loop iterations that are performed. If dimension is not None, it is the loop dimension that is returned (-1 is the inner most loop and 0 the outermost) """ total_length = 1 if dimension is not None: loops = [self._loop_stack[dimension]] else: loops = reversed(self._loop_stack)
python
{ "resource": "" }
q273701
Kernel.get_loop_stack
test
def get_loop_stack(self, subs_consts=False): """Yield loop stack dictionaries in order from outer to inner.""" for l in self._loop_stack: if subs_consts: yield {'index': l[0],
python
{ "resource": "" }
q273702
Kernel.index_order
test
def index_order(self, sources=True, destinations=True): """ Return the order of indices as they appear in array references. Use *source* and *destination* to filter output """ if sources: arefs = chain(*self.sources.values()) else: arefs = [] if destinations: arefs = chain(arefs, *self.destinations.values()) ret = [] for a in
python
{ "resource": "" }
q273703
Kernel.compile_sympy_accesses
test
def compile_sympy_accesses(self, sources=True, destinations=True): """ Return a dictionary of lists of sympy accesses, for each variable. Use *source* and *destination* to filter output """ sympy_accesses = defaultdict(list) # Compile sympy accesses for var_name in self.variables: if sources: for r in self.sources.get(var_name, []): if r is None:
python
{ "resource": "" }
q273704
Kernel.compile_relative_distances
test
def compile_relative_distances(self, sympy_accesses=None): """ Return load and store distances between accesses. :param sympy_accesses: optionally restrict accesses, default from compile_sympy_accesses() e.g. if accesses are to [+N, +1, -1, -N], relative distances are [N-1, 2, N-1]
python
{ "resource": "" }
q273705
Kernel.global_iterator_to_indices
test
def global_iterator_to_indices(self, git=None): """ Return sympy expressions translating global_iterator to loop indices. If global_iterator is given, an integer is returned """ # unwind global iteration count into loop counters: base_loop_counters = {} global_iterator = symbol_pos_int('global_iterator') idiv = implemented_function(sympy.Function(str('idiv')), lambda x, y: x//y) total_length = 1 last_incr = 1 for var_name, start, end, incr in reversed(self._loop_stack): loop_var = symbol_pos_int(var_name)
python
{ "resource": "" }
q273706
Kernel.global_iterator
test
def global_iterator(self): """ Return global iterator sympy expression """ global_iterator = sympy.Integer(0) total_length = sympy.Integer(1) for var_name, start, end, incr in reversed(self._loop_stack): loop_var = symbol_pos_int(var_name) length = end -
python
{ "resource": "" }
q273707
Kernel.indices_to_global_iterator
test
def indices_to_global_iterator(self, indices): """ Transform a dictionary of indices to a global iterator integer. Inverse of global_iterator_to_indices().
python
{ "resource": "" }
q273708
Kernel.max_global_iteration
test
def max_global_iteration(self): """Return global iterator with last iteration number"""
python
{ "resource": "" }
q273709
Kernel.print_kernel_info
test
def print_kernel_info(self, output_file=sys.stdout): """Print kernel information in human readble format.""" table = (' idx | min max step\n' + '---------+---------------------------------\n') for l in self._loop_stack: table += '{:>8} | {!r:>10} {!r:>10} {!r:>10}\n'.format(*l) print(prefix_indent('loop stack: ', table), file=output_file) table = (' name | offsets ...\n' + '---------+------------...\n') for name, offsets in list(self.sources.items()): prefix = '{:>8} | '.format(name) right_side = '\n'.join(['{!r:}'.format(o) for o in offsets]) table += prefix_indent(prefix, right_side, later_prefix=' | ') print(prefix_indent('data sources: ', table), file=output_file) table = (' name | offsets ...\n' + '---------+------------...\n') for name, offsets in list(self.destinations.items()): prefix = '{:>8} | '.format(name) right_side = '\n'.join(['{!r:}'.format(o) for o in offsets])
python
{ "resource": "" }
q273710
Kernel.print_variables_info
test
def print_variables_info(self, output_file=sys.stdout): """Print variables information in human readble format.""" table = (' name | type size \n' + '---------+-------------------------\n') for name, var_info in list(self.variables.items()):
python
{ "resource": "" }
q273711
Kernel.print_constants_info
test
def print_constants_info(self, output_file=sys.stdout): """Print constants information in human readble format.""" table = (' name | value \n' + '---------+-----------\n') for name, value in list(self.constants.items()):
python
{ "resource": "" }
q273712
KernelCode.print_kernel_code
test
def print_kernel_code(self, output_file=sys.stdout): """Print source code of kernel."""
python
{ "resource": "" }
q273713
KernelCode.conv_ast_to_sym
test
def conv_ast_to_sym(self, math_ast): """ Convert mathematical expressions to a sympy representation. May only contain paranthesis, addition, subtraction and multiplication from AST. """ if type(math_ast) is c_ast.ID: return symbol_pos_int(math_ast.name) elif type(math_ast) is c_ast.Constant: return sympy.Integer(math_ast.value) else: # elif type(dim) is c_ast.BinaryOp: op = { '*':
python
{ "resource": "" }
q273714
KernelCode._get_offsets
test
def _get_offsets(self, aref, dim=0): """ Return a tuple of offsets of an ArrayRef object in all dimensions. The index order is right to left (c-code order). e.g. c[i+1][j-2] -> (-2, +1) If aref is actually a c_ast.ID, None will be returned. """ if isinstance(aref, c_ast.ID): return None # Check for restrictions assert type(aref.name) in [c_ast.ArrayRef, c_ast.ID], \ "array references must only be used with variables or other array references" assert type(aref.subscript) in [c_ast.ID, c_ast.Constant, c_ast.BinaryOp], \ 'array subscript must only contain variables or binary operations' # Convert subscript to sympy and append
python
{ "resource": "" }
q273715
KernelCode._get_basename
test
def _get_basename(cls, aref): """ Return base name of ArrayRef object. e.g. c[i+1][j-2] -> 'c' """ if isinstance(aref.name, c_ast.ArrayRef): return cls._get_basename(aref.name)
python
{ "resource": "" }
q273716
KernelCode.get_index_type
test
def get_index_type(self, loop_nest=None): """ Return index type used in loop nest. If index type between loops differ, an exception is raised. """ if loop_nest is None: loop_nest = self.get_kernel_loop_nest() if type(loop_nest) is c_ast.For: loop_nest = [loop_nest] index_types = (None, None)
python
{ "resource": "" }
q273717
KernelCode._build_const_declartions
test
def _build_const_declartions(self, with_init=True): """ Generate constants declarations :return: list of declarations """ decls = [] # Use type as provided by user in loop indices index_type = self.get_index_type() i = 2 # subscript for cli input, 1 is reserved for repeat for k in self.constants: # const long long N = strtoul(argv[2]) # with increasing N and 1 # TODO change subscript of argv depending on constant count type_decl = c_ast.TypeDecl(k.name, ['const'], c_ast.IdentifierType(index_type)) init = None
python
{ "resource": "" }
q273718
KernelCode.get_array_declarations
test
def get_array_declarations(self): """Return array declarations.""" return [d for d in self.kernel_ast.block_items
python
{ "resource": "" }
q273719
KernelCode.get_kernel_loop_nest
test
def get_kernel_loop_nest(self): """Return kernel loop nest including any preceding pragmas and following swaps.""" loop_nest = [s for s in self.kernel_ast.block_items
python
{ "resource": "" }
q273720
KernelCode._build_array_declarations
test
def _build_array_declarations(self, with_init=True): """ Generate declaration statements for arrays. Also transforming multi-dim to 1d arrays and initializing with malloc. :param with_init: ommit malloc initialization :return: list of declarations nodes, dictionary of array names and
python
{ "resource": "" }
q273721
KernelCode._find_inner_most_loop
test
def _find_inner_most_loop(self, loop_nest): """Return inner most for loop in loop nest""" r = None for s in loop_nest:
python
{ "resource": "" }
q273722
KernelCode._build_array_initializations
test
def _build_array_initializations(self, array_dimensions): """ Generate initialization statements for arrays. :param array_dimensions: dictionary of array dimensions :return: list of nodes """ kernel = deepcopy(deepcopy(self.get_kernel_loop_nest())) # traverse to the inner most for loop: inner_most = self._find_inner_most_loop(kernel) orig_inner_stmt = inner_most.stmt inner_most.stmt = c_ast.Compound([]) rand_float_str = str(random.uniform(1.0, 0.1)) # find all array references in original orig_inner_stmt for aref in find_node_type(orig_inner_stmt, c_ast.ArrayRef):
python
{ "resource": "" }
q273723
KernelCode._build_dummy_calls
test
def _build_dummy_calls(self): """ Generate false if branch with dummy calls Requires kerncraft.h to be included, which defines dummy(...) and var_false. :return: dummy statement """ # Make sure nothing gets removed by inserting dummy calls dummy_calls = [] for d in self.kernel_ast.block_items: # Only consider toplevel declarations from kernel ast if type(d) is not c_ast.Decl: continue if type(d.type) is c_ast.ArrayDecl: dummy_calls.append(c_ast.FuncCall(
python
{ "resource": "" }
q273724
KernelCode._build_kernel_function_declaration
test
def _build_kernel_function_declaration(self, name='kernel'): """Build and return kernel function declaration""" array_declarations, array_dimensions = self._build_array_declarations(with_init=False) scalar_declarations = self._build_scalar_declarations(with_init=False) const_declarations = self._build_const_declartions(with_init=False) return c_ast.FuncDecl(args=c_ast.ParamList(params=array_declarations + scalar_declarations +
python
{ "resource": "" }
q273725
KernelCode._build_scalar_declarations
test
def _build_scalar_declarations(self, with_init=True): """Build and return scalar variable declarations""" # copy scalar declarations from from kernel ast scalar_declarations = [deepcopy(d) for d in self.kernel_ast.block_items if type(d) is c_ast.Decl and type(d.type) is c_ast.TypeDecl] # add init values to declarations if with_init: random.seed(2342) # we want reproducible random numbers for d in scalar_declarations: if d.type.type.names[0] in ['double', 'float']:
python
{ "resource": "" }
q273726
KernelCode.get_kernel_code
test
def get_kernel_code(self, openmp=False, as_filename=False, name='kernel'): """ Generate and return compilable source code with kernel function from AST. :param openmp: if true, OpenMP code will be generated :param as_filename: if true, will save to file and return filename :param name: name of kernel function """ assert self.kernel_ast is not None, "AST does not exist, this could be due to running " \ "based on a kernel description rather than code." file_name = 'kernel' if openmp: file_name += '-omp' file_name += '.c' fp, already_available = self._get_intermediate_file( file_name, machine_and_compiler_dependent=False) # Use already cached version if already_available: code = fp.read() else: array_declarations, array_dimensions = self._build_array_declarations() # Prepare actual kernel loop nest if openmp: # with OpenMP code kernel = deepcopy(self.get_kernel_loop_nest()) # find all array references in kernel for aref in find_node_type(kernel, c_ast.ArrayRef): # transform to 1d references transform_multidim_to_1d_ref(aref, array_dimensions) omp_pragmas = [p for p in find_node_type(kernel, c_ast.Pragma) if 'omp' in p.string] # TODO if omp parallel was found, remove it (also replace "parallel for" -> "for") # if no omp for pragmas are present, insert suitable ones if not omp_pragmas: kernel.insert(0, c_ast.Pragma("omp for")) # otherwise do not change anything else:
python
{ "resource": "" }
q273727
KernelCode._build_kernel_call
test
def _build_kernel_call(self, name='kernel'): """Generate and return kernel call ast.""" return c_ast.FuncCall(name=c_ast.ID(name=name), args=c_ast.ExprList(exprs=[ c_ast.ID(name=d.name) for d in (
python
{ "resource": "" }
q273728
KernelCode.get_main_code
test
def get_main_code(self, as_filename=False, kernel_function_name='kernel'): """ Generate and return compilable source code from AST. """ # TODO produce nicer code, including help text and other "comfort features". assert self.kernel_ast is not None, "AST does not exist, this could be due to running " \ "based on a kernel description rather than code." fp, already_available = self._get_intermediate_file('main.c', machine_and_compiler_dependent=False) # Use already cached version if already_available: code = fp.read() else: parser = CParser() template_code = self.CODE_TEMPLATE template_ast = parser.parse(clean_code(template_code, macros=True, comments=True, pragmas=False)) ast = deepcopy(template_ast) # Define and replace DECLARE_CONSTS replace_id(ast, "DECLARE_CONSTS", self._build_const_declartions(with_init=True)) # Define and replace DECLARE_ARRAYS array_declarations, array_dimensions = self._build_array_declarations() replace_id(ast, "DECLARE_ARRAYS", array_declarations) # Define and replace DECLARE_INIT_SCALARS replace_id(ast, "DECLARE_INIT_SCALARS", self._build_scalar_declarations()) # Define and replace DUMMY_CALLS replace_id(ast,
python
{ "resource": "" }
q273729
KernelCode.iaca_analysis
test
def iaca_analysis(self, micro_architecture, asm_block='auto', pointer_increment='auto_with_manual_fallback', verbose=False): """ Run an IACA analysis and return its outcome. *asm_block* controls how the to-be-marked block is chosen. "auto" (default) results in the largest block, "manual" results in interactive and a number in the according block. *pointer_increment* is the number of bytes the pointer is incremented after the loop or - 'auto': automatic detection, RuntimeError is raised in case of failure - 'auto_with_manual_fallback': automatic detection, fallback to manual input - 'manual': prompt user """ asm_filename = self.compile_kernel(assembly=True, verbose=verbose) asm_marked_filename =
python
{ "resource": "" }
q273730
KernelCode.build_executable
test
def build_executable(self, lflags=None, verbose=False, openmp=False): """Compile source to executable with likwid capabilities and return the executable name.""" compiler, compiler_args = self._machine.get_compiler() kernel_obj_filename = self.compile_kernel(openmp=openmp, verbose=verbose) out_filename, already_exists = self._get_intermediate_file( os.path.splitext(os.path.basename(kernel_obj_filename))[0], binary=True, fp=False) if not already_exists: main_source_filename = self.get_main_code(as_filename=True) if not (('LIKWID_INCLUDE' in os.environ or 'LIKWID_INC' in os.environ) and 'LIKWID_LIB' in os.environ): print('Could not find LIKWID_INCLUDE (e.g., "-I/app/likwid/4.1.2/include") and ' 'LIKWID_LIB (e.g., "-L/apps/likwid/4.1.2/lib") environment variables', file=sys.stderr) sys.exit(1) compiler_args += [ '-std=c99', '-I'+reduce_path(os.path.abspath(os.path.dirname( os.path.realpath(__file__)))+'/headers/'), os.environ.get('LIKWID_INCLUDE', ''), os.environ.get('LIKWID_INC', ''), '-llikwid'] # This is a special case for unittesting if os.environ.get('LIKWID_LIB') ==
python
{ "resource": "" }
q273731
KernelDescription.string_to_sympy
test
def string_to_sympy(cls, s): """Convert any string to a sympy object or None.""" if isinstance(s, int): return sympy.Integer(s) elif isinstance(s, list): return tuple([cls.string_to_sympy(e) for e in s]) elif s is None: return None else: # Step 1 build expression with the whole alphabet
python
{ "resource": "" }
q273732
MachineModel.get_identifier
test
def get_identifier(self): """Return identifier which is either the machine file name or sha256 checksum of data."""
python
{ "resource": "" }
q273733
MachineModel.get_last_modified_datetime
test
def get_last_modified_datetime(self): """Return datetime object of modified time of machine file. Return now if not a file.""" if self._path: statbuf = os.stat(self._path)
python
{ "resource": "" }
q273734
MachineModel.get_cachesim
test
def get_cachesim(self, cores=1): """ Return a cachesim.CacheSimulator object based on the machine description. :param cores: core count (default: 1) """ cache_dict = {} for c in self['memory hierarchy']: # Skip main memory if 'cache per group' not in c: continue cache_dict[c['level']] = deepcopy(c['cache per group']) # Scale size of shared caches
python
{ "resource": "" }
q273735
MachineModel.get_bandwidth
test
def get_bandwidth(self, cache_level, read_streams, write_streams, threads_per_core, cores=None): """ Return best fitting bandwidth according to number of threads, read and write streams. :param cache_level: integer of cache (0 is L1, 1 is L2 ...) :param read_streams: number of read streams expected :param write_streams: number of write streams expected :param threads_per_core: number of threads that are run on each core :param cores: if not given, will choose maximum bandwidth for single NUMA domain """ # try to find best fitting kernel (closest to read/write ratio): # write allocate has to be handled in kernel information (all writes are also reads) # TODO support for non-write-allocate architectures try: target_ratio = read_streams/write_streams except ZeroDivisionError: target_ratio = float('inf') measurement_kernel = 'load' measurement_kernel_info = self['benchmarks']['kernels'][measurement_kernel] measurement_kernel_ratio = float('inf') for kernel_name, kernel_info in sorted(self['benchmarks']['kernels'].items()): try: kernel_ratio = ((kernel_info['read streams']['streams'] + kernel_info['write streams']['streams'] - kernel_info['read+write streams']['streams']) / kernel_info['write streams']['streams']) except ZeroDivisionError: kernel_ratio = float('inf') if abs(kernel_ratio - target_ratio) < abs(measurement_kernel_ratio - target_ratio): measurement_kernel = kernel_name measurement_kernel_info = kernel_info measurement_kernel_ratio = kernel_ratio # choose smt, and then use max/saturation bw bw_level = self['memory hierarchy'][cache_level]['level'] bw_measurements = \ self['benchmarks']['measurements'][bw_level][threads_per_core] assert threads_per_core == bw_measurements['threads per core'], \ 'malformed measurement dictionary in machine file.' if cores is not None: # Used by Roofline model run_index = bw_measurements['cores'].index(cores) bw =
python
{ "resource": "" }
q273736
MachineModel.get_compiler
test
def get_compiler(self, compiler=None, flags=None): """ Return tuple of compiler and compiler flags. Selects compiler and flags from machine description file, commandline arguments or call arguements. """ if self._args: compiler = compiler or self._args.compiler flags = flags or self._args.compiler_flags if compiler is None: # Select first available compiler in machine description file's compiler dict for c in self['compiler'].keys(): # Making sure compiler is available: if find_executable(c) is not None: compiler = c break else: raise RuntimeError("No compiler ({}) was found. Add different one in machine
python
{ "resource": "" }
q273737
MachineModel.parse_perfctr_event
test
def parse_perfctr_event(perfctr): """ Parse events in machine description to tuple representation used in Benchmark module. Examples: >>> parse_perfctr_event('PERF_EVENT:REG[0-3]') ('PERF_EVENT', 'REG[0-3]') >>> parse_perfctr_event('PERF_EVENT:REG[0-3]:STAY:FOO=23:BAR=0x23') ('PERF_EVENT', 'REG[0-3]', {'STAY': None, 'FOO': 23, 'BAR': 35}) """ split_perfctr = perfctr.split(':') assert len(split_perfctr) >= 2, "Atleast one colon (:) is required in the event name" event_tuple = split_perfctr[:2] parameters = {}
python
{ "resource": "" }
q273738
Intervals._enforce_no_overlap
test
def _enforce_no_overlap(self, start_at=0): """Enforce that no ranges overlap in internal storage.""" i = start_at while i+1 < len(self.data): if self.data[i][1] >= self.data[i+1][0]:
python
{ "resource": "" }
q273739
get_header_path
test
def get_header_path() -> str: """Return local folder path of header files.""" import os
python
{ "resource": "" }
q273740
CacheSimulationPredictor._align_iteration_with_cl_boundary
test
def _align_iteration_with_cl_boundary(self, iteration, subtract=True): """Align iteration with cacheline boundary.""" # FIXME handle multiple datatypes element_size = self.kernel.datatypes_size[self.kernel.datatype] cacheline_size = self.machine['cacheline size'] elements_per_cacheline = int(cacheline_size // element_size) # Gathering some loop information: inner_loop = list(self.kernel.get_loop_stack(subs_consts=True))[-1] inner_increment = inner_loop['increment']
python
{ "resource": "" }
q273741
CacheSimulationPredictor.get_loads
test
def get_loads(self): """Return a list with number of loaded cache lines per memory hierarchy level.""" return
python
{ "resource": "" }
q273742
CacheSimulationPredictor.get_hits
test
def get_hits(self): """Return a list with number of hit cache lines per memory hierarchy level.""" return
python
{ "resource": "" }
q273743
CacheSimulationPredictor.get_misses
test
def get_misses(self): """Return a list with number of missed cache lines per memory hierarchy level.""" return
python
{ "resource": "" }
q273744
CacheSimulationPredictor.get_stores
test
def get_stores(self): """Return a list with number of stored cache lines per memory hierarchy level.""" return
python
{ "resource": "" }
q273745
CacheSimulationPredictor.get_evicts
test
def get_evicts(self): """Return a list with number of evicted cache lines per memory hierarchy level.""" return
python
{ "resource": "" }
q273746
CacheSimulationPredictor.get_infos
test
def get_infos(self): """Return verbose information about the predictor.""" first_dim_factor = self.first_dim_factor infos = {'memory hierarchy': [], 'cache stats': self.stats, 'cachelines in stats': first_dim_factor} for cache_level, cache_info in list(enumerate(self.machine['memory hierarchy'])): infos['memory hierarchy'].append({ 'index': len(infos['memory hierarchy']), 'level': '{}'.format(cache_info['level']), 'total loads': self.stats[cache_level]['LOAD_byte']/first_dim_factor, 'total misses': self.stats[cache_level]['MISS_byte']/first_dim_factor, 'total hits': self.stats[cache_level]['HIT_byte']/first_dim_factor, 'total stores': self.stats[cache_level]['STORE_byte']/first_dim_factor,
python
{ "resource": "" }
q273747
fix_env_variable
test
def fix_env_variable(name, value): """Fix environment variable to a value within context. Unset if value is None.""" orig = os.environ.get(name, None) if value is not None: # Set if value is not None os.environ[name] = value
python
{ "resource": "" }
q273748
Benchmark.configure_arggroup
test
def configure_arggroup(cls, parser): """Configure argument parser.""" parser.add_argument( '--no-phenoecm', action='store_true', help='Disables the phenomenological ECM model building.') parser.add_argument( '--iterations', type=int, default=10, help='Number of outer-loop iterations (e.g. time loop) during benchmarking. '
python
{ "resource": "" }
q273749
Benchmark.report
test
def report(self, output_file=sys.stdout): """Report gathered analysis data in human readable form.""" if self.verbose > 1: with pprint_nosort(): pprint.pprint(self.results) if self.verbose > 0: print('Runtime (per repetition): {:.2g} s'.format( self.results['Runtime (per repetition) [s]']), file=output_file) if self.verbose > 0: print('Iterations per repetition: {!s}'.format( self.results['Iterations per repetition']), file=output_file) print('Runtime (per cacheline update): {:.2f} cy/CL'.format( self.results['Runtime (per cacheline update) [cy/CL]']), file=output_file) print('MEM volume (per repetition): {:.0f} Byte'.format( self.results['MEM volume (per repetition) [B]']), file=output_file) print('Performance: {:.2f} MFLOP/s'.format(self.results['Performance [MFLOP/s]']), file=output_file) print('Performance: {:.2f} MLUP/s'.format(self.results['Performance [MLUP/s]']), file=output_file) print('Performance: {:.2f} It/s'.format(self.results['Performance [MIt/s]']), file=output_file) if self.verbose > 0: print('MEM bandwidth: {:.2f} MByte/s'.format(self.results['MEM BW [MByte/s]']), file=output_file) print('', file=output_file) if not self.no_phenoecm: print("Data Transfers:") print("{:^8} |".format("cache"), end='')
python
{ "resource": "" }
q273750
parse_description
test
def parse_description(): """ Parse the description in the README file CommandLine: python -c "import setup; print(setup.parse_description())" """ from os.path import dirname, join, exists readme_fpath = join(dirname(__file__), 'README.md') # print('readme_fpath = %r' % (readme_fpath,)) # This breaks on pip install, so check that it exists. if exists(readme_fpath): # try: # # convert markdown to rst for pypi # import pypandoc # return pypandoc.convert(readme_fpath, 'rst') # except Exception as ex: # strip out markdown to make a clean readme for pypi textlines = [] with open(readme_fpath, 'r') as f: capture = False for line in f.readlines(): if '# Purpose' in line:
python
{ "resource": "" }
q273751
schedule_retry
test
def schedule_retry(self, config): """Schedule a retry""" raise self.retry(countdown=config.get('SAILTHRU_RETRY_SECONDS'),
python
{ "resource": "" }
q273752
_build_purchase_item
test
def _build_purchase_item(course_id, course_url, cost_in_cents, mode, course_data, sku): """Build and return Sailthru purchase item object""" # build item description item = { 'id': "{}-{}".format(course_id, mode), 'url': course_url, 'price': cost_in_cents, 'qty': 1, } # get title from course info if
python
{ "resource": "" }
q273753
_record_purchase
test
def _record_purchase(sailthru_client, email, item, purchase_incomplete, message_id, options): """Record a purchase in Sailthru Arguments: sailthru_client (object): SailthruClient email (str): user's email address item (dict): Sailthru required information about the course purchase_incomplete (boolean): True if adding item to shopping cart message_id (str): Cookie used to identify marketing campaign options (dict): Sailthru purchase API options (e.g. template name) Returns: False if retryable error, else True """ try: sailthru_response = sailthru_client.purchase(email, [item], incomplete=purchase_incomplete, message_id=message_id,
python
{ "resource": "" }
q273754
_get_course_content
test
def _get_course_content(course_id, course_url, sailthru_client, site_code, config): """Get course information using the Sailthru content api or from cache. If there is an error, just return with an empty response. Arguments: course_id (str): course key of the course course_url (str): LMS url for course info page. sailthru_client (object): SailthruClient site_code (str): site code config (dict): config options Returns: course information from Sailthru """ # check cache first cache_key = "{}:{}".format(site_code, course_url) response = cache.get(cache_key) if not response: try: sailthru_response = sailthru_client.api_get("content", {"id": course_url}) if not sailthru_response.is_ok(): response = {} else: response = sailthru_response.json cache.set(cache_key, response, config.get('SAILTHRU_CACHE_TTL_SECONDS')) except SailthruClientError:
python
{ "resource": "" }
q273755
_get_course_content_from_ecommerce
test
def _get_course_content_from_ecommerce(course_id, site_code=None): """ Get course information using the Ecommerce course api. In case of error returns empty response. Arguments: course_id (str): course key of the course site_code (str): site code Returns: course information from Ecommerce """ api = get_ecommerce_client(site_code=site_code) try: api_response = api.courses(course_id).get() except Exception: # pylint: disable=broad-except
python
{ "resource": "" }
q273756
_update_unenrolled_list
test
def _update_unenrolled_list(sailthru_client, email, course_url, unenroll): """Maintain a list of courses the user has unenrolled from in the Sailthru user record Arguments: sailthru_client (object): SailthruClient email (str): user's email address course_url (str): LMS url for course info page. unenroll (boolean): True if unenrolling, False if enrolling Returns: False if retryable error, else True """ try: # get the user 'vars' values from sailthru sailthru_response = sailthru_client.api_get("user", {"id": email, "fields": {"vars": 1}}) if not sailthru_response.is_ok(): error = sailthru_response.get_error() logger.error("Error attempting to read user record from Sailthru: %s", error.get_message()) return not can_retry_sailthru_request(error) response_json = sailthru_response.json unenroll_list = [] if response_json and "vars" in response_json and response_json["vars"] \ and "unenrolled" in response_json["vars"]: unenroll_list = response_json["vars"]["unenrolled"] changed = False # if unenrolling, add course to unenroll list if unenroll: if course_url not in unenroll_list: unenroll_list.append(course_url)
python
{ "resource": "" }
q273757
send_course_refund_email
test
def send_course_refund_email(self, email, refund_id, amount, course_name, order_number, order_url, site_code=None): """ Sends the course refund email. Args: self: Ignore. email (str): Recipient's email address. refund_id (int): ID of the refund that initiated this task. amount (str): Formatted amount of the refund. course_name (str): Name of the course for which payment was refunded. order_number (str): Order number of the order that was refunded. order_url (str): Receipt URL of the refunded order. site_code (str): Identifier of the site sending the email. """ config = get_sailthru_configuration(site_code) try: sailthru_client = get_sailthru_client(site_code) except SailthruError: # NOTE: We rely on the function to log the error for us return email_vars = { 'amount': amount, 'course_name': course_name, 'order_number': order_number, 'order_url': order_url, } try: response = sailthru_client.send( template=config['templates']['course_refund'], email=email, _vars=email_vars ) except SailthruClientError: logger.exception( 'A client error occurred while attempting to send a course refund
python
{ "resource": "" }
q273758
_send_offer_assignment_notification_email
test
def _send_offer_assignment_notification_email(config, user_email, subject, email_body, site_code, task): """Handles sending offer assignment notification emails and retrying failed emails when appropriate.""" try: sailthru_client = get_sailthru_client(site_code) except SailthruError: logger.exception( '[Offer Assignment] A client error occurred while attempting to send a offer assignment notification.' ' Message: {message}'.format(message=email_body) ) return None email_vars = { 'subject': subject, 'email_body': email_body, } try: response = sailthru_client.send( template=config['templates']['assignment_email'], email=user_email, _vars=email_vars ) except SailthruClientError: logger.exception( '[Offer Assignment] A client error occurred while attempting to send a offer assignment notification.' ' Message: {message}'.format(message=email_body) ) return None if not response.is_ok(): error = response.get_error() logger.error( '[Offer Assignment] A {token_error_code} - {token_error_message} error occurred' ' while attempting to send a offer assignment notification.'
python
{ "resource": "" }
q273759
get_logger_config
test
def get_logger_config(log_dir='/var/tmp', logging_env='no_env', edx_filename='edx.log', dev_env=False, debug=False, local_loglevel='INFO', service_variant='ecomworker'): """ Returns a dictionary containing logging configuration. If dev_env is True, logging will not be done via local rsyslogd. Instead, application logs will be dropped into log_dir. 'edx_filename' is ignored unless dev_env is True. """ # Revert to INFO if an invalid string is passed in if local_loglevel not in ['DEBUG', 'INFO', 'WARNING', 'ERROR', 'CRITICAL']: local_loglevel = 'INFO' hostname = platform.node().split('.')[0] syslog_format = ( '[service_variant={service_variant}]' '[%(name)s][env:{logging_env}] %(levelname)s ' '[{hostname} %(process)d] [%(filename)s:%(lineno)d] ' '- %(message)s' ).format( service_variant=service_variant, logging_env=logging_env, hostname=hostname ) if debug: handlers = ['console'] else: handlers = ['local'] logger_config = { 'version': 1, 'disable_existing_loggers': False, 'formatters': { 'standard': { 'format': '%(asctime)s %(levelname)s %(process)d ' '[%(name)s] %(filename)s:%(lineno)d - %(message)s', }, 'syslog_format': {'format': syslog_format}, 'raw': {'format': '%(message)s'}, }, 'handlers': { 'console': { 'level': 'DEBUG' if debug else 'INFO', 'class': 'logging.StreamHandler', 'formatter': 'standard', 'stream': sys.stdout, }, }, 'loggers': { 'requests': {
python
{ "resource": "" }
q273760
_retry_order
test
def _retry_order(self, exception, max_fulfillment_retries, order_number): """ Retry with exponential backoff until fulfillment succeeds or the retry limit is reached. If the retry limit is exceeded, the exception is re-raised. """ retries = self.request.retries if retries == max_fulfillment_retries: logger.exception('Fulfillment of order [%s] failed. Giving up.', order_number) else:
python
{ "resource": "" }
q273761
fulfill_order
test
def fulfill_order(self, order_number, site_code=None, email_opt_in=False): """Fulfills an order. Arguments: order_number (str): Order number indicating which order to fulfill. Returns: None """ max_fulfillment_retries = get_configuration('MAX_FULFILLMENT_RETRIES', site_code=site_code) api = get_ecommerce_client(site_code=site_code) try: logger.info('Requesting fulfillment of order [%s].', order_number) api.orders(order_number).fulfill.put(email_opt_in=email_opt_in) except exceptions.HttpClientError as exc:
python
{ "resource": "" }
q273762
get_sailthru_client
test
def get_sailthru_client(site_code): """ Returns a Sailthru client for the specified site. Args: site_code (str): Site for which the client should be configured. Returns: SailthruClient Raises: SailthruNotEnabled: If Sailthru is not enabled for the specified site. ConfigurationError: If either the Sailthru API key or secret are not set for the site. """ # Get configuration config = get_sailthru_configuration(site_code) # Return if Sailthru integration disabled if not config.get('SAILTHRU_ENABLE'): msg = 'Sailthru is not enabled for site {}'.format(site_code) log.debug(msg) raise SailthruNotEnabled(msg) # Make sure
python
{ "resource": "" }
q273763
Cache.get
test
def get(self, key): """Get an object from the cache Arguments: key (str): Cache key Returns: Cached object """ lock.acquire() try: if key not in self: return None current_time = time.time() if self[key].expire > current_time: return self[key].value # expired key, clean out all expired keys deletes = []
python
{ "resource": "" }
q273764
Cache.set
test
def set(self, key, value, duration): """Save an object in the cache Arguments: key (str): Cache key value (object): object to cache duration (int): time in seconds to keep object in cache """
python
{ "resource": "" }
q273765
get_configuration
test
def get_configuration(variable, site_code=None): """ Get a value from configuration. Retrieves the value corresponding to the given variable from the configuration module currently in use by the app. Specify a site_code value to check for a site-specific override. Arguments: variable (str): The name of a variable from the configuration module. Keyword Arguments: site_code (str): The SITE_OVERRIDES key to inspect for site-specific values Returns: The value corresponding to the variable, or None if the variable is not found. """ name = os.environ.get(CONFIGURATION_MODULE) # __import__ performs a full import, but only returns the top-level # package, not the targeted module. sys.modules is a dictionary # mapping module names to loaded modules. __import__(name) module = sys.modules[name] # Locate the setting in the specified module, then attempt to apply a site-specific override
python
{ "resource": "" }
q273766
get_overrides_filename
test
def get_overrides_filename(variable): """ Get the name of the file containing configuration overrides from the provided environment variable.
python
{ "resource": "" }
q273767
get_value_by_version
test
def get_value_by_version(d): """ Finds the value depending in current eplus version. Parameters ---------- d: dict {(0, 0): value, (x, x): value, ...} for current version (cv), current value is the value of version v such as v <= cv < v+1 """ from oplus import
python
{ "resource": "" }
q273768
_Conf.eplus_version
test
def eplus_version(self): """ if _eplus_version is defined => _eplus_version else most recent eplus available version """ # check energy plus is installed if len(self.eplus_available_versions) == 0: raise RuntimeError("Energy plus is not install, can't use oplus package.") # see if
python
{ "resource": "" }
q273769
Simulation._file_refs
test
def _file_refs(self): """ Defined here so that we can use the class variables, in order to subclass in oplusplus """ if self._prepared_file_refs is None: self._prepared_file_refs = { FILE_REFS.idf: FileInfo( constructor=lambda path: self._epm_cls.from_idf(path, idd_or_buffer_or_path=self._idd), get_path=lambda: get_input_file_path(self.dir_path, FILE_REFS.idf) ), FILE_REFS.epw: FileInfo( constructor=lambda path: self._weather_data_cls.from_epw(path), get_path=lambda: get_input_file_path(self.dir_path, FILE_REFS.epw) ), FILE_REFS.eio: FileInfo( constructor=lambda path: self._eio_cls(path), get_path=lambda: get_output_file_path(self.dir_path, FILE_REFS.eio) ), FILE_REFS.eso: FileInfo( constructor=lambda path: self._standard_output_cls(path), get_path=lambda: get_output_file_path( self.dir_path, FILE_REFS.eso ) ), FILE_REFS.mtr: FileInfo( constructor=lambda path: self._standard_output_cls(path), get_path=lambda: get_output_file_path(self.dir_path, FILE_REFS.mtr) ), FILE_REFS.mtd: FileInfo( constructor=lambda path: self._mtd_cls(path), get_path=lambda: get_output_file_path(self.dir_path, FILE_REFS.mtd)
python
{ "resource": "" }
q273770
Epm._dev_populate_from_json_data
test
def _dev_populate_from_json_data(self, json_data): """ !! Must only be called once, when empty !! """ # workflow # -------- # (methods belonging to create/update/delete framework: # epm._dev_populate_from_json_data, table.batch_add, record.update, queryset.delete, record.delete) # 1. add inert # * data is checked # * old links are unregistered # * record is stored in table (=> pk uniqueness is checked) # 2. activate: hooks, links, external files # manage comment if any comment = json_data.pop("_comment", None)
python
{ "resource": "" }
q273771
Epm.get_external_files
test
def get_external_files(self): """ An external file manages file paths. """ external_files = [] for table in self._tables.values(): for r in table:
python
{ "resource": "" }
q273772
Epm.set_defaults
test
def set_defaults(self): """ All fields of Epm with a default value and that are null will be set to their default value. """
python
{ "resource": "" }
q273773
TableDescriptor.prepare_extensible
test
def prepare_extensible(self): """ This function finishes initialization, must be called once all field descriptors and tag have been filled. """ # see if extensible and store cycle len for k in self._tags: if "extensible" in k: cycle_len = int(k.split(":")[1]) break else: # not extensible return # find cycle start and prepare patterns cycle_start = None cycle_patterns = [] for i, field_descriptor in enumerate(self._field_descriptors): # quit if finished if (cycle_start is not None) and (i >= (cycle_start + cycle_len)): break # set cycle start if not set yet if (cycle_start is None) and ("begin-extensible" in field_descriptor.tags): cycle_start = i # leave if cycle start not reached yet if cycle_start is None: continue
python
{ "resource": "" }
q273774
TableDescriptor.get_extended_name
test
def get_extended_name(self, index): """ manages extensible names """ field_descriptor = self.get_field_descriptor(index) if self.extensible_info is None: return field_descriptor.name cycle_start, cycle_len, _ = self.extensible_info
python
{ "resource": "" }
q273775
ExternalFilesManager.short_refs
test
def short_refs(self): """ we calculate on the fly to avoid managing registrations and un-registrations Returns ------- {ref: short_ref, ... """ naive_short_refs_d = dict() # naive_short_ref: {refs, ...} for ef in self._external_files: if ef.naive_short_ref not in naive_short_refs_d: naive_short_refs_d[ef.naive_short_ref] = set() naive_short_refs_d[ef.naive_short_ref].add(ef.ref) short_refs = dict() for naive_short_ref, refs in naive_short_refs_d.items(): if len(refs) == 1:
python
{ "resource": "" }
q273776
EioTable.get_value
test
def get_value(self, column_name_or_i, filter_column_name_or_i, filter_criterion): """ Returns first occurrence of value of filter column matching filter criterion. """ # find column indexes column_i = self._get_column_index(column_name_or_i) filter_column_i = self._get_column_index(filter_column_name_or_i) filter_fct = { float: lambda x: float(x) == filter_criterion, int: lambda x: int(x) == filter_criterion, str: lambda x: x.lower() == filter_criterion.lower()
python
{ "resource": "" }
q273777
Record._update_value_inert
test
def _update_value_inert(self, index, value): """ is only called by _update_inert """ # get field descriptor field_descriptor = self._table._dev_descriptor.get_field_descriptor(index) # prepare value value = field_descriptor.deserialize(value, index) # unregister previous link if relevant if isinstance(value, Link): # de-activate current link if any current_link = self._data.get(index) if current_link is not None: current_link.unregister() # unregister previous hook if relevant if isinstance(value, RecordHook): current_record_hook = self._data.get(index) if current_record_hook is not None: current_record_hook.unregister() # unregister previous external file if relevant if isinstance(value, ExternalFile): current_external_file = self._data.get(index) if current_external_file is not None: current_external_file._dev_unregister() # if None remove and leave if value in (None, NONE_RECORD_HOOK, NONE_LINK, NONE_EXTERNAL_FILE): # we don't check required, because this method
python
{ "resource": "" }
q273778
Record.update
test
def update(self, data=None, **or_data): """ Updates simultaneously all given fields. Parameters ---------- data: dictionary containing field lowercase names or index as keys, and field values as values (dict syntax) or_data: keyword arguments containing field names as keys (kwargs syntax) """ # workflow # -------- # (methods belonging to create/update/delete framework: # epm._dev_populate_from_json_data, table.batch_add, record.update, queryset.delete, record.delete) # 1. add inert # * data is checked # * old links are
python
{ "resource": "" }
q273779
Record.set_defaults
test
def set_defaults(self): """ sets all empty fields for which a default value is defined to default value """ defaults = {} for i in range(len(self)): if i in self._data: continue
python
{ "resource": "" }
q273780
Record.add_fields
test
def add_fields(self, *args): """ This method only works for extensible fields. It allows to add values without precising their fields' names or indexes. Parameters ---------- args: field values """ if not self.is_extensible():
python
{ "resource": "" }
q273781
Record.pop
test
def pop(self, index=None): """ This method only works for extensible fields. It allows to remove a value and shift all other values to fill the gap. Parameters ---------- index: int, default None index of field to remove. Returns ------- serialize value of popped field """ # prepare index (will check for extensible) index = self._prepare_pop_insert_index(index=index) # get extensible info cycle_start, cycle_len, patterns = self.get_extensible_info()
python
{ "resource": "" }
q273782
Record.insert
test
def insert(self, index, value): """ This method only works for extensible fields. It allows to insert a value, and shifts all other following values. Parameters ---------- index: position of insertion value: value to insert """ # prepare index (will check for
python
{ "resource": "" }
q273783
Record.delete
test
def delete(self): """ Deletes record, and removes it from database. """ # workflow # -------- # (methods belonging to create/update/delete framework: # epm._dev_populate_from_json_data, table.batch_add, record.update, queryset.delete, record.delete) # 1. unregister: links, hooks and external files # 3. remove from table without unregistering # unregister links self._unregister_links() # unregister hooks self._unregister_hooks() # unregister
python
{ "resource": "" }
q273784
RelationsManager.register_record_hook
test
def register_record_hook(self, hook): """ target record must have been set """ for key in hook.keys: if key in self._record_hooks: field_descriptor = hook.target_record.get_field_descriptor(hook.target_index) raise FieldValidationError(
python
{ "resource": "" }
q273785
RelationsManager.register_link
test
def register_link(self, link): """ source record and index must have been set """ keys = tuple((ref, link.initial_hook_value) for ref in link.hook_references) # look for a record hook for k in keys: if k in self._record_hooks: # set link target link.set_target(target_record=self._record_hooks[k].target_record) break else: # look for a table hook for k in keys: if k in self._table_hooks: # set link target
python
{ "resource": "" }
q273786
IntentContainer._create_regex
test
def _create_regex(self, line, intent_name): """ Create regex and return. If error occurs returns None. """ try: return re.compile(self._create_intent_pattern(line, intent_name), re.IGNORECASE)
python
{ "resource": "" }
q273787
BaseEvent.remaining_duration
test
def remaining_duration(self, time): '''Returns the remaining duration for a recording. '''
python
{ "resource": "" }
q273788
BaseEvent.serialize
test
def serialize(self): '''Serialize this object as dictionary usable for conversion to JSON. :return: Dictionary representing this object. ''' return { 'type': 'event', 'id': self.uid, 'attributes': { 'start': self.start, 'end': self.end,
python
{ "resource": "" }
q273789
http_request
test
def http_request(url, post_data=None): '''Make an HTTP request to a given URL with optional parameters. ''' logger.debug('Requesting URL: %s' % url) buf = bio() curl = pycurl.Curl() curl.setopt(curl.URL, url.encode('ascii', 'ignore')) # Disable HTTPS verification methods if insecure is set if config()['server']['insecure']: curl.setopt(curl.SSL_VERIFYPEER, 0) curl.setopt(curl.SSL_VERIFYHOST, 0) if config()['server']['certificate']: # Make sure verification methods are turned on curl.setopt(curl.SSL_VERIFYPEER, 1) curl.setopt(curl.SSL_VERIFYHOST, 2) # Import your certificates curl.setopt(pycurl.CAINFO, config()['server']['certificate']) if post_data: curl.setopt(curl.HTTPPOST, post_data) curl.setopt(curl.WRITEFUNCTION, buf.write)
python
{ "resource": "" }
q273790
get_service
test
def get_service(service_type): '''Get available service endpoints for a given service type from the Opencast ServiceRegistry. ''' endpoint = '/services/available.json?serviceType=' + str(service_type) url = '%s%s' % (config()['server']['url'], endpoint) response = http_request(url).decode('utf-8') services = (json.loads(response).get('services') or {}).get('service', []) services
python
{ "resource": "" }
q273791
try_mkdir
test
def try_mkdir(directory): '''Try to create a directory. Pass without error if it already exists. ''' try:
python
{ "resource": "" }
q273792
configure_service
test
def configure_service(service): '''Get the location of a given service from Opencast and add it to the current configuration. ''' while not config().get('service-' + service) and not terminate(): try: config()['service-' + service] = \
python
{ "resource": "" }
q273793
register_ca
test
def register_ca(status='idle'): '''Register this capture agent at the Matterhorn admin server so that it shows up in the admin interface. :param address: Address of the capture agent web ui :param status: Current status of the capture agent ''' # If this is a backup CA we don't tell the Matterhorn core that we are # here. We will just run silently in the background: if config()['agent']['backup_mode']: return params = [('address', config()['ui']['url']), ('state', status)] name
python
{ "resource": "" }
q273794
recording_state
test
def recording_state(recording_id, status): '''Send the state of the current recording to the Matterhorn core. :param recording_id: ID of the current recording :param status: Status of the recording ''' # If this is a backup CA we do not update the recording state since the # actual CA does that and we want to interfere. We will just run silently # in the background: if config()['agent']['backup_mode']:
python
{ "resource": "" }
q273795
update_event_status
test
def update_event_status(event, status): '''Update the status of a particular event in the database. ''' dbs = db.get_session() dbs.query(db.RecordedEvent).filter(db.RecordedEvent.start == event.start)\
python
{ "resource": "" }
q273796
update_agent_state
test
def update_agent_state(): '''Update the current agent state in opencast. ''' configure_service('capture.admin') status = 'idle' # Determine reported agent state with priority list if get_service_status(db.Service.SCHEDULE) == db.ServiceStatus.STOPPED: status = 'offline' elif
python
{ "resource": "" }
q273797
configuration_file
test
def configuration_file(cfgfile): '''Find the best match for the configuration file. ''' if cfgfile is not None: return cfgfile # If no file is explicitely specified, probe for the configuration file
python
{ "resource": "" }
q273798
update_configuration
test
def update_configuration(cfgfile=None): '''Update configuration from file. :param cfgfile: Configuration file to load. ''' configobj.DEFAULT_INTERPOLATION = 'template' cfgfile = configuration_file(cfgfile) cfg = configobj.ConfigObj(cfgfile, configspec=cfgspec, encoding='utf-8') validator = Validator() val = cfg.validate(validator) if val is not True: raise ValueError('Invalid configuration: %s' % val) if len(cfg['capture']['files']) != len(cfg['capture']['flavors']): raise ValueError('List of files and flavors
python
{ "resource": "" }
q273799
check
test
def check(): '''Check configuration for sanity. ''' if config('server')['insecure']: logger.warning('HTTPS CHECKS ARE TURNED OFF. A SECURE CONNECTION IS ' 'NOT GUARANTEED') if config('server')['certificate']: # Ensure certificate exists
python
{ "resource": "" }