repo
stringlengths
7
55
path
stringlengths
4
223
func_name
stringlengths
1
134
original_string
stringlengths
75
104k
language
stringclasses
1 value
code
stringlengths
75
104k
code_tokens
listlengths
19
28.4k
docstring
stringlengths
1
46.9k
docstring_tokens
listlengths
1
1.97k
sha
stringlengths
40
40
url
stringlengths
87
315
partition
stringclasses
1 value
dmwm/DBS
Server/Python/src/dbs/business/DBSAcquisitionEra.py
DBSAcquisitionEra.listAcquisitionEras_CI
def listAcquisitionEras_CI(self, acq=''): """ Returns all acquistion eras in dbs """ try: acq = str(acq) except: dbsExceptionHandler('dbsException-invalid-input', 'aquistion_era_name given is not valid : %s'%acq) conn = self.dbi.connection() try: result = self.acqlst_ci.execute(conn, acq) return result finally: if conn:conn.close()
python
def listAcquisitionEras_CI(self, acq=''): """ Returns all acquistion eras in dbs """ try: acq = str(acq) except: dbsExceptionHandler('dbsException-invalid-input', 'aquistion_era_name given is not valid : %s'%acq) conn = self.dbi.connection() try: result = self.acqlst_ci.execute(conn, acq) return result finally: if conn:conn.close()
[ "def", "listAcquisitionEras_CI", "(", "self", ",", "acq", "=", "''", ")", ":", "try", ":", "acq", "=", "str", "(", "acq", ")", "except", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "'aquistion_era_name given is not valid : %s'", "%", "acq", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "result", "=", "self", ".", "acqlst_ci", ".", "execute", "(", "conn", ",", "acq", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Returns all acquistion eras in dbs
[ "Returns", "all", "acquistion", "eras", "in", "dbs" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSAcquisitionEra.py#L42-L55
train
dmwm/DBS
Server/Python/src/dbs/business/DBSAcquisitionEra.py
DBSAcquisitionEra.insertAcquisitionEra
def insertAcquisitionEra(self, businput): """ Input dictionary has to have the following keys: acquisition_era_name, creation_date, create_by, start_date, end_date. it builds the correct dictionary for dao input and executes the dao """ conn = self.dbi.connection() tran = conn.begin() try: businput["acquisition_era_id"] = self.sm.increment(conn, "SEQ_AQE", tran) businput["acquisition_era_name"] = businput["acquisition_era_name"] #self.logger.warning(businput) self.acqin.execute(conn, businput, tran) tran.commit() tran = None except KeyError as ke: dbsExceptionHandler('dbsException-invalid-input', "Invalid input:"+ke.args[0]) except Exception as ex: if str(ex).lower().find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1: dbsExceptionHandler('dbsException-invalid-input2', "Invalid input: acquisition_era_name already exists in DB", serverError="%s" %ex) else: raise finally: if tran: tran.rollback() if conn: conn.close()
python
def insertAcquisitionEra(self, businput): """ Input dictionary has to have the following keys: acquisition_era_name, creation_date, create_by, start_date, end_date. it builds the correct dictionary for dao input and executes the dao """ conn = self.dbi.connection() tran = conn.begin() try: businput["acquisition_era_id"] = self.sm.increment(conn, "SEQ_AQE", tran) businput["acquisition_era_name"] = businput["acquisition_era_name"] #self.logger.warning(businput) self.acqin.execute(conn, businput, tran) tran.commit() tran = None except KeyError as ke: dbsExceptionHandler('dbsException-invalid-input', "Invalid input:"+ke.args[0]) except Exception as ex: if str(ex).lower().find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1: dbsExceptionHandler('dbsException-invalid-input2', "Invalid input: acquisition_era_name already exists in DB", serverError="%s" %ex) else: raise finally: if tran: tran.rollback() if conn: conn.close()
[ "def", "insertAcquisitionEra", "(", "self", ",", "businput", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "conn", ".", "begin", "(", ")", "try", ":", "businput", "[", "\"acquisition_era_id\"", "]", "=", "self", ".", "sm", ".", "increment", "(", "conn", ",", "\"SEQ_AQE\"", ",", "tran", ")", "businput", "[", "\"acquisition_era_name\"", "]", "=", "businput", "[", "\"acquisition_era_name\"", "]", "#self.logger.warning(businput)", "self", ".", "acqin", ".", "execute", "(", "conn", ",", "businput", ",", "tran", ")", "tran", ".", "commit", "(", ")", "tran", "=", "None", "except", "KeyError", "as", "ke", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Invalid input:\"", "+", "ke", ".", "args", "[", "0", "]", ")", "except", "Exception", "as", "ex", ":", "if", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"unique constraint\"", ")", "!=", "-", "1", "or", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"duplicate\"", ")", "!=", "-", "1", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "\"Invalid input: acquisition_era_name already exists in DB\"", ",", "serverError", "=", "\"%s\"", "%", "ex", ")", "else", ":", "raise", "finally", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Input dictionary has to have the following keys: acquisition_era_name, creation_date, create_by, start_date, end_date. it builds the correct dictionary for dao input and executes the dao
[ "Input", "dictionary", "has", "to", "have", "the", "following", "keys", ":", "acquisition_era_name", "creation_date", "create_by", "start_date", "end_date", ".", "it", "builds", "the", "correct", "dictionary", "for", "dao", "input", "and", "executes", "the", "dao" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSAcquisitionEra.py#L57-L83
train
dmwm/DBS
Server/Python/src/dbs/business/DBSAcquisitionEra.py
DBSAcquisitionEra.UpdateAcqEraEndDate
def UpdateAcqEraEndDate(self, acquisition_era_name ="", end_date=0): """ Input dictionary has to have the following keys: acquisition_era_name, end_date. """ if acquisition_era_name =="" or end_date==0: dbsExceptionHandler('dbsException-invalid-input', "acquisition_era_name and end_date are required") conn = self.dbi.connection() tran = conn.begin() try: self.acqud.execute(conn, acquisition_era_name, end_date, tran) if tran:tran.commit() tran = None finally: if tran:tran.rollback() if conn:conn.close()
python
def UpdateAcqEraEndDate(self, acquisition_era_name ="", end_date=0): """ Input dictionary has to have the following keys: acquisition_era_name, end_date. """ if acquisition_era_name =="" or end_date==0: dbsExceptionHandler('dbsException-invalid-input', "acquisition_era_name and end_date are required") conn = self.dbi.connection() tran = conn.begin() try: self.acqud.execute(conn, acquisition_era_name, end_date, tran) if tran:tran.commit() tran = None finally: if tran:tran.rollback() if conn:conn.close()
[ "def", "UpdateAcqEraEndDate", "(", "self", ",", "acquisition_era_name", "=", "\"\"", ",", "end_date", "=", "0", ")", ":", "if", "acquisition_era_name", "==", "\"\"", "or", "end_date", "==", "0", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"acquisition_era_name and end_date are required\"", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "conn", ".", "begin", "(", ")", "try", ":", "self", ".", "acqud", ".", "execute", "(", "conn", ",", "acquisition_era_name", ",", "end_date", ",", "tran", ")", "if", "tran", ":", "tran", ".", "commit", "(", ")", "tran", "=", "None", "finally", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Input dictionary has to have the following keys: acquisition_era_name, end_date.
[ "Input", "dictionary", "has", "to", "have", "the", "following", "keys", ":", "acquisition_era_name", "end_date", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSAcquisitionEra.py#L85-L100
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/OutputModuleConfig/GetID.py
GetID.execute
def execute(self, conn, app="", release_version="", pset_hash="", output_label="", global_tag='', transaction = False): """ returns id for a given application """ sql = self.sql binds = {} setAnd=False if not app == "": sql += " A.APP_NAME=:app_name" binds["app_name"]=app setAnd=True if not release_version == "": if setAnd : sql += " AND " sql += " R.RELEASE_VERSION=:release_version" binds["release_version"]=release_version setAnd=True if not pset_hash == "": if setAnd : sql += " AND " sql += " P.PSET_HASH=:pset_hash" binds["pset_hash"]=pset_hash setAnd=True if not output_label == "": if setAnd : sql += " AND " sql += " O.OUTPUT_MODULE_LABEL=:output_module_label" binds["output_module_label"]=output_label setAnd=True if not global_tag == "": if setAnd : sql += " AND " sql += " O.GLOBAL_TAG=:global_tag" binds["global_tag"]=global_tag if app == release_version == pset_hash == global_tag == "": dbsExceptionHandler('dbsException-invalid-input', "%s Either app_name, release_version, pset_hash or global_tag must be provided", self.logger.exception) result = self.dbi.processData(sql, binds, conn, transaction) plist = self.formatDict(result) if len(plist) < 1: return -1 return plist[0]["output_mod_config_id"]
python
def execute(self, conn, app="", release_version="", pset_hash="", output_label="", global_tag='', transaction = False): """ returns id for a given application """ sql = self.sql binds = {} setAnd=False if not app == "": sql += " A.APP_NAME=:app_name" binds["app_name"]=app setAnd=True if not release_version == "": if setAnd : sql += " AND " sql += " R.RELEASE_VERSION=:release_version" binds["release_version"]=release_version setAnd=True if not pset_hash == "": if setAnd : sql += " AND " sql += " P.PSET_HASH=:pset_hash" binds["pset_hash"]=pset_hash setAnd=True if not output_label == "": if setAnd : sql += " AND " sql += " O.OUTPUT_MODULE_LABEL=:output_module_label" binds["output_module_label"]=output_label setAnd=True if not global_tag == "": if setAnd : sql += " AND " sql += " O.GLOBAL_TAG=:global_tag" binds["global_tag"]=global_tag if app == release_version == pset_hash == global_tag == "": dbsExceptionHandler('dbsException-invalid-input', "%s Either app_name, release_version, pset_hash or global_tag must be provided", self.logger.exception) result = self.dbi.processData(sql, binds, conn, transaction) plist = self.formatDict(result) if len(plist) < 1: return -1 return plist[0]["output_mod_config_id"]
[ "def", "execute", "(", "self", ",", "conn", ",", "app", "=", "\"\"", ",", "release_version", "=", "\"\"", ",", "pset_hash", "=", "\"\"", ",", "output_label", "=", "\"\"", ",", "global_tag", "=", "''", ",", "transaction", "=", "False", ")", ":", "sql", "=", "self", ".", "sql", "binds", "=", "{", "}", "setAnd", "=", "False", "if", "not", "app", "==", "\"\"", ":", "sql", "+=", "\" A.APP_NAME=:app_name\"", "binds", "[", "\"app_name\"", "]", "=", "app", "setAnd", "=", "True", "if", "not", "release_version", "==", "\"\"", ":", "if", "setAnd", ":", "sql", "+=", "\" AND \"", "sql", "+=", "\" R.RELEASE_VERSION=:release_version\"", "binds", "[", "\"release_version\"", "]", "=", "release_version", "setAnd", "=", "True", "if", "not", "pset_hash", "==", "\"\"", ":", "if", "setAnd", ":", "sql", "+=", "\" AND \"", "sql", "+=", "\" P.PSET_HASH=:pset_hash\"", "binds", "[", "\"pset_hash\"", "]", "=", "pset_hash", "setAnd", "=", "True", "if", "not", "output_label", "==", "\"\"", ":", "if", "setAnd", ":", "sql", "+=", "\" AND \"", "sql", "+=", "\" O.OUTPUT_MODULE_LABEL=:output_module_label\"", "binds", "[", "\"output_module_label\"", "]", "=", "output_label", "setAnd", "=", "True", "if", "not", "global_tag", "==", "\"\"", ":", "if", "setAnd", ":", "sql", "+=", "\" AND \"", "sql", "+=", "\" O.GLOBAL_TAG=:global_tag\"", "binds", "[", "\"global_tag\"", "]", "=", "global_tag", "if", "app", "==", "release_version", "==", "pset_hash", "==", "global_tag", "==", "\"\"", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"%s Either app_name, release_version, pset_hash or global_tag must be provided\"", ",", "self", ".", "logger", ".", "exception", ")", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ")", "plist", "=", "self", ".", "formatDict", "(", "result", ")", "if", "len", "(", "plist", ")", "<", "1", ":", "return", "-", "1", "return", "plist", "[", "0", "]", "[", "\"output_mod_config_id\"", "]" ]
returns id for a given application
[ "returns", "id", "for", "a", "given", "application" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/OutputModuleConfig/GetID.py#L31-L68
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.prepareDatasetMigrationList
def prepareDatasetMigrationList(self, conn, request): """ Prepare the ordered lists of blocks based on input DATASET (note Block is different) 1. Get list of blocks from source 2. Check and see if these blocks are already at DST 3. Check if dataset has parents 4. Check if parent blocks are already at DST """ ordered_dict = {} order_counter = 0 srcdataset = request["migration_input"] url = request["migration_url"] try: tmp_ordered_dict = self.processDatasetBlocks(url, conn, srcdataset, order_counter) if tmp_ordered_dict != {}: ordered_dict.update(tmp_ordered_dict) self.logger.debug("ordered_dict length at level %s" %order_counter) self.logger.debug(len(ordered_dict)) else: #return {} m = 'Requested dataset %s is already in destination' %srcdataset dbsExceptionHandler('dbsException-invalid-input2', message=m, serverError=m) # Now process the parent datasets parent_ordered_dict = self.getParentDatasetsOrderedList(url, conn, srcdataset, order_counter+1) if parent_ordered_dict != {}: ordered_dict.update(parent_ordered_dict) self.logger.debug("***** parent ordered_dict length at level %s ******" %(order_counter+1)) self.logger.debug(len(ordered_dict)) return remove_duplicated_items(ordered_dict) except dbsException: raise except Exception as ex: if 'urlopen error' in str(ex): message='Connection to source DBS server refused. Check your source url.' elif 'Bad Request' in str(ex): message='cannot get data from the source DBS server. Check your migration input.' else: message='Failed to make a dataset migration list.' dbsExceptionHandler('dbsException-invalid-input2', \ serverError="""DBSMigrate/prepareDatasetMigrationList failed to prepare ordered block list: %s""" %str(ex), message=message)
python
def prepareDatasetMigrationList(self, conn, request): """ Prepare the ordered lists of blocks based on input DATASET (note Block is different) 1. Get list of blocks from source 2. Check and see if these blocks are already at DST 3. Check if dataset has parents 4. Check if parent blocks are already at DST """ ordered_dict = {} order_counter = 0 srcdataset = request["migration_input"] url = request["migration_url"] try: tmp_ordered_dict = self.processDatasetBlocks(url, conn, srcdataset, order_counter) if tmp_ordered_dict != {}: ordered_dict.update(tmp_ordered_dict) self.logger.debug("ordered_dict length at level %s" %order_counter) self.logger.debug(len(ordered_dict)) else: #return {} m = 'Requested dataset %s is already in destination' %srcdataset dbsExceptionHandler('dbsException-invalid-input2', message=m, serverError=m) # Now process the parent datasets parent_ordered_dict = self.getParentDatasetsOrderedList(url, conn, srcdataset, order_counter+1) if parent_ordered_dict != {}: ordered_dict.update(parent_ordered_dict) self.logger.debug("***** parent ordered_dict length at level %s ******" %(order_counter+1)) self.logger.debug(len(ordered_dict)) return remove_duplicated_items(ordered_dict) except dbsException: raise except Exception as ex: if 'urlopen error' in str(ex): message='Connection to source DBS server refused. Check your source url.' elif 'Bad Request' in str(ex): message='cannot get data from the source DBS server. Check your migration input.' else: message='Failed to make a dataset migration list.' dbsExceptionHandler('dbsException-invalid-input2', \ serverError="""DBSMigrate/prepareDatasetMigrationList failed to prepare ordered block list: %s""" %str(ex), message=message)
[ "def", "prepareDatasetMigrationList", "(", "self", ",", "conn", ",", "request", ")", ":", "ordered_dict", "=", "{", "}", "order_counter", "=", "0", "srcdataset", "=", "request", "[", "\"migration_input\"", "]", "url", "=", "request", "[", "\"migration_url\"", "]", "try", ":", "tmp_ordered_dict", "=", "self", ".", "processDatasetBlocks", "(", "url", ",", "conn", ",", "srcdataset", ",", "order_counter", ")", "if", "tmp_ordered_dict", "!=", "{", "}", ":", "ordered_dict", ".", "update", "(", "tmp_ordered_dict", ")", "self", ".", "logger", ".", "debug", "(", "\"ordered_dict length at level %s\"", "%", "order_counter", ")", "self", ".", "logger", ".", "debug", "(", "len", "(", "ordered_dict", ")", ")", "else", ":", "#return {}", "m", "=", "'Requested dataset %s is already in destination'", "%", "srcdataset", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "message", "=", "m", ",", "serverError", "=", "m", ")", "# Now process the parent datasets", "parent_ordered_dict", "=", "self", ".", "getParentDatasetsOrderedList", "(", "url", ",", "conn", ",", "srcdataset", ",", "order_counter", "+", "1", ")", "if", "parent_ordered_dict", "!=", "{", "}", ":", "ordered_dict", ".", "update", "(", "parent_ordered_dict", ")", "self", ".", "logger", ".", "debug", "(", "\"***** parent ordered_dict length at level %s ******\"", "%", "(", "order_counter", "+", "1", ")", ")", "self", ".", "logger", ".", "debug", "(", "len", "(", "ordered_dict", ")", ")", "return", "remove_duplicated_items", "(", "ordered_dict", ")", "except", "dbsException", ":", "raise", "except", "Exception", "as", "ex", ":", "if", "'urlopen error'", "in", "str", "(", "ex", ")", ":", "message", "=", "'Connection to source DBS server refused. Check your source url.'", "elif", "'Bad Request'", "in", "str", "(", "ex", ")", ":", "message", "=", "'cannot get data from the source DBS server. Check your migration input.'", "else", ":", "message", "=", "'Failed to make a dataset migration list.'", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "serverError", "=", "\"\"\"DBSMigrate/prepareDatasetMigrationList failed\n to prepare ordered block list: %s\"\"\"", "%", "str", "(", "ex", ")", ",", "message", "=", "message", ")" ]
Prepare the ordered lists of blocks based on input DATASET (note Block is different) 1. Get list of blocks from source 2. Check and see if these blocks are already at DST 3. Check if dataset has parents 4. Check if parent blocks are already at DST
[ "Prepare", "the", "ordered", "lists", "of", "blocks", "based", "on", "input", "DATASET", "(", "note", "Block", "is", "different", ")", "1", ".", "Get", "list", "of", "blocks", "from", "source", "2", ".", "Check", "and", "see", "if", "these", "blocks", "are", "already", "at", "DST", "3", ".", "Check", "if", "dataset", "has", "parents", "4", ".", "Check", "if", "parent", "blocks", "are", "already", "at", "DST" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L77-L120
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.processDatasetBlocks
def processDatasetBlocks(self, url, conn, inputdataset, order_counter): """ Utility function, that comapares blocks of a dataset at source and dst and returns an ordered list of blocks not already at dst for migration """ ordered_dict = {} srcblks = self.getSrcBlocks(url, dataset=inputdataset) if len(srcblks) < 0: e = "DBSMigration: No blocks in the required dataset %s found at source %s."%(inputdataset, url) dbsExceptionHandler('dbsException-invalid-input2', e, self.logger.exception, e) dstblks = self.blocklist.execute(conn, dataset=inputdataset) self.logger.debug("******* dstblks for dataset %s ***********" %inputdataset) self.logger.debug(dstblks) blocksInSrcNames = [ y['block_name'] for y in srcblks] blocksInDstNames = [] for item in dstblks: blocksInDstNames.append(item['block_name']) ordered_dict[order_counter] = [] for ablk in blocksInSrcNames: if not ablk in blocksInDstNames: ordered_dict[order_counter].append(ablk) if ordered_dict[order_counter] != []: self.logger.debug("**** ordered_dict dict length ****") self.logger.debug(len(ordered_dict)) return ordered_dict else: return {}
python
def processDatasetBlocks(self, url, conn, inputdataset, order_counter): """ Utility function, that comapares blocks of a dataset at source and dst and returns an ordered list of blocks not already at dst for migration """ ordered_dict = {} srcblks = self.getSrcBlocks(url, dataset=inputdataset) if len(srcblks) < 0: e = "DBSMigration: No blocks in the required dataset %s found at source %s."%(inputdataset, url) dbsExceptionHandler('dbsException-invalid-input2', e, self.logger.exception, e) dstblks = self.blocklist.execute(conn, dataset=inputdataset) self.logger.debug("******* dstblks for dataset %s ***********" %inputdataset) self.logger.debug(dstblks) blocksInSrcNames = [ y['block_name'] for y in srcblks] blocksInDstNames = [] for item in dstblks: blocksInDstNames.append(item['block_name']) ordered_dict[order_counter] = [] for ablk in blocksInSrcNames: if not ablk in blocksInDstNames: ordered_dict[order_counter].append(ablk) if ordered_dict[order_counter] != []: self.logger.debug("**** ordered_dict dict length ****") self.logger.debug(len(ordered_dict)) return ordered_dict else: return {}
[ "def", "processDatasetBlocks", "(", "self", ",", "url", ",", "conn", ",", "inputdataset", ",", "order_counter", ")", ":", "ordered_dict", "=", "{", "}", "srcblks", "=", "self", ".", "getSrcBlocks", "(", "url", ",", "dataset", "=", "inputdataset", ")", "if", "len", "(", "srcblks", ")", "<", "0", ":", "e", "=", "\"DBSMigration: No blocks in the required dataset %s found at source %s.\"", "%", "(", "inputdataset", ",", "url", ")", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "e", ",", "self", ".", "logger", ".", "exception", ",", "e", ")", "dstblks", "=", "self", ".", "blocklist", ".", "execute", "(", "conn", ",", "dataset", "=", "inputdataset", ")", "self", ".", "logger", ".", "debug", "(", "\"******* dstblks for dataset %s ***********\"", "%", "inputdataset", ")", "self", ".", "logger", ".", "debug", "(", "dstblks", ")", "blocksInSrcNames", "=", "[", "y", "[", "'block_name'", "]", "for", "y", "in", "srcblks", "]", "blocksInDstNames", "=", "[", "]", "for", "item", "in", "dstblks", ":", "blocksInDstNames", ".", "append", "(", "item", "[", "'block_name'", "]", ")", "ordered_dict", "[", "order_counter", "]", "=", "[", "]", "for", "ablk", "in", "blocksInSrcNames", ":", "if", "not", "ablk", "in", "blocksInDstNames", ":", "ordered_dict", "[", "order_counter", "]", ".", "append", "(", "ablk", ")", "if", "ordered_dict", "[", "order_counter", "]", "!=", "[", "]", ":", "self", ".", "logger", ".", "debug", "(", "\"**** ordered_dict dict length ****\"", ")", "self", ".", "logger", ".", "debug", "(", "len", "(", "ordered_dict", ")", ")", "return", "ordered_dict", "else", ":", "return", "{", "}" ]
Utility function, that comapares blocks of a dataset at source and dst and returns an ordered list of blocks not already at dst for migration
[ "Utility", "function", "that", "comapares", "blocks", "of", "a", "dataset", "at", "source", "and", "dst", "and", "returns", "an", "ordered", "list", "of", "blocks", "not", "already", "at", "dst", "for", "migration" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L122-L148
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.getParentDatasetsOrderedList
def getParentDatasetsOrderedList(self, url, conn, dataset, order_counter): """ check if input dataset has parents, check if any of the blocks are already at dst, prepare the ordered list and return it. url : source DBS url dataset : to be migrated dataset order_counter: the order in which migration happends. """ ordered_dict = {} parentSrcDatasets = self.getSrcDatasetParents(url, dataset) if len(parentSrcDatasets) > 0: parentSrcDatasetNames = [y['parent_dataset'] for y in parentSrcDatasets] for aparentDataset in parentSrcDatasetNames: parent_ordered_dict = self.processDatasetBlocks(url, conn, aparentDataset, order_counter) self.logger.debug("************ dict length of parent blocks for the parent dataset %s at level %s" %(aparentDataset, order_counter)) self.logger.debug(len(parent_ordered_dict)) if parent_ordered_dict != {}: ordered_dict.update(parent_ordered_dict) self.logger.debug("**** ordered_dict length ****") self.logger.debug(len(ordered_dict)) # parents of parent pparent_ordered_dict = self.getParentDatasetsOrderedList(url, conn, aparentDataset, order_counter+1) self.logger.debug("************dict length parent parent blocks for the parent dataset %s at level %s" %(aparentDataset, order_counter+1)) self.logger.debug(len(pparent_ordered_dict)) if pparent_ordered_dict != {}: ordered_dict.update(pparent_ordered_dict) self.logger.debug("**** ordered_dict length ****") self.logger.debug(len(ordered_dict)) return ordered_dict
python
def getParentDatasetsOrderedList(self, url, conn, dataset, order_counter): """ check if input dataset has parents, check if any of the blocks are already at dst, prepare the ordered list and return it. url : source DBS url dataset : to be migrated dataset order_counter: the order in which migration happends. """ ordered_dict = {} parentSrcDatasets = self.getSrcDatasetParents(url, dataset) if len(parentSrcDatasets) > 0: parentSrcDatasetNames = [y['parent_dataset'] for y in parentSrcDatasets] for aparentDataset in parentSrcDatasetNames: parent_ordered_dict = self.processDatasetBlocks(url, conn, aparentDataset, order_counter) self.logger.debug("************ dict length of parent blocks for the parent dataset %s at level %s" %(aparentDataset, order_counter)) self.logger.debug(len(parent_ordered_dict)) if parent_ordered_dict != {}: ordered_dict.update(parent_ordered_dict) self.logger.debug("**** ordered_dict length ****") self.logger.debug(len(ordered_dict)) # parents of parent pparent_ordered_dict = self.getParentDatasetsOrderedList(url, conn, aparentDataset, order_counter+1) self.logger.debug("************dict length parent parent blocks for the parent dataset %s at level %s" %(aparentDataset, order_counter+1)) self.logger.debug(len(pparent_ordered_dict)) if pparent_ordered_dict != {}: ordered_dict.update(pparent_ordered_dict) self.logger.debug("**** ordered_dict length ****") self.logger.debug(len(ordered_dict)) return ordered_dict
[ "def", "getParentDatasetsOrderedList", "(", "self", ",", "url", ",", "conn", ",", "dataset", ",", "order_counter", ")", ":", "ordered_dict", "=", "{", "}", "parentSrcDatasets", "=", "self", ".", "getSrcDatasetParents", "(", "url", ",", "dataset", ")", "if", "len", "(", "parentSrcDatasets", ")", ">", "0", ":", "parentSrcDatasetNames", "=", "[", "y", "[", "'parent_dataset'", "]", "for", "y", "in", "parentSrcDatasets", "]", "for", "aparentDataset", "in", "parentSrcDatasetNames", ":", "parent_ordered_dict", "=", "self", ".", "processDatasetBlocks", "(", "url", ",", "conn", ",", "aparentDataset", ",", "order_counter", ")", "self", ".", "logger", ".", "debug", "(", "\"************ dict length of parent blocks for the parent dataset %s at level %s\"", "%", "(", "aparentDataset", ",", "order_counter", ")", ")", "self", ".", "logger", ".", "debug", "(", "len", "(", "parent_ordered_dict", ")", ")", "if", "parent_ordered_dict", "!=", "{", "}", ":", "ordered_dict", ".", "update", "(", "parent_ordered_dict", ")", "self", ".", "logger", ".", "debug", "(", "\"**** ordered_dict length ****\"", ")", "self", ".", "logger", ".", "debug", "(", "len", "(", "ordered_dict", ")", ")", "# parents of parent", "pparent_ordered_dict", "=", "self", ".", "getParentDatasetsOrderedList", "(", "url", ",", "conn", ",", "aparentDataset", ",", "order_counter", "+", "1", ")", "self", ".", "logger", ".", "debug", "(", "\"************dict length parent parent blocks for the parent dataset %s at level %s\"", "%", "(", "aparentDataset", ",", "order_counter", "+", "1", ")", ")", "self", ".", "logger", ".", "debug", "(", "len", "(", "pparent_ordered_dict", ")", ")", "if", "pparent_ordered_dict", "!=", "{", "}", ":", "ordered_dict", ".", "update", "(", "pparent_ordered_dict", ")", "self", ".", "logger", ".", "debug", "(", "\"**** ordered_dict length ****\"", ")", "self", ".", "logger", ".", "debug", "(", "len", "(", "ordered_dict", ")", ")", "return", "ordered_dict" ]
check if input dataset has parents, check if any of the blocks are already at dst, prepare the ordered list and return it. url : source DBS url dataset : to be migrated dataset order_counter: the order in which migration happends.
[ "check", "if", "input", "dataset", "has", "parents", "check", "if", "any", "of", "the", "blocks", "are", "already", "at", "dst", "prepare", "the", "ordered", "list", "and", "return", "it", ".", "url", ":", "source", "DBS", "url", "dataset", ":", "to", "be", "migrated", "dataset", "order_counter", ":", "the", "order", "in", "which", "migration", "happends", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L150-L182
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.prepareBlockMigrationList
def prepareBlockMigrationList(self, conn, request): """ Prepare the ordered lists of blocks based on input BLOCK 1. see if block already exists at dst (no need to migrate), raise "ALREADY EXISTS" 2. see if block exists at src & make sure the block's open_for_writing=0 3. see if block has parents 4. see if parent blocks are already at dst 5. add 'order' to parent and then this block (ascending) 6. return the ordered list """ ordered_dict = {} block_name = request["migration_input"] url = request["migration_url"] order_counter = 0 try: #1. dstblock = self.blocklist.execute(conn, block_name=block_name) for item in dstblock: if item: dbsExceptionHandler('dbsException-invalid-input', 'ALREADY EXISTS: \ Required block (%s) migration is already at destination' %block_name, self.logger.exception) #2. srcblock = self.getSrcBlocks(url, block=block_name) if len(srcblock) < 1: e = 'DBSMigration: Invalid input. Required Block %s not found at source %s.' %(block_name, url) dbsExceptionHandler('dbsException-invalid-input2', e, self.logger.exception, e) ##This block has to be migrated ordered_dict[order_counter] = [] ordered_dict[order_counter].append(block_name) parent_ordered_dict = self.getParentBlocksOrderedList(url, conn, block_name, order_counter+1) if parent_ordered_dict != {}: ordered_dict.update(parent_ordered_dict) #6. #check for duplicates return remove_duplicated_items(ordered_dict) except Exception as ex: if '500 Internal Server Error' in str(ex): #"Server Error" is the default in dbsExceptionHandler dbsExceptionHandler('Server Error', str(ex), self.logger.exception, "DBSMigrate/prepareBlockMigrationList: "+str(ex)) if isinstance(ex, pycurl.error): if ex.args[0] == 7: message = ex.args[1] dbsExceptionHandler('dbsException-failed-connect2host', message, self.logger.exception, message) if 'urlopen error' in str(ex): message='Connection to source DBS server refused. Check your source url.' elif 'Bad Request' in str(ex): message='cannot get data from the source DBS server. Check your migration input.' else: message='Failed to make a block migration list.' dbsExceptionHandler('dbsException-invalid-input2', \ """DBSMigrate/prepareBlockMigrationList failed to prepare ordered block list: %s""" %str(ex), self.logger.exception, message)
python
def prepareBlockMigrationList(self, conn, request): """ Prepare the ordered lists of blocks based on input BLOCK 1. see if block already exists at dst (no need to migrate), raise "ALREADY EXISTS" 2. see if block exists at src & make sure the block's open_for_writing=0 3. see if block has parents 4. see if parent blocks are already at dst 5. add 'order' to parent and then this block (ascending) 6. return the ordered list """ ordered_dict = {} block_name = request["migration_input"] url = request["migration_url"] order_counter = 0 try: #1. dstblock = self.blocklist.execute(conn, block_name=block_name) for item in dstblock: if item: dbsExceptionHandler('dbsException-invalid-input', 'ALREADY EXISTS: \ Required block (%s) migration is already at destination' %block_name, self.logger.exception) #2. srcblock = self.getSrcBlocks(url, block=block_name) if len(srcblock) < 1: e = 'DBSMigration: Invalid input. Required Block %s not found at source %s.' %(block_name, url) dbsExceptionHandler('dbsException-invalid-input2', e, self.logger.exception, e) ##This block has to be migrated ordered_dict[order_counter] = [] ordered_dict[order_counter].append(block_name) parent_ordered_dict = self.getParentBlocksOrderedList(url, conn, block_name, order_counter+1) if parent_ordered_dict != {}: ordered_dict.update(parent_ordered_dict) #6. #check for duplicates return remove_duplicated_items(ordered_dict) except Exception as ex: if '500 Internal Server Error' in str(ex): #"Server Error" is the default in dbsExceptionHandler dbsExceptionHandler('Server Error', str(ex), self.logger.exception, "DBSMigrate/prepareBlockMigrationList: "+str(ex)) if isinstance(ex, pycurl.error): if ex.args[0] == 7: message = ex.args[1] dbsExceptionHandler('dbsException-failed-connect2host', message, self.logger.exception, message) if 'urlopen error' in str(ex): message='Connection to source DBS server refused. Check your source url.' elif 'Bad Request' in str(ex): message='cannot get data from the source DBS server. Check your migration input.' else: message='Failed to make a block migration list.' dbsExceptionHandler('dbsException-invalid-input2', \ """DBSMigrate/prepareBlockMigrationList failed to prepare ordered block list: %s""" %str(ex), self.logger.exception, message)
[ "def", "prepareBlockMigrationList", "(", "self", ",", "conn", ",", "request", ")", ":", "ordered_dict", "=", "{", "}", "block_name", "=", "request", "[", "\"migration_input\"", "]", "url", "=", "request", "[", "\"migration_url\"", "]", "order_counter", "=", "0", "try", ":", "#1.", "dstblock", "=", "self", ".", "blocklist", ".", "execute", "(", "conn", ",", "block_name", "=", "block_name", ")", "for", "item", "in", "dstblock", ":", "if", "item", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "'ALREADY EXISTS: \\\n\t\t\tRequired block (%s) migration is already at destination'", "%", "block_name", ",", "self", ".", "logger", ".", "exception", ")", "#2.", "srcblock", "=", "self", ".", "getSrcBlocks", "(", "url", ",", "block", "=", "block_name", ")", "if", "len", "(", "srcblock", ")", "<", "1", ":", "e", "=", "'DBSMigration: Invalid input. Required Block %s not found at source %s.'", "%", "(", "block_name", ",", "url", ")", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "e", ",", "self", ".", "logger", ".", "exception", ",", "e", ")", "##This block has to be migrated", "ordered_dict", "[", "order_counter", "]", "=", "[", "]", "ordered_dict", "[", "order_counter", "]", ".", "append", "(", "block_name", ")", "parent_ordered_dict", "=", "self", ".", "getParentBlocksOrderedList", "(", "url", ",", "conn", ",", "block_name", ",", "order_counter", "+", "1", ")", "if", "parent_ordered_dict", "!=", "{", "}", ":", "ordered_dict", ".", "update", "(", "parent_ordered_dict", ")", "#6.", "#check for duplicates", "return", "remove_duplicated_items", "(", "ordered_dict", ")", "except", "Exception", "as", "ex", ":", "if", "'500 Internal Server Error'", "in", "str", "(", "ex", ")", ":", "#\"Server Error\" is the default in dbsExceptionHandler", "dbsExceptionHandler", "(", "'Server Error'", ",", "str", "(", "ex", ")", ",", "self", ".", "logger", ".", "exception", ",", "\"DBSMigrate/prepareBlockMigrationList: \"", "+", "str", "(", "ex", ")", ")", "if", "isinstance", "(", "ex", ",", "pycurl", ".", "error", ")", ":", "if", "ex", ".", "args", "[", "0", "]", "==", "7", ":", "message", "=", "ex", ".", "args", "[", "1", "]", "dbsExceptionHandler", "(", "'dbsException-failed-connect2host'", ",", "message", ",", "self", ".", "logger", ".", "exception", ",", "message", ")", "if", "'urlopen error'", "in", "str", "(", "ex", ")", ":", "message", "=", "'Connection to source DBS server refused. Check your source url.'", "elif", "'Bad Request'", "in", "str", "(", "ex", ")", ":", "message", "=", "'cannot get data from the source DBS server. Check your migration input.'", "else", ":", "message", "=", "'Failed to make a block migration list.'", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "\"\"\"DBSMigrate/prepareBlockMigrationList failed\n to prepare ordered block list: %s\"\"\"", "%", "str", "(", "ex", ")", ",", "self", ".", "logger", ".", "exception", ",", "message", ")" ]
Prepare the ordered lists of blocks based on input BLOCK 1. see if block already exists at dst (no need to migrate), raise "ALREADY EXISTS" 2. see if block exists at src & make sure the block's open_for_writing=0 3. see if block has parents 4. see if parent blocks are already at dst 5. add 'order' to parent and then this block (ascending) 6. return the ordered list
[ "Prepare", "the", "ordered", "lists", "of", "blocks", "based", "on", "input", "BLOCK", "1", ".", "see", "if", "block", "already", "exists", "at", "dst", "(", "no", "need", "to", "migrate", ")", "raise", "ALREADY", "EXISTS", "2", ".", "see", "if", "block", "exists", "at", "src", "&", "make", "sure", "the", "block", "s", "open_for_writing", "=", "0", "3", ".", "see", "if", "block", "has", "parents", "4", ".", "see", "if", "parent", "blocks", "are", "already", "at", "dst", "5", ".", "add", "order", "to", "parent", "and", "then", "this", "block", "(", "ascending", ")", "6", ".", "return", "the", "ordered", "list" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L184-L238
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.removeMigrationRequest
def removeMigrationRequest(self, migration_rqst): """ Method to remove pending or failed migration request from the queue. """ conn = self.dbi.connection() try: tran = conn.begin() self.mgrremove.execute(conn, migration_rqst) tran.commit() except dbsException as he: if conn: conn.close() raise except Exception as ex: if conn: conn.close() raise if conn: conn.close()
python
def removeMigrationRequest(self, migration_rqst): """ Method to remove pending or failed migration request from the queue. """ conn = self.dbi.connection() try: tran = conn.begin() self.mgrremove.execute(conn, migration_rqst) tran.commit() except dbsException as he: if conn: conn.close() raise except Exception as ex: if conn: conn.close() raise if conn: conn.close()
[ "def", "removeMigrationRequest", "(", "self", ",", "migration_rqst", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "tran", "=", "conn", ".", "begin", "(", ")", "self", ".", "mgrremove", ".", "execute", "(", "conn", ",", "migration_rqst", ")", "tran", ".", "commit", "(", ")", "except", "dbsException", "as", "he", ":", "if", "conn", ":", "conn", ".", "close", "(", ")", "raise", "except", "Exception", "as", "ex", ":", "if", "conn", ":", "conn", ".", "close", "(", ")", "raise", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Method to remove pending or failed migration request from the queue.
[ "Method", "to", "remove", "pending", "or", "failed", "migration", "request", "from", "the", "queue", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L285-L301
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.insertMigrationRequest
def insertMigrationRequest(self, request): """ Method to insert use requests to MIGRATION_REQUESTS table. request keys: migration_url, migration_input """ conn = self.dbi.connection() # check if already queued. #If the migration_input is the same, but the src url is different, #We will consider it as a submitted request. YG 05-18-2012 try: alreadyqueued = self.mgrlist.execute(conn, migration_input=request["migration_input"]) is_already_queued = len(alreadyqueued) > 0 # close connection before returning json object if is_already_queued and conn: conn.close() #if the queued is not failed, then we don't need to do it again. #add a new migration_status=9 (terminal failure) if is_already_queued and alreadyqueued[0]['migration_status'] == 2: return {"migration_report" : "REQUEST ALREADY QUEUED. Migration is finished", "migration_details" : alreadyqueued[0] } elif is_already_queued and alreadyqueued[0]['migration_status'] != 9: return {"migration_report" : "REQUEST ALREADY QUEUED. Migration in progress", "migration_details" : alreadyqueued[0] } elif is_already_queued and alreadyqueued[0]['migration_status'] == 9: return {"migration_report" : "REQUEST ALREADY QUEUED. Migration terminally failed. ", "migration_details" : alreadyqueued[0] } else: # not already queued #Determine if its a dataset or block migration #The prepare list calls will check if the requested blocks/dataset already in destination. if request["migration_input"].find("#") != -1: ordered_list = self.prepareBlockMigrationList(conn, request) else: ordered_list = self.prepareDatasetMigrationList(conn, request) # now we have the blocks that need to be queued (ordered) except Exception as ex: if conn: conn.close() raise tran = conn.begin() try: # insert the request #request.update(migration_status=0) request['migration_request_id'] = self.sm.increment(conn, "SEQ_MR", tran) self.mgrin.execute(conn, request, tran) # INSERT the ordered_list totalQueued = 0 k = ordered_list.keys() k.sort() k.reverse() self.logger.debug("****************** ordered_list keys **********") self.logger.debug(k) #for iter in reversed(range(len(ordered_list))): for iter in k: self.logger.debug("length for Key: %s" %iter) self.logger.debug(len(ordered_list[iter])) if len(ordered_list[iter]) > 0: daoinput = [{ "migration_block_id" : self.sm.increment(conn, "SEQ_MB", tran), "migration_request_id" : request["migration_request_id"], "migration_block_name" : blk, "migration_order" : iter, "migration_status" : 0, "creation_date" : request['creation_date'], "last_modification_date" : request['last_modification_date'], "create_by" : request['create_by'], "last_modified_by" : request['last_modified_by'] } for blk in ordered_list[iter]] self.mgrblkin.execute(conn, daoinput, tran) totalQueued += len(ordered_list[iter]) # all good ?, commit the transaction tran.commit() if conn: conn.close() # return things like (X blocks queued for migration) return { "migration_report" : "REQUEST QUEUED with total %d blocks to be migrated" %totalQueued, "migration_details" : request } except SQLAlchemyIntegrityError as ex: e = "DBSMigration: ENQUEUEING_FAILED1 from SQLAichemy Integrity Error. Reason may be (%s)" %(ex.statement + "; " + str(ex.params) + "; " + str(ex.orig)) self.logger.debug(e) import traceback tk = traceback.format_exc() self.logger.debug(tk) tran.rollback() if conn: conn.close() if (str(ex).find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1): #FIXME: Need to check which unique key. YG 2/11/13 #The unique constraints are: MIGRATION_REQUESTS(MIGRATION_INPUT) #MIGRATION_BLOCKS(MIGRATION_BLOCK_NAME, MIGRATION_REQUEST_ID) return { "migration_report" : "REQUEST ALREADY QUEUED", "migration_details" : request } else: if conn: conn.close() self.logger.error(tk) m = "DBSMigration: ENQUEUEING_FAILED1." dbsExceptionHandler('dbsException-invalid-input2', m, self.logger.exception, e) except HTTPError as he: raise he except Exception as ex: import traceback self.logger.error(traceback.format_exc()) if tran: tran.rollback() if conn: conn.close() m = "DBSMigration: ENQUEUEING_FAILED." e = "DBSMigration: ENQUEUEING_FAILED. General exception caught: Reason may be (%s)" %str(ex) dbsExceptionHandler('dbsException-invalid-input2', m, self.logger.exception, e) finally: if conn: conn.close()
python
def insertMigrationRequest(self, request): """ Method to insert use requests to MIGRATION_REQUESTS table. request keys: migration_url, migration_input """ conn = self.dbi.connection() # check if already queued. #If the migration_input is the same, but the src url is different, #We will consider it as a submitted request. YG 05-18-2012 try: alreadyqueued = self.mgrlist.execute(conn, migration_input=request["migration_input"]) is_already_queued = len(alreadyqueued) > 0 # close connection before returning json object if is_already_queued and conn: conn.close() #if the queued is not failed, then we don't need to do it again. #add a new migration_status=9 (terminal failure) if is_already_queued and alreadyqueued[0]['migration_status'] == 2: return {"migration_report" : "REQUEST ALREADY QUEUED. Migration is finished", "migration_details" : alreadyqueued[0] } elif is_already_queued and alreadyqueued[0]['migration_status'] != 9: return {"migration_report" : "REQUEST ALREADY QUEUED. Migration in progress", "migration_details" : alreadyqueued[0] } elif is_already_queued and alreadyqueued[0]['migration_status'] == 9: return {"migration_report" : "REQUEST ALREADY QUEUED. Migration terminally failed. ", "migration_details" : alreadyqueued[0] } else: # not already queued #Determine if its a dataset or block migration #The prepare list calls will check if the requested blocks/dataset already in destination. if request["migration_input"].find("#") != -1: ordered_list = self.prepareBlockMigrationList(conn, request) else: ordered_list = self.prepareDatasetMigrationList(conn, request) # now we have the blocks that need to be queued (ordered) except Exception as ex: if conn: conn.close() raise tran = conn.begin() try: # insert the request #request.update(migration_status=0) request['migration_request_id'] = self.sm.increment(conn, "SEQ_MR", tran) self.mgrin.execute(conn, request, tran) # INSERT the ordered_list totalQueued = 0 k = ordered_list.keys() k.sort() k.reverse() self.logger.debug("****************** ordered_list keys **********") self.logger.debug(k) #for iter in reversed(range(len(ordered_list))): for iter in k: self.logger.debug("length for Key: %s" %iter) self.logger.debug(len(ordered_list[iter])) if len(ordered_list[iter]) > 0: daoinput = [{ "migration_block_id" : self.sm.increment(conn, "SEQ_MB", tran), "migration_request_id" : request["migration_request_id"], "migration_block_name" : blk, "migration_order" : iter, "migration_status" : 0, "creation_date" : request['creation_date'], "last_modification_date" : request['last_modification_date'], "create_by" : request['create_by'], "last_modified_by" : request['last_modified_by'] } for blk in ordered_list[iter]] self.mgrblkin.execute(conn, daoinput, tran) totalQueued += len(ordered_list[iter]) # all good ?, commit the transaction tran.commit() if conn: conn.close() # return things like (X blocks queued for migration) return { "migration_report" : "REQUEST QUEUED with total %d blocks to be migrated" %totalQueued, "migration_details" : request } except SQLAlchemyIntegrityError as ex: e = "DBSMigration: ENQUEUEING_FAILED1 from SQLAichemy Integrity Error. Reason may be (%s)" %(ex.statement + "; " + str(ex.params) + "; " + str(ex.orig)) self.logger.debug(e) import traceback tk = traceback.format_exc() self.logger.debug(tk) tran.rollback() if conn: conn.close() if (str(ex).find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1): #FIXME: Need to check which unique key. YG 2/11/13 #The unique constraints are: MIGRATION_REQUESTS(MIGRATION_INPUT) #MIGRATION_BLOCKS(MIGRATION_BLOCK_NAME, MIGRATION_REQUEST_ID) return { "migration_report" : "REQUEST ALREADY QUEUED", "migration_details" : request } else: if conn: conn.close() self.logger.error(tk) m = "DBSMigration: ENQUEUEING_FAILED1." dbsExceptionHandler('dbsException-invalid-input2', m, self.logger.exception, e) except HTTPError as he: raise he except Exception as ex: import traceback self.logger.error(traceback.format_exc()) if tran: tran.rollback() if conn: conn.close() m = "DBSMigration: ENQUEUEING_FAILED." e = "DBSMigration: ENQUEUEING_FAILED. General exception caught: Reason may be (%s)" %str(ex) dbsExceptionHandler('dbsException-invalid-input2', m, self.logger.exception, e) finally: if conn: conn.close()
[ "def", "insertMigrationRequest", "(", "self", ",", "request", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "# check if already queued.", "#If the migration_input is the same, but the src url is different,", "#We will consider it as a submitted request. YG 05-18-2012", "try", ":", "alreadyqueued", "=", "self", ".", "mgrlist", ".", "execute", "(", "conn", ",", "migration_input", "=", "request", "[", "\"migration_input\"", "]", ")", "is_already_queued", "=", "len", "(", "alreadyqueued", ")", ">", "0", "# close connection before returning json object", "if", "is_already_queued", "and", "conn", ":", "conn", ".", "close", "(", ")", "#if the queued is not failed, then we don't need to do it again.", "#add a new migration_status=9 (terminal failure)", "if", "is_already_queued", "and", "alreadyqueued", "[", "0", "]", "[", "'migration_status'", "]", "==", "2", ":", "return", "{", "\"migration_report\"", ":", "\"REQUEST ALREADY QUEUED. Migration is finished\"", ",", "\"migration_details\"", ":", "alreadyqueued", "[", "0", "]", "}", "elif", "is_already_queued", "and", "alreadyqueued", "[", "0", "]", "[", "'migration_status'", "]", "!=", "9", ":", "return", "{", "\"migration_report\"", ":", "\"REQUEST ALREADY QUEUED. Migration in progress\"", ",", "\"migration_details\"", ":", "alreadyqueued", "[", "0", "]", "}", "elif", "is_already_queued", "and", "alreadyqueued", "[", "0", "]", "[", "'migration_status'", "]", "==", "9", ":", "return", "{", "\"migration_report\"", ":", "\"REQUEST ALREADY QUEUED. Migration terminally failed. \"", ",", "\"migration_details\"", ":", "alreadyqueued", "[", "0", "]", "}", "else", ":", "# not already queued", "#Determine if its a dataset or block migration", "#The prepare list calls will check if the requested blocks/dataset already in destination.", "if", "request", "[", "\"migration_input\"", "]", ".", "find", "(", "\"#\"", ")", "!=", "-", "1", ":", "ordered_list", "=", "self", ".", "prepareBlockMigrationList", "(", "conn", ",", "request", ")", "else", ":", "ordered_list", "=", "self", ".", "prepareDatasetMigrationList", "(", "conn", ",", "request", ")", "# now we have the blocks that need to be queued (ordered)", "except", "Exception", "as", "ex", ":", "if", "conn", ":", "conn", ".", "close", "(", ")", "raise", "tran", "=", "conn", ".", "begin", "(", ")", "try", ":", "# insert the request", "#request.update(migration_status=0)", "request", "[", "'migration_request_id'", "]", "=", "self", ".", "sm", ".", "increment", "(", "conn", ",", "\"SEQ_MR\"", ",", "tran", ")", "self", ".", "mgrin", ".", "execute", "(", "conn", ",", "request", ",", "tran", ")", "# INSERT the ordered_list", "totalQueued", "=", "0", "k", "=", "ordered_list", ".", "keys", "(", ")", "k", ".", "sort", "(", ")", "k", ".", "reverse", "(", ")", "self", ".", "logger", ".", "debug", "(", "\"****************** ordered_list keys **********\"", ")", "self", ".", "logger", ".", "debug", "(", "k", ")", "#for iter in reversed(range(len(ordered_list))):", "for", "iter", "in", "k", ":", "self", ".", "logger", ".", "debug", "(", "\"length for Key: %s\"", "%", "iter", ")", "self", ".", "logger", ".", "debug", "(", "len", "(", "ordered_list", "[", "iter", "]", ")", ")", "if", "len", "(", "ordered_list", "[", "iter", "]", ")", ">", "0", ":", "daoinput", "=", "[", "{", "\"migration_block_id\"", ":", "self", ".", "sm", ".", "increment", "(", "conn", ",", "\"SEQ_MB\"", ",", "tran", ")", ",", "\"migration_request_id\"", ":", "request", "[", "\"migration_request_id\"", "]", ",", "\"migration_block_name\"", ":", "blk", ",", "\"migration_order\"", ":", "iter", ",", "\"migration_status\"", ":", "0", ",", "\"creation_date\"", ":", "request", "[", "'creation_date'", "]", ",", "\"last_modification_date\"", ":", "request", "[", "'last_modification_date'", "]", ",", "\"create_by\"", ":", "request", "[", "'create_by'", "]", ",", "\"last_modified_by\"", ":", "request", "[", "'last_modified_by'", "]", "}", "for", "blk", "in", "ordered_list", "[", "iter", "]", "]", "self", ".", "mgrblkin", ".", "execute", "(", "conn", ",", "daoinput", ",", "tran", ")", "totalQueued", "+=", "len", "(", "ordered_list", "[", "iter", "]", ")", "# all good ?, commit the transaction", "tran", ".", "commit", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")", "# return things like (X blocks queued for migration)", "return", "{", "\"migration_report\"", ":", "\"REQUEST QUEUED with total %d blocks to be migrated\"", "%", "totalQueued", ",", "\"migration_details\"", ":", "request", "}", "except", "SQLAlchemyIntegrityError", "as", "ex", ":", "e", "=", "\"DBSMigration: ENQUEUEING_FAILED1 from SQLAichemy Integrity Error. Reason may be (%s)\"", "%", "(", "ex", ".", "statement", "+", "\"; \"", "+", "str", "(", "ex", ".", "params", ")", "+", "\"; \"", "+", "str", "(", "ex", ".", "orig", ")", ")", "self", ".", "logger", ".", "debug", "(", "e", ")", "import", "traceback", "tk", "=", "traceback", ".", "format_exc", "(", ")", "self", ".", "logger", ".", "debug", "(", "tk", ")", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")", "if", "(", "str", "(", "ex", ")", ".", "find", "(", "\"unique constraint\"", ")", "!=", "-", "1", "or", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"duplicate\"", ")", "!=", "-", "1", ")", ":", "#FIXME: Need to check which unique key. YG 2/11/13", "#The unique constraints are: MIGRATION_REQUESTS(MIGRATION_INPUT)", "#MIGRATION_BLOCKS(MIGRATION_BLOCK_NAME, MIGRATION_REQUEST_ID)", "return", "{", "\"migration_report\"", ":", "\"REQUEST ALREADY QUEUED\"", ",", "\"migration_details\"", ":", "request", "}", "else", ":", "if", "conn", ":", "conn", ".", "close", "(", ")", "self", ".", "logger", ".", "error", "(", "tk", ")", "m", "=", "\"DBSMigration: ENQUEUEING_FAILED1.\"", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "m", ",", "self", ".", "logger", ".", "exception", ",", "e", ")", "except", "HTTPError", "as", "he", ":", "raise", "he", "except", "Exception", "as", "ex", ":", "import", "traceback", "self", ".", "logger", ".", "error", "(", "traceback", ".", "format_exc", "(", ")", ")", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")", "m", "=", "\"DBSMigration: ENQUEUEING_FAILED.\"", "e", "=", "\"DBSMigration: ENQUEUEING_FAILED. General exception caught: Reason may be (%s)\"", "%", "str", "(", "ex", ")", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "m", ",", "self", ".", "logger", ".", "exception", ",", "e", ")", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Method to insert use requests to MIGRATION_REQUESTS table. request keys: migration_url, migration_input
[ "Method", "to", "insert", "use", "requests", "to", "MIGRATION_REQUESTS", "table", ".", "request", "keys", ":", "migration_url", "migration_input" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L305-L418
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.listMigrationRequests
def listMigrationRequests(self, migration_request_id="", block_name="", dataset="", user="", oldest=False): """ get the status of the migration migratee : can be dataset or block_name """ conn = self.dbi.connection() migratee = "" try: if block_name: migratee = block_name elif dataset: migratee = dataset result = self.mgrlist.execute(conn, migration_url="", migration_input=migratee, create_by=user, migration_request_id=migration_request_id, oldest=oldest) return result finally: if conn: conn.close()
python
def listMigrationRequests(self, migration_request_id="", block_name="", dataset="", user="", oldest=False): """ get the status of the migration migratee : can be dataset or block_name """ conn = self.dbi.connection() migratee = "" try: if block_name: migratee = block_name elif dataset: migratee = dataset result = self.mgrlist.execute(conn, migration_url="", migration_input=migratee, create_by=user, migration_request_id=migration_request_id, oldest=oldest) return result finally: if conn: conn.close()
[ "def", "listMigrationRequests", "(", "self", ",", "migration_request_id", "=", "\"\"", ",", "block_name", "=", "\"\"", ",", "dataset", "=", "\"\"", ",", "user", "=", "\"\"", ",", "oldest", "=", "False", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "migratee", "=", "\"\"", "try", ":", "if", "block_name", ":", "migratee", "=", "block_name", "elif", "dataset", ":", "migratee", "=", "dataset", "result", "=", "self", ".", "mgrlist", ".", "execute", "(", "conn", ",", "migration_url", "=", "\"\"", ",", "migration_input", "=", "migratee", ",", "create_by", "=", "user", ",", "migration_request_id", "=", "migration_request_id", ",", "oldest", "=", "oldest", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
get the status of the migration migratee : can be dataset or block_name
[ "get", "the", "status", "of", "the", "migration", "migratee", ":", "can", "be", "dataset", "or", "block_name" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L420-L440
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.listMigrationBlocks
def listMigrationBlocks(self, migration_request_id=""): """ get eveything of block that is has status = 0 and migration_request_id as specified. """ conn = self.dbi.connection() try: return self.mgrblklist.execute(conn, migration_request_id=migration_request_id) finally: if conn: conn.close()
python
def listMigrationBlocks(self, migration_request_id=""): """ get eveything of block that is has status = 0 and migration_request_id as specified. """ conn = self.dbi.connection() try: return self.mgrblklist.execute(conn, migration_request_id=migration_request_id) finally: if conn: conn.close()
[ "def", "listMigrationBlocks", "(", "self", ",", "migration_request_id", "=", "\"\"", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "return", "self", ".", "mgrblklist", ".", "execute", "(", "conn", ",", "migration_request_id", "=", "migration_request_id", ")", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
get eveything of block that is has status = 0 and migration_request_id as specified.
[ "get", "eveything", "of", "block", "that", "is", "has", "status", "=", "0", "and", "migration_request_id", "as", "specified", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L442-L451
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.updateMigrationRequestStatus
def updateMigrationRequestStatus(self, migration_status, migration_request_id): """ migration_status: 0=PENDING 1=IN PROGRESS 2=COMPLETED 3=FAILED (will be retried) 9=Terminally FAILED status change: 0 -> 1 1 -> 2 1 -> 3 1 -> 9 are only allowed changes for working through migration. 3 -> 1 is allowed for retrying and retry count +1. """ conn = self.dbi.connection() tran = conn.begin() try: upst = dict(migration_status=migration_status, migration_request_id=migration_request_id, last_modification_date=dbsUtils().getTime()) self.mgrRqUp.execute(conn, upst) except: if tran:tran.rollback() raise else: if tran:tran.commit() finally: #open transaction is committed when conn closed. if conn:conn.close()
python
def updateMigrationRequestStatus(self, migration_status, migration_request_id): """ migration_status: 0=PENDING 1=IN PROGRESS 2=COMPLETED 3=FAILED (will be retried) 9=Terminally FAILED status change: 0 -> 1 1 -> 2 1 -> 3 1 -> 9 are only allowed changes for working through migration. 3 -> 1 is allowed for retrying and retry count +1. """ conn = self.dbi.connection() tran = conn.begin() try: upst = dict(migration_status=migration_status, migration_request_id=migration_request_id, last_modification_date=dbsUtils().getTime()) self.mgrRqUp.execute(conn, upst) except: if tran:tran.rollback() raise else: if tran:tran.commit() finally: #open transaction is committed when conn closed. if conn:conn.close()
[ "def", "updateMigrationRequestStatus", "(", "self", ",", "migration_status", ",", "migration_request_id", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "conn", ".", "begin", "(", ")", "try", ":", "upst", "=", "dict", "(", "migration_status", "=", "migration_status", ",", "migration_request_id", "=", "migration_request_id", ",", "last_modification_date", "=", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ")", "self", ".", "mgrRqUp", ".", "execute", "(", "conn", ",", "upst", ")", "except", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "raise", "else", ":", "if", "tran", ":", "tran", ".", "commit", "(", ")", "finally", ":", "#open transaction is committed when conn closed.", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
migration_status: 0=PENDING 1=IN PROGRESS 2=COMPLETED 3=FAILED (will be retried) 9=Terminally FAILED status change: 0 -> 1 1 -> 2 1 -> 3 1 -> 9 are only allowed changes for working through migration. 3 -> 1 is allowed for retrying and retry count +1.
[ "migration_status", ":", "0", "=", "PENDING", "1", "=", "IN", "PROGRESS", "2", "=", "COMPLETED", "3", "=", "FAILED", "(", "will", "be", "retried", ")", "9", "=", "Terminally", "FAILED", "status", "change", ":", "0", "-", ">", "1", "1", "-", ">", "2", "1", "-", ">", "3", "1", "-", ">", "9", "are", "only", "allowed", "changes", "for", "working", "through", "migration", ".", "3", "-", ">", "1", "is", "allowed", "for", "retrying", "and", "retry", "count", "+", "1", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L454-L486
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.updateMigrationBlockStatus
def updateMigrationBlockStatus(self, migration_status=0, migration_block=None, migration_request=None): """ migration_status: 0=PENDING 1=IN PROGRESS 2=COMPLETED 3=FAILED (will be retried) 9=Terminally FAILED status change: 0 -> 1 1 -> 2 1 -> 3 1 -> 9 are only allowed changes for working through migration. 3 -> 1 allowed for retrying. """ conn = self.dbi.connection() tran = conn.begin() try: if migration_block: upst = dict(migration_status=migration_status, migration_block_id=migration_block, last_modification_date=dbsUtils().getTime()) elif migration_request: upst = dict(migration_status=migration_status, migration_request_id=migration_request, last_modification_date=dbsUtils().getTime()) self.mgrup.execute(conn, upst) except: if tran:tran.rollback() raise else: if tran:tran.commit() finally: if conn:conn.close()
python
def updateMigrationBlockStatus(self, migration_status=0, migration_block=None, migration_request=None): """ migration_status: 0=PENDING 1=IN PROGRESS 2=COMPLETED 3=FAILED (will be retried) 9=Terminally FAILED status change: 0 -> 1 1 -> 2 1 -> 3 1 -> 9 are only allowed changes for working through migration. 3 -> 1 allowed for retrying. """ conn = self.dbi.connection() tran = conn.begin() try: if migration_block: upst = dict(migration_status=migration_status, migration_block_id=migration_block, last_modification_date=dbsUtils().getTime()) elif migration_request: upst = dict(migration_status=migration_status, migration_request_id=migration_request, last_modification_date=dbsUtils().getTime()) self.mgrup.execute(conn, upst) except: if tran:tran.rollback() raise else: if tran:tran.commit() finally: if conn:conn.close()
[ "def", "updateMigrationBlockStatus", "(", "self", ",", "migration_status", "=", "0", ",", "migration_block", "=", "None", ",", "migration_request", "=", "None", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "conn", ".", "begin", "(", ")", "try", ":", "if", "migration_block", ":", "upst", "=", "dict", "(", "migration_status", "=", "migration_status", ",", "migration_block_id", "=", "migration_block", ",", "last_modification_date", "=", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ")", "elif", "migration_request", ":", "upst", "=", "dict", "(", "migration_status", "=", "migration_status", ",", "migration_request_id", "=", "migration_request", ",", "last_modification_date", "=", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ")", "self", ".", "mgrup", ".", "execute", "(", "conn", ",", "upst", ")", "except", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "raise", "else", ":", "if", "tran", ":", "tran", ".", "commit", "(", ")", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
migration_status: 0=PENDING 1=IN PROGRESS 2=COMPLETED 3=FAILED (will be retried) 9=Terminally FAILED status change: 0 -> 1 1 -> 2 1 -> 3 1 -> 9 are only allowed changes for working through migration. 3 -> 1 allowed for retrying.
[ "migration_status", ":", "0", "=", "PENDING", "1", "=", "IN", "PROGRESS", "2", "=", "COMPLETED", "3", "=", "FAILED", "(", "will", "be", "retried", ")", "9", "=", "Terminally", "FAILED", "status", "change", ":", "0", "-", ">", "1", "1", "-", ">", "2", "1", "-", ">", "3", "1", "-", ">", "9", "are", "only", "allowed", "changes", "for", "working", "through", "migration", ".", "3", "-", ">", "1", "allowed", "for", "retrying", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L492-L526
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.getSrcDatasetParents
def getSrcDatasetParents(self, url, dataset): """ List block at src DBS """ #resturl = "%s/datasetparents?dataset=%s" % (url, dataset) params={'dataset':dataset} return cjson.decode(self.callDBSService(url, 'datasetparents', params, {}))
python
def getSrcDatasetParents(self, url, dataset): """ List block at src DBS """ #resturl = "%s/datasetparents?dataset=%s" % (url, dataset) params={'dataset':dataset} return cjson.decode(self.callDBSService(url, 'datasetparents', params, {}))
[ "def", "getSrcDatasetParents", "(", "self", ",", "url", ",", "dataset", ")", ":", "#resturl = \"%s/datasetparents?dataset=%s\" % (url, dataset)", "params", "=", "{", "'dataset'", ":", "dataset", "}", "return", "cjson", ".", "decode", "(", "self", ".", "callDBSService", "(", "url", ",", "'datasetparents'", ",", "params", ",", "{", "}", ")", ")" ]
List block at src DBS
[ "List", "block", "at", "src", "DBS" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L554-L560
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.getSrcBlockParents
def getSrcBlockParents(self, url, block): """ List block at src DBS """ #blockname = block.replace("#", urllib.quote_plus('#')) #resturl = "%s/blockparents?block_name=%s" % (url, blockname) params={'block_name':block} return cjson.decode(self.callDBSService(url, 'blockparents', params, {}))
python
def getSrcBlockParents(self, url, block): """ List block at src DBS """ #blockname = block.replace("#", urllib.quote_plus('#')) #resturl = "%s/blockparents?block_name=%s" % (url, blockname) params={'block_name':block} return cjson.decode(self.callDBSService(url, 'blockparents', params, {}))
[ "def", "getSrcBlockParents", "(", "self", ",", "url", ",", "block", ")", ":", "#blockname = block.replace(\"#\", urllib.quote_plus('#'))", "#resturl = \"%s/blockparents?block_name=%s\" % (url, blockname)", "params", "=", "{", "'block_name'", ":", "block", "}", "return", "cjson", ".", "decode", "(", "self", ".", "callDBSService", "(", "url", ",", "'blockparents'", ",", "params", ",", "{", "}", ")", ")" ]
List block at src DBS
[ "List", "block", "at", "src", "DBS" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L562-L569
train
dmwm/DBS
Server/Python/src/dbs/business/DBSMigrate.py
DBSMigrate.getSrcBlocks
def getSrcBlocks(self, url, dataset="", block=""): """ Need to list all blocks of the dataset and its parents starting from the top For now just list the blocks from this dataset. Client type call... """ if block: params={'block_name':block, 'open_for_writing':0} elif dataset: params={'dataset':dataset, 'open_for_writing':0} else: m = 'DBSMigration: Invalid input. Either block or dataset name has to be provided' e = 'DBSMigrate/getSrcBlocks: Invalid input. Either block or dataset name has to be provided' dbsExceptionHandler('dbsException-invalid-input2', m, self.logger.exception, e ) return cjson.decode(self.callDBSService(url, 'blocks', params, {}))
python
def getSrcBlocks(self, url, dataset="", block=""): """ Need to list all blocks of the dataset and its parents starting from the top For now just list the blocks from this dataset. Client type call... """ if block: params={'block_name':block, 'open_for_writing':0} elif dataset: params={'dataset':dataset, 'open_for_writing':0} else: m = 'DBSMigration: Invalid input. Either block or dataset name has to be provided' e = 'DBSMigrate/getSrcBlocks: Invalid input. Either block or dataset name has to be provided' dbsExceptionHandler('dbsException-invalid-input2', m, self.logger.exception, e ) return cjson.decode(self.callDBSService(url, 'blocks', params, {}))
[ "def", "getSrcBlocks", "(", "self", ",", "url", ",", "dataset", "=", "\"\"", ",", "block", "=", "\"\"", ")", ":", "if", "block", ":", "params", "=", "{", "'block_name'", ":", "block", ",", "'open_for_writing'", ":", "0", "}", "elif", "dataset", ":", "params", "=", "{", "'dataset'", ":", "dataset", ",", "'open_for_writing'", ":", "0", "}", "else", ":", "m", "=", "'DBSMigration: Invalid input. Either block or dataset name has to be provided'", "e", "=", "'DBSMigrate/getSrcBlocks: Invalid input. Either block or dataset name has to be provided'", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "m", ",", "self", ".", "logger", ".", "exception", ",", "e", ")", "return", "cjson", ".", "decode", "(", "self", ".", "callDBSService", "(", "url", ",", "'blocks'", ",", "params", ",", "{", "}", ")", ")" ]
Need to list all blocks of the dataset and its parents starting from the top For now just list the blocks from this dataset. Client type call...
[ "Need", "to", "list", "all", "blocks", "of", "the", "dataset", "and", "its", "parents", "starting", "from", "the", "top", "For", "now", "just", "list", "the", "blocks", "from", "this", "dataset", ".", "Client", "type", "call", "..." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSMigrate.py#L571-L586
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/InsertTable/Insert.py
InsertSingle.executeSingle
def executeSingle( self, conn, daoinput, tablename, transaction = False): """build dynamic sql based on daoinput""" sql1 = " insert into %s%s( " %(self.owner, tablename) sql2 =" values(" "Now loop over all the input keys. We need to check if all the keys are valid !!!" for key in daoinput: sql1 += "%s," %key.upper() sql2 += ":%s," %key.lower() sql = sql1.strip(',') + ') ' + sql2.strip(',') + ' )' self.dbi.processData(sql, daoinput, conn, transaction)
python
def executeSingle( self, conn, daoinput, tablename, transaction = False): """build dynamic sql based on daoinput""" sql1 = " insert into %s%s( " %(self.owner, tablename) sql2 =" values(" "Now loop over all the input keys. We need to check if all the keys are valid !!!" for key in daoinput: sql1 += "%s," %key.upper() sql2 += ":%s," %key.lower() sql = sql1.strip(',') + ') ' + sql2.strip(',') + ' )' self.dbi.processData(sql, daoinput, conn, transaction)
[ "def", "executeSingle", "(", "self", ",", "conn", ",", "daoinput", ",", "tablename", ",", "transaction", "=", "False", ")", ":", "sql1", "=", "\" insert into %s%s( \"", "%", "(", "self", ".", "owner", ",", "tablename", ")", "sql2", "=", "\" values(\"", "\"Now loop over all the input keys. We need to check if all the keys are valid !!!\"", "for", "key", "in", "daoinput", ":", "sql1", "+=", "\"%s,\"", "%", "key", ".", "upper", "(", ")", "sql2", "+=", "\":%s,\"", "%", "key", ".", "lower", "(", ")", "sql", "=", "sql1", ".", "strip", "(", "','", ")", "+", "') '", "+", "sql2", ".", "strip", "(", "','", ")", "+", "' )'", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "daoinput", ",", "conn", ",", "transaction", ")" ]
build dynamic sql based on daoinput
[ "build", "dynamic", "sql", "based", "on", "daoinput" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/InsertTable/Insert.py#L16-L27
train
henriquebastos/django-decouple
decouple.py
Config._cast_boolean
def _cast_boolean(self, value): """ Helper to convert config values to boolean as ConfigParser do. """ if value.lower() not in self._BOOLEANS: raise ValueError('Not a boolean: %s' % value) return self._BOOLEANS[value.lower()]
python
def _cast_boolean(self, value): """ Helper to convert config values to boolean as ConfigParser do. """ if value.lower() not in self._BOOLEANS: raise ValueError('Not a boolean: %s' % value) return self._BOOLEANS[value.lower()]
[ "def", "_cast_boolean", "(", "self", ",", "value", ")", ":", "if", "value", ".", "lower", "(", ")", "not", "in", "self", ".", "_BOOLEANS", ":", "raise", "ValueError", "(", "'Not a boolean: %s'", "%", "value", ")", "return", "self", ".", "_BOOLEANS", "[", "value", ".", "lower", "(", ")", "]" ]
Helper to convert config values to boolean as ConfigParser do.
[ "Helper", "to", "convert", "config", "values", "to", "boolean", "as", "ConfigParser", "do", "." ]
fbc4a1b139565e1468b97de7e8c4276bcfc5b3e9
https://github.com/henriquebastos/django-decouple/blob/fbc4a1b139565e1468b97de7e8c4276bcfc5b3e9/decouple.py#L40-L47
train
henriquebastos/django-decouple
decouple.py
Config.get
def get(self, option, default=undefined, cast=undefined): """ Return the value for option or default if defined. """ if option in self.repository: value = self.repository.get(option) else: value = default if isinstance(value, Undefined): raise UndefinedValueError('%s option not found and default value was not defined.' % option) if isinstance(cast, Undefined): cast = lambda v: v # nop elif cast is bool: cast = self._cast_boolean return cast(value)
python
def get(self, option, default=undefined, cast=undefined): """ Return the value for option or default if defined. """ if option in self.repository: value = self.repository.get(option) else: value = default if isinstance(value, Undefined): raise UndefinedValueError('%s option not found and default value was not defined.' % option) if isinstance(cast, Undefined): cast = lambda v: v # nop elif cast is bool: cast = self._cast_boolean return cast(value)
[ "def", "get", "(", "self", ",", "option", ",", "default", "=", "undefined", ",", "cast", "=", "undefined", ")", ":", "if", "option", "in", "self", ".", "repository", ":", "value", "=", "self", ".", "repository", ".", "get", "(", "option", ")", "else", ":", "value", "=", "default", "if", "isinstance", "(", "value", ",", "Undefined", ")", ":", "raise", "UndefinedValueError", "(", "'%s option not found and default value was not defined.'", "%", "option", ")", "if", "isinstance", "(", "cast", ",", "Undefined", ")", ":", "cast", "=", "lambda", "v", ":", "v", "# nop", "elif", "cast", "is", "bool", ":", "cast", "=", "self", ".", "_cast_boolean", "return", "cast", "(", "value", ")" ]
Return the value for option or default if defined.
[ "Return", "the", "value", "for", "option", "or", "default", "if", "defined", "." ]
fbc4a1b139565e1468b97de7e8c4276bcfc5b3e9
https://github.com/henriquebastos/django-decouple/blob/fbc4a1b139565e1468b97de7e8c4276bcfc5b3e9/decouple.py#L49-L66
train
dmwm/DBS
Client/setup.py
parse_requirements
def parse_requirements(requirements_file): """ Create a list for the 'install_requires' component of the setup function by parsing a requirements file """ if os.path.exists(requirements_file): # return a list that contains each line of the requirements file return open(requirements_file, 'r').read().splitlines() else: print("ERROR: requirements file " + requirements_file + " not found.") sys.exit(1)
python
def parse_requirements(requirements_file): """ Create a list for the 'install_requires' component of the setup function by parsing a requirements file """ if os.path.exists(requirements_file): # return a list that contains each line of the requirements file return open(requirements_file, 'r').read().splitlines() else: print("ERROR: requirements file " + requirements_file + " not found.") sys.exit(1)
[ "def", "parse_requirements", "(", "requirements_file", ")", ":", "if", "os", ".", "path", ".", "exists", "(", "requirements_file", ")", ":", "# return a list that contains each line of the requirements file", "return", "open", "(", "requirements_file", ",", "'r'", ")", ".", "read", "(", ")", ".", "splitlines", "(", ")", "else", ":", "print", "(", "\"ERROR: requirements file \"", "+", "requirements_file", "+", "\" not found.\"", ")", "sys", ".", "exit", "(", "1", ")" ]
Create a list for the 'install_requires' component of the setup function by parsing a requirements file
[ "Create", "a", "list", "for", "the", "install_requires", "component", "of", "the", "setup", "function", "by", "parsing", "a", "requirements", "file" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Client/setup.py#L16-L27
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/Block/BriefList.py
BriefList.execute
def execute(self, conn, dataset="", block_name="", data_tier_name="", origin_site_name="", logical_file_name="", run_num=-1, min_cdate=0, max_cdate=0, min_ldate=0, max_ldate=0, cdate=0, ldate=0, open_for_writing=-1, transaction = False): """ dataset: /a/b/c block: /a/b/c#d """ binds = {} basesql = self.sql joinsql = "" wheresql = "" generatedsql = "" if logical_file_name and logical_file_name != "%": joinsql += " JOIN %sFILES FL ON FL.BLOCK_ID = B.BLOCK_ID " %(self.owner) op = ("=", "like")["%" in logical_file_name] wheresql += " WHERE LOGICAL_FILE_NAME %s :logical_file_name " % op binds.update( logical_file_name = logical_file_name ) if block_name and block_name !="%": andorwhere = ("WHERE", "AND")[bool(wheresql)] op = ("=", "like")["%" in block_name] wheresql += " %s B.BLOCK_NAME %s :block_name " % ((andorwhere, op)) binds.update( block_name = block_name ) if data_tier_name or (dataset and dataset!="%"): joinsql += "JOIN %sDATASETS DS ON DS.DATASET_ID = B.DATASET_ID " % (self.owner) andorwhere = ("WHERE", "AND")[bool(wheresql)] if dataset: op = ("=", "like")["%" in dataset] wheresql += " %s DS.DATASET %s :dataset " % ((andorwhere, op)) binds.update(dataset=dataset) if data_tier_name: joinsql += "JOIN {owner}DATA_TIERS DT ON DS.DATA_TIER_ID=DT.DATA_TIER_ID ".format(owner=self.owner) wheresql += " %s DT.DATA_TIER_NAME=:data_tier_name " % (andorwhere) binds.update(data_tier_name=data_tier_name) if origin_site_name and origin_site_name != "%": op = ("=", "like")["%" in origin_site_name] wheresql += " AND B.ORIGIN_SITE_NAME %s :origin_site_name " % op binds.update(origin_site_name = origin_site_name) if open_for_writing == 0 or open_for_writing == 1: wheresql += " AND B.OPEN_FOR_WRITTING = :open_for_writing " if cdate != 0: wheresql += "AND B.CREATION_DATE = :cdate " binds.update(cdate = cdate) elif min_cdate != 0 and max_cdate != 0: wheresql += "AND B.CREATION_DATE BETWEEN :min_cdate and :max_cdate " binds.update(min_cdate = min_cdate) binds.update(max_cdate = max_cdate) elif min_cdate != 0 and max_cdate == 0: wheresql += "AND B.CREATION_DATE > :min_cdate " binds.update(min_cdate = min_cdate) elif min_cdate ==0 and max_cdate != 0: wheresql += "AND B.CREATION_DATE < :max_cdate " binds.update(max_cdate = max_cdate) else: pass if ldate != 0: wheresql += "AND B.LAST_MODIFICATION_DATE = :ldate " binds.update(ldate = ldate) elif min_ldate != 0 and max_ldate != 0: wheresql += "AND B.LAST_MODIFICATION_DATE BETWEEN :min_ldate and :max_ldate " binds.update(min_ldate = min_ldate) binds.update(max_ldate = max_ldate) elif min_ldate != 0 and max_ldate == 0: wheresql += "AND B.LAST_MODIFICATION_DATE > :min_ldate " binds.update(min_ldate = min_ldate) elif min_cdate ==0 and max_cdate != 0: wheresql += "AND B.LAST_MODIFICATION_DATE < :max_ldate " binds.update(max_ldate = max_ldate) else: pass #one may provide a list of runs , so it has to be the last one in building the bind. if run_num !=-1 : basesql = basesql.replace("SELECT", "SELECT DISTINCT") + " , FLM.RUN_NUM " if not logical_file_name: joinsql += " JOIN %sFILES FL ON FL.BLOCK_ID = B.BLOCK_ID " %(self.owner) joinsql += " JOIN %sFILE_LUMIS FLM on FLM.FILE_ID = FL.FILE_ID " %(self.owner) run_list=[] wheresql_run_list='' wheresql_run_range='' # for r in parseRunRange(run_num): if isinstance(r, basestring) or isinstance(r, int) or isinstance(r, long): run_list.append(str(r)) if isinstance(r, run_tuple): if r[0] == r[1]: dbsExceptionHandler('dbsException-invalid-input', "DBS run_num range must be apart at least by 1.", self.logger.exception) wheresql_run_range = " FLM.RUN_NUM between :minrun and :maxrun " binds.update({"minrun":r[0]}) binds.update({"maxrun":r[1]}) # if run_list: wheresql_run_list = " FLM.RUN_NUM in (SELECT TOKEN FROM TOKEN_GENERATOR) " generatedsql, run_binds = create_token_generator(run_list) binds.update(run_binds) # if wheresql_run_range and wheresql_run_list: wheresql += " and (" + wheresql_run_range + " or " + wheresql_run_list + " )" elif wheresql_run_range and not wheresql_run_list: wheresql += " and " + wheresql_run_range elif not wheresql_run_range and wheresql_run_list: wheresql += " and " + wheresql_run_list # sql = " ".join((generatedsql, basesql, self.fromsql, joinsql, wheresql)) cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) for i in cursors: d = self.formatCursor(i, size=100) if isinstance(d, list) or isinstance(d, GeneratorType): for elem in d: yield elem elif d: yield d
python
def execute(self, conn, dataset="", block_name="", data_tier_name="", origin_site_name="", logical_file_name="", run_num=-1, min_cdate=0, max_cdate=0, min_ldate=0, max_ldate=0, cdate=0, ldate=0, open_for_writing=-1, transaction = False): """ dataset: /a/b/c block: /a/b/c#d """ binds = {} basesql = self.sql joinsql = "" wheresql = "" generatedsql = "" if logical_file_name and logical_file_name != "%": joinsql += " JOIN %sFILES FL ON FL.BLOCK_ID = B.BLOCK_ID " %(self.owner) op = ("=", "like")["%" in logical_file_name] wheresql += " WHERE LOGICAL_FILE_NAME %s :logical_file_name " % op binds.update( logical_file_name = logical_file_name ) if block_name and block_name !="%": andorwhere = ("WHERE", "AND")[bool(wheresql)] op = ("=", "like")["%" in block_name] wheresql += " %s B.BLOCK_NAME %s :block_name " % ((andorwhere, op)) binds.update( block_name = block_name ) if data_tier_name or (dataset and dataset!="%"): joinsql += "JOIN %sDATASETS DS ON DS.DATASET_ID = B.DATASET_ID " % (self.owner) andorwhere = ("WHERE", "AND")[bool(wheresql)] if dataset: op = ("=", "like")["%" in dataset] wheresql += " %s DS.DATASET %s :dataset " % ((andorwhere, op)) binds.update(dataset=dataset) if data_tier_name: joinsql += "JOIN {owner}DATA_TIERS DT ON DS.DATA_TIER_ID=DT.DATA_TIER_ID ".format(owner=self.owner) wheresql += " %s DT.DATA_TIER_NAME=:data_tier_name " % (andorwhere) binds.update(data_tier_name=data_tier_name) if origin_site_name and origin_site_name != "%": op = ("=", "like")["%" in origin_site_name] wheresql += " AND B.ORIGIN_SITE_NAME %s :origin_site_name " % op binds.update(origin_site_name = origin_site_name) if open_for_writing == 0 or open_for_writing == 1: wheresql += " AND B.OPEN_FOR_WRITTING = :open_for_writing " if cdate != 0: wheresql += "AND B.CREATION_DATE = :cdate " binds.update(cdate = cdate) elif min_cdate != 0 and max_cdate != 0: wheresql += "AND B.CREATION_DATE BETWEEN :min_cdate and :max_cdate " binds.update(min_cdate = min_cdate) binds.update(max_cdate = max_cdate) elif min_cdate != 0 and max_cdate == 0: wheresql += "AND B.CREATION_DATE > :min_cdate " binds.update(min_cdate = min_cdate) elif min_cdate ==0 and max_cdate != 0: wheresql += "AND B.CREATION_DATE < :max_cdate " binds.update(max_cdate = max_cdate) else: pass if ldate != 0: wheresql += "AND B.LAST_MODIFICATION_DATE = :ldate " binds.update(ldate = ldate) elif min_ldate != 0 and max_ldate != 0: wheresql += "AND B.LAST_MODIFICATION_DATE BETWEEN :min_ldate and :max_ldate " binds.update(min_ldate = min_ldate) binds.update(max_ldate = max_ldate) elif min_ldate != 0 and max_ldate == 0: wheresql += "AND B.LAST_MODIFICATION_DATE > :min_ldate " binds.update(min_ldate = min_ldate) elif min_cdate ==0 and max_cdate != 0: wheresql += "AND B.LAST_MODIFICATION_DATE < :max_ldate " binds.update(max_ldate = max_ldate) else: pass #one may provide a list of runs , so it has to be the last one in building the bind. if run_num !=-1 : basesql = basesql.replace("SELECT", "SELECT DISTINCT") + " , FLM.RUN_NUM " if not logical_file_name: joinsql += " JOIN %sFILES FL ON FL.BLOCK_ID = B.BLOCK_ID " %(self.owner) joinsql += " JOIN %sFILE_LUMIS FLM on FLM.FILE_ID = FL.FILE_ID " %(self.owner) run_list=[] wheresql_run_list='' wheresql_run_range='' # for r in parseRunRange(run_num): if isinstance(r, basestring) or isinstance(r, int) or isinstance(r, long): run_list.append(str(r)) if isinstance(r, run_tuple): if r[0] == r[1]: dbsExceptionHandler('dbsException-invalid-input', "DBS run_num range must be apart at least by 1.", self.logger.exception) wheresql_run_range = " FLM.RUN_NUM between :minrun and :maxrun " binds.update({"minrun":r[0]}) binds.update({"maxrun":r[1]}) # if run_list: wheresql_run_list = " FLM.RUN_NUM in (SELECT TOKEN FROM TOKEN_GENERATOR) " generatedsql, run_binds = create_token_generator(run_list) binds.update(run_binds) # if wheresql_run_range and wheresql_run_list: wheresql += " and (" + wheresql_run_range + " or " + wheresql_run_list + " )" elif wheresql_run_range and not wheresql_run_list: wheresql += " and " + wheresql_run_range elif not wheresql_run_range and wheresql_run_list: wheresql += " and " + wheresql_run_list # sql = " ".join((generatedsql, basesql, self.fromsql, joinsql, wheresql)) cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) for i in cursors: d = self.formatCursor(i, size=100) if isinstance(d, list) or isinstance(d, GeneratorType): for elem in d: yield elem elif d: yield d
[ "def", "execute", "(", "self", ",", "conn", ",", "dataset", "=", "\"\"", ",", "block_name", "=", "\"\"", ",", "data_tier_name", "=", "\"\"", ",", "origin_site_name", "=", "\"\"", ",", "logical_file_name", "=", "\"\"", ",", "run_num", "=", "-", "1", ",", "min_cdate", "=", "0", ",", "max_cdate", "=", "0", ",", "min_ldate", "=", "0", ",", "max_ldate", "=", "0", ",", "cdate", "=", "0", ",", "ldate", "=", "0", ",", "open_for_writing", "=", "-", "1", ",", "transaction", "=", "False", ")", ":", "binds", "=", "{", "}", "basesql", "=", "self", ".", "sql", "joinsql", "=", "\"\"", "wheresql", "=", "\"\"", "generatedsql", "=", "\"\"", "if", "logical_file_name", "and", "logical_file_name", "!=", "\"%\"", ":", "joinsql", "+=", "\" JOIN %sFILES FL ON FL.BLOCK_ID = B.BLOCK_ID \"", "%", "(", "self", ".", "owner", ")", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "logical_file_name", "]", "wheresql", "+=", "\" WHERE LOGICAL_FILE_NAME %s :logical_file_name \"", "%", "op", "binds", ".", "update", "(", "logical_file_name", "=", "logical_file_name", ")", "if", "block_name", "and", "block_name", "!=", "\"%\"", ":", "andorwhere", "=", "(", "\"WHERE\"", ",", "\"AND\"", ")", "[", "bool", "(", "wheresql", ")", "]", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "block_name", "]", "wheresql", "+=", "\" %s B.BLOCK_NAME %s :block_name \"", "%", "(", "(", "andorwhere", ",", "op", ")", ")", "binds", ".", "update", "(", "block_name", "=", "block_name", ")", "if", "data_tier_name", "or", "(", "dataset", "and", "dataset", "!=", "\"%\"", ")", ":", "joinsql", "+=", "\"JOIN %sDATASETS DS ON DS.DATASET_ID = B.DATASET_ID \"", "%", "(", "self", ".", "owner", ")", "andorwhere", "=", "(", "\"WHERE\"", ",", "\"AND\"", ")", "[", "bool", "(", "wheresql", ")", "]", "if", "dataset", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "dataset", "]", "wheresql", "+=", "\" %s DS.DATASET %s :dataset \"", "%", "(", "(", "andorwhere", ",", "op", ")", ")", "binds", ".", "update", "(", "dataset", "=", "dataset", ")", "if", "data_tier_name", ":", "joinsql", "+=", "\"JOIN {owner}DATA_TIERS DT ON DS.DATA_TIER_ID=DT.DATA_TIER_ID \"", ".", "format", "(", "owner", "=", "self", ".", "owner", ")", "wheresql", "+=", "\" %s DT.DATA_TIER_NAME=:data_tier_name \"", "%", "(", "andorwhere", ")", "binds", ".", "update", "(", "data_tier_name", "=", "data_tier_name", ")", "if", "origin_site_name", "and", "origin_site_name", "!=", "\"%\"", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "origin_site_name", "]", "wheresql", "+=", "\" AND B.ORIGIN_SITE_NAME %s :origin_site_name \"", "%", "op", "binds", ".", "update", "(", "origin_site_name", "=", "origin_site_name", ")", "if", "open_for_writing", "==", "0", "or", "open_for_writing", "==", "1", ":", "wheresql", "+=", "\" AND B.OPEN_FOR_WRITTING = :open_for_writing \"", "if", "cdate", "!=", "0", ":", "wheresql", "+=", "\"AND B.CREATION_DATE = :cdate \"", "binds", ".", "update", "(", "cdate", "=", "cdate", ")", "elif", "min_cdate", "!=", "0", "and", "max_cdate", "!=", "0", ":", "wheresql", "+=", "\"AND B.CREATION_DATE BETWEEN :min_cdate and :max_cdate \"", "binds", ".", "update", "(", "min_cdate", "=", "min_cdate", ")", "binds", ".", "update", "(", "max_cdate", "=", "max_cdate", ")", "elif", "min_cdate", "!=", "0", "and", "max_cdate", "==", "0", ":", "wheresql", "+=", "\"AND B.CREATION_DATE > :min_cdate \"", "binds", ".", "update", "(", "min_cdate", "=", "min_cdate", ")", "elif", "min_cdate", "==", "0", "and", "max_cdate", "!=", "0", ":", "wheresql", "+=", "\"AND B.CREATION_DATE < :max_cdate \"", "binds", ".", "update", "(", "max_cdate", "=", "max_cdate", ")", "else", ":", "pass", "if", "ldate", "!=", "0", ":", "wheresql", "+=", "\"AND B.LAST_MODIFICATION_DATE = :ldate \"", "binds", ".", "update", "(", "ldate", "=", "ldate", ")", "elif", "min_ldate", "!=", "0", "and", "max_ldate", "!=", "0", ":", "wheresql", "+=", "\"AND B.LAST_MODIFICATION_DATE BETWEEN :min_ldate and :max_ldate \"", "binds", ".", "update", "(", "min_ldate", "=", "min_ldate", ")", "binds", ".", "update", "(", "max_ldate", "=", "max_ldate", ")", "elif", "min_ldate", "!=", "0", "and", "max_ldate", "==", "0", ":", "wheresql", "+=", "\"AND B.LAST_MODIFICATION_DATE > :min_ldate \"", "binds", ".", "update", "(", "min_ldate", "=", "min_ldate", ")", "elif", "min_cdate", "==", "0", "and", "max_cdate", "!=", "0", ":", "wheresql", "+=", "\"AND B.LAST_MODIFICATION_DATE < :max_ldate \"", "binds", ".", "update", "(", "max_ldate", "=", "max_ldate", ")", "else", ":", "pass", "#one may provide a list of runs , so it has to be the last one in building the bind.", "if", "run_num", "!=", "-", "1", ":", "basesql", "=", "basesql", ".", "replace", "(", "\"SELECT\"", ",", "\"SELECT DISTINCT\"", ")", "+", "\" , FLM.RUN_NUM \"", "if", "not", "logical_file_name", ":", "joinsql", "+=", "\" JOIN %sFILES FL ON FL.BLOCK_ID = B.BLOCK_ID \"", "%", "(", "self", ".", "owner", ")", "joinsql", "+=", "\" JOIN %sFILE_LUMIS FLM on FLM.FILE_ID = FL.FILE_ID \"", "%", "(", "self", ".", "owner", ")", "run_list", "=", "[", "]", "wheresql_run_list", "=", "''", "wheresql_run_range", "=", "''", "#", "for", "r", "in", "parseRunRange", "(", "run_num", ")", ":", "if", "isinstance", "(", "r", ",", "basestring", ")", "or", "isinstance", "(", "r", ",", "int", ")", "or", "isinstance", "(", "r", ",", "long", ")", ":", "run_list", ".", "append", "(", "str", "(", "r", ")", ")", "if", "isinstance", "(", "r", ",", "run_tuple", ")", ":", "if", "r", "[", "0", "]", "==", "r", "[", "1", "]", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"DBS run_num range must be apart at least by 1.\"", ",", "self", ".", "logger", ".", "exception", ")", "wheresql_run_range", "=", "\" FLM.RUN_NUM between :minrun and :maxrun \"", "binds", ".", "update", "(", "{", "\"minrun\"", ":", "r", "[", "0", "]", "}", ")", "binds", ".", "update", "(", "{", "\"maxrun\"", ":", "r", "[", "1", "]", "}", ")", "#", "if", "run_list", ":", "wheresql_run_list", "=", "\" FLM.RUN_NUM in (SELECT TOKEN FROM TOKEN_GENERATOR) \"", "generatedsql", ",", "run_binds", "=", "create_token_generator", "(", "run_list", ")", "binds", ".", "update", "(", "run_binds", ")", "# ", "if", "wheresql_run_range", "and", "wheresql_run_list", ":", "wheresql", "+=", "\" and (\"", "+", "wheresql_run_range", "+", "\" or \"", "+", "wheresql_run_list", "+", "\" )\"", "elif", "wheresql_run_range", "and", "not", "wheresql_run_list", ":", "wheresql", "+=", "\" and \"", "+", "wheresql_run_range", "elif", "not", "wheresql_run_range", "and", "wheresql_run_list", ":", "wheresql", "+=", "\" and \"", "+", "wheresql_run_list", "#", "sql", "=", "\" \"", ".", "join", "(", "(", "generatedsql", ",", "basesql", ",", "self", ".", "fromsql", ",", "joinsql", ",", "wheresql", ")", ")", "cursors", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ",", "returnCursor", "=", "True", ")", "for", "i", "in", "cursors", ":", "d", "=", "self", ".", "formatCursor", "(", "i", ",", "size", "=", "100", ")", "if", "isinstance", "(", "d", ",", "list", ")", "or", "isinstance", "(", "d", ",", "GeneratorType", ")", ":", "for", "elem", "in", "d", ":", "yield", "elem", "elif", "d", ":", "yield", "d" ]
dataset: /a/b/c block: /a/b/c#d
[ "dataset", ":", "/", "a", "/", "b", "/", "c", "block", ":", "/", "a", "/", "b", "/", "c#d" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/Block/BriefList.py#L26-L144
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/PrimaryDSType/List.py
List.execute
def execute(self, conn, dsType = "", dataset="", transaction = False): """ Lists all primary dataset types if no user input is provided. """ sql = self.sql binds={} if not dsType and not dataset: pass elif dsType and dataset in ("", None, '%'): op = ("=", "like")["%" in dsType] sql += "WHERE PDT.PRIMARY_DS_TYPE %s :primdstype"%op binds = {"primdstype":dsType} elif dataset and dsType in ("", None, '%'): op = ("=", "like")["%" in dataset] sql += "JOIN %sPRIMARY_DATASETS PDS on PDS.PRIMARY_DS_TYPE_ID = PDT.PRIMARY_DS_TYPE_ID \ JOIN %sDATASETS DS ON DS.PRIMARY_DS_ID = PDS.PRIMARY_DS_ID \ WHERE DS.DATASET %s :dataset" %(self.owner, self.owner, op) binds={"dataset":dataset} elif dataset and dsType: op = ("=", "like")["%" in dsType] op1 = ("=", "like")["%" in dataset] sql += "JOIN %sPRIMARY_DATASETS PDS on PDS.PRIMARY_DS_TYPE_ID = PDT.PRIMARY_DS_TYPE_ID \ JOIN %sDATASETS DS ON DS.PRIMARY_DS_ID = PDS.PRIMARY_DS_ID \ WHERE DS.DATASET %s :dataset and PDT.PRIMARY_DS_TYPE %s :primdstype" \ %(self.owner, self.owner, op1, op) binds = {"primdstype":dsType, "dataset":dataset} else: dbsExceptionHandler('dbsException-invalid-input', "DAO Primary_DS_TYPE List accepts no input, or\ dataset,primary_ds_type as input.", self.logger.exception) cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = [] for c in cursors: result.extend(self.formatCursor(c, size=100)) return result
python
def execute(self, conn, dsType = "", dataset="", transaction = False): """ Lists all primary dataset types if no user input is provided. """ sql = self.sql binds={} if not dsType and not dataset: pass elif dsType and dataset in ("", None, '%'): op = ("=", "like")["%" in dsType] sql += "WHERE PDT.PRIMARY_DS_TYPE %s :primdstype"%op binds = {"primdstype":dsType} elif dataset and dsType in ("", None, '%'): op = ("=", "like")["%" in dataset] sql += "JOIN %sPRIMARY_DATASETS PDS on PDS.PRIMARY_DS_TYPE_ID = PDT.PRIMARY_DS_TYPE_ID \ JOIN %sDATASETS DS ON DS.PRIMARY_DS_ID = PDS.PRIMARY_DS_ID \ WHERE DS.DATASET %s :dataset" %(self.owner, self.owner, op) binds={"dataset":dataset} elif dataset and dsType: op = ("=", "like")["%" in dsType] op1 = ("=", "like")["%" in dataset] sql += "JOIN %sPRIMARY_DATASETS PDS on PDS.PRIMARY_DS_TYPE_ID = PDT.PRIMARY_DS_TYPE_ID \ JOIN %sDATASETS DS ON DS.PRIMARY_DS_ID = PDS.PRIMARY_DS_ID \ WHERE DS.DATASET %s :dataset and PDT.PRIMARY_DS_TYPE %s :primdstype" \ %(self.owner, self.owner, op1, op) binds = {"primdstype":dsType, "dataset":dataset} else: dbsExceptionHandler('dbsException-invalid-input', "DAO Primary_DS_TYPE List accepts no input, or\ dataset,primary_ds_type as input.", self.logger.exception) cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = [] for c in cursors: result.extend(self.formatCursor(c, size=100)) return result
[ "def", "execute", "(", "self", ",", "conn", ",", "dsType", "=", "\"\"", ",", "dataset", "=", "\"\"", ",", "transaction", "=", "False", ")", ":", "sql", "=", "self", ".", "sql", "binds", "=", "{", "}", "if", "not", "dsType", "and", "not", "dataset", ":", "pass", "elif", "dsType", "and", "dataset", "in", "(", "\"\"", ",", "None", ",", "'%'", ")", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "dsType", "]", "sql", "+=", "\"WHERE PDT.PRIMARY_DS_TYPE %s :primdstype\"", "%", "op", "binds", "=", "{", "\"primdstype\"", ":", "dsType", "}", "elif", "dataset", "and", "dsType", "in", "(", "\"\"", ",", "None", ",", "'%'", ")", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "dataset", "]", "sql", "+=", "\"JOIN %sPRIMARY_DATASETS PDS on PDS.PRIMARY_DS_TYPE_ID = PDT.PRIMARY_DS_TYPE_ID \\\n\t JOIN %sDATASETS DS ON DS.PRIMARY_DS_ID = PDS.PRIMARY_DS_ID \\\n\t WHERE DS.DATASET %s :dataset\"", "%", "(", "self", ".", "owner", ",", "self", ".", "owner", ",", "op", ")", "binds", "=", "{", "\"dataset\"", ":", "dataset", "}", "elif", "dataset", "and", "dsType", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "dsType", "]", "op1", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "dataset", "]", "sql", "+=", "\"JOIN %sPRIMARY_DATASETS PDS on PDS.PRIMARY_DS_TYPE_ID = PDT.PRIMARY_DS_TYPE_ID \\\n JOIN %sDATASETS DS ON DS.PRIMARY_DS_ID = PDS.PRIMARY_DS_ID \\\n WHERE DS.DATASET %s :dataset and PDT.PRIMARY_DS_TYPE %s :primdstype\"", "%", "(", "self", ".", "owner", ",", "self", ".", "owner", ",", "op1", ",", "op", ")", "binds", "=", "{", "\"primdstype\"", ":", "dsType", ",", "\"dataset\"", ":", "dataset", "}", "else", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"DAO Primary_DS_TYPE List accepts no input, or\\\n dataset,primary_ds_type as input.\"", ",", "self", ".", "logger", ".", "exception", ")", "cursors", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ",", "returnCursor", "=", "True", ")", "result", "=", "[", "]", "for", "c", "in", "cursors", ":", "result", ".", "extend", "(", "self", ".", "formatCursor", "(", "c", ",", "size", "=", "100", ")", ")", "return", "result" ]
Lists all primary dataset types if no user input is provided.
[ "Lists", "all", "primary", "dataset", "types", "if", "no", "user", "input", "is", "provided", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/PrimaryDSType/List.py#L29-L62
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/OutputModuleConfig/List.py
List.execute
def execute(self, conn, dataset="", logical_file_name="", app="", release_version="", pset_hash="", output_label ="", block_id=0, global_tag='', transaction = False): """ returns id for a given application """ #sql=self.sql binds = {} setAnd=False #add search only block id only for migration dump block. if block_id==0: sql = self.sql1 + self.sql2 if dataset: sql += " JOIN %sDATASET_OUTPUT_MOD_CONFIGS DC ON DC.OUTPUT_MOD_CONFIG_ID=O.OUTPUT_MOD_CONFIG_ID" % self.owner sql += " JOIN %sDATASETS DS ON DS.DATASET_ID=DC.DATASET_ID" % self.owner if logical_file_name: sql += " JOIN %sFILE_OUTPUT_MOD_CONFIGS FC ON FC.OUTPUT_MOD_CONFIG_ID=O.OUTPUT_MOD_CONFIG_ID" % self.owner sql += " JOIN %sFILES FS ON FS.FILE_ID=FC.FILE_ID" % self.owner if not app == "": op = ("=", "like")["%" in app] sql += " WHERE A.APP_NAME %s :app_name" % op binds["app_name"]=app setAnd=True if not release_version == "": op = ("=", "like")["%" in release_version] if setAnd : sql += " AND " else : sql += " WHERE " sql += " R.RELEASE_VERSION %s :release_version" % op binds["release_version"]=release_version setAnd=True if not pset_hash == "": op = ("=", "like")["%" in pset_hash] if setAnd : sql += " AND " else : sql += " WHERE " sql += " P.PSET_HASH %s :pset_hash" % op binds["pset_hash"]=pset_hash setAnd=True if not output_label == "": op = ("=", "like")["%" in output_label] if setAnd : sql += " AND " else : sql += " WHERE " sql += " O.OUTPUT_MODULE_LABEL %s :output_module_label" % op binds["output_module_label"]=output_label setAnd=True if not global_tag == "": op = ("=", "like")["%" in global_tag] if setAnd : sql += " AND " else : sql += " WHERE " sql += " O.GLOBAL_TAG %s :global_tag" % op binds["global_tag"]=global_tag setAnd=True if dataset: if setAnd : sql += " AND " else : sql += " WHERE " sql += "DS.DATASET=:dataset" binds["dataset"]=dataset setAnd=True if logical_file_name: if setAnd : sql += " AND " else : sql += " WHERE " sql += "FS.LOGICAL_FILE_NAME=:logical_file_name" binds["logical_file_name"]=logical_file_name setAnd=True else: #select by block id and return config along with LFN sql= self.sql1 + " , FS.LOGICAL_FILE_NAME LFN " + self.sql2 \ + " JOIN %sFILE_OUTPUT_MOD_CONFIGS FC ON FC.OUTPUT_MOD_CONFIG_ID=O.OUTPUT_MOD_CONFIG_ID" % self.owner \ + " JOIN %sFILES FS ON FS.FILE_ID=FC.FILE_ID" % self.owner \ + " WHERE FS.BLOCK_ID = :block_id " binds["block_id"]=block_id cursors = self.dbi.processData(sql, binds, conn, transaction=False, returnCursor=True) #assert len(cursors) == 1, "output module config does not exist" result = [] for c in cursors: result.extend(self.formatCursor(c, size=100)) return result
python
def execute(self, conn, dataset="", logical_file_name="", app="", release_version="", pset_hash="", output_label ="", block_id=0, global_tag='', transaction = False): """ returns id for a given application """ #sql=self.sql binds = {} setAnd=False #add search only block id only for migration dump block. if block_id==0: sql = self.sql1 + self.sql2 if dataset: sql += " JOIN %sDATASET_OUTPUT_MOD_CONFIGS DC ON DC.OUTPUT_MOD_CONFIG_ID=O.OUTPUT_MOD_CONFIG_ID" % self.owner sql += " JOIN %sDATASETS DS ON DS.DATASET_ID=DC.DATASET_ID" % self.owner if logical_file_name: sql += " JOIN %sFILE_OUTPUT_MOD_CONFIGS FC ON FC.OUTPUT_MOD_CONFIG_ID=O.OUTPUT_MOD_CONFIG_ID" % self.owner sql += " JOIN %sFILES FS ON FS.FILE_ID=FC.FILE_ID" % self.owner if not app == "": op = ("=", "like")["%" in app] sql += " WHERE A.APP_NAME %s :app_name" % op binds["app_name"]=app setAnd=True if not release_version == "": op = ("=", "like")["%" in release_version] if setAnd : sql += " AND " else : sql += " WHERE " sql += " R.RELEASE_VERSION %s :release_version" % op binds["release_version"]=release_version setAnd=True if not pset_hash == "": op = ("=", "like")["%" in pset_hash] if setAnd : sql += " AND " else : sql += " WHERE " sql += " P.PSET_HASH %s :pset_hash" % op binds["pset_hash"]=pset_hash setAnd=True if not output_label == "": op = ("=", "like")["%" in output_label] if setAnd : sql += " AND " else : sql += " WHERE " sql += " O.OUTPUT_MODULE_LABEL %s :output_module_label" % op binds["output_module_label"]=output_label setAnd=True if not global_tag == "": op = ("=", "like")["%" in global_tag] if setAnd : sql += " AND " else : sql += " WHERE " sql += " O.GLOBAL_TAG %s :global_tag" % op binds["global_tag"]=global_tag setAnd=True if dataset: if setAnd : sql += " AND " else : sql += " WHERE " sql += "DS.DATASET=:dataset" binds["dataset"]=dataset setAnd=True if logical_file_name: if setAnd : sql += " AND " else : sql += " WHERE " sql += "FS.LOGICAL_FILE_NAME=:logical_file_name" binds["logical_file_name"]=logical_file_name setAnd=True else: #select by block id and return config along with LFN sql= self.sql1 + " , FS.LOGICAL_FILE_NAME LFN " + self.sql2 \ + " JOIN %sFILE_OUTPUT_MOD_CONFIGS FC ON FC.OUTPUT_MOD_CONFIG_ID=O.OUTPUT_MOD_CONFIG_ID" % self.owner \ + " JOIN %sFILES FS ON FS.FILE_ID=FC.FILE_ID" % self.owner \ + " WHERE FS.BLOCK_ID = :block_id " binds["block_id"]=block_id cursors = self.dbi.processData(sql, binds, conn, transaction=False, returnCursor=True) #assert len(cursors) == 1, "output module config does not exist" result = [] for c in cursors: result.extend(self.formatCursor(c, size=100)) return result
[ "def", "execute", "(", "self", ",", "conn", ",", "dataset", "=", "\"\"", ",", "logical_file_name", "=", "\"\"", ",", "app", "=", "\"\"", ",", "release_version", "=", "\"\"", ",", "pset_hash", "=", "\"\"", ",", "output_label", "=", "\"\"", ",", "block_id", "=", "0", ",", "global_tag", "=", "''", ",", "transaction", "=", "False", ")", ":", "#sql=self.sql\t", "binds", "=", "{", "}", "setAnd", "=", "False", "#add search only block id only for migration dump block. ", "if", "block_id", "==", "0", ":", "sql", "=", "self", ".", "sql1", "+", "self", ".", "sql2", "if", "dataset", ":", "sql", "+=", "\" JOIN %sDATASET_OUTPUT_MOD_CONFIGS DC ON DC.OUTPUT_MOD_CONFIG_ID=O.OUTPUT_MOD_CONFIG_ID\"", "%", "self", ".", "owner", "sql", "+=", "\" JOIN %sDATASETS DS ON DS.DATASET_ID=DC.DATASET_ID\"", "%", "self", ".", "owner", "if", "logical_file_name", ":", "sql", "+=", "\" JOIN %sFILE_OUTPUT_MOD_CONFIGS FC ON FC.OUTPUT_MOD_CONFIG_ID=O.OUTPUT_MOD_CONFIG_ID\"", "%", "self", ".", "owner", "sql", "+=", "\" JOIN %sFILES FS ON FS.FILE_ID=FC.FILE_ID\"", "%", "self", ".", "owner", "if", "not", "app", "==", "\"\"", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "app", "]", "sql", "+=", "\" WHERE A.APP_NAME %s :app_name\"", "%", "op", "binds", "[", "\"app_name\"", "]", "=", "app", "setAnd", "=", "True", "if", "not", "release_version", "==", "\"\"", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "release_version", "]", "if", "setAnd", ":", "sql", "+=", "\" AND \"", "else", ":", "sql", "+=", "\" WHERE \"", "sql", "+=", "\" R.RELEASE_VERSION %s :release_version\"", "%", "op", "binds", "[", "\"release_version\"", "]", "=", "release_version", "setAnd", "=", "True", "if", "not", "pset_hash", "==", "\"\"", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "pset_hash", "]", "if", "setAnd", ":", "sql", "+=", "\" AND \"", "else", ":", "sql", "+=", "\" WHERE \"", "sql", "+=", "\" P.PSET_HASH %s :pset_hash\"", "%", "op", "binds", "[", "\"pset_hash\"", "]", "=", "pset_hash", "setAnd", "=", "True", "if", "not", "output_label", "==", "\"\"", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "output_label", "]", "if", "setAnd", ":", "sql", "+=", "\" AND \"", "else", ":", "sql", "+=", "\" WHERE \"", "sql", "+=", "\" O.OUTPUT_MODULE_LABEL %s :output_module_label\"", "%", "op", "binds", "[", "\"output_module_label\"", "]", "=", "output_label", "setAnd", "=", "True", "if", "not", "global_tag", "==", "\"\"", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "global_tag", "]", "if", "setAnd", ":", "sql", "+=", "\" AND \"", "else", ":", "sql", "+=", "\" WHERE \"", "sql", "+=", "\" O.GLOBAL_TAG %s :global_tag\"", "%", "op", "binds", "[", "\"global_tag\"", "]", "=", "global_tag", "setAnd", "=", "True", "if", "dataset", ":", "if", "setAnd", ":", "sql", "+=", "\" AND \"", "else", ":", "sql", "+=", "\" WHERE \"", "sql", "+=", "\"DS.DATASET=:dataset\"", "binds", "[", "\"dataset\"", "]", "=", "dataset", "setAnd", "=", "True", "if", "logical_file_name", ":", "if", "setAnd", ":", "sql", "+=", "\" AND \"", "else", ":", "sql", "+=", "\" WHERE \"", "sql", "+=", "\"FS.LOGICAL_FILE_NAME=:logical_file_name\"", "binds", "[", "\"logical_file_name\"", "]", "=", "logical_file_name", "setAnd", "=", "True", "else", ":", "#select by block id and return config along with LFN", "sql", "=", "self", ".", "sql1", "+", "\" , FS.LOGICAL_FILE_NAME LFN \"", "+", "self", ".", "sql2", "+", "\" JOIN %sFILE_OUTPUT_MOD_CONFIGS FC ON FC.OUTPUT_MOD_CONFIG_ID=O.OUTPUT_MOD_CONFIG_ID\"", "%", "self", ".", "owner", "+", "\" JOIN %sFILES FS ON FS.FILE_ID=FC.FILE_ID\"", "%", "self", ".", "owner", "+", "\" WHERE FS.BLOCK_ID = :block_id \"", "binds", "[", "\"block_id\"", "]", "=", "block_id", "cursors", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", "=", "False", ",", "returnCursor", "=", "True", ")", "#assert len(cursors) == 1, \"output module config does not exist\"", "result", "=", "[", "]", "for", "c", "in", "cursors", ":", "result", ".", "extend", "(", "self", ".", "formatCursor", "(", "c", ",", "size", "=", "100", ")", ")", "return", "result" ]
returns id for a given application
[ "returns", "id", "for", "a", "given", "application" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/OutputModuleConfig/List.py#L42-L116
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/BlockParent/ListChild.py
ListChild.execute
def execute(self, conn, block_name="", transaction = False): """ block: /a/b/c#d """ if not conn: msg='Oracle/BlockParent/List. No DB connection found' dbsExceptionHandler('dbsException-failed-connect2host', msg, self.logger.exception) sql = self.sql binds = {} if block_name: binds.update(block_name = block_name) else: dbsExceptionHandler("dbsException-invalid-input", "Oracle/BlockParent/ListChild. block_name must be provided.", self.logger.exception) cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = [] for c in cursors: result.extend(self.formatCursor(c, size=100)) return result
python
def execute(self, conn, block_name="", transaction = False): """ block: /a/b/c#d """ if not conn: msg='Oracle/BlockParent/List. No DB connection found' dbsExceptionHandler('dbsException-failed-connect2host', msg, self.logger.exception) sql = self.sql binds = {} if block_name: binds.update(block_name = block_name) else: dbsExceptionHandler("dbsException-invalid-input", "Oracle/BlockParent/ListChild. block_name must be provided.", self.logger.exception) cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = [] for c in cursors: result.extend(self.formatCursor(c, size=100)) return result
[ "def", "execute", "(", "self", ",", "conn", ",", "block_name", "=", "\"\"", ",", "transaction", "=", "False", ")", ":", "if", "not", "conn", ":", "msg", "=", "'Oracle/BlockParent/List. No DB connection found'", "dbsExceptionHandler", "(", "'dbsException-failed-connect2host'", ",", "msg", ",", "self", ".", "logger", ".", "exception", ")", "sql", "=", "self", ".", "sql", "binds", "=", "{", "}", "if", "block_name", ":", "binds", ".", "update", "(", "block_name", "=", "block_name", ")", "else", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"Oracle/BlockParent/ListChild. block_name must be provided.\"", ",", "self", ".", "logger", ".", "exception", ")", "cursors", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ",", "returnCursor", "=", "True", ")", "result", "=", "[", "]", "for", "c", "in", "cursors", ":", "result", ".", "extend", "(", "self", ".", "formatCursor", "(", "c", ",", "size", "=", "100", ")", ")", "return", "result" ]
block: /a/b/c#d
[ "block", ":", "/", "a", "/", "b", "/", "c#d" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/BlockParent/ListChild.py#L27-L46
train
dmwm/DBS
Server/Python/src/dbs/business/DBSFile.py
DBSFile.listFileLumis
def listFileLumis(self, logical_file_name="", block_name="", run_num=-1, validFileOnly=0, input_body=-1): """ optional parameter: logical_file_name, block_name, validFileOnly returns: logical_file_name, file_lumi_id, run_num, lumi_section_num """ if((logical_file_name=='' or '*'in logical_file_name or '%' in logical_file_name) \ and (block_name=='' or '*' in block_name or '%' in block_name) and input_body==-1 ): dbsExceptionHandler('dbsException-invalid-input', \ "Fully specified logical_file_name or block_name is required if GET is called. No wildcards are allowed.", self.logger.exception, "Fully specified logical_file_name or block_name is required if GET is called. No wildcards are allowed.") elif input_body != -1 : try: logical_file_name = input_body["logical_file_name"] run_num = input_body.get("run_num", -1) validFileOnly = input_body.get("validFileOnly", 0) block_name = "" except cjson.DecodeError as de: msg = "business/listFileLumis requires at least a list of logical_file_name. %s" % de dbsExceptionHandler('dbsException-invalid-input2', "Invalid input", self.logger.exception, msg) elif input_body != -1 and (logical_file_name is not None or block_name is not None): dbsExceptionHandler('dbsException-invalid-input', "listFileLumis may have input in the command or in the payload, not mixed.", self.logger.exception, "listFileLumis may have input in the command or in the payload, not mixed.") with self.dbi.connection() as conn: for item in self.filelumilist.execute(conn, logical_file_name, block_name, run_num, validFileOnly=validFileOnly): yield item
python
def listFileLumis(self, logical_file_name="", block_name="", run_num=-1, validFileOnly=0, input_body=-1): """ optional parameter: logical_file_name, block_name, validFileOnly returns: logical_file_name, file_lumi_id, run_num, lumi_section_num """ if((logical_file_name=='' or '*'in logical_file_name or '%' in logical_file_name) \ and (block_name=='' or '*' in block_name or '%' in block_name) and input_body==-1 ): dbsExceptionHandler('dbsException-invalid-input', \ "Fully specified logical_file_name or block_name is required if GET is called. No wildcards are allowed.", self.logger.exception, "Fully specified logical_file_name or block_name is required if GET is called. No wildcards are allowed.") elif input_body != -1 : try: logical_file_name = input_body["logical_file_name"] run_num = input_body.get("run_num", -1) validFileOnly = input_body.get("validFileOnly", 0) block_name = "" except cjson.DecodeError as de: msg = "business/listFileLumis requires at least a list of logical_file_name. %s" % de dbsExceptionHandler('dbsException-invalid-input2', "Invalid input", self.logger.exception, msg) elif input_body != -1 and (logical_file_name is not None or block_name is not None): dbsExceptionHandler('dbsException-invalid-input', "listFileLumis may have input in the command or in the payload, not mixed.", self.logger.exception, "listFileLumis may have input in the command or in the payload, not mixed.") with self.dbi.connection() as conn: for item in self.filelumilist.execute(conn, logical_file_name, block_name, run_num, validFileOnly=validFileOnly): yield item
[ "def", "listFileLumis", "(", "self", ",", "logical_file_name", "=", "\"\"", ",", "block_name", "=", "\"\"", ",", "run_num", "=", "-", "1", ",", "validFileOnly", "=", "0", ",", "input_body", "=", "-", "1", ")", ":", "if", "(", "(", "logical_file_name", "==", "''", "or", "'*'", "in", "logical_file_name", "or", "'%'", "in", "logical_file_name", ")", "and", "(", "block_name", "==", "''", "or", "'*'", "in", "block_name", "or", "'%'", "in", "block_name", ")", "and", "input_body", "==", "-", "1", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Fully specified logical_file_name or block_name is required if GET is called. No wildcards are allowed.\"", ",", "self", ".", "logger", ".", "exception", ",", "\"Fully specified logical_file_name or block_name is required if GET is called. No wildcards are allowed.\"", ")", "elif", "input_body", "!=", "-", "1", ":", "try", ":", "logical_file_name", "=", "input_body", "[", "\"logical_file_name\"", "]", "run_num", "=", "input_body", ".", "get", "(", "\"run_num\"", ",", "-", "1", ")", "validFileOnly", "=", "input_body", ".", "get", "(", "\"validFileOnly\"", ",", "0", ")", "block_name", "=", "\"\"", "except", "cjson", ".", "DecodeError", "as", "de", ":", "msg", "=", "\"business/listFileLumis requires at least a list of logical_file_name. %s\"", "%", "de", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "\"Invalid input\"", ",", "self", ".", "logger", ".", "exception", ",", "msg", ")", "elif", "input_body", "!=", "-", "1", "and", "(", "logical_file_name", "is", "not", "None", "or", "block_name", "is", "not", "None", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"listFileLumis may have input in the command or in the payload, not mixed.\"", ",", "self", ".", "logger", ".", "exception", ",", "\"listFileLumis may have input in the command or in the payload, not mixed.\"", ")", "with", "self", ".", "dbi", ".", "connection", "(", ")", "as", "conn", ":", "for", "item", "in", "self", ".", "filelumilist", ".", "execute", "(", "conn", ",", "logical_file_name", ",", "block_name", ",", "run_num", ",", "validFileOnly", "=", "validFileOnly", ")", ":", "yield", "item" ]
optional parameter: logical_file_name, block_name, validFileOnly returns: logical_file_name, file_lumi_id, run_num, lumi_section_num
[ "optional", "parameter", ":", "logical_file_name", "block_name", "validFileOnly", "returns", ":", "logical_file_name", "file_lumi_id", "run_num", "lumi_section_num" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSFile.py#L52-L76
train
dmwm/DBS
Server/Python/src/dbs/business/DBSFile.py
DBSFile.listFileSummary
def listFileSummary(self, block_name="", dataset="", run_num=-1, validFileOnly=0, sumOverLumi=0): """ required parameter: full block_name or dataset name. No wildcards allowed. run_num is optional. """ if not block_name and not dataset: msg = "Block_name or dataset is required for listFileSummary API" dbsExceptionHandler('dbsException-invalid-input', msg, self.logger.exception) if '%' in block_name or '*' in block_name or '%' in dataset or '*' in dataset: msg = "No wildcard is allowed in block_name or dataset for filesummaries API" dbsExceptionHandler('dbsException-invalid-input', msg, self.logger.exception) # with self.dbi.connection() as conn: for item in self.filesummarylist.execute(conn, block_name, dataset, run_num, validFileOnly=validFileOnly, sumOverLumi=sumOverLumi): if item['num_file']==0 and item['num_block']==0 \ and item['num_event']==0 and item['file_size']==0: pass else: yield item
python
def listFileSummary(self, block_name="", dataset="", run_num=-1, validFileOnly=0, sumOverLumi=0): """ required parameter: full block_name or dataset name. No wildcards allowed. run_num is optional. """ if not block_name and not dataset: msg = "Block_name or dataset is required for listFileSummary API" dbsExceptionHandler('dbsException-invalid-input', msg, self.logger.exception) if '%' in block_name or '*' in block_name or '%' in dataset or '*' in dataset: msg = "No wildcard is allowed in block_name or dataset for filesummaries API" dbsExceptionHandler('dbsException-invalid-input', msg, self.logger.exception) # with self.dbi.connection() as conn: for item in self.filesummarylist.execute(conn, block_name, dataset, run_num, validFileOnly=validFileOnly, sumOverLumi=sumOverLumi): if item['num_file']==0 and item['num_block']==0 \ and item['num_event']==0 and item['file_size']==0: pass else: yield item
[ "def", "listFileSummary", "(", "self", ",", "block_name", "=", "\"\"", ",", "dataset", "=", "\"\"", ",", "run_num", "=", "-", "1", ",", "validFileOnly", "=", "0", ",", "sumOverLumi", "=", "0", ")", ":", "if", "not", "block_name", "and", "not", "dataset", ":", "msg", "=", "\"Block_name or dataset is required for listFileSummary API\"", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "msg", ",", "self", ".", "logger", ".", "exception", ")", "if", "'%'", "in", "block_name", "or", "'*'", "in", "block_name", "or", "'%'", "in", "dataset", "or", "'*'", "in", "dataset", ":", "msg", "=", "\"No wildcard is allowed in block_name or dataset for filesummaries API\"", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "msg", ",", "self", ".", "logger", ".", "exception", ")", "#", "with", "self", ".", "dbi", ".", "connection", "(", ")", "as", "conn", ":", "for", "item", "in", "self", ".", "filesummarylist", ".", "execute", "(", "conn", ",", "block_name", ",", "dataset", ",", "run_num", ",", "validFileOnly", "=", "validFileOnly", ",", "sumOverLumi", "=", "sumOverLumi", ")", ":", "if", "item", "[", "'num_file'", "]", "==", "0", "and", "item", "[", "'num_block'", "]", "==", "0", "and", "item", "[", "'num_event'", "]", "==", "0", "and", "item", "[", "'file_size'", "]", "==", "0", ":", "pass", "else", ":", "yield", "item" ]
required parameter: full block_name or dataset name. No wildcards allowed. run_num is optional.
[ "required", "parameter", ":", "full", "block_name", "or", "dataset", "name", ".", "No", "wildcards", "allowed", ".", "run_num", "is", "optional", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSFile.py#L78-L96
train
dmwm/DBS
Server/Python/src/dbs/business/DBSFile.py
DBSFile.listFileParents
def listFileParents(self, logical_file_name="", block_id=0, block_name=""): """ required parameter: logical_file_name or block_name returns: this_logical_file_name, parent_logical_file_name, parent_file_id """ #self.logger.debug("lfn %s, block_name %s, block_id :%s" % (logical_file_name, block_name, block_id)) if not logical_file_name and not block_name and not block_id: dbsExceptionHandler('dbsException-invalid-input', \ "Logical_file_name, block_id or block_name is required for fileparents api", self.logger.exception ) with self.dbi.connection() as conn: sqlresult = self.fileparentlist.execute(conn, logical_file_name, block_id, block_name) d = {} #self.logger.debug(sqlresult) for i in sqlresult: k = i['this_logical_file_name'] v = i['parent_logical_file_name'] d.setdefault(k, []).append(v) for k, v in d.iteritems(): yield {'logical_file_name':k, 'parent_logical_file_name': v} del d
python
def listFileParents(self, logical_file_name="", block_id=0, block_name=""): """ required parameter: logical_file_name or block_name returns: this_logical_file_name, parent_logical_file_name, parent_file_id """ #self.logger.debug("lfn %s, block_name %s, block_id :%s" % (logical_file_name, block_name, block_id)) if not logical_file_name and not block_name and not block_id: dbsExceptionHandler('dbsException-invalid-input', \ "Logical_file_name, block_id or block_name is required for fileparents api", self.logger.exception ) with self.dbi.connection() as conn: sqlresult = self.fileparentlist.execute(conn, logical_file_name, block_id, block_name) d = {} #self.logger.debug(sqlresult) for i in sqlresult: k = i['this_logical_file_name'] v = i['parent_logical_file_name'] d.setdefault(k, []).append(v) for k, v in d.iteritems(): yield {'logical_file_name':k, 'parent_logical_file_name': v} del d
[ "def", "listFileParents", "(", "self", ",", "logical_file_name", "=", "\"\"", ",", "block_id", "=", "0", ",", "block_name", "=", "\"\"", ")", ":", "#self.logger.debug(\"lfn %s, block_name %s, block_id :%s\" % (logical_file_name, block_name, block_id))", "if", "not", "logical_file_name", "and", "not", "block_name", "and", "not", "block_id", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Logical_file_name, block_id or block_name is required for fileparents api\"", ",", "self", ".", "logger", ".", "exception", ")", "with", "self", ".", "dbi", ".", "connection", "(", ")", "as", "conn", ":", "sqlresult", "=", "self", ".", "fileparentlist", ".", "execute", "(", "conn", ",", "logical_file_name", ",", "block_id", ",", "block_name", ")", "d", "=", "{", "}", "#self.logger.debug(sqlresult)", "for", "i", "in", "sqlresult", ":", "k", "=", "i", "[", "'this_logical_file_name'", "]", "v", "=", "i", "[", "'parent_logical_file_name'", "]", "d", ".", "setdefault", "(", "k", ",", "[", "]", ")", ".", "append", "(", "v", ")", "for", "k", ",", "v", "in", "d", ".", "iteritems", "(", ")", ":", "yield", "{", "'logical_file_name'", ":", "k", ",", "'parent_logical_file_name'", ":", "v", "}", "del", "d" ]
required parameter: logical_file_name or block_name returns: this_logical_file_name, parent_logical_file_name, parent_file_id
[ "required", "parameter", ":", "logical_file_name", "or", "block_name", "returns", ":", "this_logical_file_name", "parent_logical_file_name", "parent_file_id" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSFile.py#L97-L116
train
dmwm/DBS
Server/Python/src/dbs/business/DBSFile.py
DBSFile.listFileParentsByLumi
def listFileParentsByLumi(self, block_name='', logical_file_name=[]): """ required parameter: block_name returns: [{child_parent_id_list: [(cid1, pid1), (cid2, pid2), ... (cidn, pidn)]}] """ #self.logger.debug("lfn %s, block_name %s" % (logical_file_name, block_name)) if not block_name: dbsExceptionHandler('dbsException-invalid-input', \ "Child block_name is required for fileparents/listFileParentsByLumi api", self.logger.exception ) with self.dbi.connection() as conn: sqlresult = self.fileparentbylumi.execute(conn, block_name, logical_file_name) return [{"child_parent_id_list":sqlresult}]
python
def listFileParentsByLumi(self, block_name='', logical_file_name=[]): """ required parameter: block_name returns: [{child_parent_id_list: [(cid1, pid1), (cid2, pid2), ... (cidn, pidn)]}] """ #self.logger.debug("lfn %s, block_name %s" % (logical_file_name, block_name)) if not block_name: dbsExceptionHandler('dbsException-invalid-input', \ "Child block_name is required for fileparents/listFileParentsByLumi api", self.logger.exception ) with self.dbi.connection() as conn: sqlresult = self.fileparentbylumi.execute(conn, block_name, logical_file_name) return [{"child_parent_id_list":sqlresult}]
[ "def", "listFileParentsByLumi", "(", "self", ",", "block_name", "=", "''", ",", "logical_file_name", "=", "[", "]", ")", ":", "#self.logger.debug(\"lfn %s, block_name %s\" % (logical_file_name, block_name))", "if", "not", "block_name", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Child block_name is required for fileparents/listFileParentsByLumi api\"", ",", "self", ".", "logger", ".", "exception", ")", "with", "self", ".", "dbi", ".", "connection", "(", ")", "as", "conn", ":", "sqlresult", "=", "self", ".", "fileparentbylumi", ".", "execute", "(", "conn", ",", "block_name", ",", "logical_file_name", ")", "return", "[", "{", "\"child_parent_id_list\"", ":", "sqlresult", "}", "]" ]
required parameter: block_name returns: [{child_parent_id_list: [(cid1, pid1), (cid2, pid2), ... (cidn, pidn)]}]
[ "required", "parameter", ":", "block_name", "returns", ":", "[", "{", "child_parent_id_list", ":", "[", "(", "cid1", "pid1", ")", "(", "cid2", "pid2", ")", "...", "(", "cidn", "pidn", ")", "]", "}", "]" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSFile.py#L118-L129
train
dmwm/DBS
Server/Python/src/dbs/business/DBSFile.py
DBSFile.listFileChildren
def listFileChildren(self, logical_file_name='', block_name='', block_id=0): """ required parameter: logical_file_name or block_name or block_id returns: logical_file_name, child_logical_file_name, parent_file_id """ conn = self.dbi.connection() try: if not logical_file_name and not block_name and not block_id: dbsExceptionHandler('dbsException-invalid-input',\ "Logical_file_name, block_id or block_name is required for listFileChildren api") sqlresult = self.filechildlist.execute(conn, logical_file_name, block_name, block_id) d = {} result = [] for i in range(len(sqlresult)): k = sqlresult[i]['logical_file_name'] v = sqlresult[i]['child_logical_file_name'] if k in d: d[k].append(v) else: d[k] = [v] for k, v in d.iteritems(): r = {'logical_file_name':k, 'child_logical_file_name': v} result.append(r) return result finally: if conn: conn.close()
python
def listFileChildren(self, logical_file_name='', block_name='', block_id=0): """ required parameter: logical_file_name or block_name or block_id returns: logical_file_name, child_logical_file_name, parent_file_id """ conn = self.dbi.connection() try: if not logical_file_name and not block_name and not block_id: dbsExceptionHandler('dbsException-invalid-input',\ "Logical_file_name, block_id or block_name is required for listFileChildren api") sqlresult = self.filechildlist.execute(conn, logical_file_name, block_name, block_id) d = {} result = [] for i in range(len(sqlresult)): k = sqlresult[i]['logical_file_name'] v = sqlresult[i]['child_logical_file_name'] if k in d: d[k].append(v) else: d[k] = [v] for k, v in d.iteritems(): r = {'logical_file_name':k, 'child_logical_file_name': v} result.append(r) return result finally: if conn: conn.close()
[ "def", "listFileChildren", "(", "self", ",", "logical_file_name", "=", "''", ",", "block_name", "=", "''", ",", "block_id", "=", "0", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "if", "not", "logical_file_name", "and", "not", "block_name", "and", "not", "block_id", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Logical_file_name, block_id or block_name is required for listFileChildren api\"", ")", "sqlresult", "=", "self", ".", "filechildlist", ".", "execute", "(", "conn", ",", "logical_file_name", ",", "block_name", ",", "block_id", ")", "d", "=", "{", "}", "result", "=", "[", "]", "for", "i", "in", "range", "(", "len", "(", "sqlresult", ")", ")", ":", "k", "=", "sqlresult", "[", "i", "]", "[", "'logical_file_name'", "]", "v", "=", "sqlresult", "[", "i", "]", "[", "'child_logical_file_name'", "]", "if", "k", "in", "d", ":", "d", "[", "k", "]", ".", "append", "(", "v", ")", "else", ":", "d", "[", "k", "]", "=", "[", "v", "]", "for", "k", ",", "v", "in", "d", ".", "iteritems", "(", ")", ":", "r", "=", "{", "'logical_file_name'", ":", "k", ",", "'child_logical_file_name'", ":", "v", "}", "result", ".", "append", "(", "r", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
required parameter: logical_file_name or block_name or block_id returns: logical_file_name, child_logical_file_name, parent_file_id
[ "required", "parameter", ":", "logical_file_name", "or", "block_name", "or", "block_id", "returns", ":", "logical_file_name", "child_logical_file_name", "parent_file_id" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSFile.py#L131-L157
train
dmwm/DBS
Server/Python/src/dbs/business/DBSFile.py
DBSFile.updateStatus
def updateStatus(self, logical_file_name, is_file_valid, lost, dataset): """ Used to toggle the status of a file from is_file_valid=1 (valid) to is_file_valid=0 (invalid) """ conn = self.dbi.connection() trans = conn.begin() try : self.updatestatus.execute(conn, logical_file_name, is_file_valid, lost, dataset, trans) trans.commit() trans = None except Exception as ex: if trans: trans.rollback() trans = None raise ex finally: if trans: trans.rollback() if conn: conn.close()
python
def updateStatus(self, logical_file_name, is_file_valid, lost, dataset): """ Used to toggle the status of a file from is_file_valid=1 (valid) to is_file_valid=0 (invalid) """ conn = self.dbi.connection() trans = conn.begin() try : self.updatestatus.execute(conn, logical_file_name, is_file_valid, lost, dataset, trans) trans.commit() trans = None except Exception as ex: if trans: trans.rollback() trans = None raise ex finally: if trans: trans.rollback() if conn: conn.close()
[ "def", "updateStatus", "(", "self", ",", "logical_file_name", ",", "is_file_valid", ",", "lost", ",", "dataset", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "trans", "=", "conn", ".", "begin", "(", ")", "try", ":", "self", ".", "updatestatus", ".", "execute", "(", "conn", ",", "logical_file_name", ",", "is_file_valid", ",", "lost", ",", "dataset", ",", "trans", ")", "trans", ".", "commit", "(", ")", "trans", "=", "None", "except", "Exception", "as", "ex", ":", "if", "trans", ":", "trans", ".", "rollback", "(", ")", "trans", "=", "None", "raise", "ex", "finally", ":", "if", "trans", ":", "trans", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Used to toggle the status of a file from is_file_valid=1 (valid) to is_file_valid=0 (invalid)
[ "Used", "to", "toggle", "the", "status", "of", "a", "file", "from", "is_file_valid", "=", "1", "(", "valid", ")", "to", "is_file_valid", "=", "0", "(", "invalid", ")" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSFile.py#L159-L180
train
dmwm/DBS
Server/Python/src/dbs/business/DBSFile.py
DBSFile.listFiles
def listFiles(self, dataset="", block_name="", logical_file_name="", release_version="", pset_hash="", app_name="", output_module_label="", run_num=-1, origin_site_name="", lumi_list=[], detail=False, validFileOnly=0, sumOverLumi=0, input_body=-1): """ One of below parameter groups must be present: non-patterned dataset, non-patterned block, non-patterned dataset with lfn, non-patterned block with lfn, non-patterned lfn non-patterned lfn list """ if input_body != -1 : try: logical_file_name = input_body.get("logical_file_name", "") run_num = input_body.get("run_num", -1) validFileOnly = input_body.get("validFileOnly", 0) sumOverLumi = input_body.get("sumOverLumi", 0) detail = input_body.get("detail", False) block_name = input_body.get("block_name", "") dataset = input_body.get("dataset", "") release_version = input_body.get("release_version", "") pset_hash = input_body.get("pset_hash", "") app_name = input_body.get("app_name", "") output_module_label = input_body.get("output_module_label", "") origin_site_name = input_body.get("origin_site_name", "") lumi_list = input_body.get("lumi_list", []) except cjson.DecodeError as de: msg = "business/listFilss POST call requires at least dataset, block_name, or a list of logical_file_name %s" % de dbsExceptionHandler('dbsException-invalid-input', "Invalid input", self.logger.exception, msg) if ('%' in block_name): dbsExceptionHandler('dbsException-invalid-input', "You must specify exact block name not a pattern", self.logger.exception) elif ('%' in dataset): print("***** in dataset name") dbsExceptionHandler('dbsException-invalid-input', " You must specify exact dataset name not a pattern", self.logger.exception) elif (not dataset and not block_name and (not logical_file_name or '%'in logical_file_name) ): dbsExceptionHandler('dbsException-invalid-input', """You must specify one of the parameter groups: \ non-pattern dataset, \ non-pattern block , non-pattern dataset with lfn ,\ non-pattern block with lfn or no-pattern lfn, \ non-patterned lfn list .""", self.logger.exception) elif (lumi_list and len(lumi_list) != 0): if run_num==-1: dbsExceptionHandler('dbsException-invalid-input', "Lumi list must accompany A single run number, \ use run_num=123", self.logger.exception) elif isinstance(run_num, basestring): try: run_num = int(run_num) except: dbsExceptionHandler('dbsException-invalid-input', "Lumi list must accompany A single run number,\ use run_num=123", self.logger.exception) elif isinstance(run_num, list): if len(run_num) == 1: try: run_num = int(run_num[0]) except: dbsExceptionHandler('dbsException-invalid-input', "Lumi list must accompany A single run number,\ use run_num=123", self.logger.exception) else: dbsExceptionHandler('dbsException-invalid-input', "Lumi list must accompany A single run number,\ use run_num=123", self.logger.exception) else: pass with self.dbi.connection() as conn: dao = (self.filebrieflist, self.filelist)[detail] for item in dao.execute(conn, dataset, block_name, logical_file_name, release_version, pset_hash, app_name, output_module_label, run_num, origin_site_name, lumi_list, validFileOnly, sumOverLumi): yield item # we need to yield while connection is open
python
def listFiles(self, dataset="", block_name="", logical_file_name="", release_version="", pset_hash="", app_name="", output_module_label="", run_num=-1, origin_site_name="", lumi_list=[], detail=False, validFileOnly=0, sumOverLumi=0, input_body=-1): """ One of below parameter groups must be present: non-patterned dataset, non-patterned block, non-patterned dataset with lfn, non-patterned block with lfn, non-patterned lfn non-patterned lfn list """ if input_body != -1 : try: logical_file_name = input_body.get("logical_file_name", "") run_num = input_body.get("run_num", -1) validFileOnly = input_body.get("validFileOnly", 0) sumOverLumi = input_body.get("sumOverLumi", 0) detail = input_body.get("detail", False) block_name = input_body.get("block_name", "") dataset = input_body.get("dataset", "") release_version = input_body.get("release_version", "") pset_hash = input_body.get("pset_hash", "") app_name = input_body.get("app_name", "") output_module_label = input_body.get("output_module_label", "") origin_site_name = input_body.get("origin_site_name", "") lumi_list = input_body.get("lumi_list", []) except cjson.DecodeError as de: msg = "business/listFilss POST call requires at least dataset, block_name, or a list of logical_file_name %s" % de dbsExceptionHandler('dbsException-invalid-input', "Invalid input", self.logger.exception, msg) if ('%' in block_name): dbsExceptionHandler('dbsException-invalid-input', "You must specify exact block name not a pattern", self.logger.exception) elif ('%' in dataset): print("***** in dataset name") dbsExceptionHandler('dbsException-invalid-input', " You must specify exact dataset name not a pattern", self.logger.exception) elif (not dataset and not block_name and (not logical_file_name or '%'in logical_file_name) ): dbsExceptionHandler('dbsException-invalid-input', """You must specify one of the parameter groups: \ non-pattern dataset, \ non-pattern block , non-pattern dataset with lfn ,\ non-pattern block with lfn or no-pattern lfn, \ non-patterned lfn list .""", self.logger.exception) elif (lumi_list and len(lumi_list) != 0): if run_num==-1: dbsExceptionHandler('dbsException-invalid-input', "Lumi list must accompany A single run number, \ use run_num=123", self.logger.exception) elif isinstance(run_num, basestring): try: run_num = int(run_num) except: dbsExceptionHandler('dbsException-invalid-input', "Lumi list must accompany A single run number,\ use run_num=123", self.logger.exception) elif isinstance(run_num, list): if len(run_num) == 1: try: run_num = int(run_num[0]) except: dbsExceptionHandler('dbsException-invalid-input', "Lumi list must accompany A single run number,\ use run_num=123", self.logger.exception) else: dbsExceptionHandler('dbsException-invalid-input', "Lumi list must accompany A single run number,\ use run_num=123", self.logger.exception) else: pass with self.dbi.connection() as conn: dao = (self.filebrieflist, self.filelist)[detail] for item in dao.execute(conn, dataset, block_name, logical_file_name, release_version, pset_hash, app_name, output_module_label, run_num, origin_site_name, lumi_list, validFileOnly, sumOverLumi): yield item # we need to yield while connection is open
[ "def", "listFiles", "(", "self", ",", "dataset", "=", "\"\"", ",", "block_name", "=", "\"\"", ",", "logical_file_name", "=", "\"\"", ",", "release_version", "=", "\"\"", ",", "pset_hash", "=", "\"\"", ",", "app_name", "=", "\"\"", ",", "output_module_label", "=", "\"\"", ",", "run_num", "=", "-", "1", ",", "origin_site_name", "=", "\"\"", ",", "lumi_list", "=", "[", "]", ",", "detail", "=", "False", ",", "validFileOnly", "=", "0", ",", "sumOverLumi", "=", "0", ",", "input_body", "=", "-", "1", ")", ":", "if", "input_body", "!=", "-", "1", ":", "try", ":", "logical_file_name", "=", "input_body", ".", "get", "(", "\"logical_file_name\"", ",", "\"\"", ")", "run_num", "=", "input_body", ".", "get", "(", "\"run_num\"", ",", "-", "1", ")", "validFileOnly", "=", "input_body", ".", "get", "(", "\"validFileOnly\"", ",", "0", ")", "sumOverLumi", "=", "input_body", ".", "get", "(", "\"sumOverLumi\"", ",", "0", ")", "detail", "=", "input_body", ".", "get", "(", "\"detail\"", ",", "False", ")", "block_name", "=", "input_body", ".", "get", "(", "\"block_name\"", ",", "\"\"", ")", "dataset", "=", "input_body", ".", "get", "(", "\"dataset\"", ",", "\"\"", ")", "release_version", "=", "input_body", ".", "get", "(", "\"release_version\"", ",", "\"\"", ")", "pset_hash", "=", "input_body", ".", "get", "(", "\"pset_hash\"", ",", "\"\"", ")", "app_name", "=", "input_body", ".", "get", "(", "\"app_name\"", ",", "\"\"", ")", "output_module_label", "=", "input_body", ".", "get", "(", "\"output_module_label\"", ",", "\"\"", ")", "origin_site_name", "=", "input_body", ".", "get", "(", "\"origin_site_name\"", ",", "\"\"", ")", "lumi_list", "=", "input_body", ".", "get", "(", "\"lumi_list\"", ",", "[", "]", ")", "except", "cjson", ".", "DecodeError", "as", "de", ":", "msg", "=", "\"business/listFilss POST call requires at least dataset, block_name, or a list of logical_file_name %s\"", "%", "de", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Invalid input\"", ",", "self", ".", "logger", ".", "exception", ",", "msg", ")", "if", "(", "'%'", "in", "block_name", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"You must specify exact block name not a pattern\"", ",", "self", ".", "logger", ".", "exception", ")", "elif", "(", "'%'", "in", "dataset", ")", ":", "print", "(", "\"***** in dataset name\"", ")", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\" You must specify exact dataset name not a pattern\"", ",", "self", ".", "logger", ".", "exception", ")", "elif", "(", "not", "dataset", "and", "not", "block_name", "and", "(", "not", "logical_file_name", "or", "'%'", "in", "logical_file_name", ")", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"\"\"You must specify one of the parameter groups: \\\n non-pattern dataset, \\\n non-pattern block , non-pattern dataset with lfn ,\\\n non-pattern block with lfn or no-pattern lfn, \\\n\t\t non-patterned lfn list .\"\"\"", ",", "self", ".", "logger", ".", "exception", ")", "elif", "(", "lumi_list", "and", "len", "(", "lumi_list", ")", "!=", "0", ")", ":", "if", "run_num", "==", "-", "1", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Lumi list must accompany A single run number, \\\n use run_num=123\"", ",", "self", ".", "logger", ".", "exception", ")", "elif", "isinstance", "(", "run_num", ",", "basestring", ")", ":", "try", ":", "run_num", "=", "int", "(", "run_num", ")", "except", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Lumi list must accompany A single run number,\\\n use run_num=123\"", ",", "self", ".", "logger", ".", "exception", ")", "elif", "isinstance", "(", "run_num", ",", "list", ")", ":", "if", "len", "(", "run_num", ")", "==", "1", ":", "try", ":", "run_num", "=", "int", "(", "run_num", "[", "0", "]", ")", "except", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Lumi list must accompany A single run number,\\\n use run_num=123\"", ",", "self", ".", "logger", ".", "exception", ")", "else", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Lumi list must accompany A single run number,\\\n use run_num=123\"", ",", "self", ".", "logger", ".", "exception", ")", "else", ":", "pass", "with", "self", ".", "dbi", ".", "connection", "(", ")", "as", "conn", ":", "dao", "=", "(", "self", ".", "filebrieflist", ",", "self", ".", "filelist", ")", "[", "detail", "]", "for", "item", "in", "dao", ".", "execute", "(", "conn", ",", "dataset", ",", "block_name", ",", "logical_file_name", ",", "release_version", ",", "pset_hash", ",", "app_name", ",", "output_module_label", ",", "run_num", ",", "origin_site_name", ",", "lumi_list", ",", "validFileOnly", ",", "sumOverLumi", ")", ":", "yield", "item", "# we need to yield while connection is open" ]
One of below parameter groups must be present: non-patterned dataset, non-patterned block, non-patterned dataset with lfn, non-patterned block with lfn, non-patterned lfn non-patterned lfn list
[ "One", "of", "below", "parameter", "groups", "must", "be", "present", ":", "non", "-", "patterned", "dataset", "non", "-", "patterned", "block", "non", "-", "patterned", "dataset", "with", "lfn", "non", "-", "patterned", "block", "with", "lfn", "non", "-", "patterned", "lfn", "non", "-", "patterned", "lfn", "list" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSFile.py#L182-L249
train
dmwm/DBS
Server/Python/src/dbs/business/DBSFile.py
DBSFile.insertFile
def insertFile(self, businput, qInserts=False): """ This method supports bulk insert of files performing other operations such as setting Block and Dataset parentages, setting mapping between OutputConfigModules and File(s) etc. :param qInserts: True means that inserts will be queued instead of done immediately. INSERT QUEUE Manager will perform the inserts, within few minutes. :type qInserts: bool :param logical_file_name (required) : string :param is_file_valid: (optional, default = 1): 1/0 :param block, required: /a/b/c#d :param dataset, required: /a/b/c :param file_type (optional, default = EDM): one of the predefined types, :param check_sum (optional): string :param event_count (optional, default = -1): int :param file_size (optional, default = -1.): float :param adler32 (optional): string :param md5 (optional): string :param auto_cross_section (optional, default = -1.): float :param file_lumi_list (optional, default = []): [{'run_num': 123, 'lumi_section_num': 12},{}....] :param file_parent_list(optional, default = []) :[{'file_parent_lfn': 'mylfn'},{}....] :param file_assoc_list(optional, default = []) :[{'file_parent_lfn': 'mylfn'},{}....] :param file_output_config_list(optional, default = []) : [{'app_name':..., 'release_version':..., 'pset_hash':...., output_module_label':...},{}.....] """ # We do not want to go be beyond 10 files at a time # If user wants to insert over 10 files in one shot, we run into risks of locking the database # tables for longer time, and in case of error, it will be hard to see where error occured if len(businput) > 10: dbsExceptionHandler('dbsException-input-too-large', "DBS cannot insert \ more than 10 files in one bulk call") return conn = self.dbi.connection() tran = conn.begin() try: #Now we are dealing with independent files that have different dataset/block and so on. #See Trac #358. #The expected input data format is a list of dictionary to insert independent files into DBS, #inputdata={'files':[{}, {}, {}]} #YG 09/15/2011 # AA- 01/06/2010 -- we have to do this file-by-file, there is no real good way to do this complex operation otherwise #files2insert = [] #fidl = [] fileInserted = False dataset = "" block_name = "" dataset_id = -1 block_id = -1 dsconfigs = [] for f in businput: if not ("logical_file_name" in f and "block_name" in f and "dataset" in f ): dbsExceptionHandler('dbsException-invalid-input', "DBSFile/insertFile must have logical_file_name, block_name and dataset as input") if f["block_name"].split('#')[0] != f["dataset"]: dbsExceptionHandler('dbsException-invalid-input', "DBSFile/insertFile: dataset and block_name NOT match") # first check if the dataset exists # and block exists that files are suppose to be going to and is OPEN for writing if dataset != f["dataset"]: dataset_id = self.datasetid.execute(conn, dataset=f["dataset"]) dataset = f["dataset"] if dataset_id == -1 : dbsExceptionHandler('dbsException-missing-data', "Required Dataset Not Found.", None, "Required Dataset %s does not exist"%f["dataset"] ) # get the list of configs in for this dataset dsconfigs = [x['output_mod_config_id'] for x in self.dsconfigids.execute(conn, dataset=f["dataset"])] fileconfigs = [] # this will hold file configs that we will list in the insert file logic below if block_name != f["block_name"]: block_info = self.blocklist.execute(conn, block_name=f["block_name"]) for b in block_info: if not b : dbsExceptionHandler( "dbsException-missing-data", "Required block not found", None, "Cannot found required block %s in DB" %f["block_name"]) else: if b["open_for_writing"] != 1 : dbsExceptionHandler("dbsException-conflict-data", "Block closed", None, "Block %s is not open for writting" %f["block_name"]) if "block_id" in b: block_id = b["block_id"] else: dbsExceptionHandler("dbsException-missing-data", "Block not found", None, "Cannot found required block %s in DB" %f["block_name"]) else: dbsExceptionHandler('dbsException-missing-data', "Required block name Not Found in input.", None, "Required block Not Found in input.") #make the default file_type=EDM file_type_id = self.ftypeid.execute( conn, f.get("file_type", "EDM")) if file_type_id == -1: dbsExceptionHandler('dbsException-missing-data', "File type not found.", None, "Required file type %s not found in DBS"%f.get("file_type", "EDM") ) iFile = 0 fileIncrement = 40 fID = self.sm.increment(conn, "SEQ_FL", incCount=fileIncrement) #looping over the files, everytime create a new object 'filein' as you never know #whats in the original object and we do not want to know #for f in businput: file_clob = {} fparents2insert = [] flumis2insert = [] fconfigs2insert = [] # create the file object from the original # taking care of defaults, and required filein = { "logical_file_name" : f["logical_file_name"], "is_file_valid" : f.get("is_file_valid", 1), "check_sum" : f.get("check_sum", None), "event_count" : f.get("event_count", -1), "file_size" : f.get("file_size", -1), "adler32" : f.get("adler32", None), "md5" : f.get("md5", None), "auto_cross_section" : f.get("auto_cross_section", -1), #"creation_date" : f.get("creation_date", None), See Ticket #965 YG. #"create_by": f.get("create_by", None), "last_modification_date": f.get("last_modification_date", None), #"last_modified_by" : f.get("last_modified_by", None) "last_modified_by" : dbsUtils().getCreateBy() } if filein["md5"] is None and filein["check_sum"] is None and filein["adler32"] is None: dbsExceptionHandler('dbsException-invalid-input', "Missing check_sum or adler32, or md5") if iFile == fileIncrement: fID = self.sm.increment(conn, "SEQ_FL", incCount=fileIncrement) iFile = 0 filein["file_id"] = fID + iFile iFile += 1 filein["dataset_id"] = dataset_id filein["block_id"] = block_id filein["file_type_id"] = file_type_id #FIXME: Add this later if f.get("branch_hash", "") not in ("", None): #filein["branch_hash"]=self.fbranchid.execute( f.get("branch_hash"), conn, transaction=tran) # insert file -- as decided, one file at a time # filein will be what goes into database try: if not qInserts: self.filein.execute(conn, filein, transaction=tran) fileInserted = True else: file_clob['file'] = filein except SQLAlchemyIntegrityError as ex: if str(ex).find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1: # Lets move on to NEXT file, we do not want to continue processing this file #Nothing about this file is updated when it is already in DB. No file parentage, block parentage, dataset parentage and so on. #Is this right? YG Oct. 24 self.logger.warning("DBSFile/insertFile. File already exists in DBS, not changing it: %s" %filein["logical_file_name"] ) continue else: raise #process file parents, file lumi, file outputmodconfigs, ... #file lumi sections if "file_lumi_list" in f: fllist = f["file_lumi_list"] if len(fllist) > 0: for fl in fllist: fldao = { "run_num" : fl["run_num"], "lumi_section_num" : fl["lumi_section_num"] } if "event_count" in fl: fldao["event_count"] = fl["event_count"] fldao["file_id"] = filein["file_id"] flumis2insert.append(fldao) if "file_parent_list" in f: #file parents fplist = f["file_parent_list"] for fp in fplist: fpdao = {} fpdao["this_file_id"] = filein["file_id"] fpdao["parent_logical_file_name"] = fp["file_parent_lfn"] fparents2insert.append(fpdao) if "file_output_config_list" in f: #file output config modules foutconfigs = f["file_output_config_list"] if(len(foutconfigs) > 0): for fc in foutconfigs: fcdao = {} fcdao["file_id"] = filein["file_id"] fcdao["output_mod_config_id"] = self.outconfigid.execute(conn, fc["app_name"], fc["release_version"], fc["pset_hash"], fc["output_module_label"], fc["global_tag"]) if fcdao["output_mod_config_id"] == -1 : dbsExceptionHandler('dbsException-missing-data', 'Config Not found.', None, "DBSFile/insertFile.\ Output module config (%s, %s, %s, %s) \ not found" % (fc["app_name"], fc["release_version"], fc["pset_hash"], fc["output_module_label"]) ) fileconfigs.append(fcdao["output_mod_config_id"]) fconfigs2insert.append(fcdao) #FIXME: file associations?-- in a later release # # insert file - lumi if flumis2insert: file_clob['file_lumi_list'] = flumis2insert if not qInserts: self.flumiin.execute(conn, flumis2insert, transaction=tran) # insert file parent mapping if fparents2insert: file_clob['file_parent_list'] = fparents2insert if not qInserts: self.fparentin.execute(conn, fparents2insert, transaction=tran) # First check to see if these output configs are mapped to THIS dataset as well, if not raise an exception if not set(fileconfigs).issubset(set(dsconfigs)) : dbsExceptionHandler('dbsException-conflict-data', 'Mismatched configure. ', None, "DBSFile/insertFile. Output configs mismatch, \ output configs known to dataset: \ %s are different from what are being mapped to file : %s " \ %(f["dataset"], filein["logical_file_name"]) ) # insert output module config mapping if fconfigs2insert: file_clob['file_output_config_list'] = fconfigs2insert if not qInserts: self.fconfigin.execute(conn, fconfigs2insert, transaction=tran) if qInserts: try: self.logger.warning(file_clob) self.filebufin.execute(conn, filein['logical_file_name'], block_id, file_clob, transaction=tran) except SQLAlchemyIntegrityError as ex: if str(ex).find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1: pass else: raise #insert block parentages and dataset parentages based on file parentages # Do this one by one, as it is sure to have duplicate in dest table if fileInserted and fparents2insert: for fp in fparents2insert: try: bkParentage2insert={'this_block_id' : filein["block_id"], 'parent_logical_file_name': fp['parent_logical_file_name']} self.blkparentin.execute(conn, bkParentage2insert, transaction=tran) dsParentage2insert={'this_dataset_id': filein["dataset_id"], 'parent_logical_file_name' : fp['parent_logical_file_name']} self.dsparentin.execute(conn, dsParentage2insert, transaction=tran) except SQLAlchemyIntegrityError as ex: #ORA-00001 if (str(ex).find("ORA-00001") != -1 and str(ex).find("PK_DP") != -1) or str(ex).find("PK_BP") != -1 or str(ex).lower().find("duplicate") != -1: pass elif str(ex).find("ORA-01400") != -1: raise else: raise # Update block parameters, file_count, block_size if not qInserts: blkParams = self.blkstats.execute(conn, block_id, transaction=tran) blkParams['block_size'] = long(blkParams['block_size']) self.blkstatsin.execute(conn, blkParams, transaction=tran) # All good ? tran.commit() tran = None except Exception as ex: if tran: tran.rollback() tran = None raise finally: if tran: tran.rollback() if conn: conn.close()
python
def insertFile(self, businput, qInserts=False): """ This method supports bulk insert of files performing other operations such as setting Block and Dataset parentages, setting mapping between OutputConfigModules and File(s) etc. :param qInserts: True means that inserts will be queued instead of done immediately. INSERT QUEUE Manager will perform the inserts, within few minutes. :type qInserts: bool :param logical_file_name (required) : string :param is_file_valid: (optional, default = 1): 1/0 :param block, required: /a/b/c#d :param dataset, required: /a/b/c :param file_type (optional, default = EDM): one of the predefined types, :param check_sum (optional): string :param event_count (optional, default = -1): int :param file_size (optional, default = -1.): float :param adler32 (optional): string :param md5 (optional): string :param auto_cross_section (optional, default = -1.): float :param file_lumi_list (optional, default = []): [{'run_num': 123, 'lumi_section_num': 12},{}....] :param file_parent_list(optional, default = []) :[{'file_parent_lfn': 'mylfn'},{}....] :param file_assoc_list(optional, default = []) :[{'file_parent_lfn': 'mylfn'},{}....] :param file_output_config_list(optional, default = []) : [{'app_name':..., 'release_version':..., 'pset_hash':...., output_module_label':...},{}.....] """ # We do not want to go be beyond 10 files at a time # If user wants to insert over 10 files in one shot, we run into risks of locking the database # tables for longer time, and in case of error, it will be hard to see where error occured if len(businput) > 10: dbsExceptionHandler('dbsException-input-too-large', "DBS cannot insert \ more than 10 files in one bulk call") return conn = self.dbi.connection() tran = conn.begin() try: #Now we are dealing with independent files that have different dataset/block and so on. #See Trac #358. #The expected input data format is a list of dictionary to insert independent files into DBS, #inputdata={'files':[{}, {}, {}]} #YG 09/15/2011 # AA- 01/06/2010 -- we have to do this file-by-file, there is no real good way to do this complex operation otherwise #files2insert = [] #fidl = [] fileInserted = False dataset = "" block_name = "" dataset_id = -1 block_id = -1 dsconfigs = [] for f in businput: if not ("logical_file_name" in f and "block_name" in f and "dataset" in f ): dbsExceptionHandler('dbsException-invalid-input', "DBSFile/insertFile must have logical_file_name, block_name and dataset as input") if f["block_name"].split('#')[0] != f["dataset"]: dbsExceptionHandler('dbsException-invalid-input', "DBSFile/insertFile: dataset and block_name NOT match") # first check if the dataset exists # and block exists that files are suppose to be going to and is OPEN for writing if dataset != f["dataset"]: dataset_id = self.datasetid.execute(conn, dataset=f["dataset"]) dataset = f["dataset"] if dataset_id == -1 : dbsExceptionHandler('dbsException-missing-data', "Required Dataset Not Found.", None, "Required Dataset %s does not exist"%f["dataset"] ) # get the list of configs in for this dataset dsconfigs = [x['output_mod_config_id'] for x in self.dsconfigids.execute(conn, dataset=f["dataset"])] fileconfigs = [] # this will hold file configs that we will list in the insert file logic below if block_name != f["block_name"]: block_info = self.blocklist.execute(conn, block_name=f["block_name"]) for b in block_info: if not b : dbsExceptionHandler( "dbsException-missing-data", "Required block not found", None, "Cannot found required block %s in DB" %f["block_name"]) else: if b["open_for_writing"] != 1 : dbsExceptionHandler("dbsException-conflict-data", "Block closed", None, "Block %s is not open for writting" %f["block_name"]) if "block_id" in b: block_id = b["block_id"] else: dbsExceptionHandler("dbsException-missing-data", "Block not found", None, "Cannot found required block %s in DB" %f["block_name"]) else: dbsExceptionHandler('dbsException-missing-data', "Required block name Not Found in input.", None, "Required block Not Found in input.") #make the default file_type=EDM file_type_id = self.ftypeid.execute( conn, f.get("file_type", "EDM")) if file_type_id == -1: dbsExceptionHandler('dbsException-missing-data', "File type not found.", None, "Required file type %s not found in DBS"%f.get("file_type", "EDM") ) iFile = 0 fileIncrement = 40 fID = self.sm.increment(conn, "SEQ_FL", incCount=fileIncrement) #looping over the files, everytime create a new object 'filein' as you never know #whats in the original object and we do not want to know #for f in businput: file_clob = {} fparents2insert = [] flumis2insert = [] fconfigs2insert = [] # create the file object from the original # taking care of defaults, and required filein = { "logical_file_name" : f["logical_file_name"], "is_file_valid" : f.get("is_file_valid", 1), "check_sum" : f.get("check_sum", None), "event_count" : f.get("event_count", -1), "file_size" : f.get("file_size", -1), "adler32" : f.get("adler32", None), "md5" : f.get("md5", None), "auto_cross_section" : f.get("auto_cross_section", -1), #"creation_date" : f.get("creation_date", None), See Ticket #965 YG. #"create_by": f.get("create_by", None), "last_modification_date": f.get("last_modification_date", None), #"last_modified_by" : f.get("last_modified_by", None) "last_modified_by" : dbsUtils().getCreateBy() } if filein["md5"] is None and filein["check_sum"] is None and filein["adler32"] is None: dbsExceptionHandler('dbsException-invalid-input', "Missing check_sum or adler32, or md5") if iFile == fileIncrement: fID = self.sm.increment(conn, "SEQ_FL", incCount=fileIncrement) iFile = 0 filein["file_id"] = fID + iFile iFile += 1 filein["dataset_id"] = dataset_id filein["block_id"] = block_id filein["file_type_id"] = file_type_id #FIXME: Add this later if f.get("branch_hash", "") not in ("", None): #filein["branch_hash"]=self.fbranchid.execute( f.get("branch_hash"), conn, transaction=tran) # insert file -- as decided, one file at a time # filein will be what goes into database try: if not qInserts: self.filein.execute(conn, filein, transaction=tran) fileInserted = True else: file_clob['file'] = filein except SQLAlchemyIntegrityError as ex: if str(ex).find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1: # Lets move on to NEXT file, we do not want to continue processing this file #Nothing about this file is updated when it is already in DB. No file parentage, block parentage, dataset parentage and so on. #Is this right? YG Oct. 24 self.logger.warning("DBSFile/insertFile. File already exists in DBS, not changing it: %s" %filein["logical_file_name"] ) continue else: raise #process file parents, file lumi, file outputmodconfigs, ... #file lumi sections if "file_lumi_list" in f: fllist = f["file_lumi_list"] if len(fllist) > 0: for fl in fllist: fldao = { "run_num" : fl["run_num"], "lumi_section_num" : fl["lumi_section_num"] } if "event_count" in fl: fldao["event_count"] = fl["event_count"] fldao["file_id"] = filein["file_id"] flumis2insert.append(fldao) if "file_parent_list" in f: #file parents fplist = f["file_parent_list"] for fp in fplist: fpdao = {} fpdao["this_file_id"] = filein["file_id"] fpdao["parent_logical_file_name"] = fp["file_parent_lfn"] fparents2insert.append(fpdao) if "file_output_config_list" in f: #file output config modules foutconfigs = f["file_output_config_list"] if(len(foutconfigs) > 0): for fc in foutconfigs: fcdao = {} fcdao["file_id"] = filein["file_id"] fcdao["output_mod_config_id"] = self.outconfigid.execute(conn, fc["app_name"], fc["release_version"], fc["pset_hash"], fc["output_module_label"], fc["global_tag"]) if fcdao["output_mod_config_id"] == -1 : dbsExceptionHandler('dbsException-missing-data', 'Config Not found.', None, "DBSFile/insertFile.\ Output module config (%s, %s, %s, %s) \ not found" % (fc["app_name"], fc["release_version"], fc["pset_hash"], fc["output_module_label"]) ) fileconfigs.append(fcdao["output_mod_config_id"]) fconfigs2insert.append(fcdao) #FIXME: file associations?-- in a later release # # insert file - lumi if flumis2insert: file_clob['file_lumi_list'] = flumis2insert if not qInserts: self.flumiin.execute(conn, flumis2insert, transaction=tran) # insert file parent mapping if fparents2insert: file_clob['file_parent_list'] = fparents2insert if not qInserts: self.fparentin.execute(conn, fparents2insert, transaction=tran) # First check to see if these output configs are mapped to THIS dataset as well, if not raise an exception if not set(fileconfigs).issubset(set(dsconfigs)) : dbsExceptionHandler('dbsException-conflict-data', 'Mismatched configure. ', None, "DBSFile/insertFile. Output configs mismatch, \ output configs known to dataset: \ %s are different from what are being mapped to file : %s " \ %(f["dataset"], filein["logical_file_name"]) ) # insert output module config mapping if fconfigs2insert: file_clob['file_output_config_list'] = fconfigs2insert if not qInserts: self.fconfigin.execute(conn, fconfigs2insert, transaction=tran) if qInserts: try: self.logger.warning(file_clob) self.filebufin.execute(conn, filein['logical_file_name'], block_id, file_clob, transaction=tran) except SQLAlchemyIntegrityError as ex: if str(ex).find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1: pass else: raise #insert block parentages and dataset parentages based on file parentages # Do this one by one, as it is sure to have duplicate in dest table if fileInserted and fparents2insert: for fp in fparents2insert: try: bkParentage2insert={'this_block_id' : filein["block_id"], 'parent_logical_file_name': fp['parent_logical_file_name']} self.blkparentin.execute(conn, bkParentage2insert, transaction=tran) dsParentage2insert={'this_dataset_id': filein["dataset_id"], 'parent_logical_file_name' : fp['parent_logical_file_name']} self.dsparentin.execute(conn, dsParentage2insert, transaction=tran) except SQLAlchemyIntegrityError as ex: #ORA-00001 if (str(ex).find("ORA-00001") != -1 and str(ex).find("PK_DP") != -1) or str(ex).find("PK_BP") != -1 or str(ex).lower().find("duplicate") != -1: pass elif str(ex).find("ORA-01400") != -1: raise else: raise # Update block parameters, file_count, block_size if not qInserts: blkParams = self.blkstats.execute(conn, block_id, transaction=tran) blkParams['block_size'] = long(blkParams['block_size']) self.blkstatsin.execute(conn, blkParams, transaction=tran) # All good ? tran.commit() tran = None except Exception as ex: if tran: tran.rollback() tran = None raise finally: if tran: tran.rollback() if conn: conn.close()
[ "def", "insertFile", "(", "self", ",", "businput", ",", "qInserts", "=", "False", ")", ":", "# We do not want to go be beyond 10 files at a time", "# If user wants to insert over 10 files in one shot, we run into risks of locking the database", "# tables for longer time, and in case of error, it will be hard to see where error occured", "if", "len", "(", "businput", ")", ">", "10", ":", "dbsExceptionHandler", "(", "'dbsException-input-too-large'", ",", "\"DBS cannot insert \\\n more than 10 files in one bulk call\"", ")", "return", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "conn", ".", "begin", "(", ")", "try", ":", "#Now we are dealing with independent files that have different dataset/block and so on.", "#See Trac #358.", "#The expected input data format is a list of dictionary to insert independent files into DBS,", "#inputdata={'files':[{}, {}, {}]}", "#YG 09/15/2011", "# AA- 01/06/2010 -- we have to do this file-by-file, there is no real good way to do this complex operation otherwise", "#files2insert = []", "#fidl = []", "fileInserted", "=", "False", "dataset", "=", "\"\"", "block_name", "=", "\"\"", "dataset_id", "=", "-", "1", "block_id", "=", "-", "1", "dsconfigs", "=", "[", "]", "for", "f", "in", "businput", ":", "if", "not", "(", "\"logical_file_name\"", "in", "f", "and", "\"block_name\"", "in", "f", "and", "\"dataset\"", "in", "f", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"DBSFile/insertFile must have logical_file_name, block_name and dataset as input\"", ")", "if", "f", "[", "\"block_name\"", "]", ".", "split", "(", "'#'", ")", "[", "0", "]", "!=", "f", "[", "\"dataset\"", "]", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"DBSFile/insertFile: dataset and block_name NOT match\"", ")", "# first check if the dataset exists", "# and block exists that files are suppose to be going to and is OPEN for writing", "if", "dataset", "!=", "f", "[", "\"dataset\"", "]", ":", "dataset_id", "=", "self", ".", "datasetid", ".", "execute", "(", "conn", ",", "dataset", "=", "f", "[", "\"dataset\"", "]", ")", "dataset", "=", "f", "[", "\"dataset\"", "]", "if", "dataset_id", "==", "-", "1", ":", "dbsExceptionHandler", "(", "'dbsException-missing-data'", ",", "\"Required Dataset Not Found.\"", ",", "None", ",", "\"Required Dataset %s does not exist\"", "%", "f", "[", "\"dataset\"", "]", ")", "# get the list of configs in for this dataset", "dsconfigs", "=", "[", "x", "[", "'output_mod_config_id'", "]", "for", "x", "in", "self", ".", "dsconfigids", ".", "execute", "(", "conn", ",", "dataset", "=", "f", "[", "\"dataset\"", "]", ")", "]", "fileconfigs", "=", "[", "]", "# this will hold file configs that we will list in the insert file logic below", "if", "block_name", "!=", "f", "[", "\"block_name\"", "]", ":", "block_info", "=", "self", ".", "blocklist", ".", "execute", "(", "conn", ",", "block_name", "=", "f", "[", "\"block_name\"", "]", ")", "for", "b", "in", "block_info", ":", "if", "not", "b", ":", "dbsExceptionHandler", "(", "\"dbsException-missing-data\"", ",", "\"Required block not found\"", ",", "None", ",", "\"Cannot found required block %s in DB\"", "%", "f", "[", "\"block_name\"", "]", ")", "else", ":", "if", "b", "[", "\"open_for_writing\"", "]", "!=", "1", ":", "dbsExceptionHandler", "(", "\"dbsException-conflict-data\"", ",", "\"Block closed\"", ",", "None", ",", "\"Block %s is not open for writting\"", "%", "f", "[", "\"block_name\"", "]", ")", "if", "\"block_id\"", "in", "b", ":", "block_id", "=", "b", "[", "\"block_id\"", "]", "else", ":", "dbsExceptionHandler", "(", "\"dbsException-missing-data\"", ",", "\"Block not found\"", ",", "None", ",", "\"Cannot found required block %s in DB\"", "%", "f", "[", "\"block_name\"", "]", ")", "else", ":", "dbsExceptionHandler", "(", "'dbsException-missing-data'", ",", "\"Required block name Not Found in input.\"", ",", "None", ",", "\"Required block Not Found in input.\"", ")", "#make the default file_type=EDM", "file_type_id", "=", "self", ".", "ftypeid", ".", "execute", "(", "conn", ",", "f", ".", "get", "(", "\"file_type\"", ",", "\"EDM\"", ")", ")", "if", "file_type_id", "==", "-", "1", ":", "dbsExceptionHandler", "(", "'dbsException-missing-data'", ",", "\"File type not found.\"", ",", "None", ",", "\"Required file type %s not found in DBS\"", "%", "f", ".", "get", "(", "\"file_type\"", ",", "\"EDM\"", ")", ")", "iFile", "=", "0", "fileIncrement", "=", "40", "fID", "=", "self", ".", "sm", ".", "increment", "(", "conn", ",", "\"SEQ_FL\"", ",", "incCount", "=", "fileIncrement", ")", "#looping over the files, everytime create a new object 'filein' as you never know", "#whats in the original object and we do not want to know", "#for f in businput:", "file_clob", "=", "{", "}", "fparents2insert", "=", "[", "]", "flumis2insert", "=", "[", "]", "fconfigs2insert", "=", "[", "]", "# create the file object from the original", "# taking care of defaults, and required", "filein", "=", "{", "\"logical_file_name\"", ":", "f", "[", "\"logical_file_name\"", "]", ",", "\"is_file_valid\"", ":", "f", ".", "get", "(", "\"is_file_valid\"", ",", "1", ")", ",", "\"check_sum\"", ":", "f", ".", "get", "(", "\"check_sum\"", ",", "None", ")", ",", "\"event_count\"", ":", "f", ".", "get", "(", "\"event_count\"", ",", "-", "1", ")", ",", "\"file_size\"", ":", "f", ".", "get", "(", "\"file_size\"", ",", "-", "1", ")", ",", "\"adler32\"", ":", "f", ".", "get", "(", "\"adler32\"", ",", "None", ")", ",", "\"md5\"", ":", "f", ".", "get", "(", "\"md5\"", ",", "None", ")", ",", "\"auto_cross_section\"", ":", "f", ".", "get", "(", "\"auto_cross_section\"", ",", "-", "1", ")", ",", "#\"creation_date\" : f.get(\"creation_date\", None), See Ticket #965 YG.", "#\"create_by\": f.get(\"create_by\", None),", "\"last_modification_date\"", ":", "f", ".", "get", "(", "\"last_modification_date\"", ",", "None", ")", ",", "#\"last_modified_by\" : f.get(\"last_modified_by\", None)", "\"last_modified_by\"", ":", "dbsUtils", "(", ")", ".", "getCreateBy", "(", ")", "}", "if", "filein", "[", "\"md5\"", "]", "is", "None", "and", "filein", "[", "\"check_sum\"", "]", "is", "None", "and", "filein", "[", "\"adler32\"", "]", "is", "None", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Missing check_sum or adler32, or md5\"", ")", "if", "iFile", "==", "fileIncrement", ":", "fID", "=", "self", ".", "sm", ".", "increment", "(", "conn", ",", "\"SEQ_FL\"", ",", "incCount", "=", "fileIncrement", ")", "iFile", "=", "0", "filein", "[", "\"file_id\"", "]", "=", "fID", "+", "iFile", "iFile", "+=", "1", "filein", "[", "\"dataset_id\"", "]", "=", "dataset_id", "filein", "[", "\"block_id\"", "]", "=", "block_id", "filein", "[", "\"file_type_id\"", "]", "=", "file_type_id", "#FIXME: Add this later if f.get(\"branch_hash\", \"\") not in (\"\", None):", "#filein[\"branch_hash\"]=self.fbranchid.execute( f.get(\"branch_hash\"), conn, transaction=tran)", "# insert file -- as decided, one file at a time", "# filein will be what goes into database", "try", ":", "if", "not", "qInserts", ":", "self", ".", "filein", ".", "execute", "(", "conn", ",", "filein", ",", "transaction", "=", "tran", ")", "fileInserted", "=", "True", "else", ":", "file_clob", "[", "'file'", "]", "=", "filein", "except", "SQLAlchemyIntegrityError", "as", "ex", ":", "if", "str", "(", "ex", ")", ".", "find", "(", "\"unique constraint\"", ")", "!=", "-", "1", "or", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"duplicate\"", ")", "!=", "-", "1", ":", "# Lets move on to NEXT file, we do not want to continue processing this file", "#Nothing about this file is updated when it is already in DB. No file parentage, block parentage, dataset parentage and so on.", "#Is this right? YG Oct. 24", "self", ".", "logger", ".", "warning", "(", "\"DBSFile/insertFile. File already exists in DBS, not changing it: %s\"", "%", "filein", "[", "\"logical_file_name\"", "]", ")", "continue", "else", ":", "raise", "#process file parents, file lumi, file outputmodconfigs, ...", "#file lumi sections", "if", "\"file_lumi_list\"", "in", "f", ":", "fllist", "=", "f", "[", "\"file_lumi_list\"", "]", "if", "len", "(", "fllist", ")", ">", "0", ":", "for", "fl", "in", "fllist", ":", "fldao", "=", "{", "\"run_num\"", ":", "fl", "[", "\"run_num\"", "]", ",", "\"lumi_section_num\"", ":", "fl", "[", "\"lumi_section_num\"", "]", "}", "if", "\"event_count\"", "in", "fl", ":", "fldao", "[", "\"event_count\"", "]", "=", "fl", "[", "\"event_count\"", "]", "fldao", "[", "\"file_id\"", "]", "=", "filein", "[", "\"file_id\"", "]", "flumis2insert", ".", "append", "(", "fldao", ")", "if", "\"file_parent_list\"", "in", "f", ":", "#file parents", "fplist", "=", "f", "[", "\"file_parent_list\"", "]", "for", "fp", "in", "fplist", ":", "fpdao", "=", "{", "}", "fpdao", "[", "\"this_file_id\"", "]", "=", "filein", "[", "\"file_id\"", "]", "fpdao", "[", "\"parent_logical_file_name\"", "]", "=", "fp", "[", "\"file_parent_lfn\"", "]", "fparents2insert", ".", "append", "(", "fpdao", ")", "if", "\"file_output_config_list\"", "in", "f", ":", "#file output config modules", "foutconfigs", "=", "f", "[", "\"file_output_config_list\"", "]", "if", "(", "len", "(", "foutconfigs", ")", ">", "0", ")", ":", "for", "fc", "in", "foutconfigs", ":", "fcdao", "=", "{", "}", "fcdao", "[", "\"file_id\"", "]", "=", "filein", "[", "\"file_id\"", "]", "fcdao", "[", "\"output_mod_config_id\"", "]", "=", "self", ".", "outconfigid", ".", "execute", "(", "conn", ",", "fc", "[", "\"app_name\"", "]", ",", "fc", "[", "\"release_version\"", "]", ",", "fc", "[", "\"pset_hash\"", "]", ",", "fc", "[", "\"output_module_label\"", "]", ",", "fc", "[", "\"global_tag\"", "]", ")", "if", "fcdao", "[", "\"output_mod_config_id\"", "]", "==", "-", "1", ":", "dbsExceptionHandler", "(", "'dbsException-missing-data'", ",", "'Config Not found.'", ",", "None", ",", "\"DBSFile/insertFile.\\\n Output module config (%s, %s, %s, %s) \\\n not found\"", "%", "(", "fc", "[", "\"app_name\"", "]", ",", "fc", "[", "\"release_version\"", "]", ",", "fc", "[", "\"pset_hash\"", "]", ",", "fc", "[", "\"output_module_label\"", "]", ")", ")", "fileconfigs", ".", "append", "(", "fcdao", "[", "\"output_mod_config_id\"", "]", ")", "fconfigs2insert", ".", "append", "(", "fcdao", ")", "#FIXME: file associations?-- in a later release", "#", "# insert file - lumi", "if", "flumis2insert", ":", "file_clob", "[", "'file_lumi_list'", "]", "=", "flumis2insert", "if", "not", "qInserts", ":", "self", ".", "flumiin", ".", "execute", "(", "conn", ",", "flumis2insert", ",", "transaction", "=", "tran", ")", "# insert file parent mapping", "if", "fparents2insert", ":", "file_clob", "[", "'file_parent_list'", "]", "=", "fparents2insert", "if", "not", "qInserts", ":", "self", ".", "fparentin", ".", "execute", "(", "conn", ",", "fparents2insert", ",", "transaction", "=", "tran", ")", "# First check to see if these output configs are mapped to THIS dataset as well, if not raise an exception", "if", "not", "set", "(", "fileconfigs", ")", ".", "issubset", "(", "set", "(", "dsconfigs", ")", ")", ":", "dbsExceptionHandler", "(", "'dbsException-conflict-data'", ",", "'Mismatched configure. '", ",", "None", ",", "\"DBSFile/insertFile. Output configs mismatch, \\\n output configs known to dataset: \\\n %s are different from what are being mapped to file : %s \"", "%", "(", "f", "[", "\"dataset\"", "]", ",", "filein", "[", "\"logical_file_name\"", "]", ")", ")", "# insert output module config mapping", "if", "fconfigs2insert", ":", "file_clob", "[", "'file_output_config_list'", "]", "=", "fconfigs2insert", "if", "not", "qInserts", ":", "self", ".", "fconfigin", ".", "execute", "(", "conn", ",", "fconfigs2insert", ",", "transaction", "=", "tran", ")", "if", "qInserts", ":", "try", ":", "self", ".", "logger", ".", "warning", "(", "file_clob", ")", "self", ".", "filebufin", ".", "execute", "(", "conn", ",", "filein", "[", "'logical_file_name'", "]", ",", "block_id", ",", "file_clob", ",", "transaction", "=", "tran", ")", "except", "SQLAlchemyIntegrityError", "as", "ex", ":", "if", "str", "(", "ex", ")", ".", "find", "(", "\"unique constraint\"", ")", "!=", "-", "1", "or", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"duplicate\"", ")", "!=", "-", "1", ":", "pass", "else", ":", "raise", "#insert block parentages and dataset parentages based on file parentages", "# Do this one by one, as it is sure to have duplicate in dest table", "if", "fileInserted", "and", "fparents2insert", ":", "for", "fp", "in", "fparents2insert", ":", "try", ":", "bkParentage2insert", "=", "{", "'this_block_id'", ":", "filein", "[", "\"block_id\"", "]", ",", "'parent_logical_file_name'", ":", "fp", "[", "'parent_logical_file_name'", "]", "}", "self", ".", "blkparentin", ".", "execute", "(", "conn", ",", "bkParentage2insert", ",", "transaction", "=", "tran", ")", "dsParentage2insert", "=", "{", "'this_dataset_id'", ":", "filein", "[", "\"dataset_id\"", "]", ",", "'parent_logical_file_name'", ":", "fp", "[", "'parent_logical_file_name'", "]", "}", "self", ".", "dsparentin", ".", "execute", "(", "conn", ",", "dsParentage2insert", ",", "transaction", "=", "tran", ")", "except", "SQLAlchemyIntegrityError", "as", "ex", ":", "#ORA-00001", "if", "(", "str", "(", "ex", ")", ".", "find", "(", "\"ORA-00001\"", ")", "!=", "-", "1", "and", "str", "(", "ex", ")", ".", "find", "(", "\"PK_DP\"", ")", "!=", "-", "1", ")", "or", "str", "(", "ex", ")", ".", "find", "(", "\"PK_BP\"", ")", "!=", "-", "1", "or", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"duplicate\"", ")", "!=", "-", "1", ":", "pass", "elif", "str", "(", "ex", ")", ".", "find", "(", "\"ORA-01400\"", ")", "!=", "-", "1", ":", "raise", "else", ":", "raise", "# Update block parameters, file_count, block_size", "if", "not", "qInserts", ":", "blkParams", "=", "self", ".", "blkstats", ".", "execute", "(", "conn", ",", "block_id", ",", "transaction", "=", "tran", ")", "blkParams", "[", "'block_size'", "]", "=", "long", "(", "blkParams", "[", "'block_size'", "]", ")", "self", ".", "blkstatsin", ".", "execute", "(", "conn", ",", "blkParams", ",", "transaction", "=", "tran", ")", "# All good ?", "tran", ".", "commit", "(", ")", "tran", "=", "None", "except", "Exception", "as", "ex", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "tran", "=", "None", "raise", "finally", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
This method supports bulk insert of files performing other operations such as setting Block and Dataset parentages, setting mapping between OutputConfigModules and File(s) etc. :param qInserts: True means that inserts will be queued instead of done immediately. INSERT QUEUE Manager will perform the inserts, within few minutes. :type qInserts: bool :param logical_file_name (required) : string :param is_file_valid: (optional, default = 1): 1/0 :param block, required: /a/b/c#d :param dataset, required: /a/b/c :param file_type (optional, default = EDM): one of the predefined types, :param check_sum (optional): string :param event_count (optional, default = -1): int :param file_size (optional, default = -1.): float :param adler32 (optional): string :param md5 (optional): string :param auto_cross_section (optional, default = -1.): float :param file_lumi_list (optional, default = []): [{'run_num': 123, 'lumi_section_num': 12},{}....] :param file_parent_list(optional, default = []) :[{'file_parent_lfn': 'mylfn'},{}....] :param file_assoc_list(optional, default = []) :[{'file_parent_lfn': 'mylfn'},{}....] :param file_output_config_list(optional, default = []) : [{'app_name':..., 'release_version':..., 'pset_hash':...., output_module_label':...},{}.....]
[ "This", "method", "supports", "bulk", "insert", "of", "files", "performing", "other", "operations", "such", "as", "setting", "Block", "and", "Dataset", "parentages", "setting", "mapping", "between", "OutputConfigModules", "and", "File", "(", "s", ")", "etc", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSFile.py#L250-L513
train
dmwm/DBS
Server/Python/src/dbs/business/DBSFile.py
DBSFile.insertFileParents
def insertFileParents(self, businput): """ This is a special function for WMAgent only. input block_name: is a child block name. input chils_parent_id_list: is a list of file id of child, parent pair: [[cid1, pid1],[cid2,pid2],[cid3,pid3],...] The requirment for this API is 1. All the child files belong to the block. 2. All the child-parent pairs are not already in DBS. 3. The dataset parentage is already in DBS. We will fill the block parentage here using the file parentage info. Y. Guo July 18, 2018 """ if "block_name" not in businput.keys() or "child_parent_id_list" not in businput.keys() or not businput["child_parent_id_list"] or not businput["block_name"]: dbsExceptionHandler("dbsException-invalid-input2", "DBSFile/insertFileParents: require child block_name and list of child/parent file id pairs" , self.logger.exception, "DBSFile/insertFileParents: require child block_name and list of child/parent file id pairs") tran = None conn = None try: #We should get clean insert for both file/block parentage. #block parent duplication is handled at dao level. File parent should not have deplication. conn = self.dbi.connection() tran = conn.begin() self.logger.info("Insert File parentage mapping") self.fparentin2.execute(conn, businput, tran) self.logger.info("Insert block parentage mapping") self.blkparentin3.execute(conn, businput, tran) if tran:tran.commit() if conn:conn.close() except SQLAlchemyIntegrityError as ex: if tran:tran.rollback() if conn:conn.close() if str(ex).find("ORA-01400") > -1: dbsExceptionHandler('dbsException-missing-data', 'Missing data when insert filei/block parent. ', self.logger.exception, 'Missing data when insert file/block parent. '+ str(ex)) else: dbsExceptionHandler('dbsException-invalid-input2', 'Invalid data when insert file/block parent. ', self.logger.exception, 'Invalid data when insert file/block parent. '+ str(ex)) finally: if tran:tran.rollback() if conn:conn.close()
python
def insertFileParents(self, businput): """ This is a special function for WMAgent only. input block_name: is a child block name. input chils_parent_id_list: is a list of file id of child, parent pair: [[cid1, pid1],[cid2,pid2],[cid3,pid3],...] The requirment for this API is 1. All the child files belong to the block. 2. All the child-parent pairs are not already in DBS. 3. The dataset parentage is already in DBS. We will fill the block parentage here using the file parentage info. Y. Guo July 18, 2018 """ if "block_name" not in businput.keys() or "child_parent_id_list" not in businput.keys() or not businput["child_parent_id_list"] or not businput["block_name"]: dbsExceptionHandler("dbsException-invalid-input2", "DBSFile/insertFileParents: require child block_name and list of child/parent file id pairs" , self.logger.exception, "DBSFile/insertFileParents: require child block_name and list of child/parent file id pairs") tran = None conn = None try: #We should get clean insert for both file/block parentage. #block parent duplication is handled at dao level. File parent should not have deplication. conn = self.dbi.connection() tran = conn.begin() self.logger.info("Insert File parentage mapping") self.fparentin2.execute(conn, businput, tran) self.logger.info("Insert block parentage mapping") self.blkparentin3.execute(conn, businput, tran) if tran:tran.commit() if conn:conn.close() except SQLAlchemyIntegrityError as ex: if tran:tran.rollback() if conn:conn.close() if str(ex).find("ORA-01400") > -1: dbsExceptionHandler('dbsException-missing-data', 'Missing data when insert filei/block parent. ', self.logger.exception, 'Missing data when insert file/block parent. '+ str(ex)) else: dbsExceptionHandler('dbsException-invalid-input2', 'Invalid data when insert file/block parent. ', self.logger.exception, 'Invalid data when insert file/block parent. '+ str(ex)) finally: if tran:tran.rollback() if conn:conn.close()
[ "def", "insertFileParents", "(", "self", ",", "businput", ")", ":", "if", "\"block_name\"", "not", "in", "businput", ".", "keys", "(", ")", "or", "\"child_parent_id_list\"", "not", "in", "businput", ".", "keys", "(", ")", "or", "not", "businput", "[", "\"child_parent_id_list\"", "]", "or", "not", "businput", "[", "\"block_name\"", "]", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input2\"", ",", "\"DBSFile/insertFileParents: require child block_name and list of child/parent file id pairs\"", ",", "self", ".", "logger", ".", "exception", ",", "\"DBSFile/insertFileParents: require child block_name and list of child/parent file id pairs\"", ")", "tran", "=", "None", "conn", "=", "None", "try", ":", "#We should get clean insert for both file/block parentage.", "#block parent duplication is handled at dao level. File parent should not have deplication. ", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "conn", ".", "begin", "(", ")", "self", ".", "logger", ".", "info", "(", "\"Insert File parentage mapping\"", ")", "self", ".", "fparentin2", ".", "execute", "(", "conn", ",", "businput", ",", "tran", ")", "self", ".", "logger", ".", "info", "(", "\"Insert block parentage mapping\"", ")", "self", ".", "blkparentin3", ".", "execute", "(", "conn", ",", "businput", ",", "tran", ")", "if", "tran", ":", "tran", ".", "commit", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")", "except", "SQLAlchemyIntegrityError", "as", "ex", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")", "if", "str", "(", "ex", ")", ".", "find", "(", "\"ORA-01400\"", ")", ">", "-", "1", ":", "dbsExceptionHandler", "(", "'dbsException-missing-data'", ",", "'Missing data when insert filei/block parent. '", ",", "self", ".", "logger", ".", "exception", ",", "'Missing data when insert file/block parent. '", "+", "str", "(", "ex", ")", ")", "else", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "'Invalid data when insert file/block parent. '", ",", "self", ".", "logger", ".", "exception", ",", "'Invalid data when insert file/block parent. '", "+", "str", "(", "ex", ")", ")", "finally", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
This is a special function for WMAgent only. input block_name: is a child block name. input chils_parent_id_list: is a list of file id of child, parent pair: [[cid1, pid1],[cid2,pid2],[cid3,pid3],...] The requirment for this API is 1. All the child files belong to the block. 2. All the child-parent pairs are not already in DBS. 3. The dataset parentage is already in DBS. We will fill the block parentage here using the file parentage info. Y. Guo July 18, 2018
[ "This", "is", "a", "special", "function", "for", "WMAgent", "only", ".", "input", "block_name", ":", "is", "a", "child", "block", "name", ".", "input", "chils_parent_id_list", ":", "is", "a", "list", "of", "file", "id", "of", "child", "parent", "pair", ":", "[[", "cid1", "pid1", "]", "[", "cid2", "pid2", "]", "[", "cid3", "pid3", "]", "...", "]", "The", "requirment", "for", "this", "API", "is", "1", ".", "All", "the", "child", "files", "belong", "to", "the", "block", ".", "2", ".", "All", "the", "child", "-", "parent", "pairs", "are", "not", "already", "in", "DBS", ".", "3", ".", "The", "dataset", "parentage", "is", "already", "in", "DBS", ".", "We", "will", "fill", "the", "block", "parentage", "here", "using", "the", "file", "parentage", "info", ".", "Y", ".", "Guo", "July", "18", "2018" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSFile.py#L515-L557
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/SequenceManager.py
SequenceManager.increment
def increment(self, conn, seqName, transaction = False, incCount=1): """ increments the sequence `seqName` by default `Incremented by` and returns its value incCount: is UNUSED variable in Oracle implementation """ #FIXME: Do we need to lock the tables here? sql = "select %s%s.nextval as val from dual" % (self.owner, seqName) result = self.dbi.processData(sql, conn=conn, transaction=transaction) resultlist = self.formatDict(result) return resultlist[0]['val']
python
def increment(self, conn, seqName, transaction = False, incCount=1): """ increments the sequence `seqName` by default `Incremented by` and returns its value incCount: is UNUSED variable in Oracle implementation """ #FIXME: Do we need to lock the tables here? sql = "select %s%s.nextval as val from dual" % (self.owner, seqName) result = self.dbi.processData(sql, conn=conn, transaction=transaction) resultlist = self.formatDict(result) return resultlist[0]['val']
[ "def", "increment", "(", "self", ",", "conn", ",", "seqName", ",", "transaction", "=", "False", ",", "incCount", "=", "1", ")", ":", "#FIXME: Do we need to lock the tables here?", "sql", "=", "\"select %s%s.nextval as val from dual\"", "%", "(", "self", ".", "owner", ",", "seqName", ")", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "conn", "=", "conn", ",", "transaction", "=", "transaction", ")", "resultlist", "=", "self", ".", "formatDict", "(", "result", ")", "return", "resultlist", "[", "0", "]", "[", "'val'", "]" ]
increments the sequence `seqName` by default `Incremented by` and returns its value incCount: is UNUSED variable in Oracle implementation
[ "increments", "the", "sequence", "seqName", "by", "default", "Incremented", "by", "and", "returns", "its", "value", "incCount", ":", "is", "UNUSED", "variable", "in", "Oracle", "implementation" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/SequenceManager.py#L17-L29
train
dmwm/DBS
Server/Python/src/dbs/business/DBSReleaseVersion.py
DBSReleaseVersion.listReleaseVersions
def listReleaseVersions(self, release_version="", dataset='', logical_file_name=''): """ List release versions """ if dataset and ('%' in dataset or '*' in dataset): dbsExceptionHandler('dbsException-invalid-input', " DBSReleaseVersion/listReleaseVersions. No wildcards are" + " allowed in dataset.\n.") if logical_file_name and ('%' in logical_file_name or '*' in logical_file_name): dbsExceptionHandler('dbsException-invalid-input', " DBSReleaseVersion/listReleaseVersions. No wildcards are" + " allowed in logical_file_name.\n.") conn = self.dbi.connection() try: plist = self.releaseVersion.execute(conn, release_version.upper(), dataset, logical_file_name) result = [{}] if plist: t = [] for i in plist: for k, v in i.iteritems(): t.append(v) result[0]['release_version'] = t return result finally: if conn: conn.close()
python
def listReleaseVersions(self, release_version="", dataset='', logical_file_name=''): """ List release versions """ if dataset and ('%' in dataset or '*' in dataset): dbsExceptionHandler('dbsException-invalid-input', " DBSReleaseVersion/listReleaseVersions. No wildcards are" + " allowed in dataset.\n.") if logical_file_name and ('%' in logical_file_name or '*' in logical_file_name): dbsExceptionHandler('dbsException-invalid-input', " DBSReleaseVersion/listReleaseVersions. No wildcards are" + " allowed in logical_file_name.\n.") conn = self.dbi.connection() try: plist = self.releaseVersion.execute(conn, release_version.upper(), dataset, logical_file_name) result = [{}] if plist: t = [] for i in plist: for k, v in i.iteritems(): t.append(v) result[0]['release_version'] = t return result finally: if conn: conn.close()
[ "def", "listReleaseVersions", "(", "self", ",", "release_version", "=", "\"\"", ",", "dataset", "=", "''", ",", "logical_file_name", "=", "''", ")", ":", "if", "dataset", "and", "(", "'%'", "in", "dataset", "or", "'*'", "in", "dataset", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\" DBSReleaseVersion/listReleaseVersions. No wildcards are\"", "+", "\" allowed in dataset.\\n.\"", ")", "if", "logical_file_name", "and", "(", "'%'", "in", "logical_file_name", "or", "'*'", "in", "logical_file_name", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\" DBSReleaseVersion/listReleaseVersions. No wildcards are\"", "+", "\" allowed in logical_file_name.\\n.\"", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "plist", "=", "self", ".", "releaseVersion", ".", "execute", "(", "conn", ",", "release_version", ".", "upper", "(", ")", ",", "dataset", ",", "logical_file_name", ")", "result", "=", "[", "{", "}", "]", "if", "plist", ":", "t", "=", "[", "]", "for", "i", "in", "plist", ":", "for", "k", ",", "v", "in", "i", ".", "iteritems", "(", ")", ":", "t", ".", "append", "(", "v", ")", "result", "[", "0", "]", "[", "'release_version'", "]", "=", "t", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
List release versions
[ "List", "release", "versions" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSReleaseVersion.py#L23-L50
train
dmwm/DBS
PycurlClient/src/python/RestClient/AuthHandling/X509Auth.py
X509Auth.__search_ca_path
def __search_ca_path(self): """ Get CA Path to check the validity of the server host certificate on the client side """ if "X509_CERT_DIR" in os.environ: self._ca_path = os.environ['X509_CERT_DIR'] elif os.path.exists('/etc/grid-security/certificates'): self._ca_path = '/etc/grid-security/certificates' else: raise ClientAuthException("Could not find a valid CA path")
python
def __search_ca_path(self): """ Get CA Path to check the validity of the server host certificate on the client side """ if "X509_CERT_DIR" in os.environ: self._ca_path = os.environ['X509_CERT_DIR'] elif os.path.exists('/etc/grid-security/certificates'): self._ca_path = '/etc/grid-security/certificates' else: raise ClientAuthException("Could not find a valid CA path")
[ "def", "__search_ca_path", "(", "self", ")", ":", "if", "\"X509_CERT_DIR\"", "in", "os", ".", "environ", ":", "self", ".", "_ca_path", "=", "os", ".", "environ", "[", "'X509_CERT_DIR'", "]", "elif", "os", ".", "path", ".", "exists", "(", "'/etc/grid-security/certificates'", ")", ":", "self", ".", "_ca_path", "=", "'/etc/grid-security/certificates'", "else", ":", "raise", "ClientAuthException", "(", "\"Could not find a valid CA path\"", ")" ]
Get CA Path to check the validity of the server host certificate on the client side
[ "Get", "CA", "Path", "to", "check", "the", "validity", "of", "the", "server", "host", "certificate", "on", "the", "client", "side" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/PycurlClient/src/python/RestClient/AuthHandling/X509Auth.py#L26-L37
train
dmwm/DBS
PycurlClient/src/python/RestClient/AuthHandling/X509Auth.py
X509Auth.__search_cert_key
def __search_cert_key(self): """ Get the user credentials if they exist, otherwise throw an exception. This code was modified from DBSAPI/dbsHttpService.py and WMCore/Services/Requests.py """ # Now we're trying to guess what the right cert/key combo is... # First preference to HOST Certificate, This is how it set in Tier0 if 'X509_HOST_CERT' in os.environ: self._ssl_cert = os.environ['X509_HOST_CERT'] self._ssl_key = os.environ['X509_HOST_KEY'] # Second preference to User Proxy, very common elif 'X509_USER_PROXY' in os.environ and os.path.exists(os.environ['X509_USER_PROXY']): self._ssl_cert = os.environ['X509_USER_PROXY'] self._ssl_key = self._ssl_cert # Third preference to User Cert/Proxy combinition elif 'X509_USER_CERT' in os.environ and 'X509_USER_KEY' in os.environ: self._ssl_cert = os.environ['X509_USER_CERT'] self._ssl_key = os.environ['X509_USER_KEY'] # TODO: only in linux, unix case, add other os case # look for proxy at default location /tmp/x509up_u$uid elif os.path.exists('/tmp/x509up_u%s' % str(os.getuid())): self._ssl_cert = '/tmp/x509up_u%s' % str(os.getuid()) self._ssl_key = self._ssl_cert elif sys.stdin.isatty(): home_dir = os.environ['HOME'] user_cert = os.path.join(home_dir, '.globus/usercert.pem') user_key = os.path.join(home_dir, '.globus/userkey.pem') if os.path.exists(user_cert): self._ssl_cert = user_cert if os.path.exists(user_key): self._ssl_key = user_key #store password for convenience self._ssl_key_pass = getpass("Password for %s: " % self._ssl_key) else: self._ssl_key = self._ssl_cert else: raise ClientAuthException("No valid X509 cert-key-pair found.") else: raise ClientAuthException("No valid X509 cert-key-pair found.")
python
def __search_cert_key(self): """ Get the user credentials if they exist, otherwise throw an exception. This code was modified from DBSAPI/dbsHttpService.py and WMCore/Services/Requests.py """ # Now we're trying to guess what the right cert/key combo is... # First preference to HOST Certificate, This is how it set in Tier0 if 'X509_HOST_CERT' in os.environ: self._ssl_cert = os.environ['X509_HOST_CERT'] self._ssl_key = os.environ['X509_HOST_KEY'] # Second preference to User Proxy, very common elif 'X509_USER_PROXY' in os.environ and os.path.exists(os.environ['X509_USER_PROXY']): self._ssl_cert = os.environ['X509_USER_PROXY'] self._ssl_key = self._ssl_cert # Third preference to User Cert/Proxy combinition elif 'X509_USER_CERT' in os.environ and 'X509_USER_KEY' in os.environ: self._ssl_cert = os.environ['X509_USER_CERT'] self._ssl_key = os.environ['X509_USER_KEY'] # TODO: only in linux, unix case, add other os case # look for proxy at default location /tmp/x509up_u$uid elif os.path.exists('/tmp/x509up_u%s' % str(os.getuid())): self._ssl_cert = '/tmp/x509up_u%s' % str(os.getuid()) self._ssl_key = self._ssl_cert elif sys.stdin.isatty(): home_dir = os.environ['HOME'] user_cert = os.path.join(home_dir, '.globus/usercert.pem') user_key = os.path.join(home_dir, '.globus/userkey.pem') if os.path.exists(user_cert): self._ssl_cert = user_cert if os.path.exists(user_key): self._ssl_key = user_key #store password for convenience self._ssl_key_pass = getpass("Password for %s: " % self._ssl_key) else: self._ssl_key = self._ssl_cert else: raise ClientAuthException("No valid X509 cert-key-pair found.") else: raise ClientAuthException("No valid X509 cert-key-pair found.")
[ "def", "__search_cert_key", "(", "self", ")", ":", "# Now we're trying to guess what the right cert/key combo is...", "# First preference to HOST Certificate, This is how it set in Tier0", "if", "'X509_HOST_CERT'", "in", "os", ".", "environ", ":", "self", ".", "_ssl_cert", "=", "os", ".", "environ", "[", "'X509_HOST_CERT'", "]", "self", ".", "_ssl_key", "=", "os", ".", "environ", "[", "'X509_HOST_KEY'", "]", "# Second preference to User Proxy, very common", "elif", "'X509_USER_PROXY'", "in", "os", ".", "environ", "and", "os", ".", "path", ".", "exists", "(", "os", ".", "environ", "[", "'X509_USER_PROXY'", "]", ")", ":", "self", ".", "_ssl_cert", "=", "os", ".", "environ", "[", "'X509_USER_PROXY'", "]", "self", ".", "_ssl_key", "=", "self", ".", "_ssl_cert", "# Third preference to User Cert/Proxy combinition", "elif", "'X509_USER_CERT'", "in", "os", ".", "environ", "and", "'X509_USER_KEY'", "in", "os", ".", "environ", ":", "self", ".", "_ssl_cert", "=", "os", ".", "environ", "[", "'X509_USER_CERT'", "]", "self", ".", "_ssl_key", "=", "os", ".", "environ", "[", "'X509_USER_KEY'", "]", "# TODO: only in linux, unix case, add other os case", "# look for proxy at default location /tmp/x509up_u$uid", "elif", "os", ".", "path", ".", "exists", "(", "'/tmp/x509up_u%s'", "%", "str", "(", "os", ".", "getuid", "(", ")", ")", ")", ":", "self", ".", "_ssl_cert", "=", "'/tmp/x509up_u%s'", "%", "str", "(", "os", ".", "getuid", "(", ")", ")", "self", ".", "_ssl_key", "=", "self", ".", "_ssl_cert", "elif", "sys", ".", "stdin", ".", "isatty", "(", ")", ":", "home_dir", "=", "os", ".", "environ", "[", "'HOME'", "]", "user_cert", "=", "os", ".", "path", ".", "join", "(", "home_dir", ",", "'.globus/usercert.pem'", ")", "user_key", "=", "os", ".", "path", ".", "join", "(", "home_dir", ",", "'.globus/userkey.pem'", ")", "if", "os", ".", "path", ".", "exists", "(", "user_cert", ")", ":", "self", ".", "_ssl_cert", "=", "user_cert", "if", "os", ".", "path", ".", "exists", "(", "user_key", ")", ":", "self", ".", "_ssl_key", "=", "user_key", "#store password for convenience", "self", ".", "_ssl_key_pass", "=", "getpass", "(", "\"Password for %s: \"", "%", "self", ".", "_ssl_key", ")", "else", ":", "self", ".", "_ssl_key", "=", "self", ".", "_ssl_cert", "else", ":", "raise", "ClientAuthException", "(", "\"No valid X509 cert-key-pair found.\"", ")", "else", ":", "raise", "ClientAuthException", "(", "\"No valid X509 cert-key-pair found.\"", ")" ]
Get the user credentials if they exist, otherwise throw an exception. This code was modified from DBSAPI/dbsHttpService.py and WMCore/Services/Requests.py
[ "Get", "the", "user", "credentials", "if", "they", "exist", "otherwise", "throw", "an", "exception", ".", "This", "code", "was", "modified", "from", "DBSAPI", "/", "dbsHttpService", ".", "py", "and", "WMCore", "/", "Services", "/", "Requests", ".", "py" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/PycurlClient/src/python/RestClient/AuthHandling/X509Auth.py#L39-L83
train
dmwm/DBS
Server/Python/src/dbs/web/DBSMigrateModel.py
authInsert
def authInsert(user, role, group, site): """ Authorization function for general insert """ if not role: return True for k, v in user['roles'].iteritems(): for g in v['group']: if k in role.get(g, '').split(':'): return True return False
python
def authInsert(user, role, group, site): """ Authorization function for general insert """ if not role: return True for k, v in user['roles'].iteritems(): for g in v['group']: if k in role.get(g, '').split(':'): return True return False
[ "def", "authInsert", "(", "user", ",", "role", ",", "group", ",", "site", ")", ":", "if", "not", "role", ":", "return", "True", "for", "k", ",", "v", "in", "user", "[", "'roles'", "]", ".", "iteritems", "(", ")", ":", "for", "g", "in", "v", "[", "'group'", "]", ":", "if", "k", "in", "role", ".", "get", "(", "g", ",", "''", ")", ".", "split", "(", "':'", ")", ":", "return", "True", "return", "False" ]
Authorization function for general insert
[ "Authorization", "function", "for", "general", "insert" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSMigrateModel.py#L32-L41
train
dmwm/DBS
Server/Python/src/dbs/web/DBSMigrateModel.py
DBSMigrateModel.submit
def submit(self): """ Interface for submitting a migration request. Required input keys: MIGRATION_URL: The source DBS url for migration. MIGRATION_INPUT: The block or dataset names to be migrated. """ body = request.body.read() indata = cjson.decode(body) try: indata = validateJSONInputNoCopy("migration_rqst", indata) indata.update({"creation_date": dbsUtils().getTime(), "last_modification_date" : dbsUtils().getTime(), "create_by" : dbsUtils().getCreateBy() , "last_modified_by" : dbsUtils().getCreateBy(), "migration_status": 0}) return self.dbsMigrate.insertMigrationRequest(indata) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.serverError) except Exception as ex: sError = "DBSMigrateModle/submit. %s\n Exception trace: \n %s." \ % (ex, traceback.format_exc() ) if hasattr(ex, 'status') and ex.status == 400: dbsExceptionHandler('dbsException-invalid-input2', str(ex), self.logger.exception, sError) else: dbsExceptionHandler('dbsException-server-error', str(ex), self.logger.exception, sError)
python
def submit(self): """ Interface for submitting a migration request. Required input keys: MIGRATION_URL: The source DBS url for migration. MIGRATION_INPUT: The block or dataset names to be migrated. """ body = request.body.read() indata = cjson.decode(body) try: indata = validateJSONInputNoCopy("migration_rqst", indata) indata.update({"creation_date": dbsUtils().getTime(), "last_modification_date" : dbsUtils().getTime(), "create_by" : dbsUtils().getCreateBy() , "last_modified_by" : dbsUtils().getCreateBy(), "migration_status": 0}) return self.dbsMigrate.insertMigrationRequest(indata) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.serverError) except Exception as ex: sError = "DBSMigrateModle/submit. %s\n Exception trace: \n %s." \ % (ex, traceback.format_exc() ) if hasattr(ex, 'status') and ex.status == 400: dbsExceptionHandler('dbsException-invalid-input2', str(ex), self.logger.exception, sError) else: dbsExceptionHandler('dbsException-server-error', str(ex), self.logger.exception, sError)
[ "def", "submit", "(", "self", ")", ":", "body", "=", "request", ".", "body", ".", "read", "(", ")", "indata", "=", "cjson", ".", "decode", "(", "body", ")", "try", ":", "indata", "=", "validateJSONInputNoCopy", "(", "\"migration_rqst\"", ",", "indata", ")", "indata", ".", "update", "(", "{", "\"creation_date\"", ":", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ",", "\"last_modification_date\"", ":", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ",", "\"create_by\"", ":", "dbsUtils", "(", ")", ".", "getCreateBy", "(", ")", ",", "\"last_modified_by\"", ":", "dbsUtils", "(", ")", ".", "getCreateBy", "(", ")", ",", "\"migration_status\"", ":", "0", "}", ")", "return", "self", ".", "dbsMigrate", ".", "insertMigrationRequest", "(", "indata", ")", "except", "dbsException", "as", "de", ":", "dbsExceptionHandler", "(", "de", ".", "eCode", ",", "de", ".", "message", ",", "self", ".", "logger", ".", "exception", ",", "de", ".", "serverError", ")", "except", "Exception", "as", "ex", ":", "sError", "=", "\"DBSMigrateModle/submit. %s\\n Exception trace: \\n %s.\"", "%", "(", "ex", ",", "traceback", ".", "format_exc", "(", ")", ")", "if", "hasattr", "(", "ex", ",", "'status'", ")", "and", "ex", ".", "status", "==", "400", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "str", "(", "ex", ")", ",", "self", ".", "logger", ".", "exception", ",", "sError", ")", "else", ":", "dbsExceptionHandler", "(", "'dbsException-server-error'", ",", "str", "(", "ex", ")", ",", "self", ".", "logger", ".", "exception", ",", "sError", ")" ]
Interface for submitting a migration request. Required input keys: MIGRATION_URL: The source DBS url for migration. MIGRATION_INPUT: The block or dataset names to be migrated.
[ "Interface", "for", "submitting", "a", "migration", "request", ".", "Required", "input", "keys", ":", "MIGRATION_URL", ":", "The", "source", "DBS", "url", "for", "migration", ".", "MIGRATION_INPUT", ":", "The", "block", "or", "dataset", "names", "to", "be", "migrated", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSMigrateModel.py#L68-L93
train
dmwm/DBS
Server/Python/src/dbs/web/DBSMigrateModel.py
DBSMigrateModel.status
def status(self, migration_rqst_id="", block_name="", dataset="", user=""): """ Interface to query status of a migration request In this preference order of input parameters : migration_rqst_id, block, dataset, user (if multi parameters are provided, only the precedence order is followed) """ try: return self.dbsMigrate.listMigrationRequests(migration_rqst_id, block_name, dataset, user) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.serverError) except Exception as ex: sError = "DBSMigrateModle/status. %s\n Exception trace: \n %s." \ % (ex, traceback.format_exc() ) if hasattr(ex, 'status') and ex.status == 400: dbsExceptionHandler('dbsException-invalid-input2', str(ex), self.logger.exception, sError) else: dbsExceptionHandler('dbsException-server-error', str(ex), self.logger.exception, sError)
python
def status(self, migration_rqst_id="", block_name="", dataset="", user=""): """ Interface to query status of a migration request In this preference order of input parameters : migration_rqst_id, block, dataset, user (if multi parameters are provided, only the precedence order is followed) """ try: return self.dbsMigrate.listMigrationRequests(migration_rqst_id, block_name, dataset, user) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.serverError) except Exception as ex: sError = "DBSMigrateModle/status. %s\n Exception trace: \n %s." \ % (ex, traceback.format_exc() ) if hasattr(ex, 'status') and ex.status == 400: dbsExceptionHandler('dbsException-invalid-input2', str(ex), self.logger.exception, sError) else: dbsExceptionHandler('dbsException-server-error', str(ex), self.logger.exception, sError)
[ "def", "status", "(", "self", ",", "migration_rqst_id", "=", "\"\"", ",", "block_name", "=", "\"\"", ",", "dataset", "=", "\"\"", ",", "user", "=", "\"\"", ")", ":", "try", ":", "return", "self", ".", "dbsMigrate", ".", "listMigrationRequests", "(", "migration_rqst_id", ",", "block_name", ",", "dataset", ",", "user", ")", "except", "dbsException", "as", "de", ":", "dbsExceptionHandler", "(", "de", ".", "eCode", ",", "de", ".", "message", ",", "self", ".", "logger", ".", "exception", ",", "de", ".", "serverError", ")", "except", "Exception", "as", "ex", ":", "sError", "=", "\"DBSMigrateModle/status. %s\\n Exception trace: \\n %s.\"", "%", "(", "ex", ",", "traceback", ".", "format_exc", "(", ")", ")", "if", "hasattr", "(", "ex", ",", "'status'", ")", "and", "ex", ".", "status", "==", "400", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "str", "(", "ex", ")", ",", "self", ".", "logger", ".", "exception", ",", "sError", ")", "else", ":", "dbsExceptionHandler", "(", "'dbsException-server-error'", ",", "str", "(", "ex", ")", ",", "self", ".", "logger", ".", "exception", ",", "sError", ")" ]
Interface to query status of a migration request In this preference order of input parameters : migration_rqst_id, block, dataset, user (if multi parameters are provided, only the precedence order is followed)
[ "Interface", "to", "query", "status", "of", "a", "migration", "request", "In", "this", "preference", "order", "of", "input", "parameters", ":", "migration_rqst_id", "block", "dataset", "user", "(", "if", "multi", "parameters", "are", "provided", "only", "the", "precedence", "order", "is", "followed", ")" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSMigrateModel.py#L96-L114
train
dmwm/DBS
Server/Python/src/dbs/web/DBSMigrateModel.py
DBSMigrateModel.remove
def remove(self): """ Interface to remove a migration request from the queue. Only Permanent FAILED/9 and PENDING/0 requests can be removed (running and sucessed requests cannot be removed) """ body = request.body.read() indata = cjson.decode(body) try: indata = validateJSONInputNoCopy("migration_rqst", indata) return self.dbsMigrate.removeMigrationRequest(indata) except dbsException as he: dbsExceptionHandler(he.eCode, he.message, self.logger.exception, he.message) except Exception as e: if e.code == 400: dbsExceptionHandler('dbsException-invalid-input2', str(e), self.logger.exception, str(e)) else: dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, str(e))
python
def remove(self): """ Interface to remove a migration request from the queue. Only Permanent FAILED/9 and PENDING/0 requests can be removed (running and sucessed requests cannot be removed) """ body = request.body.read() indata = cjson.decode(body) try: indata = validateJSONInputNoCopy("migration_rqst", indata) return self.dbsMigrate.removeMigrationRequest(indata) except dbsException as he: dbsExceptionHandler(he.eCode, he.message, self.logger.exception, he.message) except Exception as e: if e.code == 400: dbsExceptionHandler('dbsException-invalid-input2', str(e), self.logger.exception, str(e)) else: dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, str(e))
[ "def", "remove", "(", "self", ")", ":", "body", "=", "request", ".", "body", ".", "read", "(", ")", "indata", "=", "cjson", ".", "decode", "(", "body", ")", "try", ":", "indata", "=", "validateJSONInputNoCopy", "(", "\"migration_rqst\"", ",", "indata", ")", "return", "self", ".", "dbsMigrate", ".", "removeMigrationRequest", "(", "indata", ")", "except", "dbsException", "as", "he", ":", "dbsExceptionHandler", "(", "he", ".", "eCode", ",", "he", ".", "message", ",", "self", ".", "logger", ".", "exception", ",", "he", ".", "message", ")", "except", "Exception", "as", "e", ":", "if", "e", ".", "code", "==", "400", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "str", "(", "e", ")", ",", "self", ".", "logger", ".", "exception", ",", "str", "(", "e", ")", ")", "else", ":", "dbsExceptionHandler", "(", "'dbsException-server-error'", ",", "dbsExceptionCode", "[", "'dbsException-server-error'", "]", ",", "self", ".", "logger", ".", "exception", ",", "str", "(", "e", ")", ")" ]
Interface to remove a migration request from the queue. Only Permanent FAILED/9 and PENDING/0 requests can be removed (running and sucessed requests cannot be removed)
[ "Interface", "to", "remove", "a", "migration", "request", "from", "the", "queue", ".", "Only", "Permanent", "FAILED", "/", "9", "and", "PENDING", "/", "0", "requests", "can", "be", "removed", "(", "running", "and", "sucessed", "requests", "cannot", "be", "removed", ")" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSMigrateModel.py#L116-L134
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/FileParent/List.py
List.execute
def execute(self, conn, logical_file_name='', block_id=0, block_name='', transaction=False): """ return {} if condition is not provided. """ sql = '' binds = {} if logical_file_name: if isinstance(logical_file_name, basestring): wheresql = "WHERE F.LOGICAL_FILE_NAME = :logical_file_name" binds = {"logical_file_name": logical_file_name} sql = "{sql} {wheresql}".format(sql=self.sql, wheresql=wheresql) elif isinstance(logical_file_name, list): wheresql = "WHERE F.LOGICAL_FILE_NAME in (SELECT TOKEN FROM TOKEN_GENERATOR)" lfn_generator, binds = create_token_generator(logical_file_name) sql = "{lfn_generator} {sql} {wheresql}".format(lfn_generator=lfn_generator, sql=self.sql, wheresql=wheresql) elif block_id != 0: wheresql = "WHERE F.BLOCK_ID = :block_id" binds ={'block_id': block_id} sql = "{sql} {wheresql}".format(sql=self.sql, wheresql=wheresql) elif block_name: joins = "JOIN {owner}BLOCKS B on B.BLOCK_ID = F.BLOCK_ID".format(owner=self.owner) wheresql = "WHERE B.BLOCK_NAME= :block_name" binds ={'block_name': block_name} sql = "{sql} {joins} {wheresql}".format(sql=self.sql, joins=joins, wheresql=wheresql) else: return cursors = self.dbi.processData(sql, binds, conn, transaction=transaction, returnCursor=True) for i in cursors: d = self.formatCursor(i, size=100) if isinstance(d, list) or isinstance(d, GeneratorType): for elem in d: yield elem elif d: yield d
python
def execute(self, conn, logical_file_name='', block_id=0, block_name='', transaction=False): """ return {} if condition is not provided. """ sql = '' binds = {} if logical_file_name: if isinstance(logical_file_name, basestring): wheresql = "WHERE F.LOGICAL_FILE_NAME = :logical_file_name" binds = {"logical_file_name": logical_file_name} sql = "{sql} {wheresql}".format(sql=self.sql, wheresql=wheresql) elif isinstance(logical_file_name, list): wheresql = "WHERE F.LOGICAL_FILE_NAME in (SELECT TOKEN FROM TOKEN_GENERATOR)" lfn_generator, binds = create_token_generator(logical_file_name) sql = "{lfn_generator} {sql} {wheresql}".format(lfn_generator=lfn_generator, sql=self.sql, wheresql=wheresql) elif block_id != 0: wheresql = "WHERE F.BLOCK_ID = :block_id" binds ={'block_id': block_id} sql = "{sql} {wheresql}".format(sql=self.sql, wheresql=wheresql) elif block_name: joins = "JOIN {owner}BLOCKS B on B.BLOCK_ID = F.BLOCK_ID".format(owner=self.owner) wheresql = "WHERE B.BLOCK_NAME= :block_name" binds ={'block_name': block_name} sql = "{sql} {joins} {wheresql}".format(sql=self.sql, joins=joins, wheresql=wheresql) else: return cursors = self.dbi.processData(sql, binds, conn, transaction=transaction, returnCursor=True) for i in cursors: d = self.formatCursor(i, size=100) if isinstance(d, list) or isinstance(d, GeneratorType): for elem in d: yield elem elif d: yield d
[ "def", "execute", "(", "self", ",", "conn", ",", "logical_file_name", "=", "''", ",", "block_id", "=", "0", ",", "block_name", "=", "''", ",", "transaction", "=", "False", ")", ":", "sql", "=", "''", "binds", "=", "{", "}", "if", "logical_file_name", ":", "if", "isinstance", "(", "logical_file_name", ",", "basestring", ")", ":", "wheresql", "=", "\"WHERE F.LOGICAL_FILE_NAME = :logical_file_name\"", "binds", "=", "{", "\"logical_file_name\"", ":", "logical_file_name", "}", "sql", "=", "\"{sql} {wheresql}\"", ".", "format", "(", "sql", "=", "self", ".", "sql", ",", "wheresql", "=", "wheresql", ")", "elif", "isinstance", "(", "logical_file_name", ",", "list", ")", ":", "wheresql", "=", "\"WHERE F.LOGICAL_FILE_NAME in (SELECT TOKEN FROM TOKEN_GENERATOR)\"", "lfn_generator", ",", "binds", "=", "create_token_generator", "(", "logical_file_name", ")", "sql", "=", "\"{lfn_generator} {sql} {wheresql}\"", ".", "format", "(", "lfn_generator", "=", "lfn_generator", ",", "sql", "=", "self", ".", "sql", ",", "wheresql", "=", "wheresql", ")", "elif", "block_id", "!=", "0", ":", "wheresql", "=", "\"WHERE F.BLOCK_ID = :block_id\"", "binds", "=", "{", "'block_id'", ":", "block_id", "}", "sql", "=", "\"{sql} {wheresql}\"", ".", "format", "(", "sql", "=", "self", ".", "sql", ",", "wheresql", "=", "wheresql", ")", "elif", "block_name", ":", "joins", "=", "\"JOIN {owner}BLOCKS B on B.BLOCK_ID = F.BLOCK_ID\"", ".", "format", "(", "owner", "=", "self", ".", "owner", ")", "wheresql", "=", "\"WHERE B.BLOCK_NAME= :block_name\"", "binds", "=", "{", "'block_name'", ":", "block_name", "}", "sql", "=", "\"{sql} {joins} {wheresql}\"", ".", "format", "(", "sql", "=", "self", ".", "sql", ",", "joins", "=", "joins", ",", "wheresql", "=", "wheresql", ")", "else", ":", "return", "cursors", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", "=", "transaction", ",", "returnCursor", "=", "True", ")", "for", "i", "in", "cursors", ":", "d", "=", "self", ".", "formatCursor", "(", "i", ",", "size", "=", "100", ")", "if", "isinstance", "(", "d", ",", "list", ")", "or", "isinstance", "(", "d", ",", "GeneratorType", ")", ":", "for", "elem", "in", "d", ":", "yield", "elem", "elif", "d", ":", "yield", "d" ]
return {} if condition is not provided.
[ "return", "{}", "if", "condition", "is", "not", "provided", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/FileParent/List.py#L31-L67
train
dmwm/DBS
Server/Python/src/dbs/business/DBSDataset.py
DBSDataset.listDatasetParents
def listDatasetParents(self, dataset=""): """ takes required dataset parameter returns only parent dataset name """ if( dataset == "" ): dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/listDatasetParents. Child Dataset name is required.") conn = self.dbi.connection() try: result = self.datasetparentlist.execute(conn, dataset) return result finally: if conn: conn.close()
python
def listDatasetParents(self, dataset=""): """ takes required dataset parameter returns only parent dataset name """ if( dataset == "" ): dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/listDatasetParents. Child Dataset name is required.") conn = self.dbi.connection() try: result = self.datasetparentlist.execute(conn, dataset) return result finally: if conn: conn.close()
[ "def", "listDatasetParents", "(", "self", ",", "dataset", "=", "\"\"", ")", ":", "if", "(", "dataset", "==", "\"\"", ")", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSDataset/listDatasetParents. Child Dataset name is required.\"", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "result", "=", "self", ".", "datasetparentlist", ".", "execute", "(", "conn", ",", "dataset", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
takes required dataset parameter returns only parent dataset name
[ "takes", "required", "dataset", "parameter", "returns", "only", "parent", "dataset", "name" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSDataset.py#L42-L55
train
dmwm/DBS
Server/Python/src/dbs/business/DBSDataset.py
DBSDataset.listDatasetChildren
def listDatasetChildren(self, dataset): """ takes required dataset parameter returns only children dataset name """ if( dataset == "" ): dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/listDatasetChildren. Parent Dataset name is required.") conn = self.dbi.connection() try: result = self.datasetchildlist.execute(conn, dataset) return result finally: if conn: conn.close()
python
def listDatasetChildren(self, dataset): """ takes required dataset parameter returns only children dataset name """ if( dataset == "" ): dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/listDatasetChildren. Parent Dataset name is required.") conn = self.dbi.connection() try: result = self.datasetchildlist.execute(conn, dataset) return result finally: if conn: conn.close()
[ "def", "listDatasetChildren", "(", "self", ",", "dataset", ")", ":", "if", "(", "dataset", "==", "\"\"", ")", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSDataset/listDatasetChildren. Parent Dataset name is required.\"", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "result", "=", "self", ".", "datasetchildlist", ".", "execute", "(", "conn", ",", "dataset", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
takes required dataset parameter returns only children dataset name
[ "takes", "required", "dataset", "parameter", "returns", "only", "children", "dataset", "name" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSDataset.py#L57-L70
train
dmwm/DBS
Server/Python/src/dbs/business/DBSDataset.py
DBSDataset.updateStatus
def updateStatus(self, dataset, is_dataset_valid): """ Used to toggle the status of a dataset is_dataset_valid=0/1 (invalid/valid) """ if( dataset == "" ): dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/updateStatus. dataset is required.") conn = self.dbi.connection() trans = conn.begin() try: self.updatestatus.execute(conn, dataset, is_dataset_valid, trans) trans.commit() trans = None except Exception as ex: if trans: trans.rollback() raise ex finally: if trans: trans.rollback() if conn: conn.close()
python
def updateStatus(self, dataset, is_dataset_valid): """ Used to toggle the status of a dataset is_dataset_valid=0/1 (invalid/valid) """ if( dataset == "" ): dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/updateStatus. dataset is required.") conn = self.dbi.connection() trans = conn.begin() try: self.updatestatus.execute(conn, dataset, is_dataset_valid, trans) trans.commit() trans = None except Exception as ex: if trans: trans.rollback() raise ex finally: if trans: trans.rollback() if conn: conn.close()
[ "def", "updateStatus", "(", "self", ",", "dataset", ",", "is_dataset_valid", ")", ":", "if", "(", "dataset", "==", "\"\"", ")", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSDataset/updateStatus. dataset is required.\"", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "trans", "=", "conn", ".", "begin", "(", ")", "try", ":", "self", ".", "updatestatus", ".", "execute", "(", "conn", ",", "dataset", ",", "is_dataset_valid", ",", "trans", ")", "trans", ".", "commit", "(", ")", "trans", "=", "None", "except", "Exception", "as", "ex", ":", "if", "trans", ":", "trans", ".", "rollback", "(", ")", "raise", "ex", "finally", ":", "if", "trans", ":", "trans", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Used to toggle the status of a dataset is_dataset_valid=0/1 (invalid/valid)
[ "Used", "to", "toggle", "the", "status", "of", "a", "dataset", "is_dataset_valid", "=", "0", "/", "1", "(", "invalid", "/", "valid", ")" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSDataset.py#L72-L94
train
dmwm/DBS
Server/Python/src/dbs/business/DBSDataset.py
DBSDataset.updateType
def updateType(self, dataset, dataset_access_type): """ Used to change the status of a dataset type (production/etc.) """ if( dataset == "" ): dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/updateType. dataset is required.") conn = self.dbi.connection() trans = conn.begin() try : self.updatetype.execute(conn, dataset, dataset_access_type.upper(), trans) trans.commit() trans = None except SQLAlchemyDatabaseError as ex: if str(ex).find("ORA-01407") != -1: dbsExceptionHandler("dbsException-invalid-input2", "Invalid Input", None, "DBSDataset/updateType. A Valid dataset_access_type is required.") finally: if trans: trans.rollback() if conn: conn.close()
python
def updateType(self, dataset, dataset_access_type): """ Used to change the status of a dataset type (production/etc.) """ if( dataset == "" ): dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/updateType. dataset is required.") conn = self.dbi.connection() trans = conn.begin() try : self.updatetype.execute(conn, dataset, dataset_access_type.upper(), trans) trans.commit() trans = None except SQLAlchemyDatabaseError as ex: if str(ex).find("ORA-01407") != -1: dbsExceptionHandler("dbsException-invalid-input2", "Invalid Input", None, "DBSDataset/updateType. A Valid dataset_access_type is required.") finally: if trans: trans.rollback() if conn: conn.close()
[ "def", "updateType", "(", "self", ",", "dataset", ",", "dataset_access_type", ")", ":", "if", "(", "dataset", "==", "\"\"", ")", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSDataset/updateType. dataset is required.\"", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "trans", "=", "conn", ".", "begin", "(", ")", "try", ":", "self", ".", "updatetype", ".", "execute", "(", "conn", ",", "dataset", ",", "dataset_access_type", ".", "upper", "(", ")", ",", "trans", ")", "trans", ".", "commit", "(", ")", "trans", "=", "None", "except", "SQLAlchemyDatabaseError", "as", "ex", ":", "if", "str", "(", "ex", ")", ".", "find", "(", "\"ORA-01407\"", ")", "!=", "-", "1", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input2\"", ",", "\"Invalid Input\"", ",", "None", ",", "\"DBSDataset/updateType. A Valid dataset_access_type is required.\"", ")", "finally", ":", "if", "trans", ":", "trans", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Used to change the status of a dataset type (production/etc.)
[ "Used", "to", "change", "the", "status", "of", "a", "dataset", "type", "(", "production", "/", "etc", ".", ")" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSDataset.py#L96-L117
train
dmwm/DBS
Server/Python/src/dbs/business/DBSDataset.py
DBSDataset.listDatasets
def listDatasets(self, dataset="", parent_dataset="", is_dataset_valid=1, release_version="", pset_hash="", app_name="", output_module_label="", global_tag="", processing_version=0, acquisition_era="", run_num=-1, physics_group_name="", logical_file_name="", primary_ds_name="", primary_ds_type="", processed_ds_name="", data_tier_name="", dataset_access_type="VALID", prep_id="", create_by='', last_modified_by='', min_cdate=0, max_cdate=0, min_ldate=0, max_ldate=0, cdate=0, ldate=0, detail=False, dataset_id=-1): """ lists all datasets if dataset parameter is not given. The parameter can include % character. all other parameters are not wild card ones. """ if(logical_file_name and logical_file_name.find("%")!=-1): dbsExceptionHandler('dbsException-invalid-input', 'DBSDataset/listDatasets API requires \ fullly qualified logical_file_name. NO wildcard is allowed in logical_file_name.') if(dataset and dataset.find("/%/%/%")!=-1): dataset='' with self.dbi.connection() as conn: dao = (self.datasetbrieflist, self.datasetlist)[detail] if dataset_access_type: dataset_access_type = dataset_access_type.upper() if data_tier_name: data_tier_name = data_tier_name.upper() #if processing_version: processing_version = processing_version.upper() #if acquisition_era: acquisition_era = acquisition_era.upper() for item in dao.execute(conn, dataset, is_dataset_valid, parent_dataset, release_version, pset_hash, app_name, output_module_label, global_tag, processing_version, acquisition_era, run_num, physics_group_name, logical_file_name, primary_ds_name, primary_ds_type, processed_ds_name, data_tier_name, dataset_access_type, prep_id, create_by, last_modified_by, min_cdate, max_cdate, min_ldate, max_ldate, cdate, ldate, dataset_id): yield item
python
def listDatasets(self, dataset="", parent_dataset="", is_dataset_valid=1, release_version="", pset_hash="", app_name="", output_module_label="", global_tag="", processing_version=0, acquisition_era="", run_num=-1, physics_group_name="", logical_file_name="", primary_ds_name="", primary_ds_type="", processed_ds_name="", data_tier_name="", dataset_access_type="VALID", prep_id="", create_by='', last_modified_by='', min_cdate=0, max_cdate=0, min_ldate=0, max_ldate=0, cdate=0, ldate=0, detail=False, dataset_id=-1): """ lists all datasets if dataset parameter is not given. The parameter can include % character. all other parameters are not wild card ones. """ if(logical_file_name and logical_file_name.find("%")!=-1): dbsExceptionHandler('dbsException-invalid-input', 'DBSDataset/listDatasets API requires \ fullly qualified logical_file_name. NO wildcard is allowed in logical_file_name.') if(dataset and dataset.find("/%/%/%")!=-1): dataset='' with self.dbi.connection() as conn: dao = (self.datasetbrieflist, self.datasetlist)[detail] if dataset_access_type: dataset_access_type = dataset_access_type.upper() if data_tier_name: data_tier_name = data_tier_name.upper() #if processing_version: processing_version = processing_version.upper() #if acquisition_era: acquisition_era = acquisition_era.upper() for item in dao.execute(conn, dataset, is_dataset_valid, parent_dataset, release_version, pset_hash, app_name, output_module_label, global_tag, processing_version, acquisition_era, run_num, physics_group_name, logical_file_name, primary_ds_name, primary_ds_type, processed_ds_name, data_tier_name, dataset_access_type, prep_id, create_by, last_modified_by, min_cdate, max_cdate, min_ldate, max_ldate, cdate, ldate, dataset_id): yield item
[ "def", "listDatasets", "(", "self", ",", "dataset", "=", "\"\"", ",", "parent_dataset", "=", "\"\"", ",", "is_dataset_valid", "=", "1", ",", "release_version", "=", "\"\"", ",", "pset_hash", "=", "\"\"", ",", "app_name", "=", "\"\"", ",", "output_module_label", "=", "\"\"", ",", "global_tag", "=", "\"\"", ",", "processing_version", "=", "0", ",", "acquisition_era", "=", "\"\"", ",", "run_num", "=", "-", "1", ",", "physics_group_name", "=", "\"\"", ",", "logical_file_name", "=", "\"\"", ",", "primary_ds_name", "=", "\"\"", ",", "primary_ds_type", "=", "\"\"", ",", "processed_ds_name", "=", "\"\"", ",", "data_tier_name", "=", "\"\"", ",", "dataset_access_type", "=", "\"VALID\"", ",", "prep_id", "=", "\"\"", ",", "create_by", "=", "''", ",", "last_modified_by", "=", "''", ",", "min_cdate", "=", "0", ",", "max_cdate", "=", "0", ",", "min_ldate", "=", "0", ",", "max_ldate", "=", "0", ",", "cdate", "=", "0", ",", "ldate", "=", "0", ",", "detail", "=", "False", ",", "dataset_id", "=", "-", "1", ")", ":", "if", "(", "logical_file_name", "and", "logical_file_name", ".", "find", "(", "\"%\"", ")", "!=", "-", "1", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "'DBSDataset/listDatasets API requires \\\n fullly qualified logical_file_name. NO wildcard is allowed in logical_file_name.'", ")", "if", "(", "dataset", "and", "dataset", ".", "find", "(", "\"/%/%/%\"", ")", "!=", "-", "1", ")", ":", "dataset", "=", "''", "with", "self", ".", "dbi", ".", "connection", "(", ")", "as", "conn", ":", "dao", "=", "(", "self", ".", "datasetbrieflist", ",", "self", ".", "datasetlist", ")", "[", "detail", "]", "if", "dataset_access_type", ":", "dataset_access_type", "=", "dataset_access_type", ".", "upper", "(", ")", "if", "data_tier_name", ":", "data_tier_name", "=", "data_tier_name", ".", "upper", "(", ")", "#if processing_version: processing_version = processing_version.upper()", "#if acquisition_era: acquisition_era = acquisition_era.upper()", "for", "item", "in", "dao", ".", "execute", "(", "conn", ",", "dataset", ",", "is_dataset_valid", ",", "parent_dataset", ",", "release_version", ",", "pset_hash", ",", "app_name", ",", "output_module_label", ",", "global_tag", ",", "processing_version", ",", "acquisition_era", ",", "run_num", ",", "physics_group_name", ",", "logical_file_name", ",", "primary_ds_name", ",", "primary_ds_type", ",", "processed_ds_name", ",", "data_tier_name", ",", "dataset_access_type", ",", "prep_id", ",", "create_by", ",", "last_modified_by", ",", "min_cdate", ",", "max_cdate", ",", "min_ldate", ",", "max_ldate", ",", "cdate", ",", "ldate", ",", "dataset_id", ")", ":", "yield", "item" ]
lists all datasets if dataset parameter is not given. The parameter can include % character. all other parameters are not wild card ones.
[ "lists", "all", "datasets", "if", "dataset", "parameter", "is", "not", "given", ".", "The", "parameter", "can", "include", "%", "character", ".", "all", "other", "parameters", "are", "not", "wild", "card", "ones", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSDataset.py#L119-L160
train
dmwm/DBS
Server/Python/src/dbs/business/DBSDataset.py
DBSDataset.insertDataset
def insertDataset(self, businput): """ input dictionary must have the following keys: dataset, primary_ds_name(name), processed_ds(name), data_tier(name), acquisition_era(name), processing_version It may have following keys: physics_group(name), xtcrosssection, creation_date, create_by, last_modification_date, last_modified_by """ if not ("primary_ds_name" in businput and "dataset" in businput and "dataset_access_type" in businput and "processed_ds_name" in businput ): dbsExceptionHandler('dbsException-invalid-input', "business/DBSDataset/insertDataset must have dataset,\ dataset_access_type, primary_ds_name, processed_ds_name as input") if "data_tier_name" not in businput: dbsExceptionHandler('dbsException-invalid-input', "insertDataset must have data_tier_name as input.") conn = self.dbi.connection() tran = conn.begin() try: dsdaoinput = {} dsdaoinput["primary_ds_name"] = businput["primary_ds_name"] dsdaoinput["data_tier_name"] = businput["data_tier_name"].upper() dsdaoinput["dataset_access_type"] = businput["dataset_access_type"].upper() #not required pre-exist in the db. will insert with the dataset if not in yet #processed_ds_name=acquisition_era_name[-fileter_name][-processing_str]-vprocessing_version Changed as 4/30/2012 YG. #althrough acquisition era and processing version is not required for a dataset in the schema(the schema is build this way because #we need to accomdate the DBS2 data), but we impose the requirement on the API. So both acquisition and processing eras are required #YG 12/07/2011 TK-362 if "acquisition_era_name" in businput and "processing_version" in businput: erals=businput["processed_ds_name"].rsplit('-') if erals[0]==businput["acquisition_era_name"] and erals[len(erals)-1]=="%s%s"%("v", businput["processing_version"]): dsdaoinput["processed_ds_name"] = businput["processed_ds_name"] else: dbsExceptionHandler('dbsException-invalid-input', "insertDataset:\ processed_ds_name=acquisition_era_name[-filter_name][-processing_str]-vprocessing_version must be satisified.") else: dbsExceptionHandler("dbsException-missing-data", "insertDataset: Required acquisition_era_name or processing_version is not found in the input") if "physics_group_name" in businput: dsdaoinput["physics_group_id"] = self.phygrpid.execute(conn, businput["physics_group_name"]) if dsdaoinput["physics_group_id"] == -1: dbsExceptionHandler("dbsException-missing-data", "insertDataset. physics_group_name not found in DB") else: dsdaoinput["physics_group_id"] = None dsdaoinput["dataset_id"] = self.sm.increment(conn, "SEQ_DS") # we are better off separating out what we need for the dataset DAO dsdaoinput.update({ "dataset" : "/%s/%s/%s" % (businput["primary_ds_name"], businput["processed_ds_name"], businput["data_tier_name"].upper()), "prep_id" : businput.get("prep_id", None), "xtcrosssection" : businput.get("xtcrosssection", None), "creation_date" : businput.get("creation_date", dbsUtils().getTime() ), "create_by" : businput.get("create_by", dbsUtils().getCreateBy()) , "last_modification_date" : businput.get("last_modification_date", dbsUtils().getTime()), #"last_modified_by" : businput.get("last_modified_by", dbsUtils().getModifiedBy()) "last_modified_by" : dbsUtils().getModifiedBy() }) """ repeated again, why? comment out by YG 3/14/2012 #physics group if "physics_group_name" in businput: dsdaoinput["physics_group_id"] = self.phygrpid.execute(conn, businput["physics_group_name"]) if dsdaoinput["physics_group_id"] == -1: dbsExceptionHandler("dbsException-missing-data", "insertDataset. Physics Group : %s Not found" % businput["physics_group_name"]) else: dsdaoinput["physics_group_id"] = None """ # See if Processing Era exists if "processing_version" in businput and businput["processing_version"] != 0: dsdaoinput["processing_era_id"] = self.proceraid.execute(conn, businput["processing_version"]) if dsdaoinput["processing_era_id"] == -1 : dbsExceptionHandler("dbsException-missing-data", "DBSDataset/insertDataset: processing_version not found in DB") else: dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/insertDataset: processing_version is required") # See if Acquisition Era exists if "acquisition_era_name" in businput: dsdaoinput["acquisition_era_id"] = self.acqeraid.execute(conn, businput["acquisition_era_name"]) if dsdaoinput["acquisition_era_id"] == -1: dbsExceptionHandler("dbsException-missing-data", "DBSDataset/insertDataset: acquisition_era_name not found in DB") else: dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/insertDataset: acquisition_era_name is required") try: # insert the dataset self.datasetin.execute(conn, dsdaoinput, tran) except SQLAlchemyIntegrityError as ex: if (str(ex).lower().find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1): # dataset already exists, lets fetch the ID self.logger.warning( "Unique constraint violation being ignored...") self.logger.warning("%s" % ex) ds = "/%s/%s/%s" % (businput["primary_ds_name"], businput["processed_ds_name"], businput["data_tier_name"].upper()) dsdaoinput["dataset_id"] = self.datasetid.execute(conn, ds ) if dsdaoinput["dataset_id"] == -1 : dbsExceptionHandler("dbsException-missing-data", "DBSDataset/insertDataset. Strange error, the dataset %s does not exist ?" % ds ) if (str(ex).find("ORA-01400") ) != -1 : dbsExceptionHandler("dbsException-missing-data", "insertDataset must have: dataset,\ primary_ds_name, processed_ds_name, data_tier_name ") except Exception as e: raise #FIXME : What about the READ-only status of the dataset #There is no READ-oly status for a dataset. # Create dataset_output_mod_mod_configs mapping if "output_configs" in businput: for anOutConfig in businput["output_configs"]: dsoutconfdaoin = {} dsoutconfdaoin["dataset_id"] = dsdaoinput["dataset_id"] dsoutconfdaoin["output_mod_config_id"] = self.outconfigid.execute(conn, anOutConfig["app_name"], anOutConfig["release_version"], anOutConfig["pset_hash"], anOutConfig["output_module_label"], anOutConfig["global_tag"]) if dsoutconfdaoin["output_mod_config_id"] == -1 : dbsExceptionHandler("dbsException-missing-data", "DBSDataset/insertDataset: Output config (%s, %s, %s, %s, %s) not found" % (anOutConfig["app_name"], anOutConfig["release_version"], anOutConfig["pset_hash"], anOutConfig["output_module_label"], anOutConfig["global_tag"])) try: self.datasetoutmodconfigin.execute(conn, dsoutconfdaoin, tran) except Exception as ex: if str(ex).lower().find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1: pass else: raise # Dataset parentage will NOT be added by this API it will be set by insertFiles()--deduced by insertFiles # Dataset runs will NOT be added by this API they will be set by insertFiles()--deduced by insertFiles OR insertRun API call tran.commit() tran = None except Exception: if tran: tran.rollback() tran = None raise finally: if tran: tran.rollback() if conn: conn.close()
python
def insertDataset(self, businput): """ input dictionary must have the following keys: dataset, primary_ds_name(name), processed_ds(name), data_tier(name), acquisition_era(name), processing_version It may have following keys: physics_group(name), xtcrosssection, creation_date, create_by, last_modification_date, last_modified_by """ if not ("primary_ds_name" in businput and "dataset" in businput and "dataset_access_type" in businput and "processed_ds_name" in businput ): dbsExceptionHandler('dbsException-invalid-input', "business/DBSDataset/insertDataset must have dataset,\ dataset_access_type, primary_ds_name, processed_ds_name as input") if "data_tier_name" not in businput: dbsExceptionHandler('dbsException-invalid-input', "insertDataset must have data_tier_name as input.") conn = self.dbi.connection() tran = conn.begin() try: dsdaoinput = {} dsdaoinput["primary_ds_name"] = businput["primary_ds_name"] dsdaoinput["data_tier_name"] = businput["data_tier_name"].upper() dsdaoinput["dataset_access_type"] = businput["dataset_access_type"].upper() #not required pre-exist in the db. will insert with the dataset if not in yet #processed_ds_name=acquisition_era_name[-fileter_name][-processing_str]-vprocessing_version Changed as 4/30/2012 YG. #althrough acquisition era and processing version is not required for a dataset in the schema(the schema is build this way because #we need to accomdate the DBS2 data), but we impose the requirement on the API. So both acquisition and processing eras are required #YG 12/07/2011 TK-362 if "acquisition_era_name" in businput and "processing_version" in businput: erals=businput["processed_ds_name"].rsplit('-') if erals[0]==businput["acquisition_era_name"] and erals[len(erals)-1]=="%s%s"%("v", businput["processing_version"]): dsdaoinput["processed_ds_name"] = businput["processed_ds_name"] else: dbsExceptionHandler('dbsException-invalid-input', "insertDataset:\ processed_ds_name=acquisition_era_name[-filter_name][-processing_str]-vprocessing_version must be satisified.") else: dbsExceptionHandler("dbsException-missing-data", "insertDataset: Required acquisition_era_name or processing_version is not found in the input") if "physics_group_name" in businput: dsdaoinput["physics_group_id"] = self.phygrpid.execute(conn, businput["physics_group_name"]) if dsdaoinput["physics_group_id"] == -1: dbsExceptionHandler("dbsException-missing-data", "insertDataset. physics_group_name not found in DB") else: dsdaoinput["physics_group_id"] = None dsdaoinput["dataset_id"] = self.sm.increment(conn, "SEQ_DS") # we are better off separating out what we need for the dataset DAO dsdaoinput.update({ "dataset" : "/%s/%s/%s" % (businput["primary_ds_name"], businput["processed_ds_name"], businput["data_tier_name"].upper()), "prep_id" : businput.get("prep_id", None), "xtcrosssection" : businput.get("xtcrosssection", None), "creation_date" : businput.get("creation_date", dbsUtils().getTime() ), "create_by" : businput.get("create_by", dbsUtils().getCreateBy()) , "last_modification_date" : businput.get("last_modification_date", dbsUtils().getTime()), #"last_modified_by" : businput.get("last_modified_by", dbsUtils().getModifiedBy()) "last_modified_by" : dbsUtils().getModifiedBy() }) """ repeated again, why? comment out by YG 3/14/2012 #physics group if "physics_group_name" in businput: dsdaoinput["physics_group_id"] = self.phygrpid.execute(conn, businput["physics_group_name"]) if dsdaoinput["physics_group_id"] == -1: dbsExceptionHandler("dbsException-missing-data", "insertDataset. Physics Group : %s Not found" % businput["physics_group_name"]) else: dsdaoinput["physics_group_id"] = None """ # See if Processing Era exists if "processing_version" in businput and businput["processing_version"] != 0: dsdaoinput["processing_era_id"] = self.proceraid.execute(conn, businput["processing_version"]) if dsdaoinput["processing_era_id"] == -1 : dbsExceptionHandler("dbsException-missing-data", "DBSDataset/insertDataset: processing_version not found in DB") else: dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/insertDataset: processing_version is required") # See if Acquisition Era exists if "acquisition_era_name" in businput: dsdaoinput["acquisition_era_id"] = self.acqeraid.execute(conn, businput["acquisition_era_name"]) if dsdaoinput["acquisition_era_id"] == -1: dbsExceptionHandler("dbsException-missing-data", "DBSDataset/insertDataset: acquisition_era_name not found in DB") else: dbsExceptionHandler("dbsException-invalid-input", "DBSDataset/insertDataset: acquisition_era_name is required") try: # insert the dataset self.datasetin.execute(conn, dsdaoinput, tran) except SQLAlchemyIntegrityError as ex: if (str(ex).lower().find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1): # dataset already exists, lets fetch the ID self.logger.warning( "Unique constraint violation being ignored...") self.logger.warning("%s" % ex) ds = "/%s/%s/%s" % (businput["primary_ds_name"], businput["processed_ds_name"], businput["data_tier_name"].upper()) dsdaoinput["dataset_id"] = self.datasetid.execute(conn, ds ) if dsdaoinput["dataset_id"] == -1 : dbsExceptionHandler("dbsException-missing-data", "DBSDataset/insertDataset. Strange error, the dataset %s does not exist ?" % ds ) if (str(ex).find("ORA-01400") ) != -1 : dbsExceptionHandler("dbsException-missing-data", "insertDataset must have: dataset,\ primary_ds_name, processed_ds_name, data_tier_name ") except Exception as e: raise #FIXME : What about the READ-only status of the dataset #There is no READ-oly status for a dataset. # Create dataset_output_mod_mod_configs mapping if "output_configs" in businput: for anOutConfig in businput["output_configs"]: dsoutconfdaoin = {} dsoutconfdaoin["dataset_id"] = dsdaoinput["dataset_id"] dsoutconfdaoin["output_mod_config_id"] = self.outconfigid.execute(conn, anOutConfig["app_name"], anOutConfig["release_version"], anOutConfig["pset_hash"], anOutConfig["output_module_label"], anOutConfig["global_tag"]) if dsoutconfdaoin["output_mod_config_id"] == -1 : dbsExceptionHandler("dbsException-missing-data", "DBSDataset/insertDataset: Output config (%s, %s, %s, %s, %s) not found" % (anOutConfig["app_name"], anOutConfig["release_version"], anOutConfig["pset_hash"], anOutConfig["output_module_label"], anOutConfig["global_tag"])) try: self.datasetoutmodconfigin.execute(conn, dsoutconfdaoin, tran) except Exception as ex: if str(ex).lower().find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1: pass else: raise # Dataset parentage will NOT be added by this API it will be set by insertFiles()--deduced by insertFiles # Dataset runs will NOT be added by this API they will be set by insertFiles()--deduced by insertFiles OR insertRun API call tran.commit() tran = None except Exception: if tran: tran.rollback() tran = None raise finally: if tran: tran.rollback() if conn: conn.close()
[ "def", "insertDataset", "(", "self", ",", "businput", ")", ":", "if", "not", "(", "\"primary_ds_name\"", "in", "businput", "and", "\"dataset\"", "in", "businput", "and", "\"dataset_access_type\"", "in", "businput", "and", "\"processed_ds_name\"", "in", "businput", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"business/DBSDataset/insertDataset must have dataset,\\\n dataset_access_type, primary_ds_name, processed_ds_name as input\"", ")", "if", "\"data_tier_name\"", "not", "in", "businput", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"insertDataset must have data_tier_name as input.\"", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "conn", ".", "begin", "(", ")", "try", ":", "dsdaoinput", "=", "{", "}", "dsdaoinput", "[", "\"primary_ds_name\"", "]", "=", "businput", "[", "\"primary_ds_name\"", "]", "dsdaoinput", "[", "\"data_tier_name\"", "]", "=", "businput", "[", "\"data_tier_name\"", "]", ".", "upper", "(", ")", "dsdaoinput", "[", "\"dataset_access_type\"", "]", "=", "businput", "[", "\"dataset_access_type\"", "]", ".", "upper", "(", ")", "#not required pre-exist in the db. will insert with the dataset if not in yet", "#processed_ds_name=acquisition_era_name[-fileter_name][-processing_str]-vprocessing_version Changed as 4/30/2012 YG.", "#althrough acquisition era and processing version is not required for a dataset in the schema(the schema is build this way because", "#we need to accomdate the DBS2 data), but we impose the requirement on the API. So both acquisition and processing eras are required ", "#YG 12/07/2011 TK-362", "if", "\"acquisition_era_name\"", "in", "businput", "and", "\"processing_version\"", "in", "businput", ":", "erals", "=", "businput", "[", "\"processed_ds_name\"", "]", ".", "rsplit", "(", "'-'", ")", "if", "erals", "[", "0", "]", "==", "businput", "[", "\"acquisition_era_name\"", "]", "and", "erals", "[", "len", "(", "erals", ")", "-", "1", "]", "==", "\"%s%s\"", "%", "(", "\"v\"", ",", "businput", "[", "\"processing_version\"", "]", ")", ":", "dsdaoinput", "[", "\"processed_ds_name\"", "]", "=", "businput", "[", "\"processed_ds_name\"", "]", "else", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"insertDataset:\\\n processed_ds_name=acquisition_era_name[-filter_name][-processing_str]-vprocessing_version must be satisified.\"", ")", "else", ":", "dbsExceptionHandler", "(", "\"dbsException-missing-data\"", ",", "\"insertDataset: Required acquisition_era_name or processing_version is not found in the input\"", ")", "if", "\"physics_group_name\"", "in", "businput", ":", "dsdaoinput", "[", "\"physics_group_id\"", "]", "=", "self", ".", "phygrpid", ".", "execute", "(", "conn", ",", "businput", "[", "\"physics_group_name\"", "]", ")", "if", "dsdaoinput", "[", "\"physics_group_id\"", "]", "==", "-", "1", ":", "dbsExceptionHandler", "(", "\"dbsException-missing-data\"", ",", "\"insertDataset. physics_group_name not found in DB\"", ")", "else", ":", "dsdaoinput", "[", "\"physics_group_id\"", "]", "=", "None", "dsdaoinput", "[", "\"dataset_id\"", "]", "=", "self", ".", "sm", ".", "increment", "(", "conn", ",", "\"SEQ_DS\"", ")", "# we are better off separating out what we need for the dataset DAO", "dsdaoinput", ".", "update", "(", "{", "\"dataset\"", ":", "\"/%s/%s/%s\"", "%", "(", "businput", "[", "\"primary_ds_name\"", "]", ",", "businput", "[", "\"processed_ds_name\"", "]", ",", "businput", "[", "\"data_tier_name\"", "]", ".", "upper", "(", ")", ")", ",", "\"prep_id\"", ":", "businput", ".", "get", "(", "\"prep_id\"", ",", "None", ")", ",", "\"xtcrosssection\"", ":", "businput", ".", "get", "(", "\"xtcrosssection\"", ",", "None", ")", ",", "\"creation_date\"", ":", "businput", ".", "get", "(", "\"creation_date\"", ",", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ")", ",", "\"create_by\"", ":", "businput", ".", "get", "(", "\"create_by\"", ",", "dbsUtils", "(", ")", ".", "getCreateBy", "(", ")", ")", ",", "\"last_modification_date\"", ":", "businput", ".", "get", "(", "\"last_modification_date\"", ",", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ")", ",", "#\"last_modified_by\" : businput.get(\"last_modified_by\", dbsUtils().getModifiedBy())", "\"last_modified_by\"", ":", "dbsUtils", "(", ")", ".", "getModifiedBy", "(", ")", "}", ")", "\"\"\"\n repeated again, why? comment out by YG 3/14/2012\n #physics group\n if \"physics_group_name\" in businput:\n dsdaoinput[\"physics_group_id\"] = self.phygrpid.execute(conn, businput[\"physics_group_name\"])\n if dsdaoinput[\"physics_group_id\"] == -1:\n dbsExceptionHandler(\"dbsException-missing-data\", \"insertDataset. Physics Group : %s Not found\"\n % businput[\"physics_group_name\"])\n else: dsdaoinput[\"physics_group_id\"] = None\n \"\"\"", "# See if Processing Era exists", "if", "\"processing_version\"", "in", "businput", "and", "businput", "[", "\"processing_version\"", "]", "!=", "0", ":", "dsdaoinput", "[", "\"processing_era_id\"", "]", "=", "self", ".", "proceraid", ".", "execute", "(", "conn", ",", "businput", "[", "\"processing_version\"", "]", ")", "if", "dsdaoinput", "[", "\"processing_era_id\"", "]", "==", "-", "1", ":", "dbsExceptionHandler", "(", "\"dbsException-missing-data\"", ",", "\"DBSDataset/insertDataset: processing_version not found in DB\"", ")", "else", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSDataset/insertDataset: processing_version is required\"", ")", "# See if Acquisition Era exists", "if", "\"acquisition_era_name\"", "in", "businput", ":", "dsdaoinput", "[", "\"acquisition_era_id\"", "]", "=", "self", ".", "acqeraid", ".", "execute", "(", "conn", ",", "businput", "[", "\"acquisition_era_name\"", "]", ")", "if", "dsdaoinput", "[", "\"acquisition_era_id\"", "]", "==", "-", "1", ":", "dbsExceptionHandler", "(", "\"dbsException-missing-data\"", ",", "\"DBSDataset/insertDataset: acquisition_era_name not found in DB\"", ")", "else", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSDataset/insertDataset: acquisition_era_name is required\"", ")", "try", ":", "# insert the dataset", "self", ".", "datasetin", ".", "execute", "(", "conn", ",", "dsdaoinput", ",", "tran", ")", "except", "SQLAlchemyIntegrityError", "as", "ex", ":", "if", "(", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"unique constraint\"", ")", "!=", "-", "1", "or", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"duplicate\"", ")", "!=", "-", "1", ")", ":", "# dataset already exists, lets fetch the ID", "self", ".", "logger", ".", "warning", "(", "\"Unique constraint violation being ignored...\"", ")", "self", ".", "logger", ".", "warning", "(", "\"%s\"", "%", "ex", ")", "ds", "=", "\"/%s/%s/%s\"", "%", "(", "businput", "[", "\"primary_ds_name\"", "]", ",", "businput", "[", "\"processed_ds_name\"", "]", ",", "businput", "[", "\"data_tier_name\"", "]", ".", "upper", "(", ")", ")", "dsdaoinput", "[", "\"dataset_id\"", "]", "=", "self", ".", "datasetid", ".", "execute", "(", "conn", ",", "ds", ")", "if", "dsdaoinput", "[", "\"dataset_id\"", "]", "==", "-", "1", ":", "dbsExceptionHandler", "(", "\"dbsException-missing-data\"", ",", "\"DBSDataset/insertDataset. Strange error, the dataset %s does not exist ?\"", "%", "ds", ")", "if", "(", "str", "(", "ex", ")", ".", "find", "(", "\"ORA-01400\"", ")", ")", "!=", "-", "1", ":", "dbsExceptionHandler", "(", "\"dbsException-missing-data\"", ",", "\"insertDataset must have: dataset,\\\n primary_ds_name, processed_ds_name, data_tier_name \"", ")", "except", "Exception", "as", "e", ":", "raise", "#FIXME : What about the READ-only status of the dataset", "#There is no READ-oly status for a dataset.", "# Create dataset_output_mod_mod_configs mapping", "if", "\"output_configs\"", "in", "businput", ":", "for", "anOutConfig", "in", "businput", "[", "\"output_configs\"", "]", ":", "dsoutconfdaoin", "=", "{", "}", "dsoutconfdaoin", "[", "\"dataset_id\"", "]", "=", "dsdaoinput", "[", "\"dataset_id\"", "]", "dsoutconfdaoin", "[", "\"output_mod_config_id\"", "]", "=", "self", ".", "outconfigid", ".", "execute", "(", "conn", ",", "anOutConfig", "[", "\"app_name\"", "]", ",", "anOutConfig", "[", "\"release_version\"", "]", ",", "anOutConfig", "[", "\"pset_hash\"", "]", ",", "anOutConfig", "[", "\"output_module_label\"", "]", ",", "anOutConfig", "[", "\"global_tag\"", "]", ")", "if", "dsoutconfdaoin", "[", "\"output_mod_config_id\"", "]", "==", "-", "1", ":", "dbsExceptionHandler", "(", "\"dbsException-missing-data\"", ",", "\"DBSDataset/insertDataset: Output config (%s, %s, %s, %s, %s) not found\"", "%", "(", "anOutConfig", "[", "\"app_name\"", "]", ",", "anOutConfig", "[", "\"release_version\"", "]", ",", "anOutConfig", "[", "\"pset_hash\"", "]", ",", "anOutConfig", "[", "\"output_module_label\"", "]", ",", "anOutConfig", "[", "\"global_tag\"", "]", ")", ")", "try", ":", "self", ".", "datasetoutmodconfigin", ".", "execute", "(", "conn", ",", "dsoutconfdaoin", ",", "tran", ")", "except", "Exception", "as", "ex", ":", "if", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"unique constraint\"", ")", "!=", "-", "1", "or", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"duplicate\"", ")", "!=", "-", "1", ":", "pass", "else", ":", "raise", "# Dataset parentage will NOT be added by this API it will be set by insertFiles()--deduced by insertFiles", "# Dataset runs will NOT be added by this API they will be set by insertFiles()--deduced by insertFiles OR insertRun API call", "tran", ".", "commit", "(", ")", "tran", "=", "None", "except", "Exception", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "tran", "=", "None", "raise", "finally", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
input dictionary must have the following keys: dataset, primary_ds_name(name), processed_ds(name), data_tier(name), acquisition_era(name), processing_version It may have following keys: physics_group(name), xtcrosssection, creation_date, create_by, last_modification_date, last_modified_by
[ "input", "dictionary", "must", "have", "the", "following", "keys", ":", "dataset", "primary_ds_name", "(", "name", ")", "processed_ds", "(", "name", ")", "data_tier", "(", "name", ")", "acquisition_era", "(", "name", ")", "processing_version", "It", "may", "have", "following", "keys", ":", "physics_group", "(", "name", ")", "xtcrosssection", "creation_date", "create_by", "last_modification_date", "last_modified_by" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSDataset.py#L188-L337
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/Block/UpdateSiteName.py
UpdateSiteName.execute
def execute(self, conn, block_name, origin_site_name, transaction=False): """ Update origin_site_name for a given block_name """ if not conn: dbsExceptionHandler("dbsException-failed-connect2host", "Oracle/Block/UpdateStatus. \ Expects db connection from upper layer.", self.logger.exception) binds = {"block_name": block_name, "origin_site_name": origin_site_name, "mtime": dbsUtils().getTime(), "myuser": dbsUtils().getCreateBy()} self.dbi.processData(self.sql, binds, conn, transaction)
python
def execute(self, conn, block_name, origin_site_name, transaction=False): """ Update origin_site_name for a given block_name """ if not conn: dbsExceptionHandler("dbsException-failed-connect2host", "Oracle/Block/UpdateStatus. \ Expects db connection from upper layer.", self.logger.exception) binds = {"block_name": block_name, "origin_site_name": origin_site_name, "mtime": dbsUtils().getTime(), "myuser": dbsUtils().getCreateBy()} self.dbi.processData(self.sql, binds, conn, transaction)
[ "def", "execute", "(", "self", ",", "conn", ",", "block_name", ",", "origin_site_name", ",", "transaction", "=", "False", ")", ":", "if", "not", "conn", ":", "dbsExceptionHandler", "(", "\"dbsException-failed-connect2host\"", ",", "\"Oracle/Block/UpdateStatus. \\\nExpects db connection from upper layer.\"", ",", "self", ".", "logger", ".", "exception", ")", "binds", "=", "{", "\"block_name\"", ":", "block_name", ",", "\"origin_site_name\"", ":", "origin_site_name", ",", "\"mtime\"", ":", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ",", "\"myuser\"", ":", "dbsUtils", "(", ")", ".", "getCreateBy", "(", ")", "}", "self", ".", "dbi", ".", "processData", "(", "self", ".", "sql", ",", "binds", ",", "conn", ",", "transaction", ")" ]
Update origin_site_name for a given block_name
[ "Update", "origin_site_name", "for", "a", "given", "block_name" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/Block/UpdateSiteName.py#L25-L34
train
dmwm/DBS
Server/Python/src/dbs/dao/MySQL/SequenceManager.py
SequenceManager.increment
def increment(self, conn, seqName, transaction = False, incCount=1): """ increments the sequence `seqName` by default `Incremented by one` and returns its value """ try: seqTable = "%sS" %seqName tlock = "lock tables %s write" %seqTable self.dbi.processData(tlock, [], conn, transaction) sql = "select ID from %s" % seqTable result = self.dbi.processData(sql, [], conn, transaction) resultlist = self.formatDict(result) newSeq = resultlist[0]['id']+incCount sql = "UPDATE %s SET ID=:seq_count" % seqTable seqparms={"seq_count" : newSeq} self.dbi.processData(sql, seqparms, conn, transaction) tunlock = "unlock tables" self.dbi.processData(tunlock, [], conn, transaction) return newSeq except: #FIXME tunlock = "unlock tables" self.dbi.processData(tunlock, [], conn, transaction) raise
python
def increment(self, conn, seqName, transaction = False, incCount=1): """ increments the sequence `seqName` by default `Incremented by one` and returns its value """ try: seqTable = "%sS" %seqName tlock = "lock tables %s write" %seqTable self.dbi.processData(tlock, [], conn, transaction) sql = "select ID from %s" % seqTable result = self.dbi.processData(sql, [], conn, transaction) resultlist = self.formatDict(result) newSeq = resultlist[0]['id']+incCount sql = "UPDATE %s SET ID=:seq_count" % seqTable seqparms={"seq_count" : newSeq} self.dbi.processData(sql, seqparms, conn, transaction) tunlock = "unlock tables" self.dbi.processData(tunlock, [], conn, transaction) return newSeq except: #FIXME tunlock = "unlock tables" self.dbi.processData(tunlock, [], conn, transaction) raise
[ "def", "increment", "(", "self", ",", "conn", ",", "seqName", ",", "transaction", "=", "False", ",", "incCount", "=", "1", ")", ":", "try", ":", "seqTable", "=", "\"%sS\"", "%", "seqName", "tlock", "=", "\"lock tables %s write\"", "%", "seqTable", "self", ".", "dbi", ".", "processData", "(", "tlock", ",", "[", "]", ",", "conn", ",", "transaction", ")", "sql", "=", "\"select ID from %s\"", "%", "seqTable", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "[", "]", ",", "conn", ",", "transaction", ")", "resultlist", "=", "self", ".", "formatDict", "(", "result", ")", "newSeq", "=", "resultlist", "[", "0", "]", "[", "'id'", "]", "+", "incCount", "sql", "=", "\"UPDATE %s SET ID=:seq_count\"", "%", "seqTable", "seqparms", "=", "{", "\"seq_count\"", ":", "newSeq", "}", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "seqparms", ",", "conn", ",", "transaction", ")", "tunlock", "=", "\"unlock tables\"", "self", ".", "dbi", ".", "processData", "(", "tunlock", ",", "[", "]", ",", "conn", ",", "transaction", ")", "return", "newSeq", "except", ":", "#FIXME", "tunlock", "=", "\"unlock tables\"", "self", ".", "dbi", ".", "processData", "(", "tunlock", ",", "[", "]", ",", "conn", ",", "transaction", ")", "raise" ]
increments the sequence `seqName` by default `Incremented by one` and returns its value
[ "increments", "the", "sequence", "seqName", "by", "default", "Incremented", "by", "one", "and", "returns", "its", "value" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/MySQL/SequenceManager.py#L15-L38
train
dmwm/DBS
Server/Python/src/dbs/business/DBSRun.py
DBSRun.listRuns
def listRuns(self, run_num=-1, logical_file_name="", block_name="", dataset=""): """ List run known to DBS. """ if( '%' in logical_file_name or '%' in block_name or '%' in dataset ): dbsExceptionHandler('dbsException-invalid-input', " DBSDatasetRun/listRuns. No wildcards are allowed in logical_file_name, block_name or dataset.\n.") conn = self.dbi.connection() tran = False try: ret = self.runlist.execute(conn, run_num, logical_file_name, block_name, dataset, tran) result = [] rnum = [] for i in ret: rnum.append(i['run_num']) result.append({'run_num' : rnum}) return result finally: if conn: conn.close()
python
def listRuns(self, run_num=-1, logical_file_name="", block_name="", dataset=""): """ List run known to DBS. """ if( '%' in logical_file_name or '%' in block_name or '%' in dataset ): dbsExceptionHandler('dbsException-invalid-input', " DBSDatasetRun/listRuns. No wildcards are allowed in logical_file_name, block_name or dataset.\n.") conn = self.dbi.connection() tran = False try: ret = self.runlist.execute(conn, run_num, logical_file_name, block_name, dataset, tran) result = [] rnum = [] for i in ret: rnum.append(i['run_num']) result.append({'run_num' : rnum}) return result finally: if conn: conn.close()
[ "def", "listRuns", "(", "self", ",", "run_num", "=", "-", "1", ",", "logical_file_name", "=", "\"\"", ",", "block_name", "=", "\"\"", ",", "dataset", "=", "\"\"", ")", ":", "if", "(", "'%'", "in", "logical_file_name", "or", "'%'", "in", "block_name", "or", "'%'", "in", "dataset", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\" DBSDatasetRun/listRuns. No wildcards are allowed in logical_file_name, block_name or dataset.\\n.\"", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "False", "try", ":", "ret", "=", "self", ".", "runlist", ".", "execute", "(", "conn", ",", "run_num", ",", "logical_file_name", ",", "block_name", ",", "dataset", ",", "tran", ")", "result", "=", "[", "]", "rnum", "=", "[", "]", "for", "i", "in", "ret", ":", "rnum", ".", "append", "(", "i", "[", "'run_num'", "]", ")", "result", ".", "append", "(", "{", "'run_num'", ":", "rnum", "}", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
List run known to DBS.
[ "List", "run", "known", "to", "DBS", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSRun.py#L27-L48
train
dmwm/DBS
Server/Python/src/dbs/web/DBSWriterModel.py
insertPrimaryDataset
def insertPrimaryDataset(self): """ API to insert A primary dataset in DBS :param primaryDSObj: primary dataset object :type primaryDSObj: dict :key primary_ds_type: TYPE (out of valid types in DBS, MC, DATA) (Required) :key primary_ds_name: Name of the primary dataset (Required) """ try : body = request.body.read() indata = cjson.decode(body) indata = validateJSONInputNoCopy("primds", indata) indata.update({"creation_date": dbsUtils().getTime(), "create_by": dbsUtils().getCreateBy() }) self.dbsPrimaryDataset.insertPrimaryDataset(indata) except cjson.DecodeError as dc: dbsExceptionHandler("dbsException-invalid-input2", "Wrong format/data from insert PrimaryDataset input", self.logger.exception, str(dc)) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.message) except HTTPError as he: raise he except Exception as ex: sError = "DBSWriterModel/insertPrimaryDataset. %s\n Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
python
def insertPrimaryDataset(self): """ API to insert A primary dataset in DBS :param primaryDSObj: primary dataset object :type primaryDSObj: dict :key primary_ds_type: TYPE (out of valid types in DBS, MC, DATA) (Required) :key primary_ds_name: Name of the primary dataset (Required) """ try : body = request.body.read() indata = cjson.decode(body) indata = validateJSONInputNoCopy("primds", indata) indata.update({"creation_date": dbsUtils().getTime(), "create_by": dbsUtils().getCreateBy() }) self.dbsPrimaryDataset.insertPrimaryDataset(indata) except cjson.DecodeError as dc: dbsExceptionHandler("dbsException-invalid-input2", "Wrong format/data from insert PrimaryDataset input", self.logger.exception, str(dc)) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.message) except HTTPError as he: raise he except Exception as ex: sError = "DBSWriterModel/insertPrimaryDataset. %s\n Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
[ "def", "insertPrimaryDataset", "(", "self", ")", ":", "try", ":", "body", "=", "request", ".", "body", ".", "read", "(", ")", "indata", "=", "cjson", ".", "decode", "(", "body", ")", "indata", "=", "validateJSONInputNoCopy", "(", "\"primds\"", ",", "indata", ")", "indata", ".", "update", "(", "{", "\"creation_date\"", ":", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ",", "\"create_by\"", ":", "dbsUtils", "(", ")", ".", "getCreateBy", "(", ")", "}", ")", "self", ".", "dbsPrimaryDataset", ".", "insertPrimaryDataset", "(", "indata", ")", "except", "cjson", ".", "DecodeError", "as", "dc", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input2\"", ",", "\"Wrong format/data from insert PrimaryDataset input\"", ",", "self", ".", "logger", ".", "exception", ",", "str", "(", "dc", ")", ")", "except", "dbsException", "as", "de", ":", "dbsExceptionHandler", "(", "de", ".", "eCode", ",", "de", ".", "message", ",", "self", ".", "logger", ".", "exception", ",", "de", ".", "message", ")", "except", "HTTPError", "as", "he", ":", "raise", "he", "except", "Exception", "as", "ex", ":", "sError", "=", "\"DBSWriterModel/insertPrimaryDataset. %s\\n Exception trace: \\n %s\"", "%", "(", "ex", ",", "traceback", ".", "format_exc", "(", ")", ")", "dbsExceptionHandler", "(", "'dbsException-server-error'", ",", "dbsExceptionCode", "[", "'dbsException-server-error'", "]", ",", "self", ".", "logger", ".", "exception", ",", "sError", ")" ]
API to insert A primary dataset in DBS :param primaryDSObj: primary dataset object :type primaryDSObj: dict :key primary_ds_type: TYPE (out of valid types in DBS, MC, DATA) (Required) :key primary_ds_name: Name of the primary dataset (Required)
[ "API", "to", "insert", "A", "primary", "dataset", "in", "DBS" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSWriterModel.py#L73-L98
train
dmwm/DBS
Server/Python/src/dbs/web/DBSWriterModel.py
updateAcqEraEndDate
def updateAcqEraEndDate(self, acquisition_era_name ="", end_date=0): """ API to update the end_date of an acquisition era :param acquisition_era_name: acquisition_era_name to update (Required) :type acquisition_era_name: str :param end_date: end_date not zero (Required) :type end_date: int """ try: self.dbsAcqEra.UpdateAcqEraEndDate( acquisition_era_name, end_date) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.message) except HTTPError as he: raise he except Exception as ex: sError = "DBSWriterModel/update.AcqEraEndDate %s\n. Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
python
def updateAcqEraEndDate(self, acquisition_era_name ="", end_date=0): """ API to update the end_date of an acquisition era :param acquisition_era_name: acquisition_era_name to update (Required) :type acquisition_era_name: str :param end_date: end_date not zero (Required) :type end_date: int """ try: self.dbsAcqEra.UpdateAcqEraEndDate( acquisition_era_name, end_date) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.message) except HTTPError as he: raise he except Exception as ex: sError = "DBSWriterModel/update.AcqEraEndDate %s\n. Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
[ "def", "updateAcqEraEndDate", "(", "self", ",", "acquisition_era_name", "=", "\"\"", ",", "end_date", "=", "0", ")", ":", "try", ":", "self", ".", "dbsAcqEra", ".", "UpdateAcqEraEndDate", "(", "acquisition_era_name", ",", "end_date", ")", "except", "dbsException", "as", "de", ":", "dbsExceptionHandler", "(", "de", ".", "eCode", ",", "de", ".", "message", ",", "self", ".", "logger", ".", "exception", ",", "de", ".", "message", ")", "except", "HTTPError", "as", "he", ":", "raise", "he", "except", "Exception", "as", "ex", ":", "sError", "=", "\"DBSWriterModel/update.AcqEraEndDate %s\\n. Exception trace: \\n %s\"", "%", "(", "ex", ",", "traceback", ".", "format_exc", "(", ")", ")", "dbsExceptionHandler", "(", "'dbsException-server-error'", ",", "dbsExceptionCode", "[", "'dbsException-server-error'", "]", ",", "self", ".", "logger", ".", "exception", ",", "sError", ")" ]
API to update the end_date of an acquisition era :param acquisition_era_name: acquisition_era_name to update (Required) :type acquisition_era_name: str :param end_date: end_date not zero (Required) :type end_date: int
[ "API", "to", "update", "the", "end_date", "of", "an", "acquisition", "era" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSWriterModel.py#L134-L153
train
dmwm/DBS
Server/Python/src/dbs/web/DBSWriterModel.py
insertBulkBlock
def insertBulkBlock(self): """ API to insert a bulk block :param blockDump: Output of the block dump command :type blockDump: dict """ try: body = request.body.read() indata = cjson.decode(body) if (indata.get("file_parent_list", []) and indata.get("dataset_parent_list", [])): dbsExceptionHandler("dbsException-invalid-input2", "insertBulkBlock: dataset and file parentages cannot be in the input at the same time", self.logger.exception, "insertBulkBlock: datset and file parentages cannot be in the input at the same time.") indata = validateJSONInputNoCopy("blockBulk", indata) self.dbsBlockInsert.putBlock(indata) except cjson.DecodeError as dc: dbsExceptionHandler("dbsException-invalid-input2", "Wrong format/data from insert BulkBlock input", self.logger.exception, str(dc)) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.message) except HTTPError as he: raise he except Exception as ex: #illegal variable name/number if str(ex).find("ORA-01036") != -1: dbsExceptionHandler("dbsException-invalid-input2", "illegal variable name/number from input", self.logger.exception, str(ex)) else: sError = "DBSWriterModel/insertBulkBlock. %s\n. Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
python
def insertBulkBlock(self): """ API to insert a bulk block :param blockDump: Output of the block dump command :type blockDump: dict """ try: body = request.body.read() indata = cjson.decode(body) if (indata.get("file_parent_list", []) and indata.get("dataset_parent_list", [])): dbsExceptionHandler("dbsException-invalid-input2", "insertBulkBlock: dataset and file parentages cannot be in the input at the same time", self.logger.exception, "insertBulkBlock: datset and file parentages cannot be in the input at the same time.") indata = validateJSONInputNoCopy("blockBulk", indata) self.dbsBlockInsert.putBlock(indata) except cjson.DecodeError as dc: dbsExceptionHandler("dbsException-invalid-input2", "Wrong format/data from insert BulkBlock input", self.logger.exception, str(dc)) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.message) except HTTPError as he: raise he except Exception as ex: #illegal variable name/number if str(ex).find("ORA-01036") != -1: dbsExceptionHandler("dbsException-invalid-input2", "illegal variable name/number from input", self.logger.exception, str(ex)) else: sError = "DBSWriterModel/insertBulkBlock. %s\n. Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
[ "def", "insertBulkBlock", "(", "self", ")", ":", "try", ":", "body", "=", "request", ".", "body", ".", "read", "(", ")", "indata", "=", "cjson", ".", "decode", "(", "body", ")", "if", "(", "indata", ".", "get", "(", "\"file_parent_list\"", ",", "[", "]", ")", "and", "indata", ".", "get", "(", "\"dataset_parent_list\"", ",", "[", "]", ")", ")", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input2\"", ",", "\"insertBulkBlock: dataset and file parentages cannot be in the input at the same time\"", ",", "self", ".", "logger", ".", "exception", ",", "\"insertBulkBlock: datset and file parentages cannot be in the input at the same time.\"", ")", "indata", "=", "validateJSONInputNoCopy", "(", "\"blockBulk\"", ",", "indata", ")", "self", ".", "dbsBlockInsert", ".", "putBlock", "(", "indata", ")", "except", "cjson", ".", "DecodeError", "as", "dc", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input2\"", ",", "\"Wrong format/data from insert BulkBlock input\"", ",", "self", ".", "logger", ".", "exception", ",", "str", "(", "dc", ")", ")", "except", "dbsException", "as", "de", ":", "dbsExceptionHandler", "(", "de", ".", "eCode", ",", "de", ".", "message", ",", "self", ".", "logger", ".", "exception", ",", "de", ".", "message", ")", "except", "HTTPError", "as", "he", ":", "raise", "he", "except", "Exception", "as", "ex", ":", "#illegal variable name/number", "if", "str", "(", "ex", ")", ".", "find", "(", "\"ORA-01036\"", ")", "!=", "-", "1", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input2\"", ",", "\"illegal variable name/number from input\"", ",", "self", ".", "logger", ".", "exception", ",", "str", "(", "ex", ")", ")", "else", ":", "sError", "=", "\"DBSWriterModel/insertBulkBlock. %s\\n. Exception trace: \\n %s\"", "%", "(", "ex", ",", "traceback", ".", "format_exc", "(", ")", ")", "dbsExceptionHandler", "(", "'dbsException-server-error'", ",", "dbsExceptionCode", "[", "'dbsException-server-error'", "]", ",", "self", ".", "logger", ".", "exception", ",", "sError", ")" ]
API to insert a bulk block :param blockDump: Output of the block dump command :type blockDump: dict
[ "API", "to", "insert", "a", "bulk", "block" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSWriterModel.py#L254-L283
train
dmwm/DBS
Server/Python/src/dbs/web/DBSWriterModel.py
insertBlock
def insertBlock(self): """ API to insert a block into DBS :param blockObj: Block object :type blockObj: dict :key open_for_writing: Open For Writing (1/0) (Optional, default 1) :key block_size: Block Size (Optional, default 0) :key file_count: File Count (Optional, default 0) :key block_name: Block Name (Required) :key origin_site_name: Origin Site Name (Required) """ try: body = request.body.read() indata = cjson.decode(body) indata = validateJSONInputNoCopy("block", indata) self.dbsBlock.insertBlock(indata) except cjson.DecodeError as dc: dbsExceptionHandler("dbsException-invalid-input2", "Wrong format/data from insert Block input", self.logger.exception, str(dc)) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.message) except Exception as ex: sError = "DBSWriterModel/insertBlock. %s\n. Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
python
def insertBlock(self): """ API to insert a block into DBS :param blockObj: Block object :type blockObj: dict :key open_for_writing: Open For Writing (1/0) (Optional, default 1) :key block_size: Block Size (Optional, default 0) :key file_count: File Count (Optional, default 0) :key block_name: Block Name (Required) :key origin_site_name: Origin Site Name (Required) """ try: body = request.body.read() indata = cjson.decode(body) indata = validateJSONInputNoCopy("block", indata) self.dbsBlock.insertBlock(indata) except cjson.DecodeError as dc: dbsExceptionHandler("dbsException-invalid-input2", "Wrong format/data from insert Block input", self.logger.exception, str(dc)) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.message) except Exception as ex: sError = "DBSWriterModel/insertBlock. %s\n. Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
[ "def", "insertBlock", "(", "self", ")", ":", "try", ":", "body", "=", "request", ".", "body", ".", "read", "(", ")", "indata", "=", "cjson", ".", "decode", "(", "body", ")", "indata", "=", "validateJSONInputNoCopy", "(", "\"block\"", ",", "indata", ")", "self", ".", "dbsBlock", ".", "insertBlock", "(", "indata", ")", "except", "cjson", ".", "DecodeError", "as", "dc", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input2\"", ",", "\"Wrong format/data from insert Block input\"", ",", "self", ".", "logger", ".", "exception", ",", "str", "(", "dc", ")", ")", "except", "dbsException", "as", "de", ":", "dbsExceptionHandler", "(", "de", ".", "eCode", ",", "de", ".", "message", ",", "self", ".", "logger", ".", "exception", ",", "de", ".", "message", ")", "except", "Exception", "as", "ex", ":", "sError", "=", "\"DBSWriterModel/insertBlock. %s\\n. Exception trace: \\n %s\"", "%", "(", "ex", ",", "traceback", ".", "format_exc", "(", ")", ")", "dbsExceptionHandler", "(", "'dbsException-server-error'", ",", "dbsExceptionCode", "[", "'dbsException-server-error'", "]", ",", "self", ".", "logger", ".", "exception", ",", "sError", ")" ]
API to insert a block into DBS :param blockObj: Block object :type blockObj: dict :key open_for_writing: Open For Writing (1/0) (Optional, default 1) :key block_size: Block Size (Optional, default 0) :key file_count: File Count (Optional, default 0) :key block_name: Block Name (Required) :key origin_site_name: Origin Site Name (Required)
[ "API", "to", "insert", "a", "block", "into", "DBS" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSWriterModel.py#L315-L340
train
dmwm/DBS
Server/Python/src/dbs/web/DBSWriterModel.py
insertFile
def insertFile(self, qInserts=False): """ API to insert a list of file into DBS in DBS. Up to 10 files can be inserted in one request. :param qInserts: True means that inserts will be queued instead of done immediately. INSERT QUEUE Manager will perform the inserts, within few minutes. :type qInserts: bool :param filesList: List of dictionaries containing following information :type filesList: list of dicts :key logical_file_name: File to be inserted (str) (Required) :key is_file_valid: (optional, default = 1): (bool) :key block: required: /a/b/c#d (str) :key dataset: required: /a/b/c (str) :key file_type: (optional, default = EDM) one of the predefined types, (str) :key check_sum: (optional, default = '-1') (str) :key event_count: (optional, default = -1) (int) :key file_size: (optional, default = -1.) (float) :key adler32: (optional, default = '') (str) :key md5: (optional, default = '') (str) :key auto_cross_section: (optional, default = -1.) (float) :key file_lumi_list: (optional, default = []) [{'run_num': 123, 'lumi_section_num': 12},{}....] :key file_parent_list: (optional, default = []) [{'file_parent_lfn': 'mylfn'},{}....] :key file_assoc_list: (optional, default = []) [{'file_parent_lfn': 'mylfn'},{}....] :key file_output_config_list: (optional, default = []) [{'app_name':..., 'release_version':..., 'pset_hash':...., output_module_label':...},{}.....] """ if qInserts in (False, 'False'): qInserts=False try: body = request.body.read() indata = cjson.decode(body)["files"] if not isinstance(indata, (list, dict)): dbsExceptionHandler("dbsException-invalid-input", "Invalid Input DataType", self.logger.exception, \ "insertFile expects input as list or dirc") businput = [] if isinstance(indata, dict): indata = [indata] indata = validateJSONInputNoCopy("files", indata) for f in indata: f.update({ #"dataset":f["dataset"], "creation_date": f.get("creation_date", dbsUtils().getTime()), "create_by" : dbsUtils().getCreateBy(), "last_modification_date": f.get("last_modification_date", dbsUtils().getTime()), "last_modified_by": f.get("last_modified_by", dbsUtils().getCreateBy()), "file_lumi_list":f.get("file_lumi_list", []), "file_parent_list":f.get("file_parent_list", []), "file_assoc_list":f.get("assoc_list", []), "file_output_config_list":f.get("file_output_config_list", [])}) businput.append(f) self.dbsFile.insertFile(businput, qInserts) except cjson.DecodeError as dc: dbsExceptionHandler("dbsException-invalid-input2", "Wrong format/data from insert File input", self.logger.exception, str(dc)) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.message) except HTTPError as he: raise he except Exception as ex: sError = "DBSWriterModel/insertFile. %s\n. Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
python
def insertFile(self, qInserts=False): """ API to insert a list of file into DBS in DBS. Up to 10 files can be inserted in one request. :param qInserts: True means that inserts will be queued instead of done immediately. INSERT QUEUE Manager will perform the inserts, within few minutes. :type qInserts: bool :param filesList: List of dictionaries containing following information :type filesList: list of dicts :key logical_file_name: File to be inserted (str) (Required) :key is_file_valid: (optional, default = 1): (bool) :key block: required: /a/b/c#d (str) :key dataset: required: /a/b/c (str) :key file_type: (optional, default = EDM) one of the predefined types, (str) :key check_sum: (optional, default = '-1') (str) :key event_count: (optional, default = -1) (int) :key file_size: (optional, default = -1.) (float) :key adler32: (optional, default = '') (str) :key md5: (optional, default = '') (str) :key auto_cross_section: (optional, default = -1.) (float) :key file_lumi_list: (optional, default = []) [{'run_num': 123, 'lumi_section_num': 12},{}....] :key file_parent_list: (optional, default = []) [{'file_parent_lfn': 'mylfn'},{}....] :key file_assoc_list: (optional, default = []) [{'file_parent_lfn': 'mylfn'},{}....] :key file_output_config_list: (optional, default = []) [{'app_name':..., 'release_version':..., 'pset_hash':...., output_module_label':...},{}.....] """ if qInserts in (False, 'False'): qInserts=False try: body = request.body.read() indata = cjson.decode(body)["files"] if not isinstance(indata, (list, dict)): dbsExceptionHandler("dbsException-invalid-input", "Invalid Input DataType", self.logger.exception, \ "insertFile expects input as list or dirc") businput = [] if isinstance(indata, dict): indata = [indata] indata = validateJSONInputNoCopy("files", indata) for f in indata: f.update({ #"dataset":f["dataset"], "creation_date": f.get("creation_date", dbsUtils().getTime()), "create_by" : dbsUtils().getCreateBy(), "last_modification_date": f.get("last_modification_date", dbsUtils().getTime()), "last_modified_by": f.get("last_modified_by", dbsUtils().getCreateBy()), "file_lumi_list":f.get("file_lumi_list", []), "file_parent_list":f.get("file_parent_list", []), "file_assoc_list":f.get("assoc_list", []), "file_output_config_list":f.get("file_output_config_list", [])}) businput.append(f) self.dbsFile.insertFile(businput, qInserts) except cjson.DecodeError as dc: dbsExceptionHandler("dbsException-invalid-input2", "Wrong format/data from insert File input", self.logger.exception, str(dc)) except dbsException as de: dbsExceptionHandler(de.eCode, de.message, self.logger.exception, de.message) except HTTPError as he: raise he except Exception as ex: sError = "DBSWriterModel/insertFile. %s\n. Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
[ "def", "insertFile", "(", "self", ",", "qInserts", "=", "False", ")", ":", "if", "qInserts", "in", "(", "False", ",", "'False'", ")", ":", "qInserts", "=", "False", "try", ":", "body", "=", "request", ".", "body", ".", "read", "(", ")", "indata", "=", "cjson", ".", "decode", "(", "body", ")", "[", "\"files\"", "]", "if", "not", "isinstance", "(", "indata", ",", "(", "list", ",", "dict", ")", ")", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"Invalid Input DataType\"", ",", "self", ".", "logger", ".", "exception", ",", "\"insertFile expects input as list or dirc\"", ")", "businput", "=", "[", "]", "if", "isinstance", "(", "indata", ",", "dict", ")", ":", "indata", "=", "[", "indata", "]", "indata", "=", "validateJSONInputNoCopy", "(", "\"files\"", ",", "indata", ")", "for", "f", "in", "indata", ":", "f", ".", "update", "(", "{", "#\"dataset\":f[\"dataset\"],", "\"creation_date\"", ":", "f", ".", "get", "(", "\"creation_date\"", ",", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ")", ",", "\"create_by\"", ":", "dbsUtils", "(", ")", ".", "getCreateBy", "(", ")", ",", "\"last_modification_date\"", ":", "f", ".", "get", "(", "\"last_modification_date\"", ",", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ")", ",", "\"last_modified_by\"", ":", "f", ".", "get", "(", "\"last_modified_by\"", ",", "dbsUtils", "(", ")", ".", "getCreateBy", "(", ")", ")", ",", "\"file_lumi_list\"", ":", "f", ".", "get", "(", "\"file_lumi_list\"", ",", "[", "]", ")", ",", "\"file_parent_list\"", ":", "f", ".", "get", "(", "\"file_parent_list\"", ",", "[", "]", ")", ",", "\"file_assoc_list\"", ":", "f", ".", "get", "(", "\"assoc_list\"", ",", "[", "]", ")", ",", "\"file_output_config_list\"", ":", "f", ".", "get", "(", "\"file_output_config_list\"", ",", "[", "]", ")", "}", ")", "businput", ".", "append", "(", "f", ")", "self", ".", "dbsFile", ".", "insertFile", "(", "businput", ",", "qInserts", ")", "except", "cjson", ".", "DecodeError", "as", "dc", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input2\"", ",", "\"Wrong format/data from insert File input\"", ",", "self", ".", "logger", ".", "exception", ",", "str", "(", "dc", ")", ")", "except", "dbsException", "as", "de", ":", "dbsExceptionHandler", "(", "de", ".", "eCode", ",", "de", ".", "message", ",", "self", ".", "logger", ".", "exception", ",", "de", ".", "message", ")", "except", "HTTPError", "as", "he", ":", "raise", "he", "except", "Exception", "as", "ex", ":", "sError", "=", "\"DBSWriterModel/insertFile. %s\\n. Exception trace: \\n %s\"", "%", "(", "ex", ",", "traceback", ".", "format_exc", "(", ")", ")", "dbsExceptionHandler", "(", "'dbsException-server-error'", ",", "dbsExceptionCode", "[", "'dbsException-server-error'", "]", ",", "self", ".", "logger", ".", "exception", ",", "sError", ")" ]
API to insert a list of file into DBS in DBS. Up to 10 files can be inserted in one request. :param qInserts: True means that inserts will be queued instead of done immediately. INSERT QUEUE Manager will perform the inserts, within few minutes. :type qInserts: bool :param filesList: List of dictionaries containing following information :type filesList: list of dicts :key logical_file_name: File to be inserted (str) (Required) :key is_file_valid: (optional, default = 1): (bool) :key block: required: /a/b/c#d (str) :key dataset: required: /a/b/c (str) :key file_type: (optional, default = EDM) one of the predefined types, (str) :key check_sum: (optional, default = '-1') (str) :key event_count: (optional, default = -1) (int) :key file_size: (optional, default = -1.) (float) :key adler32: (optional, default = '') (str) :key md5: (optional, default = '') (str) :key auto_cross_section: (optional, default = -1.) (float) :key file_lumi_list: (optional, default = []) [{'run_num': 123, 'lumi_section_num': 12},{}....] :key file_parent_list: (optional, default = []) [{'file_parent_lfn': 'mylfn'},{}....] :key file_assoc_list: (optional, default = []) [{'file_parent_lfn': 'mylfn'},{}....] :key file_output_config_list: (optional, default = []) [{'app_name':..., 'release_version':..., 'pset_hash':...., output_module_label':...},{}.....]
[ "API", "to", "insert", "a", "list", "of", "file", "into", "DBS", "in", "DBS", ".", "Up", "to", "10", "files", "can", "be", "inserted", "in", "one", "request", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSWriterModel.py#L342-L400
train
dmwm/DBS
Server/Python/src/dbs/web/DBSWriterModel.py
updateFile
def updateFile(self, logical_file_name=[], is_file_valid=1, lost=0, dataset=''): """ API to update file status :param logical_file_name: logical_file_name to update (optional), but must have either a fln or a dataset :type logical_file_name: str :param is_file_valid: valid=1, invalid=0 (Required) :type is_file_valid: bool :param lost: default lost=0 (optional) :type lost: bool :param dataset: default dataset='' (optional),but must have either a fln or a dataset :type dataset: basestring """ if lost in [1, True, 'True', 'true', '1', 'y', 'yes']: lost = 1 if is_file_valid in [1, True, 'True', 'true', '1', 'y', 'yes']: dbsExceptionHandler("dbsException-invalid-input2", dbsExceptionCode["dbsException-invalid-input2"], self.logger.exception,\ "Lost file must set to invalid" ) else: lost = 0 for f in logical_file_name, dataset: if '*' in f or '%' in f: dbsExceptionHandler("dbsException-invalid-input2", dbsExceptionCode["dbsException-invalid-input2"], self.logger.exception, "No \ wildcard allow in LFN or dataset for updatefile API." ) try: self.dbsFile.updateStatus(logical_file_name, is_file_valid, lost, dataset) except HTTPError as he: raise he except Exception as ex: sError = "DBSWriterModel/updateFile. %s\n. Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
python
def updateFile(self, logical_file_name=[], is_file_valid=1, lost=0, dataset=''): """ API to update file status :param logical_file_name: logical_file_name to update (optional), but must have either a fln or a dataset :type logical_file_name: str :param is_file_valid: valid=1, invalid=0 (Required) :type is_file_valid: bool :param lost: default lost=0 (optional) :type lost: bool :param dataset: default dataset='' (optional),but must have either a fln or a dataset :type dataset: basestring """ if lost in [1, True, 'True', 'true', '1', 'y', 'yes']: lost = 1 if is_file_valid in [1, True, 'True', 'true', '1', 'y', 'yes']: dbsExceptionHandler("dbsException-invalid-input2", dbsExceptionCode["dbsException-invalid-input2"], self.logger.exception,\ "Lost file must set to invalid" ) else: lost = 0 for f in logical_file_name, dataset: if '*' in f or '%' in f: dbsExceptionHandler("dbsException-invalid-input2", dbsExceptionCode["dbsException-invalid-input2"], self.logger.exception, "No \ wildcard allow in LFN or dataset for updatefile API." ) try: self.dbsFile.updateStatus(logical_file_name, is_file_valid, lost, dataset) except HTTPError as he: raise he except Exception as ex: sError = "DBSWriterModel/updateFile. %s\n. Exception trace: \n %s" \ % (ex, traceback.format_exc()) dbsExceptionHandler('dbsException-server-error', dbsExceptionCode['dbsException-server-error'], self.logger.exception, sError)
[ "def", "updateFile", "(", "self", ",", "logical_file_name", "=", "[", "]", ",", "is_file_valid", "=", "1", ",", "lost", "=", "0", ",", "dataset", "=", "''", ")", ":", "if", "lost", "in", "[", "1", ",", "True", ",", "'True'", ",", "'true'", ",", "'1'", ",", "'y'", ",", "'yes'", "]", ":", "lost", "=", "1", "if", "is_file_valid", "in", "[", "1", ",", "True", ",", "'True'", ",", "'true'", ",", "'1'", ",", "'y'", ",", "'yes'", "]", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input2\"", ",", "dbsExceptionCode", "[", "\"dbsException-invalid-input2\"", "]", ",", "self", ".", "logger", ".", "exception", ",", "\"Lost file must set to invalid\"", ")", "else", ":", "lost", "=", "0", "for", "f", "in", "logical_file_name", ",", "dataset", ":", "if", "'*'", "in", "f", "or", "'%'", "in", "f", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input2\"", ",", "dbsExceptionCode", "[", "\"dbsException-invalid-input2\"", "]", ",", "self", ".", "logger", ".", "exception", ",", "\"No \\\n wildcard allow in LFN or dataset for updatefile API.\"", ")", "try", ":", "self", ".", "dbsFile", ".", "updateStatus", "(", "logical_file_name", ",", "is_file_valid", ",", "lost", ",", "dataset", ")", "except", "HTTPError", "as", "he", ":", "raise", "he", "except", "Exception", "as", "ex", ":", "sError", "=", "\"DBSWriterModel/updateFile. %s\\n. Exception trace: \\n %s\"", "%", "(", "ex", ",", "traceback", ".", "format_exc", "(", ")", ")", "dbsExceptionHandler", "(", "'dbsException-server-error'", ",", "dbsExceptionCode", "[", "'dbsException-server-error'", "]", ",", "self", ".", "logger", ".", "exception", ",", "sError", ")" ]
API to update file status :param logical_file_name: logical_file_name to update (optional), but must have either a fln or a dataset :type logical_file_name: str :param is_file_valid: valid=1, invalid=0 (Required) :type is_file_valid: bool :param lost: default lost=0 (optional) :type lost: bool :param dataset: default dataset='' (optional),but must have either a fln or a dataset :type dataset: basestring
[ "API", "to", "update", "file", "status" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSWriterModel.py#L404-L437
train
closeio/flask-ipblock
flask_ipblock/documents.py
IPNetwork.create_from_string
def create_from_string(cls, cidr, label=None, whitelist=False): """ Converts a CIDR like 192.168.0.0/24 into 2 parts: start: 3232235520 stop: 3232235775 """ network = netaddr.IPNetwork(cidr) start = network.first stop = start + network.size - 1 obj = cls.objects.create(label=label, start=start, stop=stop, whitelist=whitelist) return obj
python
def create_from_string(cls, cidr, label=None, whitelist=False): """ Converts a CIDR like 192.168.0.0/24 into 2 parts: start: 3232235520 stop: 3232235775 """ network = netaddr.IPNetwork(cidr) start = network.first stop = start + network.size - 1 obj = cls.objects.create(label=label, start=start, stop=stop, whitelist=whitelist) return obj
[ "def", "create_from_string", "(", "cls", ",", "cidr", ",", "label", "=", "None", ",", "whitelist", "=", "False", ")", ":", "network", "=", "netaddr", ".", "IPNetwork", "(", "cidr", ")", "start", "=", "network", ".", "first", "stop", "=", "start", "+", "network", ".", "size", "-", "1", "obj", "=", "cls", ".", "objects", ".", "create", "(", "label", "=", "label", ",", "start", "=", "start", ",", "stop", "=", "stop", ",", "whitelist", "=", "whitelist", ")", "return", "obj" ]
Converts a CIDR like 192.168.0.0/24 into 2 parts: start: 3232235520 stop: 3232235775
[ "Converts", "a", "CIDR", "like", "192", ".", "168", ".", "0", ".", "0", "/", "24", "into", "2", "parts", ":", "start", ":", "3232235520", "stop", ":", "3232235775" ]
d4d6d50e4a7d1ba7aa0f8d7298ee8d63d25b644f
https://github.com/closeio/flask-ipblock/blob/d4d6d50e4a7d1ba7aa0f8d7298ee8d63d25b644f/flask_ipblock/documents.py#L20-L31
train
closeio/flask-ipblock
flask_ipblock/documents.py
IPNetwork.qs_for_ip
def qs_for_ip(cls, ip_str): """ Returns a queryset with matching IPNetwork objects for the given IP. """ ip = int(netaddr.IPAddress(ip_str)) # ignore IPv6 addresses for now (4294967295 is 0xffffffff, aka the # biggest 32-bit number) if ip > 4294967295: return cls.objects.none() ip_range_query = { 'start__lte': ip, 'stop__gte': ip } return cls.objects.filter(**ip_range_query)
python
def qs_for_ip(cls, ip_str): """ Returns a queryset with matching IPNetwork objects for the given IP. """ ip = int(netaddr.IPAddress(ip_str)) # ignore IPv6 addresses for now (4294967295 is 0xffffffff, aka the # biggest 32-bit number) if ip > 4294967295: return cls.objects.none() ip_range_query = { 'start__lte': ip, 'stop__gte': ip } return cls.objects.filter(**ip_range_query)
[ "def", "qs_for_ip", "(", "cls", ",", "ip_str", ")", ":", "ip", "=", "int", "(", "netaddr", ".", "IPAddress", "(", "ip_str", ")", ")", "# ignore IPv6 addresses for now (4294967295 is 0xffffffff, aka the", "# biggest 32-bit number)", "if", "ip", ">", "4294967295", ":", "return", "cls", ".", "objects", ".", "none", "(", ")", "ip_range_query", "=", "{", "'start__lte'", ":", "ip", ",", "'stop__gte'", ":", "ip", "}", "return", "cls", ".", "objects", ".", "filter", "(", "*", "*", "ip_range_query", ")" ]
Returns a queryset with matching IPNetwork objects for the given IP.
[ "Returns", "a", "queryset", "with", "matching", "IPNetwork", "objects", "for", "the", "given", "IP", "." ]
d4d6d50e4a7d1ba7aa0f8d7298ee8d63d25b644f
https://github.com/closeio/flask-ipblock/blob/d4d6d50e4a7d1ba7aa0f8d7298ee8d63d25b644f/flask_ipblock/documents.py#L40-L56
train
closeio/flask-ipblock
flask_ipblock/documents.py
IPNetwork.matches_ip
def matches_ip(cls, ip_str, read_preference=None): """ Return True if provided IP exists in the blacklist and doesn't exist in the whitelist. Otherwise, return False. """ qs = cls.qs_for_ip(ip_str).only('whitelist') if read_preference: qs = qs.read_preference(read_preference) # Return True if any docs match the IP and none of them represent # a whitelist return bool(qs) and not any(obj.whitelist for obj in qs)
python
def matches_ip(cls, ip_str, read_preference=None): """ Return True if provided IP exists in the blacklist and doesn't exist in the whitelist. Otherwise, return False. """ qs = cls.qs_for_ip(ip_str).only('whitelist') if read_preference: qs = qs.read_preference(read_preference) # Return True if any docs match the IP and none of them represent # a whitelist return bool(qs) and not any(obj.whitelist for obj in qs)
[ "def", "matches_ip", "(", "cls", ",", "ip_str", ",", "read_preference", "=", "None", ")", ":", "qs", "=", "cls", ".", "qs_for_ip", "(", "ip_str", ")", ".", "only", "(", "'whitelist'", ")", "if", "read_preference", ":", "qs", "=", "qs", ".", "read_preference", "(", "read_preference", ")", "# Return True if any docs match the IP and none of them represent", "# a whitelist", "return", "bool", "(", "qs", ")", "and", "not", "any", "(", "obj", ".", "whitelist", "for", "obj", "in", "qs", ")" ]
Return True if provided IP exists in the blacklist and doesn't exist in the whitelist. Otherwise, return False.
[ "Return", "True", "if", "provided", "IP", "exists", "in", "the", "blacklist", "and", "doesn", "t", "exist", "in", "the", "whitelist", ".", "Otherwise", "return", "False", "." ]
d4d6d50e4a7d1ba7aa0f8d7298ee8d63d25b644f
https://github.com/closeio/flask-ipblock/blob/d4d6d50e4a7d1ba7aa0f8d7298ee8d63d25b644f/flask_ipblock/documents.py#L59-L70
train
dmwm/DBS
Server/Python/src/dbs/utils/dbsExceptionHandler.py
dbsExceptionHandler
def dbsExceptionHandler(eCode='', message='', logger=None , serverError=''): """ This utility function handles all dbs exceptions. It will log , raise exception based on input condition. It loggs the traceback on the server log. Send HTTPError 400 for invalid client input and HTTPError 404 for NOT FOUND required pre-existing condition. """ if logger: #HTTP Error if eCode == "dbsException-invalid-input": #logger(eCode + ": " + serverError) raise HTTPError(400, message) elif eCode == "dbsException-missing-data": logger( time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) #print (eCode + ": " + serverError) raise HTTPError(412, message) elif eCode == "dbsException-input-too-large": logger(time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) raise HTTPError(413, message) elif eCode == "dbsException-invalid-input2": logger( time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) raise HTTPError(400, message) elif eCode == "dbsException-conflict-data": logger( time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) raise HTTPError(409, message) elif eCode == "dbsException-failed-connect2host": logger( time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) raise HTTPError(443, message) else: #client gets httperror 500 for server internal error #print eCode + ": " + serverError logger( time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) raise HTTPError(500, message) else: #not HTTP Error raise dbsException(eCode, message, serverError)
python
def dbsExceptionHandler(eCode='', message='', logger=None , serverError=''): """ This utility function handles all dbs exceptions. It will log , raise exception based on input condition. It loggs the traceback on the server log. Send HTTPError 400 for invalid client input and HTTPError 404 for NOT FOUND required pre-existing condition. """ if logger: #HTTP Error if eCode == "dbsException-invalid-input": #logger(eCode + ": " + serverError) raise HTTPError(400, message) elif eCode == "dbsException-missing-data": logger( time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) #print (eCode + ": " + serverError) raise HTTPError(412, message) elif eCode == "dbsException-input-too-large": logger(time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) raise HTTPError(413, message) elif eCode == "dbsException-invalid-input2": logger( time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) raise HTTPError(400, message) elif eCode == "dbsException-conflict-data": logger( time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) raise HTTPError(409, message) elif eCode == "dbsException-failed-connect2host": logger( time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) raise HTTPError(443, message) else: #client gets httperror 500 for server internal error #print eCode + ": " + serverError logger( time.asctime(time.gmtime()) + " " + eCode + ": " + serverError) raise HTTPError(500, message) else: #not HTTP Error raise dbsException(eCode, message, serverError)
[ "def", "dbsExceptionHandler", "(", "eCode", "=", "''", ",", "message", "=", "''", ",", "logger", "=", "None", ",", "serverError", "=", "''", ")", ":", "if", "logger", ":", "#HTTP Error", "if", "eCode", "==", "\"dbsException-invalid-input\"", ":", "#logger(eCode + \": \" + serverError)", "raise", "HTTPError", "(", "400", ",", "message", ")", "elif", "eCode", "==", "\"dbsException-missing-data\"", ":", "logger", "(", "time", ".", "asctime", "(", "time", ".", "gmtime", "(", ")", ")", "+", "\" \"", "+", "eCode", "+", "\": \"", "+", "serverError", ")", "#print (eCode + \": \" + serverError)", "raise", "HTTPError", "(", "412", ",", "message", ")", "elif", "eCode", "==", "\"dbsException-input-too-large\"", ":", "logger", "(", "time", ".", "asctime", "(", "time", ".", "gmtime", "(", ")", ")", "+", "\" \"", "+", "eCode", "+", "\": \"", "+", "serverError", ")", "raise", "HTTPError", "(", "413", ",", "message", ")", "elif", "eCode", "==", "\"dbsException-invalid-input2\"", ":", "logger", "(", "time", ".", "asctime", "(", "time", ".", "gmtime", "(", ")", ")", "+", "\" \"", "+", "eCode", "+", "\": \"", "+", "serverError", ")", "raise", "HTTPError", "(", "400", ",", "message", ")", "elif", "eCode", "==", "\"dbsException-conflict-data\"", ":", "logger", "(", "time", ".", "asctime", "(", "time", ".", "gmtime", "(", ")", ")", "+", "\" \"", "+", "eCode", "+", "\": \"", "+", "serverError", ")", "raise", "HTTPError", "(", "409", ",", "message", ")", "elif", "eCode", "==", "\"dbsException-failed-connect2host\"", ":", "logger", "(", "time", ".", "asctime", "(", "time", ".", "gmtime", "(", ")", ")", "+", "\" \"", "+", "eCode", "+", "\": \"", "+", "serverError", ")", "raise", "HTTPError", "(", "443", ",", "message", ")", "else", ":", "#client gets httperror 500 for server internal error", "#print eCode + \": \" + serverError", "logger", "(", "time", ".", "asctime", "(", "time", ".", "gmtime", "(", ")", ")", "+", "\" \"", "+", "eCode", "+", "\": \"", "+", "serverError", ")", "raise", "HTTPError", "(", "500", ",", "message", ")", "else", ":", "#not HTTP Error", "raise", "dbsException", "(", "eCode", ",", "message", ",", "serverError", ")" ]
This utility function handles all dbs exceptions. It will log , raise exception based on input condition. It loggs the traceback on the server log. Send HTTPError 400 for invalid client input and HTTPError 404 for NOT FOUND required pre-existing condition.
[ "This", "utility", "function", "handles", "all", "dbs", "exceptions", ".", "It", "will", "log", "raise", "exception", "based", "on", "input", "condition", ".", "It", "loggs", "the", "traceback", "on", "the", "server", "log", ".", "Send", "HTTPError", "400", "for", "invalid", "client", "input", "and", "HTTPError", "404", "for", "NOT", "FOUND", "required", "pre", "-", "existing", "condition", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/utils/dbsExceptionHandler.py#L6-L40
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/FileBuffer/List.py
List.execute
def execute(self, conn, block_id="", transaction=False): """ simple execute """ if not conn: dbsExceptionHandler("dbsException-db-conn-failed", "Oracle/FileBuffer/List. Expects db connection from upper layer.") sql = self.sql binds = { "block_id" : block_id} cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = self.formatCursor(cursors[0]) return result
python
def execute(self, conn, block_id="", transaction=False): """ simple execute """ if not conn: dbsExceptionHandler("dbsException-db-conn-failed", "Oracle/FileBuffer/List. Expects db connection from upper layer.") sql = self.sql binds = { "block_id" : block_id} cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = self.formatCursor(cursors[0]) return result
[ "def", "execute", "(", "self", ",", "conn", ",", "block_id", "=", "\"\"", ",", "transaction", "=", "False", ")", ":", "if", "not", "conn", ":", "dbsExceptionHandler", "(", "\"dbsException-db-conn-failed\"", ",", "\"Oracle/FileBuffer/List. Expects db connection from upper layer.\"", ")", "sql", "=", "self", ".", "sql", "binds", "=", "{", "\"block_id\"", ":", "block_id", "}", "cursors", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ",", "returnCursor", "=", "True", ")", "result", "=", "self", ".", "formatCursor", "(", "cursors", "[", "0", "]", ")", "return", "result" ]
simple execute
[ "simple", "execute" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/FileBuffer/List.py#L20-L31
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/PrimaryDataset/List.py
List.execute
def execute(self, conn, primary_ds_name="", primary_ds_type="", transaction=False): """ Lists all primary datasets if pattern is not provided. """ sql = self.sql binds = {} #import pdb #pdb.set_trace() if primary_ds_name and primary_ds_type in ('', None, '%'): op = ("=", "like")["%" in primary_ds_name] sql += "WHERE P.PRIMARY_DS_NAME %s :primary_ds_name" % op binds.update(primary_ds_name=primary_ds_name) elif primary_ds_type and primary_ds_name in ('', None, '%'): op = ("=", "like")["%" in primary_ds_type] sql += "WHERE PT.PRIMARY_DS_TYPE %s :primary_ds_type" % op binds.update(primary_ds_type=primary_ds_type) elif primary_ds_name and primary_ds_type: op = ("=", "like")["%" in primary_ds_name] op1 = ("=", "like")["%" in primary_ds_type] sql += "WHERE P.PRIMARY_DS_NAME %s :primary_ds_name and PT.PRIMARY_DS_TYPE %s :primary_ds_type"\ %(op, op1) binds.update(primary_ds_name=primary_ds_name) binds.update(primary_ds_type=primary_ds_type) else: pass cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = [] for c in cursors: result.extend(self.formatCursor(c, size=100)) return result
python
def execute(self, conn, primary_ds_name="", primary_ds_type="", transaction=False): """ Lists all primary datasets if pattern is not provided. """ sql = self.sql binds = {} #import pdb #pdb.set_trace() if primary_ds_name and primary_ds_type in ('', None, '%'): op = ("=", "like")["%" in primary_ds_name] sql += "WHERE P.PRIMARY_DS_NAME %s :primary_ds_name" % op binds.update(primary_ds_name=primary_ds_name) elif primary_ds_type and primary_ds_name in ('', None, '%'): op = ("=", "like")["%" in primary_ds_type] sql += "WHERE PT.PRIMARY_DS_TYPE %s :primary_ds_type" % op binds.update(primary_ds_type=primary_ds_type) elif primary_ds_name and primary_ds_type: op = ("=", "like")["%" in primary_ds_name] op1 = ("=", "like")["%" in primary_ds_type] sql += "WHERE P.PRIMARY_DS_NAME %s :primary_ds_name and PT.PRIMARY_DS_TYPE %s :primary_ds_type"\ %(op, op1) binds.update(primary_ds_name=primary_ds_name) binds.update(primary_ds_type=primary_ds_type) else: pass cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = [] for c in cursors: result.extend(self.formatCursor(c, size=100)) return result
[ "def", "execute", "(", "self", ",", "conn", ",", "primary_ds_name", "=", "\"\"", ",", "primary_ds_type", "=", "\"\"", ",", "transaction", "=", "False", ")", ":", "sql", "=", "self", ".", "sql", "binds", "=", "{", "}", "#import pdb", "#pdb.set_trace()", "if", "primary_ds_name", "and", "primary_ds_type", "in", "(", "''", ",", "None", ",", "'%'", ")", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "primary_ds_name", "]", "sql", "+=", "\"WHERE P.PRIMARY_DS_NAME %s :primary_ds_name\"", "%", "op", "binds", ".", "update", "(", "primary_ds_name", "=", "primary_ds_name", ")", "elif", "primary_ds_type", "and", "primary_ds_name", "in", "(", "''", ",", "None", ",", "'%'", ")", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "primary_ds_type", "]", "sql", "+=", "\"WHERE PT.PRIMARY_DS_TYPE %s :primary_ds_type\"", "%", "op", "binds", ".", "update", "(", "primary_ds_type", "=", "primary_ds_type", ")", "elif", "primary_ds_name", "and", "primary_ds_type", ":", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "primary_ds_name", "]", "op1", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "primary_ds_type", "]", "sql", "+=", "\"WHERE P.PRIMARY_DS_NAME %s :primary_ds_name and PT.PRIMARY_DS_TYPE %s :primary_ds_type\"", "%", "(", "op", ",", "op1", ")", "binds", ".", "update", "(", "primary_ds_name", "=", "primary_ds_name", ")", "binds", ".", "update", "(", "primary_ds_type", "=", "primary_ds_type", ")", "else", ":", "pass", "cursors", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ",", "returnCursor", "=", "True", ")", "result", "=", "[", "]", "for", "c", "in", "cursors", ":", "result", ".", "extend", "(", "self", ".", "formatCursor", "(", "c", ",", "size", "=", "100", ")", ")", "return", "result" ]
Lists all primary datasets if pattern is not provided.
[ "Lists", "all", "primary", "datasets", "if", "pattern", "is", "not", "provided", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/PrimaryDataset/List.py#L31-L60
train
dmwm/DBS
PycurlClient/src/python/RestClient/ProxyPlugins/Socks5Proxy.py
Socks5Proxy.configure_proxy
def configure_proxy(self, curl_object): """configure pycurl proxy settings""" curl_object.setopt(curl_object.PROXY, self._proxy_hostname) curl_object.setopt(curl_object.PROXYPORT, self._proxy_port) curl_object.setopt(curl_object.PROXYTYPE, curl_object.PROXYTYPE_SOCKS5) if self._proxy_user and self._proxy_passwd: curl_object.setopt(curl_object.PROXYUSERPWD, '%s:%s' % (self._proxy_user, self._proxy_port))
python
def configure_proxy(self, curl_object): """configure pycurl proxy settings""" curl_object.setopt(curl_object.PROXY, self._proxy_hostname) curl_object.setopt(curl_object.PROXYPORT, self._proxy_port) curl_object.setopt(curl_object.PROXYTYPE, curl_object.PROXYTYPE_SOCKS5) if self._proxy_user and self._proxy_passwd: curl_object.setopt(curl_object.PROXYUSERPWD, '%s:%s' % (self._proxy_user, self._proxy_port))
[ "def", "configure_proxy", "(", "self", ",", "curl_object", ")", ":", "curl_object", ".", "setopt", "(", "curl_object", ".", "PROXY", ",", "self", ".", "_proxy_hostname", ")", "curl_object", ".", "setopt", "(", "curl_object", ".", "PROXYPORT", ",", "self", ".", "_proxy_port", ")", "curl_object", ".", "setopt", "(", "curl_object", ".", "PROXYTYPE", ",", "curl_object", ".", "PROXYTYPE_SOCKS5", ")", "if", "self", ".", "_proxy_user", "and", "self", ".", "_proxy_passwd", ":", "curl_object", ".", "setopt", "(", "curl_object", ".", "PROXYUSERPWD", ",", "'%s:%s'", "%", "(", "self", ".", "_proxy_user", ",", "self", ".", "_proxy_port", ")", ")" ]
configure pycurl proxy settings
[ "configure", "pycurl", "proxy", "settings" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/PycurlClient/src/python/RestClient/ProxyPlugins/Socks5Proxy.py#L13-L19
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/FileParent/Insert2.py
Insert2.execute
def execute( self, conn, daoinput, transaction = False ): """ daoinput must be validated to have the following keys: child_parent_id__list[[cid, pid],...], block_name """ binds = {} bindlist=[] if isinstance(daoinput, dict) and "block_name" in daoinput.keys(): binds = {"block_name": daoinput["block_name"]} r = self.dbi.processData(self.sql_sel, binds, conn, False) bfile = self.format(r) bfile_list = [] for f in bfile: bfile_list.append(f[0]) if "child_parent_id_list" in daoinput.keys(): files = [] for i in daoinput["child_parent_id_list"]: files.append(i[0]) if set(files)-set(bfile_list): dbsExceptionHandler('dbsException-invalid-input2', "Files required in the same block for FileParent/insert2 dao.", self.logger.exception) else: dbsExceptionHandler('dbsException-invalid-input2', "child_parent_id_list required for FileParent/insert2 dao.", self.logger.exception) else: dbsExceptionHandler('dbsException-invalid-input2', "Block_name required in the same block for FileParent/insert2 dao.", self.logger.exception) binds = {} for pf in daoinput["child_parent_id_list"]: binds = {"this_file_id":pf[0], "parent_file_id": pf[1]} bindlist.append(binds) self.dbi.processData(self.sql, bindlist, conn, transaction)
python
def execute( self, conn, daoinput, transaction = False ): """ daoinput must be validated to have the following keys: child_parent_id__list[[cid, pid],...], block_name """ binds = {} bindlist=[] if isinstance(daoinput, dict) and "block_name" in daoinput.keys(): binds = {"block_name": daoinput["block_name"]} r = self.dbi.processData(self.sql_sel, binds, conn, False) bfile = self.format(r) bfile_list = [] for f in bfile: bfile_list.append(f[0]) if "child_parent_id_list" in daoinput.keys(): files = [] for i in daoinput["child_parent_id_list"]: files.append(i[0]) if set(files)-set(bfile_list): dbsExceptionHandler('dbsException-invalid-input2', "Files required in the same block for FileParent/insert2 dao.", self.logger.exception) else: dbsExceptionHandler('dbsException-invalid-input2', "child_parent_id_list required for FileParent/insert2 dao.", self.logger.exception) else: dbsExceptionHandler('dbsException-invalid-input2', "Block_name required in the same block for FileParent/insert2 dao.", self.logger.exception) binds = {} for pf in daoinput["child_parent_id_list"]: binds = {"this_file_id":pf[0], "parent_file_id": pf[1]} bindlist.append(binds) self.dbi.processData(self.sql, bindlist, conn, transaction)
[ "def", "execute", "(", "self", ",", "conn", ",", "daoinput", ",", "transaction", "=", "False", ")", ":", "binds", "=", "{", "}", "bindlist", "=", "[", "]", "if", "isinstance", "(", "daoinput", ",", "dict", ")", "and", "\"block_name\"", "in", "daoinput", ".", "keys", "(", ")", ":", "binds", "=", "{", "\"block_name\"", ":", "daoinput", "[", "\"block_name\"", "]", "}", "r", "=", "self", ".", "dbi", ".", "processData", "(", "self", ".", "sql_sel", ",", "binds", ",", "conn", ",", "False", ")", "bfile", "=", "self", ".", "format", "(", "r", ")", "bfile_list", "=", "[", "]", "for", "f", "in", "bfile", ":", "bfile_list", ".", "append", "(", "f", "[", "0", "]", ")", "if", "\"child_parent_id_list\"", "in", "daoinput", ".", "keys", "(", ")", ":", "files", "=", "[", "]", "for", "i", "in", "daoinput", "[", "\"child_parent_id_list\"", "]", ":", "files", ".", "append", "(", "i", "[", "0", "]", ")", "if", "set", "(", "files", ")", "-", "set", "(", "bfile_list", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "\"Files required in the same block for FileParent/insert2 dao.\"", ",", "self", ".", "logger", ".", "exception", ")", "else", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "\"child_parent_id_list required for FileParent/insert2 dao.\"", ",", "self", ".", "logger", ".", "exception", ")", "else", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "\"Block_name required in the same block for FileParent/insert2 dao.\"", ",", "self", ".", "logger", ".", "exception", ")", "binds", "=", "{", "}", "for", "pf", "in", "daoinput", "[", "\"child_parent_id_list\"", "]", ":", "binds", "=", "{", "\"this_file_id\"", ":", "pf", "[", "0", "]", ",", "\"parent_file_id\"", ":", "pf", "[", "1", "]", "}", "bindlist", ".", "append", "(", "binds", ")", "self", ".", "dbi", ".", "processData", "(", "self", ".", "sql", ",", "bindlist", ",", "conn", ",", "transaction", ")" ]
daoinput must be validated to have the following keys: child_parent_id__list[[cid, pid],...], block_name
[ "daoinput", "must", "be", "validated", "to", "have", "the", "following", "keys", ":", "child_parent_id__list", "[[", "cid", "pid", "]", "...", "]", "block_name" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/FileParent/Insert2.py#L26-L55
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/AcquisitionEra/UpdateEndDate.py
UpdateEndDate.execute
def execute(self, conn, acquisition_era_name,end_date, transaction = False): """ for a given block_id """ if not conn: dbsExceptionHandler("dbsException-failed-connect2host", "dbs/dao/Oracle/AcquisitionEra/updateEndDate expects db connection from upper layer.", self.logger.exception) binds = { "acquisition_era_name" :acquisition_era_name , "end_date" : end_date } result = self.dbi.processData(self.sql, binds, conn, transaction)
python
def execute(self, conn, acquisition_era_name,end_date, transaction = False): """ for a given block_id """ if not conn: dbsExceptionHandler("dbsException-failed-connect2host", "dbs/dao/Oracle/AcquisitionEra/updateEndDate expects db connection from upper layer.", self.logger.exception) binds = { "acquisition_era_name" :acquisition_era_name , "end_date" : end_date } result = self.dbi.processData(self.sql, binds, conn, transaction)
[ "def", "execute", "(", "self", ",", "conn", ",", "acquisition_era_name", ",", "end_date", ",", "transaction", "=", "False", ")", ":", "if", "not", "conn", ":", "dbsExceptionHandler", "(", "\"dbsException-failed-connect2host\"", ",", "\"dbs/dao/Oracle/AcquisitionEra/updateEndDate expects db connection from upper layer.\"", ",", "self", ".", "logger", ".", "exception", ")", "binds", "=", "{", "\"acquisition_era_name\"", ":", "acquisition_era_name", ",", "\"end_date\"", ":", "end_date", "}", "result", "=", "self", ".", "dbi", ".", "processData", "(", "self", ".", "sql", ",", "binds", ",", "conn", ",", "transaction", ")" ]
for a given block_id
[ "for", "a", "given", "block_id" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/AcquisitionEra/UpdateEndDate.py#L22-L29
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/MigrationBlock/Update.py
Update.execute
def execute(self, conn, daoinput, transaction = False): """ daoinput keys: migration_status, migration_block_id, migration_request_id """ #print daoinput['migration_block_id'] if not conn: dbsExceptionHandler("dbsException-failed-connect2host", "Oracle/MigrationBlock/Update. Expects db connection from upper layer." ,self.logger.exception) if daoinput['migration_status'] == 1: sql = self.sql + " (MIGRATION_STATUS = 0 or MIGRATION_STATUS = 3)" elif daoinput['migration_status'] == 2 or daoinput['migration_status'] == 3 or daoinput['migration_status'] == 9: sql = self.sql + " MIGRATION_STATUS = 1 " else: dbsExceptionHandler("dbsException-conflict-data", "Oracle/MigrationBlock/Update. Expected migration status to be 1, 2, 3, 0r 9" ,self.logger.exception ) #print sql if 'migration_request_id' in daoinput: sql3 = sql + "and MIGRATION_REQUEST_ID =:migration_request_id" result = self.dbi.processData(sql3, daoinput, conn, transaction) elif 'migration_block_id' in daoinput: if type(daoinput['migration_block_id']) is not list: sql2 = sql+ " and MIGRATION_BLOCK_ID =:migration_block_id" result = self.dbi.processData(sql2, daoinput, conn, transaction) else: bk_id_generator, binds2 = create_token_generator(daoinput['migration_block_id']) newdaoinput = {} newdaoinput.update({"migration_status":daoinput["migration_status"], "last_modification_date":daoinput["last_modification_date"]}) newdaoinput.update(binds2) sql2 = sql+ """ and MIGRATION_BLOCK_ID in ({bk_id_generator} SELECT TOKEN FROM TOKEN_GENERATOR) """.format(bk_id_generator=bk_id_generator) result = self.dbi.processData(sql2, newdaoinput, conn, transaction) else: dbsExceptionHandler("dbsException-conflict-data", "Oracle/MigrationBlock/Update. Required IDs not in the input", self.logger.exception)
python
def execute(self, conn, daoinput, transaction = False): """ daoinput keys: migration_status, migration_block_id, migration_request_id """ #print daoinput['migration_block_id'] if not conn: dbsExceptionHandler("dbsException-failed-connect2host", "Oracle/MigrationBlock/Update. Expects db connection from upper layer." ,self.logger.exception) if daoinput['migration_status'] == 1: sql = self.sql + " (MIGRATION_STATUS = 0 or MIGRATION_STATUS = 3)" elif daoinput['migration_status'] == 2 or daoinput['migration_status'] == 3 or daoinput['migration_status'] == 9: sql = self.sql + " MIGRATION_STATUS = 1 " else: dbsExceptionHandler("dbsException-conflict-data", "Oracle/MigrationBlock/Update. Expected migration status to be 1, 2, 3, 0r 9" ,self.logger.exception ) #print sql if 'migration_request_id' in daoinput: sql3 = sql + "and MIGRATION_REQUEST_ID =:migration_request_id" result = self.dbi.processData(sql3, daoinput, conn, transaction) elif 'migration_block_id' in daoinput: if type(daoinput['migration_block_id']) is not list: sql2 = sql+ " and MIGRATION_BLOCK_ID =:migration_block_id" result = self.dbi.processData(sql2, daoinput, conn, transaction) else: bk_id_generator, binds2 = create_token_generator(daoinput['migration_block_id']) newdaoinput = {} newdaoinput.update({"migration_status":daoinput["migration_status"], "last_modification_date":daoinput["last_modification_date"]}) newdaoinput.update(binds2) sql2 = sql+ """ and MIGRATION_BLOCK_ID in ({bk_id_generator} SELECT TOKEN FROM TOKEN_GENERATOR) """.format(bk_id_generator=bk_id_generator) result = self.dbi.processData(sql2, newdaoinput, conn, transaction) else: dbsExceptionHandler("dbsException-conflict-data", "Oracle/MigrationBlock/Update. Required IDs not in the input", self.logger.exception)
[ "def", "execute", "(", "self", ",", "conn", ",", "daoinput", ",", "transaction", "=", "False", ")", ":", "#print daoinput['migration_block_id']", "if", "not", "conn", ":", "dbsExceptionHandler", "(", "\"dbsException-failed-connect2host\"", ",", "\"Oracle/MigrationBlock/Update. Expects db connection from upper layer.\"", ",", "self", ".", "logger", ".", "exception", ")", "if", "daoinput", "[", "'migration_status'", "]", "==", "1", ":", "sql", "=", "self", ".", "sql", "+", "\" (MIGRATION_STATUS = 0 or MIGRATION_STATUS = 3)\"", "elif", "daoinput", "[", "'migration_status'", "]", "==", "2", "or", "daoinput", "[", "'migration_status'", "]", "==", "3", "or", "daoinput", "[", "'migration_status'", "]", "==", "9", ":", "sql", "=", "self", ".", "sql", "+", "\" MIGRATION_STATUS = 1 \"", "else", ":", "dbsExceptionHandler", "(", "\"dbsException-conflict-data\"", ",", "\"Oracle/MigrationBlock/Update. Expected migration status to be 1, 2, 3, 0r 9\"", ",", "self", ".", "logger", ".", "exception", ")", "#print sql", "if", "'migration_request_id'", "in", "daoinput", ":", "sql3", "=", "sql", "+", "\"and MIGRATION_REQUEST_ID =:migration_request_id\"", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql3", ",", "daoinput", ",", "conn", ",", "transaction", ")", "elif", "'migration_block_id'", "in", "daoinput", ":", "if", "type", "(", "daoinput", "[", "'migration_block_id'", "]", ")", "is", "not", "list", ":", "sql2", "=", "sql", "+", "\" and MIGRATION_BLOCK_ID =:migration_block_id\"", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql2", ",", "daoinput", ",", "conn", ",", "transaction", ")", "else", ":", "bk_id_generator", ",", "binds2", "=", "create_token_generator", "(", "daoinput", "[", "'migration_block_id'", "]", ")", "newdaoinput", "=", "{", "}", "newdaoinput", ".", "update", "(", "{", "\"migration_status\"", ":", "daoinput", "[", "\"migration_status\"", "]", ",", "\"last_modification_date\"", ":", "daoinput", "[", "\"last_modification_date\"", "]", "}", ")", "newdaoinput", ".", "update", "(", "binds2", ")", "sql2", "=", "sql", "+", "\"\"\" and MIGRATION_BLOCK_ID in ({bk_id_generator} SELECT TOKEN FROM TOKEN_GENERATOR)\n \"\"\"", ".", "format", "(", "bk_id_generator", "=", "bk_id_generator", ")", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql2", ",", "newdaoinput", ",", "conn", ",", "transaction", ")", "else", ":", "dbsExceptionHandler", "(", "\"dbsException-conflict-data\"", ",", "\"Oracle/MigrationBlock/Update. Required IDs not in the input\"", ",", "self", ".", "logger", ".", "exception", ")" ]
daoinput keys: migration_status, migration_block_id, migration_request_id
[ "daoinput", "keys", ":", "migration_status", "migration_block_id", "migration_request_id" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/MigrationBlock/Update.py#L40-L72
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/FileParentBlock/List.py
List.execute
def execute(self, conn, file_id_list, transaction=False): """ file_id_list : file_id_list """ sql=self.sql binds={} if file_id_list: count=0 for an_id in file_id_list: if count > 0: sql += ", " sql += ":file_id_%s" %count binds.update({"file_id_%s" %count : an_id}) count+=1 sql += ")" else: dbsExceptionHandler('dbsException-invalid-input', "Oracle/FileParentBlock/List. this_file_id not provided", self.logger.exception) result = self.dbi.processData(sql, binds, conn, transaction) plist = self.formatDict(result) return plist
python
def execute(self, conn, file_id_list, transaction=False): """ file_id_list : file_id_list """ sql=self.sql binds={} if file_id_list: count=0 for an_id in file_id_list: if count > 0: sql += ", " sql += ":file_id_%s" %count binds.update({"file_id_%s" %count : an_id}) count+=1 sql += ")" else: dbsExceptionHandler('dbsException-invalid-input', "Oracle/FileParentBlock/List. this_file_id not provided", self.logger.exception) result = self.dbi.processData(sql, binds, conn, transaction) plist = self.formatDict(result) return plist
[ "def", "execute", "(", "self", ",", "conn", ",", "file_id_list", ",", "transaction", "=", "False", ")", ":", "sql", "=", "self", ".", "sql", "binds", "=", "{", "}", "if", "file_id_list", ":", "count", "=", "0", "for", "an_id", "in", "file_id_list", ":", "if", "count", ">", "0", ":", "sql", "+=", "\", \"", "sql", "+=", "\":file_id_%s\"", "%", "count", "binds", ".", "update", "(", "{", "\"file_id_%s\"", "%", "count", ":", "an_id", "}", ")", "count", "+=", "1", "sql", "+=", "\")\"", "else", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Oracle/FileParentBlock/List. this_file_id not provided\"", ",", "self", ".", "logger", ".", "exception", ")", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ")", "plist", "=", "self", ".", "formatDict", "(", "result", ")", "return", "plist" ]
file_id_list : file_id_list
[ "file_id_list", ":", "file_id_list" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/FileParentBlock/List.py#L25-L44
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/OutputModuleConfig/GetIDForBlockInsert.py
GetIDForBlockInsert.execute
def execute(self, conn, app, release_version, pset_hash, output_label, global_tag, transaction = False): """ returns id for a given application This always requires all four variables to be set, because you better have them in blockInsert """ binds = {} binds["app_name"]=app binds["release_version"]=release_version binds["pset_hash"]=pset_hash binds["output_module_label"]=output_label binds["global_tag"]=global_tag result = self.dbi.processData(self.sql, binds, conn, transaction) plist = self.formatDict(result) if len(plist) < 1: return -1 return plist[0]["output_mod_config_id"]
python
def execute(self, conn, app, release_version, pset_hash, output_label, global_tag, transaction = False): """ returns id for a given application This always requires all four variables to be set, because you better have them in blockInsert """ binds = {} binds["app_name"]=app binds["release_version"]=release_version binds["pset_hash"]=pset_hash binds["output_module_label"]=output_label binds["global_tag"]=global_tag result = self.dbi.processData(self.sql, binds, conn, transaction) plist = self.formatDict(result) if len(plist) < 1: return -1 return plist[0]["output_mod_config_id"]
[ "def", "execute", "(", "self", ",", "conn", ",", "app", ",", "release_version", ",", "pset_hash", ",", "output_label", ",", "global_tag", ",", "transaction", "=", "False", ")", ":", "binds", "=", "{", "}", "binds", "[", "\"app_name\"", "]", "=", "app", "binds", "[", "\"release_version\"", "]", "=", "release_version", "binds", "[", "\"pset_hash\"", "]", "=", "pset_hash", "binds", "[", "\"output_module_label\"", "]", "=", "output_label", "binds", "[", "\"global_tag\"", "]", "=", "global_tag", "result", "=", "self", ".", "dbi", ".", "processData", "(", "self", ".", "sql", ",", "binds", ",", "conn", ",", "transaction", ")", "plist", "=", "self", ".", "formatDict", "(", "result", ")", "if", "len", "(", "plist", ")", "<", "1", ":", "return", "-", "1", "return", "plist", "[", "0", "]", "[", "\"output_mod_config_id\"", "]" ]
returns id for a given application This always requires all four variables to be set, because you better have them in blockInsert
[ "returns", "id", "for", "a", "given", "application" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/OutputModuleConfig/GetIDForBlockInsert.py#L32-L51
train
dmwm/DBS
Server/Python/src/dbs/business/DBSBlock.py
DBSBlock.dumpBlock
def dumpBlock(self, block_name): """ This method is used at source server and gets the information on a single block that is being migrated. Try to return in a format to be ready for insert calls""" if '%' in block_name or '*' in block_name: msg = "No wildcard is allowed in block_name for dumpBlock API" dbsExceptionHandler('dbsException-invalid-input', msg, self.logger.exception) conn = self.dbi.connection() try : #block name is unique block1 = self.blocklist.execute(conn, block_name=block_name) block = [] for b1 in block1: if not b1: return {} else: block = b1 #a block only has one dataset and one primary dataset #in order to reduce the number of dao objects, we will not write #a special migration one. However, we will have to remove the #extras #block1 is a generator. When it is empty, it will skip the for loop above. why? #we cannot test on b1 to decide if the generator is empty or not. #so have to do below: if not block: return {} dataset1 = self.datasetlist.execute(conn, dataset=block["dataset"], dataset_access_type="") dataset = [] for d in dataset1: if d: dataset = d dconfig_list = self.outputCoflist.execute(conn, dataset=dataset['dataset']) else: return {} #get block parentage bparent = self.blockparentlist.execute(conn, block['block_name']) #get dataset parentage dsparent = self.dsparentlist.execute(conn, dataset['dataset']) for p in dsparent: del p['parent_dataset_id'] if 'dataset'in p: del p['dataset'] elif 'this_dataset' in p: del p['this_dataset'] else: pass fparent_list = self.fplist.execute(conn, block_id=block['block_id']) fparent_list2 = [] for fp in fparent_list: fparent_list2.append(fp) #print "---YG file Parent List--" #print fparent_list2 fconfig_list = self.outputCoflist.execute(conn, block_id=block['block_id']) acqEra = {} prsEra = {} if dataset["acquisition_era_name"] not in ( "", None): acqEra = self.aelist.execute(conn, acquisitionEra=dataset["acquisition_era_name"])[0] if dataset["processing_version"] not in ("", None): prsEra = self.pelist.execute(conn, processingV=dataset["processing_version"])[0] primds = self.primdslist.execute(conn, primary_ds_name=dataset["primary_ds_name"])[0] del dataset["primary_ds_name"], dataset['primary_ds_type'] files = self.filelist.execute(conn, block_name=block_name) for f in files: #There are a trade off between json sorting and db query. #We keep lumi sec in a file, but the file parentage seperate #from file file_lumi_list = [] for item in self.fllist.execute(conn, logical_file_name=f['logical_file_name'], migration=True): file_lumi_list.append(item) #print "---YG file lumi list---" f.update(file_lumi_list = file_lumi_list) del file_lumi_list #YG 09/2015 del f['branch_hash_id'] del dataset["acquisition_era_name"], dataset["processing_version"] del block["dataset"] result = dict(block=block, dataset=dataset, primds=primds, files=files, block_parent_list=bparent, ds_parent_list=dsparent, file_conf_list=fconfig_list, file_parent_list=fparent_list2, dataset_conf_list=dconfig_list) if acqEra: result["acquisition_era"] = acqEra if prsEra: result["processing_era"] = prsEra return result finally: if conn: conn.close()
python
def dumpBlock(self, block_name): """ This method is used at source server and gets the information on a single block that is being migrated. Try to return in a format to be ready for insert calls""" if '%' in block_name or '*' in block_name: msg = "No wildcard is allowed in block_name for dumpBlock API" dbsExceptionHandler('dbsException-invalid-input', msg, self.logger.exception) conn = self.dbi.connection() try : #block name is unique block1 = self.blocklist.execute(conn, block_name=block_name) block = [] for b1 in block1: if not b1: return {} else: block = b1 #a block only has one dataset and one primary dataset #in order to reduce the number of dao objects, we will not write #a special migration one. However, we will have to remove the #extras #block1 is a generator. When it is empty, it will skip the for loop above. why? #we cannot test on b1 to decide if the generator is empty or not. #so have to do below: if not block: return {} dataset1 = self.datasetlist.execute(conn, dataset=block["dataset"], dataset_access_type="") dataset = [] for d in dataset1: if d: dataset = d dconfig_list = self.outputCoflist.execute(conn, dataset=dataset['dataset']) else: return {} #get block parentage bparent = self.blockparentlist.execute(conn, block['block_name']) #get dataset parentage dsparent = self.dsparentlist.execute(conn, dataset['dataset']) for p in dsparent: del p['parent_dataset_id'] if 'dataset'in p: del p['dataset'] elif 'this_dataset' in p: del p['this_dataset'] else: pass fparent_list = self.fplist.execute(conn, block_id=block['block_id']) fparent_list2 = [] for fp in fparent_list: fparent_list2.append(fp) #print "---YG file Parent List--" #print fparent_list2 fconfig_list = self.outputCoflist.execute(conn, block_id=block['block_id']) acqEra = {} prsEra = {} if dataset["acquisition_era_name"] not in ( "", None): acqEra = self.aelist.execute(conn, acquisitionEra=dataset["acquisition_era_name"])[0] if dataset["processing_version"] not in ("", None): prsEra = self.pelist.execute(conn, processingV=dataset["processing_version"])[0] primds = self.primdslist.execute(conn, primary_ds_name=dataset["primary_ds_name"])[0] del dataset["primary_ds_name"], dataset['primary_ds_type'] files = self.filelist.execute(conn, block_name=block_name) for f in files: #There are a trade off between json sorting and db query. #We keep lumi sec in a file, but the file parentage seperate #from file file_lumi_list = [] for item in self.fllist.execute(conn, logical_file_name=f['logical_file_name'], migration=True): file_lumi_list.append(item) #print "---YG file lumi list---" f.update(file_lumi_list = file_lumi_list) del file_lumi_list #YG 09/2015 del f['branch_hash_id'] del dataset["acquisition_era_name"], dataset["processing_version"] del block["dataset"] result = dict(block=block, dataset=dataset, primds=primds, files=files, block_parent_list=bparent, ds_parent_list=dsparent, file_conf_list=fconfig_list, file_parent_list=fparent_list2, dataset_conf_list=dconfig_list) if acqEra: result["acquisition_era"] = acqEra if prsEra: result["processing_era"] = prsEra return result finally: if conn: conn.close()
[ "def", "dumpBlock", "(", "self", ",", "block_name", ")", ":", "if", "'%'", "in", "block_name", "or", "'*'", "in", "block_name", ":", "msg", "=", "\"No wildcard is allowed in block_name for dumpBlock API\"", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "msg", ",", "self", ".", "logger", ".", "exception", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "#block name is unique", "block1", "=", "self", ".", "blocklist", ".", "execute", "(", "conn", ",", "block_name", "=", "block_name", ")", "block", "=", "[", "]", "for", "b1", "in", "block1", ":", "if", "not", "b1", ":", "return", "{", "}", "else", ":", "block", "=", "b1", "#a block only has one dataset and one primary dataset", "#in order to reduce the number of dao objects, we will not write", "#a special migration one. However, we will have to remove the", "#extras", "#block1 is a generator. When it is empty, it will skip the for loop above. why? ", "#we cannot test on b1 to decide if the generator is empty or not.", "#so have to do below:", "if", "not", "block", ":", "return", "{", "}", "dataset1", "=", "self", ".", "datasetlist", ".", "execute", "(", "conn", ",", "dataset", "=", "block", "[", "\"dataset\"", "]", ",", "dataset_access_type", "=", "\"\"", ")", "dataset", "=", "[", "]", "for", "d", "in", "dataset1", ":", "if", "d", ":", "dataset", "=", "d", "dconfig_list", "=", "self", ".", "outputCoflist", ".", "execute", "(", "conn", ",", "dataset", "=", "dataset", "[", "'dataset'", "]", ")", "else", ":", "return", "{", "}", "#get block parentage", "bparent", "=", "self", ".", "blockparentlist", ".", "execute", "(", "conn", ",", "block", "[", "'block_name'", "]", ")", "#get dataset parentage", "dsparent", "=", "self", ".", "dsparentlist", ".", "execute", "(", "conn", ",", "dataset", "[", "'dataset'", "]", ")", "for", "p", "in", "dsparent", ":", "del", "p", "[", "'parent_dataset_id'", "]", "if", "'dataset'", "in", "p", ":", "del", "p", "[", "'dataset'", "]", "elif", "'this_dataset'", "in", "p", ":", "del", "p", "[", "'this_dataset'", "]", "else", ":", "pass", "fparent_list", "=", "self", ".", "fplist", ".", "execute", "(", "conn", ",", "block_id", "=", "block", "[", "'block_id'", "]", ")", "fparent_list2", "=", "[", "]", "for", "fp", "in", "fparent_list", ":", "fparent_list2", ".", "append", "(", "fp", ")", "#print \"---YG file Parent List--\"", "#print fparent_list2\t", "fconfig_list", "=", "self", ".", "outputCoflist", ".", "execute", "(", "conn", ",", "block_id", "=", "block", "[", "'block_id'", "]", ")", "acqEra", "=", "{", "}", "prsEra", "=", "{", "}", "if", "dataset", "[", "\"acquisition_era_name\"", "]", "not", "in", "(", "\"\"", ",", "None", ")", ":", "acqEra", "=", "self", ".", "aelist", ".", "execute", "(", "conn", ",", "acquisitionEra", "=", "dataset", "[", "\"acquisition_era_name\"", "]", ")", "[", "0", "]", "if", "dataset", "[", "\"processing_version\"", "]", "not", "in", "(", "\"\"", ",", "None", ")", ":", "prsEra", "=", "self", ".", "pelist", ".", "execute", "(", "conn", ",", "processingV", "=", "dataset", "[", "\"processing_version\"", "]", ")", "[", "0", "]", "primds", "=", "self", ".", "primdslist", ".", "execute", "(", "conn", ",", "primary_ds_name", "=", "dataset", "[", "\"primary_ds_name\"", "]", ")", "[", "0", "]", "del", "dataset", "[", "\"primary_ds_name\"", "]", ",", "dataset", "[", "'primary_ds_type'", "]", "files", "=", "self", ".", "filelist", ".", "execute", "(", "conn", ",", "block_name", "=", "block_name", ")", "for", "f", "in", "files", ":", "#There are a trade off between json sorting and db query.", "#We keep lumi sec in a file, but the file parentage seperate", "#from file", "file_lumi_list", "=", "[", "]", "for", "item", "in", "self", ".", "fllist", ".", "execute", "(", "conn", ",", "logical_file_name", "=", "f", "[", "'logical_file_name'", "]", ",", "migration", "=", "True", ")", ":", "file_lumi_list", ".", "append", "(", "item", ")", "#print \"---YG file lumi list---\"\t", "f", ".", "update", "(", "file_lumi_list", "=", "file_lumi_list", ")", "del", "file_lumi_list", "#YG 09/2015", "del", "f", "[", "'branch_hash_id'", "]", "del", "dataset", "[", "\"acquisition_era_name\"", "]", ",", "dataset", "[", "\"processing_version\"", "]", "del", "block", "[", "\"dataset\"", "]", "result", "=", "dict", "(", "block", "=", "block", ",", "dataset", "=", "dataset", ",", "primds", "=", "primds", ",", "files", "=", "files", ",", "block_parent_list", "=", "bparent", ",", "ds_parent_list", "=", "dsparent", ",", "file_conf_list", "=", "fconfig_list", ",", "file_parent_list", "=", "fparent_list2", ",", "dataset_conf_list", "=", "dconfig_list", ")", "if", "acqEra", ":", "result", "[", "\"acquisition_era\"", "]", "=", "acqEra", "if", "prsEra", ":", "result", "[", "\"processing_era\"", "]", "=", "prsEra", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
This method is used at source server and gets the information on a single block that is being migrated. Try to return in a format to be ready for insert calls
[ "This", "method", "is", "used", "at", "source", "server", "and", "gets", "the", "information", "on", "a", "single", "block", "that", "is", "being", "migrated", ".", "Try", "to", "return", "in", "a", "format", "to", "be", "ready", "for", "insert", "calls" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSBlock.py#L46-L138
train
dmwm/DBS
Server/Python/src/dbs/business/DBSBlock.py
DBSBlock.updateStatus
def updateStatus(self, block_name="", open_for_writing=0): """ Used to toggle the status of a block open_for_writing=1, open for writing, open_for_writing=0, closed """ if open_for_writing not in [1, 0, '1', '0']: msg = "DBSBlock/updateStatus. open_for_writing can only be 0 or 1 : passed %s."\ % open_for_writing dbsExceptionHandler('dbsException-invalid-input', msg) conn = self.dbi.connection() trans = conn.begin() try : open_for_writing = int(open_for_writing) self.updatestatus.execute(conn, block_name, open_for_writing, dbsUtils().getTime(), trans) trans.commit() trans = None except Exception as ex: if trans: trans.rollback() if conn:conn.close() raise ex finally: if conn:conn.close()
python
def updateStatus(self, block_name="", open_for_writing=0): """ Used to toggle the status of a block open_for_writing=1, open for writing, open_for_writing=0, closed """ if open_for_writing not in [1, 0, '1', '0']: msg = "DBSBlock/updateStatus. open_for_writing can only be 0 or 1 : passed %s."\ % open_for_writing dbsExceptionHandler('dbsException-invalid-input', msg) conn = self.dbi.connection() trans = conn.begin() try : open_for_writing = int(open_for_writing) self.updatestatus.execute(conn, block_name, open_for_writing, dbsUtils().getTime(), trans) trans.commit() trans = None except Exception as ex: if trans: trans.rollback() if conn:conn.close() raise ex finally: if conn:conn.close()
[ "def", "updateStatus", "(", "self", ",", "block_name", "=", "\"\"", ",", "open_for_writing", "=", "0", ")", ":", "if", "open_for_writing", "not", "in", "[", "1", ",", "0", ",", "'1'", ",", "'0'", "]", ":", "msg", "=", "\"DBSBlock/updateStatus. open_for_writing can only be 0 or 1 : passed %s.\"", "%", "open_for_writing", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "msg", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "trans", "=", "conn", ".", "begin", "(", ")", "try", ":", "open_for_writing", "=", "int", "(", "open_for_writing", ")", "self", ".", "updatestatus", ".", "execute", "(", "conn", ",", "block_name", ",", "open_for_writing", ",", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ",", "trans", ")", "trans", ".", "commit", "(", ")", "trans", "=", "None", "except", "Exception", "as", "ex", ":", "if", "trans", ":", "trans", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")", "raise", "ex", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Used to toggle the status of a block open_for_writing=1, open for writing, open_for_writing=0, closed
[ "Used", "to", "toggle", "the", "status", "of", "a", "block", "open_for_writing", "=", "1", "open", "for", "writing", "open_for_writing", "=", "0", "closed" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSBlock.py#L140-L161
train
dmwm/DBS
Server/Python/src/dbs/business/DBSBlock.py
DBSBlock.updateSiteName
def updateSiteName(self, block_name, origin_site_name): """ Update the origin_site_name for a given block name """ if not origin_site_name: dbsExceptionHandler('dbsException-invalid-input', "DBSBlock/updateSiteName. origin_site_name is mandatory.") conn = self.dbi.connection() trans = conn.begin() try: self.updatesitename.execute(conn, block_name, origin_site_name) except: if trans: trans.rollback() raise else: if trans: trans.commit() finally: if conn: conn.close()
python
def updateSiteName(self, block_name, origin_site_name): """ Update the origin_site_name for a given block name """ if not origin_site_name: dbsExceptionHandler('dbsException-invalid-input', "DBSBlock/updateSiteName. origin_site_name is mandatory.") conn = self.dbi.connection() trans = conn.begin() try: self.updatesitename.execute(conn, block_name, origin_site_name) except: if trans: trans.rollback() raise else: if trans: trans.commit() finally: if conn: conn.close()
[ "def", "updateSiteName", "(", "self", ",", "block_name", ",", "origin_site_name", ")", ":", "if", "not", "origin_site_name", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"DBSBlock/updateSiteName. origin_site_name is mandatory.\"", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "trans", "=", "conn", ".", "begin", "(", ")", "try", ":", "self", ".", "updatesitename", ".", "execute", "(", "conn", ",", "block_name", ",", "origin_site_name", ")", "except", ":", "if", "trans", ":", "trans", ".", "rollback", "(", ")", "raise", "else", ":", "if", "trans", ":", "trans", ".", "commit", "(", ")", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Update the origin_site_name for a given block name
[ "Update", "the", "origin_site_name", "for", "a", "given", "block", "name" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSBlock.py#L163-L183
train
dmwm/DBS
Server/Python/src/dbs/business/DBSBlock.py
DBSBlock.listBlockParents
def listBlockParents(self, block_name=""): """ list parents of a block """ if not block_name: msg = " DBSBlock/listBlockParents. Block_name must be provided as a string or a list. \ No wildcards allowed in block_name/s." dbsExceptionHandler('dbsException-invalid-input', msg) elif isinstance(block_name, basestring): try: block_name = str(block_name) if '%' in block_name or '*' in block_name: dbsExceptionHandler("dbsException-invalid-input", "DBSReaderModel/listBlocksParents: \ NO WILDCARDS allowed in block_name.") except: dbsExceptionHandler("dbsException-invalid-input", "DBSBlock/listBlockParents. Block_name must be \ provided as a string or a list. No wildcards allowed in block_name/s .") elif type(block_name) is list: for b in block_name: if '%' in b or '*' in b: dbsExceptionHandler("dbsException-invalid-input", "DBSReaderModel/listBlocksParents: \ NO WILDCARDS allowed in block_name.") else: msg = "DBSBlock/listBlockParents. Block_name must be provided as a string or a list. \ No wildcards allowed in block_name/s ." dbsExceptionHandler("dbsException-invalid-input", msg) conn = self.dbi.connection() try: results = self.blockparentlist.execute(conn, block_name) return results finally: if conn: conn.close()
python
def listBlockParents(self, block_name=""): """ list parents of a block """ if not block_name: msg = " DBSBlock/listBlockParents. Block_name must be provided as a string or a list. \ No wildcards allowed in block_name/s." dbsExceptionHandler('dbsException-invalid-input', msg) elif isinstance(block_name, basestring): try: block_name = str(block_name) if '%' in block_name or '*' in block_name: dbsExceptionHandler("dbsException-invalid-input", "DBSReaderModel/listBlocksParents: \ NO WILDCARDS allowed in block_name.") except: dbsExceptionHandler("dbsException-invalid-input", "DBSBlock/listBlockParents. Block_name must be \ provided as a string or a list. No wildcards allowed in block_name/s .") elif type(block_name) is list: for b in block_name: if '%' in b or '*' in b: dbsExceptionHandler("dbsException-invalid-input", "DBSReaderModel/listBlocksParents: \ NO WILDCARDS allowed in block_name.") else: msg = "DBSBlock/listBlockParents. Block_name must be provided as a string or a list. \ No wildcards allowed in block_name/s ." dbsExceptionHandler("dbsException-invalid-input", msg) conn = self.dbi.connection() try: results = self.blockparentlist.execute(conn, block_name) return results finally: if conn: conn.close()
[ "def", "listBlockParents", "(", "self", ",", "block_name", "=", "\"\"", ")", ":", "if", "not", "block_name", ":", "msg", "=", "\" DBSBlock/listBlockParents. Block_name must be provided as a string or a list. \\\n No wildcards allowed in block_name/s.\"", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "msg", ")", "elif", "isinstance", "(", "block_name", ",", "basestring", ")", ":", "try", ":", "block_name", "=", "str", "(", "block_name", ")", "if", "'%'", "in", "block_name", "or", "'*'", "in", "block_name", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSReaderModel/listBlocksParents: \\\n NO WILDCARDS allowed in block_name.\"", ")", "except", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSBlock/listBlockParents. Block_name must be \\\n provided as a string or a list. No wildcards allowed in block_name/s .\"", ")", "elif", "type", "(", "block_name", ")", "is", "list", ":", "for", "b", "in", "block_name", ":", "if", "'%'", "in", "b", "or", "'*'", "in", "b", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSReaderModel/listBlocksParents: \\\n NO WILDCARDS allowed in block_name.\"", ")", "else", ":", "msg", "=", "\"DBSBlock/listBlockParents. Block_name must be provided as a string or a list. \\\n No wildcards allowed in block_name/s .\"", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "msg", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "results", "=", "self", ".", "blockparentlist", ".", "execute", "(", "conn", ",", "block_name", ")", "return", "results", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
list parents of a block
[ "list", "parents", "of", "a", "block" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSBlock.py#L185-L217
train
dmwm/DBS
Server/Python/src/dbs/business/DBSBlock.py
DBSBlock.listBlockChildren
def listBlockChildren(self, block_name=""): """ list parents of a block """ if (not block_name) or re.search("['%','*']", block_name): dbsExceptionHandler("dbsException-invalid-input", "DBSBlock/listBlockChildren. Block_name must be provided." ) conn = self.dbi.connection() try: results = self.blockchildlist.execute(conn, block_name) return results finally: if conn: conn.close()
python
def listBlockChildren(self, block_name=""): """ list parents of a block """ if (not block_name) or re.search("['%','*']", block_name): dbsExceptionHandler("dbsException-invalid-input", "DBSBlock/listBlockChildren. Block_name must be provided." ) conn = self.dbi.connection() try: results = self.blockchildlist.execute(conn, block_name) return results finally: if conn: conn.close()
[ "def", "listBlockChildren", "(", "self", ",", "block_name", "=", "\"\"", ")", ":", "if", "(", "not", "block_name", ")", "or", "re", ".", "search", "(", "\"['%','*']\"", ",", "block_name", ")", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSBlock/listBlockChildren. Block_name must be provided.\"", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "results", "=", "self", ".", "blockchildlist", ".", "execute", "(", "conn", ",", "block_name", ")", "return", "results", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
list parents of a block
[ "list", "parents", "of", "a", "block" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSBlock.py#L219-L231
train
dmwm/DBS
Server/Python/src/dbs/business/DBSBlock.py
DBSBlock.listBlocks
def listBlocks(self, dataset="", block_name="", data_tier_name="", origin_site_name="", logical_file_name="", run_num=-1, min_cdate=0, max_cdate=0, min_ldate=0, max_ldate=0, cdate=0, ldate=0, open_for_writing=-1, detail=False): """ dataset, block_name, data_tier_name or logical_file_name must be passed. """ if (not dataset) or re.search("['%','*']", dataset): if (not block_name) or re.search("['%','*']", block_name): if (not logical_file_name) or re.search("['%','*']", logical_file_name): if not data_tier_name or re.search("['%','*']", data_tier_name): msg = "DBSBlock/listBlock. You must specify at least one parameter(dataset, block_name,\ data_tier_name, logical_file_name) with listBlocks api" dbsExceptionHandler('dbsException-invalid-input2', msg, self.logger.exception, msg) if data_tier_name: if not (min_cdate and max_cdate) or (max_cdate-min_cdate)>32*24*3600: msg = "min_cdate and max_cdate are mandatory parameters. If data_tier_name parameter is used \ the maximal time range allowed is 31 days" dbsExceptionHandler('dbsException-invalid-input2', msg, self.logger.exception, msg) if detail: msg = "DBSBlock/listBlock. Detail parameter not allowed togther with data_tier_name" dbsExceptionHandler('dbsException-invalid-input2', msg, self.logger.exception, msg) with self.dbi.connection() as conn: dao = (self.blockbrieflist, self.blocklist)[detail] for item in dao.execute(conn, dataset, block_name, data_tier_name, origin_site_name, logical_file_name, run_num, min_cdate, max_cdate, min_ldate, max_ldate, cdate, ldate): yield item
python
def listBlocks(self, dataset="", block_name="", data_tier_name="", origin_site_name="", logical_file_name="", run_num=-1, min_cdate=0, max_cdate=0, min_ldate=0, max_ldate=0, cdate=0, ldate=0, open_for_writing=-1, detail=False): """ dataset, block_name, data_tier_name or logical_file_name must be passed. """ if (not dataset) or re.search("['%','*']", dataset): if (not block_name) or re.search("['%','*']", block_name): if (not logical_file_name) or re.search("['%','*']", logical_file_name): if not data_tier_name or re.search("['%','*']", data_tier_name): msg = "DBSBlock/listBlock. You must specify at least one parameter(dataset, block_name,\ data_tier_name, logical_file_name) with listBlocks api" dbsExceptionHandler('dbsException-invalid-input2', msg, self.logger.exception, msg) if data_tier_name: if not (min_cdate and max_cdate) or (max_cdate-min_cdate)>32*24*3600: msg = "min_cdate and max_cdate are mandatory parameters. If data_tier_name parameter is used \ the maximal time range allowed is 31 days" dbsExceptionHandler('dbsException-invalid-input2', msg, self.logger.exception, msg) if detail: msg = "DBSBlock/listBlock. Detail parameter not allowed togther with data_tier_name" dbsExceptionHandler('dbsException-invalid-input2', msg, self.logger.exception, msg) with self.dbi.connection() as conn: dao = (self.blockbrieflist, self.blocklist)[detail] for item in dao.execute(conn, dataset, block_name, data_tier_name, origin_site_name, logical_file_name, run_num, min_cdate, max_cdate, min_ldate, max_ldate, cdate, ldate): yield item
[ "def", "listBlocks", "(", "self", ",", "dataset", "=", "\"\"", ",", "block_name", "=", "\"\"", ",", "data_tier_name", "=", "\"\"", ",", "origin_site_name", "=", "\"\"", ",", "logical_file_name", "=", "\"\"", ",", "run_num", "=", "-", "1", ",", "min_cdate", "=", "0", ",", "max_cdate", "=", "0", ",", "min_ldate", "=", "0", ",", "max_ldate", "=", "0", ",", "cdate", "=", "0", ",", "ldate", "=", "0", ",", "open_for_writing", "=", "-", "1", ",", "detail", "=", "False", ")", ":", "if", "(", "not", "dataset", ")", "or", "re", ".", "search", "(", "\"['%','*']\"", ",", "dataset", ")", ":", "if", "(", "not", "block_name", ")", "or", "re", ".", "search", "(", "\"['%','*']\"", ",", "block_name", ")", ":", "if", "(", "not", "logical_file_name", ")", "or", "re", ".", "search", "(", "\"['%','*']\"", ",", "logical_file_name", ")", ":", "if", "not", "data_tier_name", "or", "re", ".", "search", "(", "\"['%','*']\"", ",", "data_tier_name", ")", ":", "msg", "=", "\"DBSBlock/listBlock. You must specify at least one parameter(dataset, block_name,\\\n\t\t\t \tdata_tier_name, logical_file_name) with listBlocks api\"", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "msg", ",", "self", ".", "logger", ".", "exception", ",", "msg", ")", "if", "data_tier_name", ":", "if", "not", "(", "min_cdate", "and", "max_cdate", ")", "or", "(", "max_cdate", "-", "min_cdate", ")", ">", "32", "*", "24", "*", "3600", ":", "msg", "=", "\"min_cdate and max_cdate are mandatory parameters. If data_tier_name parameter is used \\\n the maximal time range allowed is 31 days\"", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "msg", ",", "self", ".", "logger", ".", "exception", ",", "msg", ")", "if", "detail", ":", "msg", "=", "\"DBSBlock/listBlock. Detail parameter not allowed togther with data_tier_name\"", "dbsExceptionHandler", "(", "'dbsException-invalid-input2'", ",", "msg", ",", "self", ".", "logger", ".", "exception", ",", "msg", ")", "with", "self", ".", "dbi", ".", "connection", "(", ")", "as", "conn", ":", "dao", "=", "(", "self", ".", "blockbrieflist", ",", "self", ".", "blocklist", ")", "[", "detail", "]", "for", "item", "in", "dao", ".", "execute", "(", "conn", ",", "dataset", ",", "block_name", ",", "data_tier_name", ",", "origin_site_name", ",", "logical_file_name", ",", "run_num", ",", "min_cdate", ",", "max_cdate", ",", "min_ldate", ",", "max_ldate", ",", "cdate", ",", "ldate", ")", ":", "yield", "item" ]
dataset, block_name, data_tier_name or logical_file_name must be passed.
[ "dataset", "block_name", "data_tier_name", "or", "logical_file_name", "must", "be", "passed", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSBlock.py#L233-L260
train
dmwm/DBS
Server/Python/src/dbs/business/DBSBlock.py
DBSBlock.listBlocksOrigin
def listBlocksOrigin(self, origin_site_name="", dataset="", block_name=""): """ This is the API to list all the blocks/datasets first generated in the site called origin_site_name, if origin_site_name is provided w/ no wildcards allow. If a fully spelled dataset is provided, then it will only list the blocks first generated from origin_site_name under the given dataset. """ if not (dataset or block_name): dbsExceptionHandler("dbsException-invalid-input", "DBSBlock/listBlocksOrigin: dataset or block_name must be provided.") if re.search("['%', '*']", dataset) or re.search("['%', '*']", block_name): dbsExceptionHandler("dbsException-invalid-input", "DBSBlock/listBlocksOrigin: dataset or block_name with wildcard is not supported.") try: conn = self.dbi.connection() result = self.bkOriginlist.execute(conn, origin_site_name, dataset, block_name) return result finally: if conn: conn.close()
python
def listBlocksOrigin(self, origin_site_name="", dataset="", block_name=""): """ This is the API to list all the blocks/datasets first generated in the site called origin_site_name, if origin_site_name is provided w/ no wildcards allow. If a fully spelled dataset is provided, then it will only list the blocks first generated from origin_site_name under the given dataset. """ if not (dataset or block_name): dbsExceptionHandler("dbsException-invalid-input", "DBSBlock/listBlocksOrigin: dataset or block_name must be provided.") if re.search("['%', '*']", dataset) or re.search("['%', '*']", block_name): dbsExceptionHandler("dbsException-invalid-input", "DBSBlock/listBlocksOrigin: dataset or block_name with wildcard is not supported.") try: conn = self.dbi.connection() result = self.bkOriginlist.execute(conn, origin_site_name, dataset, block_name) return result finally: if conn: conn.close()
[ "def", "listBlocksOrigin", "(", "self", ",", "origin_site_name", "=", "\"\"", ",", "dataset", "=", "\"\"", ",", "block_name", "=", "\"\"", ")", ":", "if", "not", "(", "dataset", "or", "block_name", ")", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSBlock/listBlocksOrigin: dataset or block_name must be provided.\"", ")", "if", "re", ".", "search", "(", "\"['%', '*']\"", ",", "dataset", ")", "or", "re", ".", "search", "(", "\"['%', '*']\"", ",", "block_name", ")", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\"DBSBlock/listBlocksOrigin: dataset or block_name with wildcard is not supported.\"", ")", "try", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "result", "=", "self", ".", "bkOriginlist", ".", "execute", "(", "conn", ",", "origin_site_name", ",", "dataset", ",", "block_name", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
This is the API to list all the blocks/datasets first generated in the site called origin_site_name, if origin_site_name is provided w/ no wildcards allow. If a fully spelled dataset is provided, then it will only list the blocks first generated from origin_site_name under the given dataset.
[ "This", "is", "the", "API", "to", "list", "all", "the", "blocks", "/", "datasets", "first", "generated", "in", "the", "site", "called", "origin_site_name", "if", "origin_site_name", "is", "provided", "w", "/", "no", "wildcards", "allow", ".", "If", "a", "fully", "spelled", "dataset", "is", "provided", "then", "it", "will", "only", "list", "the", "blocks", "first", "generated", "from", "origin_site_name", "under", "the", "given", "dataset", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSBlock.py#L262-L280
train
dmwm/DBS
Server/Python/src/dbs/business/DBSBlock.py
DBSBlock.insertBlock
def insertBlock(self, businput): """ Input dictionary has to have the following keys: blockname It may have: open_for_writing, origin_site(name), block_size, file_count, creation_date, create_by, last_modification_date, last_modified_by it builds the correct dictionary for dao input and executes the dao NEED to validate there are no extra keys in the businput """ if not ("block_name" in businput and "origin_site_name" in businput ): dbsExceptionHandler('dbsException-invalid-input', "business/DBSBlock/insertBlock must have block_name and origin_site_name as input") conn = self.dbi.connection() tran = conn.begin() try: blkinput = { "last_modification_date":businput.get("last_modification_date", dbsUtils().getTime()), #"last_modified_by":businput.get("last_modified_by", dbsUtils().getCreateBy()), "last_modified_by":dbsUtils().getCreateBy(), #"create_by":businput.get("create_by", dbsUtils().getCreateBy()), "create_by":dbsUtils().getCreateBy(), "creation_date":businput.get("creation_date", dbsUtils().getTime()), "open_for_writing":businput.get("open_for_writing", 1), "block_size":businput.get("block_size", 0), "file_count":businput.get("file_count", 0), "block_name":businput.get("block_name"), "origin_site_name":businput.get("origin_site_name") } ds_name = businput["block_name"].split('#')[0] blkinput["dataset_id"] = self.datasetid.execute(conn, ds_name, tran) if blkinput["dataset_id"] == -1 : msg = "DBSBlock/insertBlock. Dataset %s does not exists" % ds_name dbsExceptionHandler('dbsException-missing-data', msg) blkinput["block_id"] = self.sm.increment(conn, "SEQ_BK", tran) self.blockin.execute(conn, blkinput, tran) tran.commit() tran = None except Exception as e: if str(e).lower().find("unique constraint") != -1 or str(e).lower().find("duplicate") != -1: pass else: if tran: tran.rollback() if conn: conn.close() raise finally: if tran: tran.rollback() if conn: conn.close()
python
def insertBlock(self, businput): """ Input dictionary has to have the following keys: blockname It may have: open_for_writing, origin_site(name), block_size, file_count, creation_date, create_by, last_modification_date, last_modified_by it builds the correct dictionary for dao input and executes the dao NEED to validate there are no extra keys in the businput """ if not ("block_name" in businput and "origin_site_name" in businput ): dbsExceptionHandler('dbsException-invalid-input', "business/DBSBlock/insertBlock must have block_name and origin_site_name as input") conn = self.dbi.connection() tran = conn.begin() try: blkinput = { "last_modification_date":businput.get("last_modification_date", dbsUtils().getTime()), #"last_modified_by":businput.get("last_modified_by", dbsUtils().getCreateBy()), "last_modified_by":dbsUtils().getCreateBy(), #"create_by":businput.get("create_by", dbsUtils().getCreateBy()), "create_by":dbsUtils().getCreateBy(), "creation_date":businput.get("creation_date", dbsUtils().getTime()), "open_for_writing":businput.get("open_for_writing", 1), "block_size":businput.get("block_size", 0), "file_count":businput.get("file_count", 0), "block_name":businput.get("block_name"), "origin_site_name":businput.get("origin_site_name") } ds_name = businput["block_name"].split('#')[0] blkinput["dataset_id"] = self.datasetid.execute(conn, ds_name, tran) if blkinput["dataset_id"] == -1 : msg = "DBSBlock/insertBlock. Dataset %s does not exists" % ds_name dbsExceptionHandler('dbsException-missing-data', msg) blkinput["block_id"] = self.sm.increment(conn, "SEQ_BK", tran) self.blockin.execute(conn, blkinput, tran) tran.commit() tran = None except Exception as e: if str(e).lower().find("unique constraint") != -1 or str(e).lower().find("duplicate") != -1: pass else: if tran: tran.rollback() if conn: conn.close() raise finally: if tran: tran.rollback() if conn: conn.close()
[ "def", "insertBlock", "(", "self", ",", "businput", ")", ":", "if", "not", "(", "\"block_name\"", "in", "businput", "and", "\"origin_site_name\"", "in", "businput", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"business/DBSBlock/insertBlock must have block_name and origin_site_name as input\"", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "conn", ".", "begin", "(", ")", "try", ":", "blkinput", "=", "{", "\"last_modification_date\"", ":", "businput", ".", "get", "(", "\"last_modification_date\"", ",", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ")", ",", "#\"last_modified_by\":businput.get(\"last_modified_by\", dbsUtils().getCreateBy()),", "\"last_modified_by\"", ":", "dbsUtils", "(", ")", ".", "getCreateBy", "(", ")", ",", "#\"create_by\":businput.get(\"create_by\", dbsUtils().getCreateBy()),", "\"create_by\"", ":", "dbsUtils", "(", ")", ".", "getCreateBy", "(", ")", ",", "\"creation_date\"", ":", "businput", ".", "get", "(", "\"creation_date\"", ",", "dbsUtils", "(", ")", ".", "getTime", "(", ")", ")", ",", "\"open_for_writing\"", ":", "businput", ".", "get", "(", "\"open_for_writing\"", ",", "1", ")", ",", "\"block_size\"", ":", "businput", ".", "get", "(", "\"block_size\"", ",", "0", ")", ",", "\"file_count\"", ":", "businput", ".", "get", "(", "\"file_count\"", ",", "0", ")", ",", "\"block_name\"", ":", "businput", ".", "get", "(", "\"block_name\"", ")", ",", "\"origin_site_name\"", ":", "businput", ".", "get", "(", "\"origin_site_name\"", ")", "}", "ds_name", "=", "businput", "[", "\"block_name\"", "]", ".", "split", "(", "'#'", ")", "[", "0", "]", "blkinput", "[", "\"dataset_id\"", "]", "=", "self", ".", "datasetid", ".", "execute", "(", "conn", ",", "ds_name", ",", "tran", ")", "if", "blkinput", "[", "\"dataset_id\"", "]", "==", "-", "1", ":", "msg", "=", "\"DBSBlock/insertBlock. Dataset %s does not exists\"", "%", "ds_name", "dbsExceptionHandler", "(", "'dbsException-missing-data'", ",", "msg", ")", "blkinput", "[", "\"block_id\"", "]", "=", "self", ".", "sm", ".", "increment", "(", "conn", ",", "\"SEQ_BK\"", ",", "tran", ")", "self", ".", "blockin", ".", "execute", "(", "conn", ",", "blkinput", ",", "tran", ")", "tran", ".", "commit", "(", ")", "tran", "=", "None", "except", "Exception", "as", "e", ":", "if", "str", "(", "e", ")", ".", "lower", "(", ")", ".", "find", "(", "\"unique constraint\"", ")", "!=", "-", "1", "or", "str", "(", "e", ")", ".", "lower", "(", ")", ".", "find", "(", "\"duplicate\"", ")", "!=", "-", "1", ":", "pass", "else", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")", "raise", "finally", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Input dictionary has to have the following keys: blockname It may have: open_for_writing, origin_site(name), block_size, file_count, creation_date, create_by, last_modification_date, last_modified_by it builds the correct dictionary for dao input and executes the dao NEED to validate there are no extra keys in the businput
[ "Input", "dictionary", "has", "to", "have", "the", "following", "keys", ":", "blockname", "It", "may", "have", ":", "open_for_writing", "origin_site", "(", "name", ")", "block_size", "file_count", "creation_date", "create_by", "last_modification_date", "last_modified_by", "it", "builds", "the", "correct", "dictionary", "for", "dao", "input", "and", "executes", "the", "dao" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSBlock.py#L283-L337
train
dmwm/DBS
Server/Python/src/dbs/business/DBSDataTier.py
DBSDataTier.listDataTiers
def listDataTiers(self, data_tier_name=""): """ List data tier(s) """ if not isinstance(data_tier_name, basestring) : dbsExceptionHandler('dbsException-invalid-input', 'data_tier_name given is not valid : %s' % data_tier_name) else: try: data_tier_name = str(data_tier_name) except: dbsExceptionHandler('dbsException-invalid-input', 'data_tier_name given is not valid : %s' % data_tier_name) conn = self.dbi.connection() try: result = self.dataTier.execute(conn, data_tier_name.upper()) return result finally: if conn: conn.close()
python
def listDataTiers(self, data_tier_name=""): """ List data tier(s) """ if not isinstance(data_tier_name, basestring) : dbsExceptionHandler('dbsException-invalid-input', 'data_tier_name given is not valid : %s' % data_tier_name) else: try: data_tier_name = str(data_tier_name) except: dbsExceptionHandler('dbsException-invalid-input', 'data_tier_name given is not valid : %s' % data_tier_name) conn = self.dbi.connection() try: result = self.dataTier.execute(conn, data_tier_name.upper()) return result finally: if conn: conn.close()
[ "def", "listDataTiers", "(", "self", ",", "data_tier_name", "=", "\"\"", ")", ":", "if", "not", "isinstance", "(", "data_tier_name", ",", "basestring", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "'data_tier_name given is not valid : %s'", "%", "data_tier_name", ")", "else", ":", "try", ":", "data_tier_name", "=", "str", "(", "data_tier_name", ")", "except", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "'data_tier_name given is not valid : %s'", "%", "data_tier_name", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "result", "=", "self", ".", "dataTier", ".", "execute", "(", "conn", ",", "data_tier_name", ".", "upper", "(", ")", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
List data tier(s)
[ "List", "data", "tier", "(", "s", ")" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSDataTier.py#L25-L44
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/Site/List.py
List.execute
def execute(self, conn, site_name= "", transaction = False): """ Lists all sites types if site_name is not provided. """ sql = self.sql if site_name == "": result = self.dbi.processData(sql, conn=conn, transaction=transaction) else: sql += "WHERE S.SITE_NAME = :site_name" binds = { "site_name" : site_name } result = self.dbi.processData(sql, binds, conn, transaction) return self.formatDict(result)
python
def execute(self, conn, site_name= "", transaction = False): """ Lists all sites types if site_name is not provided. """ sql = self.sql if site_name == "": result = self.dbi.processData(sql, conn=conn, transaction=transaction) else: sql += "WHERE S.SITE_NAME = :site_name" binds = { "site_name" : site_name } result = self.dbi.processData(sql, binds, conn, transaction) return self.formatDict(result)
[ "def", "execute", "(", "self", ",", "conn", ",", "site_name", "=", "\"\"", ",", "transaction", "=", "False", ")", ":", "sql", "=", "self", ".", "sql", "if", "site_name", "==", "\"\"", ":", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "conn", "=", "conn", ",", "transaction", "=", "transaction", ")", "else", ":", "sql", "+=", "\"WHERE S.SITE_NAME = :site_name\"", "binds", "=", "{", "\"site_name\"", ":", "site_name", "}", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ")", "return", "self", ".", "formatDict", "(", "result", ")" ]
Lists all sites types if site_name is not provided.
[ "Lists", "all", "sites", "types", "if", "site_name", "is", "not", "provided", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/Site/List.py#L24-L35
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/FileBuffer/DeleteFiles.py
DeleteFiles.execute
def execute(self, conn, logical_file_name={}, transaction=False): """ simple execute """ if not conn: dbsExceptionHandler("dbsException-db-conn-failed", "Oracle/FileBuffer/DeleteFiles. Expects db connection from upper layer.") self.dbi.processData(self.sql, logical_file_name, conn, transaction, returnCursor=True)
python
def execute(self, conn, logical_file_name={}, transaction=False): """ simple execute """ if not conn: dbsExceptionHandler("dbsException-db-conn-failed", "Oracle/FileBuffer/DeleteFiles. Expects db connection from upper layer.") self.dbi.processData(self.sql, logical_file_name, conn, transaction, returnCursor=True)
[ "def", "execute", "(", "self", ",", "conn", ",", "logical_file_name", "=", "{", "}", ",", "transaction", "=", "False", ")", ":", "if", "not", "conn", ":", "dbsExceptionHandler", "(", "\"dbsException-db-conn-failed\"", ",", "\"Oracle/FileBuffer/DeleteFiles. Expects db connection from upper layer.\"", ")", "self", ".", "dbi", ".", "processData", "(", "self", ".", "sql", ",", "logical_file_name", ",", "conn", ",", "transaction", ",", "returnCursor", "=", "True", ")" ]
simple execute
[ "simple", "execute" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/FileBuffer/DeleteFiles.py#L20-L27
train
dmwm/DBS
Server/Python/src/dbs/business/DBSDataType.py
DBSDataType.listDataType
def listDataType(self, dataType="", dataset=""): """ List data-type/primary-ds-type """ conn = self.dbi.connection() try: if dataset and dataType: dbsExceptionHandler('dbsException-invalid-input', "DBSDataType/listDataType. Data Type can be only searched by data_type or by dataset, not both.") else: result = self.dataType.execute(conn, dataType, dataset) return result finally: if conn: conn.close()
python
def listDataType(self, dataType="", dataset=""): """ List data-type/primary-ds-type """ conn = self.dbi.connection() try: if dataset and dataType: dbsExceptionHandler('dbsException-invalid-input', "DBSDataType/listDataType. Data Type can be only searched by data_type or by dataset, not both.") else: result = self.dataType.execute(conn, dataType, dataset) return result finally: if conn: conn.close()
[ "def", "listDataType", "(", "self", ",", "dataType", "=", "\"\"", ",", "dataset", "=", "\"\"", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "if", "dataset", "and", "dataType", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"DBSDataType/listDataType. Data Type can be only searched by data_type or by dataset, not both.\"", ")", "else", ":", "result", "=", "self", ".", "dataType", ".", "execute", "(", "conn", ",", "dataType", ",", "dataset", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
List data-type/primary-ds-type
[ "List", "data", "-", "type", "/", "primary", "-", "ds", "-", "type" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSDataType.py#L22-L36
train
dmwm/DBS
Server/Python/src/dbs/business/DBSFileBuffer.py
DBSFileBuffer.getBlocks
def getBlocks(self): """ Get the blocks that need to be migrated """ try: conn = self.dbi.connection() result = self.buflistblks.execute(conn) return result finally: if conn: conn.close()
python
def getBlocks(self): """ Get the blocks that need to be migrated """ try: conn = self.dbi.connection() result = self.buflistblks.execute(conn) return result finally: if conn: conn.close()
[ "def", "getBlocks", "(", "self", ")", ":", "try", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "result", "=", "self", ".", "buflistblks", ".", "execute", "(", "conn", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Get the blocks that need to be migrated
[ "Get", "the", "blocks", "that", "need", "to", "be", "migrated" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSFileBuffer.py#L35-L45
train
dmwm/DBS
Server/Python/src/dbs/business/DBSFileBuffer.py
DBSFileBuffer.getBufferedFiles
def getBufferedFiles(self, block_id): """ Get some files from the insert buffer """ try: conn = self.dbi.connection() result = self.buflist.execute(conn, block_id) return result finally: if conn: conn.close()
python
def getBufferedFiles(self, block_id): """ Get some files from the insert buffer """ try: conn = self.dbi.connection() result = self.buflist.execute(conn, block_id) return result finally: if conn: conn.close()
[ "def", "getBufferedFiles", "(", "self", ",", "block_id", ")", ":", "try", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "result", "=", "self", ".", "buflist", ".", "execute", "(", "conn", ",", "block_id", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Get some files from the insert buffer
[ "Get", "some", "files", "from", "the", "insert", "buffer" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSFileBuffer.py#L47-L58
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/DataTier/List.py
List.execute
def execute(self, conn, data_tier_name='', transaction = False, cache=None): """ returns id for a given datatier name """ if cache: ret=cache.get("DATA_TIERS") if not ret==None: return ret sql = self.sql binds={} if data_tier_name: op = ('=', 'like')['%' in data_tier_name] sql += "WHERE DT.DATA_TIER_NAME %s :datatier" %op binds = {"datatier":data_tier_name} result = self.dbi.processData(sql, binds, conn, transaction) plist = self.formatDict(result) return plist
python
def execute(self, conn, data_tier_name='', transaction = False, cache=None): """ returns id for a given datatier name """ if cache: ret=cache.get("DATA_TIERS") if not ret==None: return ret sql = self.sql binds={} if data_tier_name: op = ('=', 'like')['%' in data_tier_name] sql += "WHERE DT.DATA_TIER_NAME %s :datatier" %op binds = {"datatier":data_tier_name} result = self.dbi.processData(sql, binds, conn, transaction) plist = self.formatDict(result) return plist
[ "def", "execute", "(", "self", ",", "conn", ",", "data_tier_name", "=", "''", ",", "transaction", "=", "False", ",", "cache", "=", "None", ")", ":", "if", "cache", ":", "ret", "=", "cache", ".", "get", "(", "\"DATA_TIERS\"", ")", "if", "not", "ret", "==", "None", ":", "return", "ret", "sql", "=", "self", ".", "sql", "binds", "=", "{", "}", "if", "data_tier_name", ":", "op", "=", "(", "'='", ",", "'like'", ")", "[", "'%'", "in", "data_tier_name", "]", "sql", "+=", "\"WHERE DT.DATA_TIER_NAME %s :datatier\"", "%", "op", "binds", "=", "{", "\"datatier\"", ":", "data_tier_name", "}", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ")", "plist", "=", "self", ".", "formatDict", "(", "result", ")", "return", "plist" ]
returns id for a given datatier name
[ "returns", "id", "for", "a", "given", "datatier", "name" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/DataTier/List.py#L27-L43
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/MigrationRequests/ListOldest.py
ListOldest.execute
def execute(self, conn, migration_url="", migration_input="", create_by="", migration_request_id="", transaction=False): """ Lists the oldest request queued """ binds = {} result = self.dbi.processData(self.sql, binds, conn, transaction) result = self.formatDict(result) if len(result) == 0 : return [] if result[0]["migration_request_id"] in ('', None) : return [] return result
python
def execute(self, conn, migration_url="", migration_input="", create_by="", migration_request_id="", transaction=False): """ Lists the oldest request queued """ binds = {} result = self.dbi.processData(self.sql, binds, conn, transaction) result = self.formatDict(result) if len(result) == 0 : return [] if result[0]["migration_request_id"] in ('', None) : return [] return result
[ "def", "execute", "(", "self", ",", "conn", ",", "migration_url", "=", "\"\"", ",", "migration_input", "=", "\"\"", ",", "create_by", "=", "\"\"", ",", "migration_request_id", "=", "\"\"", ",", "transaction", "=", "False", ")", ":", "binds", "=", "{", "}", "result", "=", "self", ".", "dbi", ".", "processData", "(", "self", ".", "sql", ",", "binds", ",", "conn", ",", "transaction", ")", "result", "=", "self", ".", "formatDict", "(", "result", ")", "if", "len", "(", "result", ")", "==", "0", ":", "return", "[", "]", "if", "result", "[", "0", "]", "[", "\"migration_request_id\"", "]", "in", "(", "''", ",", "None", ")", ":", "return", "[", "]", "return", "result" ]
Lists the oldest request queued
[ "Lists", "the", "oldest", "request", "queued" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/MigrationRequests/ListOldest.py#L21-L32
train
dmwm/DBS
Server/Python/src/dbs/business/DBSProcessingEra.py
DBSProcessingEra.listProcessingEras
def listProcessingEras(self, processing_version=''): """ Returns all processing eras in dbs """ conn = self.dbi.connection() try: result = self.pelst.execute(conn, processing_version) return result finally: if conn: conn.close()
python
def listProcessingEras(self, processing_version=''): """ Returns all processing eras in dbs """ conn = self.dbi.connection() try: result = self.pelst.execute(conn, processing_version) return result finally: if conn: conn.close()
[ "def", "listProcessingEras", "(", "self", ",", "processing_version", "=", "''", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "result", "=", "self", ".", "pelst", ".", "execute", "(", "conn", ",", "processing_version", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Returns all processing eras in dbs
[ "Returns", "all", "processing", "eras", "in", "dbs" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSProcessingEra.py#L25-L35
train
dmwm/DBS
Server/Python/src/dbs/business/DBSProcessingEra.py
DBSProcessingEra.insertProcessingEra
def insertProcessingEra(self, businput): """ Input dictionary has to have the following keys: processing_version, creation_date, create_by, description it builds the correct dictionary for dao input and executes the dao """ conn = self.dbi.connection() tran = conn.begin() try: businput["processing_era_id"] = self.sm.increment(conn, "SEQ_PE", tran) businput["processing_version"] = businput["processing_version"] self.pein.execute(conn, businput, tran) tran.commit() tran = None except KeyError as ke: dbsExceptionHandler('dbsException-invalid-input', "Invalid input:" + ke.args[0]) except Exception as ex: if (str(ex).lower().find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1): # already exist self.logger.warning("DBSProcessingEra/insertProcessingEras. " + "Unique constraint violation being ignored...") self.logger.warning(ex) else: if tran: tran.rollback() tran = None raise finally: if tran: tran.rollback() if conn: conn.close()
python
def insertProcessingEra(self, businput): """ Input dictionary has to have the following keys: processing_version, creation_date, create_by, description it builds the correct dictionary for dao input and executes the dao """ conn = self.dbi.connection() tran = conn.begin() try: businput["processing_era_id"] = self.sm.increment(conn, "SEQ_PE", tran) businput["processing_version"] = businput["processing_version"] self.pein.execute(conn, businput, tran) tran.commit() tran = None except KeyError as ke: dbsExceptionHandler('dbsException-invalid-input', "Invalid input:" + ke.args[0]) except Exception as ex: if (str(ex).lower().find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1): # already exist self.logger.warning("DBSProcessingEra/insertProcessingEras. " + "Unique constraint violation being ignored...") self.logger.warning(ex) else: if tran: tran.rollback() tran = None raise finally: if tran: tran.rollback() if conn: conn.close()
[ "def", "insertProcessingEra", "(", "self", ",", "businput", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "conn", ".", "begin", "(", ")", "try", ":", "businput", "[", "\"processing_era_id\"", "]", "=", "self", ".", "sm", ".", "increment", "(", "conn", ",", "\"SEQ_PE\"", ",", "tran", ")", "businput", "[", "\"processing_version\"", "]", "=", "businput", "[", "\"processing_version\"", "]", "self", ".", "pein", ".", "execute", "(", "conn", ",", "businput", ",", "tran", ")", "tran", ".", "commit", "(", ")", "tran", "=", "None", "except", "KeyError", "as", "ke", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "\"Invalid input:\"", "+", "ke", ".", "args", "[", "0", "]", ")", "except", "Exception", "as", "ex", ":", "if", "(", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"unique constraint\"", ")", "!=", "-", "1", "or", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"duplicate\"", ")", "!=", "-", "1", ")", ":", "# already exist", "self", ".", "logger", ".", "warning", "(", "\"DBSProcessingEra/insertProcessingEras. \"", "+", "\"Unique constraint violation being ignored...\"", ")", "self", ".", "logger", ".", "warning", "(", "ex", ")", "else", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "tran", "=", "None", "raise", "finally", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Input dictionary has to have the following keys: processing_version, creation_date, create_by, description it builds the correct dictionary for dao input and executes the dao
[ "Input", "dictionary", "has", "to", "have", "the", "following", "keys", ":", "processing_version", "creation_date", "create_by", "description", "it", "builds", "the", "correct", "dictionary", "for", "dao", "input", "and", "executes", "the", "dao" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSProcessingEra.py#L37-L70
train
dmwm/DBS
Server/Python/src/dbs/business/DBSPhysicsGroup.py
DBSPhysicsGroup.listPhysicsGroups
def listPhysicsGroups(self, physics_group_name=""): """ Returns all physics groups if physics group names are not passed. """ if not isinstance(physics_group_name, basestring): dbsExceptionHandler('dbsException-invalid-input', 'physics group name given is not valid : %s' % physics_group_name) else: try: physics_group_name = str(physics_group_name) except: dbsExceptionHandler('dbsException-invalid-input', 'physics group name given is not valid : %s' % physics_group_name) conn = self.dbi.connection() try: result = self.pglist.execute(conn, physics_group_name) return result finally: if conn: conn.close()
python
def listPhysicsGroups(self, physics_group_name=""): """ Returns all physics groups if physics group names are not passed. """ if not isinstance(physics_group_name, basestring): dbsExceptionHandler('dbsException-invalid-input', 'physics group name given is not valid : %s' % physics_group_name) else: try: physics_group_name = str(physics_group_name) except: dbsExceptionHandler('dbsException-invalid-input', 'physics group name given is not valid : %s' % physics_group_name) conn = self.dbi.connection() try: result = self.pglist.execute(conn, physics_group_name) return result finally: if conn: conn.close()
[ "def", "listPhysicsGroups", "(", "self", ",", "physics_group_name", "=", "\"\"", ")", ":", "if", "not", "isinstance", "(", "physics_group_name", ",", "basestring", ")", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "'physics group name given is not valid : %s'", "%", "physics_group_name", ")", "else", ":", "try", ":", "physics_group_name", "=", "str", "(", "physics_group_name", ")", "except", ":", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "'physics group name given is not valid : %s'", "%", "physics_group_name", ")", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "result", "=", "self", ".", "pglist", ".", "execute", "(", "conn", ",", "physics_group_name", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Returns all physics groups if physics group names are not passed.
[ "Returns", "all", "physics", "groups", "if", "physics", "group", "names", "are", "not", "passed", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSPhysicsGroup.py#L23-L45
train
dmwm/DBS
Server/Python/src/dbs/utils/DBSDaoTools.py
create_token_generator
def create_token_generator(input_list): """SQL Generator to select from list of values in Oracle""" ###Generator trick from http://betteratoracle.com/posts/20-how-do-i-bind-a-variable-in-list ###The maximum length of the comma separated list is 4000 characters, therefore we need to split the list ###ORA-01460: unimplemented or unreasonable conversion requested will thrown if list is larger oracle_limit = 4000 grp_list = [] if type(input_list[0]) == int : input_str = ','.join(map(str, input_list)) else: input_str = ','.join(input_list) if len(input_str) >= oracle_limit: index = 0 while True: begin, end = index, index+oracle_limit if end > len(input_str): end = len(input_str) grp_list.append(input_str[begin:end]) break else: index = input_str.rfind(',', begin, end) if index == -1: break grp_list.append(input_str[begin:index]) index += 1 #to remove the leading comma else: grp_list.append(input_str) token_generator = """ WITH TOKEN_GENERATOR AS ( """ binds = {} for index, chunk in enumerate(grp_list): if index: token_generator += """ UNION ALL """ bind = "token_%s" % index token_generator += """SELECT REGEXP_SUBSTR(:{bind}, '[^,]+', 1, LEVEL) token FROM DUAL CONNECT BY LEVEL <= LENGTH(:{bind}) - LENGTH(REPLACE(:{bind}, ',', '')) + 1 """.format(bind=bind) binds.update({bind: chunk}) token_generator += ")" return token_generator, binds
python
def create_token_generator(input_list): """SQL Generator to select from list of values in Oracle""" ###Generator trick from http://betteratoracle.com/posts/20-how-do-i-bind-a-variable-in-list ###The maximum length of the comma separated list is 4000 characters, therefore we need to split the list ###ORA-01460: unimplemented or unreasonable conversion requested will thrown if list is larger oracle_limit = 4000 grp_list = [] if type(input_list[0]) == int : input_str = ','.join(map(str, input_list)) else: input_str = ','.join(input_list) if len(input_str) >= oracle_limit: index = 0 while True: begin, end = index, index+oracle_limit if end > len(input_str): end = len(input_str) grp_list.append(input_str[begin:end]) break else: index = input_str.rfind(',', begin, end) if index == -1: break grp_list.append(input_str[begin:index]) index += 1 #to remove the leading comma else: grp_list.append(input_str) token_generator = """ WITH TOKEN_GENERATOR AS ( """ binds = {} for index, chunk in enumerate(grp_list): if index: token_generator += """ UNION ALL """ bind = "token_%s" % index token_generator += """SELECT REGEXP_SUBSTR(:{bind}, '[^,]+', 1, LEVEL) token FROM DUAL CONNECT BY LEVEL <= LENGTH(:{bind}) - LENGTH(REPLACE(:{bind}, ',', '')) + 1 """.format(bind=bind) binds.update({bind: chunk}) token_generator += ")" return token_generator, binds
[ "def", "create_token_generator", "(", "input_list", ")", ":", "###Generator trick from http://betteratoracle.com/posts/20-how-do-i-bind-a-variable-in-list", "###The maximum length of the comma separated list is 4000 characters, therefore we need to split the list", "###ORA-01460: unimplemented or unreasonable conversion requested will thrown if list is larger", "oracle_limit", "=", "4000", "grp_list", "=", "[", "]", "if", "type", "(", "input_list", "[", "0", "]", ")", "==", "int", ":", "input_str", "=", "','", ".", "join", "(", "map", "(", "str", ",", "input_list", ")", ")", "else", ":", "input_str", "=", "','", ".", "join", "(", "input_list", ")", "if", "len", "(", "input_str", ")", ">=", "oracle_limit", ":", "index", "=", "0", "while", "True", ":", "begin", ",", "end", "=", "index", ",", "index", "+", "oracle_limit", "if", "end", ">", "len", "(", "input_str", ")", ":", "end", "=", "len", "(", "input_str", ")", "grp_list", ".", "append", "(", "input_str", "[", "begin", ":", "end", "]", ")", "break", "else", ":", "index", "=", "input_str", ".", "rfind", "(", "','", ",", "begin", ",", "end", ")", "if", "index", "==", "-", "1", ":", "break", "grp_list", ".", "append", "(", "input_str", "[", "begin", ":", "index", "]", ")", "index", "+=", "1", "#to remove the leading comma", "else", ":", "grp_list", ".", "append", "(", "input_str", ")", "token_generator", "=", "\"\"\"\n WITH TOKEN_GENERATOR AS (\n \"\"\"", "binds", "=", "{", "}", "for", "index", ",", "chunk", "in", "enumerate", "(", "grp_list", ")", ":", "if", "index", ":", "token_generator", "+=", "\"\"\"\n UNION ALL\n \"\"\"", "bind", "=", "\"token_%s\"", "%", "index", "token_generator", "+=", "\"\"\"SELECT REGEXP_SUBSTR(:{bind}, '[^,]+', 1, LEVEL) token\n FROM DUAL\n CONNECT BY LEVEL <= LENGTH(:{bind}) - LENGTH(REPLACE(:{bind}, ',', '')) + 1\n \"\"\"", ".", "format", "(", "bind", "=", "bind", ")", "binds", ".", "update", "(", "{", "bind", ":", "chunk", "}", ")", "token_generator", "+=", "\")\"", "return", "token_generator", ",", "binds" ]
SQL Generator to select from list of values in Oracle
[ "SQL", "Generator", "to", "select", "from", "list", "of", "values", "in", "Oracle" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/utils/DBSDaoTools.py#L4-L50
train
dmwm/DBS
Server/Python/src/dbs/web/DBSServicesRegistry.py
DBSServicesRegistry.getServices
def getServices(self): """ Simple method that returs list of all know DBS instances, instances known to this registry """ try: conn = self.dbi.connection() result = self.serviceslist.execute(conn) return result except Exception as ex: msg = (("%s DBSServicesRegistry/getServices." + " %s\n. Exception trace: \n %s") % (DBSEXCEPTIONS['dbsException-3'], ex, traceback.format_exc())) self.logger.exception(msg ) raise Exception ("dbsException-3", msg ) finally: conn.close()
python
def getServices(self): """ Simple method that returs list of all know DBS instances, instances known to this registry """ try: conn = self.dbi.connection() result = self.serviceslist.execute(conn) return result except Exception as ex: msg = (("%s DBSServicesRegistry/getServices." + " %s\n. Exception trace: \n %s") % (DBSEXCEPTIONS['dbsException-3'], ex, traceback.format_exc())) self.logger.exception(msg ) raise Exception ("dbsException-3", msg ) finally: conn.close()
[ "def", "getServices", "(", "self", ")", ":", "try", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "result", "=", "self", ".", "serviceslist", ".", "execute", "(", "conn", ")", "return", "result", "except", "Exception", "as", "ex", ":", "msg", "=", "(", "(", "\"%s DBSServicesRegistry/getServices.\"", "+", "\" %s\\n. Exception trace: \\n %s\"", ")", "%", "(", "DBSEXCEPTIONS", "[", "'dbsException-3'", "]", ",", "ex", ",", "traceback", ".", "format_exc", "(", ")", ")", ")", "self", ".", "logger", ".", "exception", "(", "msg", ")", "raise", "Exception", "(", "\"dbsException-3\"", ",", "msg", ")", "finally", ":", "conn", ".", "close", "(", ")" ]
Simple method that returs list of all know DBS instances, instances known to this registry
[ "Simple", "method", "that", "returs", "list", "of", "all", "know", "DBS", "instances", "instances", "known", "to", "this", "registry" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSServicesRegistry.py#L37-L53
train
dmwm/DBS
Server/Python/src/dbs/web/DBSServicesRegistry.py
DBSServicesRegistry.addService
def addService(self): """ Add a service to service registry """ conn = self.dbi.connection() tran = conn.begin() try: body = request.body.read() service = cjson.decode(body) addthis = {} addthis['service_id'] = self.sm.increment(conn, "SEQ_RS", tran) addthis['name'] = service.get('NAME', '') if addthis['name'] == '': msg = (("%s DBSServicesRegistry/addServices." + " Service Must be Named\n") % DBSEXCEPTIONS['dbsException-3']) raise Exception("dbsException-3", msg) addthis['type'] = service.get('TYPE', 'GENERIC') addthis['location'] = service.get('LOCATION', 'HYPERSPACE') addthis['status'] = service.get('STATUS', 'UNKNOWN') addthis['admin'] = service.get('ADMIN', 'UNADMINISTRATED') addthis['uri'] = service.get('URI', '') if addthis['uri'] == '': msg = (("%s DBSServicesRegistry/addServices." + " Service URI must be provided.\n") % DBSEXCEPTIONS['dbsException-3']) self.logger.exception(msg) raise Exception("dbsException-3", msg) addthis['db'] = service.get('DB', 'NO_DATABASE') addthis['version'] = service.get('VERSION', 'UNKNOWN' ) addthis['last_contact'] = dbsUtils().getTime() addthis['comments'] = service.get('COMMENTS', 'NO COMMENTS') addthis['alias'] = service.get('ALIAS', 'No Alias') self.servicesadd.execute(conn, addthis, tran) tran.commit() except exceptions.IntegrityError as ex: if (str(ex).find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1) : #Update the service instead try: self.servicesupdate.execute(conn, addthis, tran) tran.commit() except Exception as ex: msg = (("%s DBSServiceRegistry/addServices." + " %s\n. Exception trace: \n %s") % (DBSEXCEPTIONS['dbsException-3'], ex, traceback.format_exc())) self.logger.exception(msg ) raise Exception ("dbsException-3", msg ) except Exception as ex: tran.rollback() msg = (("%s DBSServiceRegistry/addServices." + " %s\n. Exception trace: \n %s") % (DBSEXCEPTIONS['dbsException-3'], ex, traceback.format_exc())) self.logger.exception(msg ) raise Exception ("dbsException-3", msg ) finally: conn.close()
python
def addService(self): """ Add a service to service registry """ conn = self.dbi.connection() tran = conn.begin() try: body = request.body.read() service = cjson.decode(body) addthis = {} addthis['service_id'] = self.sm.increment(conn, "SEQ_RS", tran) addthis['name'] = service.get('NAME', '') if addthis['name'] == '': msg = (("%s DBSServicesRegistry/addServices." + " Service Must be Named\n") % DBSEXCEPTIONS['dbsException-3']) raise Exception("dbsException-3", msg) addthis['type'] = service.get('TYPE', 'GENERIC') addthis['location'] = service.get('LOCATION', 'HYPERSPACE') addthis['status'] = service.get('STATUS', 'UNKNOWN') addthis['admin'] = service.get('ADMIN', 'UNADMINISTRATED') addthis['uri'] = service.get('URI', '') if addthis['uri'] == '': msg = (("%s DBSServicesRegistry/addServices." + " Service URI must be provided.\n") % DBSEXCEPTIONS['dbsException-3']) self.logger.exception(msg) raise Exception("dbsException-3", msg) addthis['db'] = service.get('DB', 'NO_DATABASE') addthis['version'] = service.get('VERSION', 'UNKNOWN' ) addthis['last_contact'] = dbsUtils().getTime() addthis['comments'] = service.get('COMMENTS', 'NO COMMENTS') addthis['alias'] = service.get('ALIAS', 'No Alias') self.servicesadd.execute(conn, addthis, tran) tran.commit() except exceptions.IntegrityError as ex: if (str(ex).find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1) : #Update the service instead try: self.servicesupdate.execute(conn, addthis, tran) tran.commit() except Exception as ex: msg = (("%s DBSServiceRegistry/addServices." + " %s\n. Exception trace: \n %s") % (DBSEXCEPTIONS['dbsException-3'], ex, traceback.format_exc())) self.logger.exception(msg ) raise Exception ("dbsException-3", msg ) except Exception as ex: tran.rollback() msg = (("%s DBSServiceRegistry/addServices." + " %s\n. Exception trace: \n %s") % (DBSEXCEPTIONS['dbsException-3'], ex, traceback.format_exc())) self.logger.exception(msg ) raise Exception ("dbsException-3", msg ) finally: conn.close()
[ "def", "addService", "(", "self", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "conn", ".", "begin", "(", ")", "try", ":", "body", "=", "request", ".", "body", ".", "read", "(", ")", "service", "=", "cjson", ".", "decode", "(", "body", ")", "addthis", "=", "{", "}", "addthis", "[", "'service_id'", "]", "=", "self", ".", "sm", ".", "increment", "(", "conn", ",", "\"SEQ_RS\"", ",", "tran", ")", "addthis", "[", "'name'", "]", "=", "service", ".", "get", "(", "'NAME'", ",", "''", ")", "if", "addthis", "[", "'name'", "]", "==", "''", ":", "msg", "=", "(", "(", "\"%s DBSServicesRegistry/addServices.\"", "+", "\" Service Must be Named\\n\"", ")", "%", "DBSEXCEPTIONS", "[", "'dbsException-3'", "]", ")", "raise", "Exception", "(", "\"dbsException-3\"", ",", "msg", ")", "addthis", "[", "'type'", "]", "=", "service", ".", "get", "(", "'TYPE'", ",", "'GENERIC'", ")", "addthis", "[", "'location'", "]", "=", "service", ".", "get", "(", "'LOCATION'", ",", "'HYPERSPACE'", ")", "addthis", "[", "'status'", "]", "=", "service", ".", "get", "(", "'STATUS'", ",", "'UNKNOWN'", ")", "addthis", "[", "'admin'", "]", "=", "service", ".", "get", "(", "'ADMIN'", ",", "'UNADMINISTRATED'", ")", "addthis", "[", "'uri'", "]", "=", "service", ".", "get", "(", "'URI'", ",", "''", ")", "if", "addthis", "[", "'uri'", "]", "==", "''", ":", "msg", "=", "(", "(", "\"%s DBSServicesRegistry/addServices.\"", "+", "\" Service URI must be provided.\\n\"", ")", "%", "DBSEXCEPTIONS", "[", "'dbsException-3'", "]", ")", "self", ".", "logger", ".", "exception", "(", "msg", ")", "raise", "Exception", "(", "\"dbsException-3\"", ",", "msg", ")", "addthis", "[", "'db'", "]", "=", "service", ".", "get", "(", "'DB'", ",", "'NO_DATABASE'", ")", "addthis", "[", "'version'", "]", "=", "service", ".", "get", "(", "'VERSION'", ",", "'UNKNOWN'", ")", "addthis", "[", "'last_contact'", "]", "=", "dbsUtils", "(", ")", ".", "getTime", "(", ")", "addthis", "[", "'comments'", "]", "=", "service", ".", "get", "(", "'COMMENTS'", ",", "'NO COMMENTS'", ")", "addthis", "[", "'alias'", "]", "=", "service", ".", "get", "(", "'ALIAS'", ",", "'No Alias'", ")", "self", ".", "servicesadd", ".", "execute", "(", "conn", ",", "addthis", ",", "tran", ")", "tran", ".", "commit", "(", ")", "except", "exceptions", ".", "IntegrityError", "as", "ex", ":", "if", "(", "str", "(", "ex", ")", ".", "find", "(", "\"unique constraint\"", ")", "!=", "-", "1", "or", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"duplicate\"", ")", "!=", "-", "1", ")", ":", "#Update the service instead", "try", ":", "self", ".", "servicesupdate", ".", "execute", "(", "conn", ",", "addthis", ",", "tran", ")", "tran", ".", "commit", "(", ")", "except", "Exception", "as", "ex", ":", "msg", "=", "(", "(", "\"%s DBSServiceRegistry/addServices.\"", "+", "\" %s\\n. Exception trace: \\n %s\"", ")", "%", "(", "DBSEXCEPTIONS", "[", "'dbsException-3'", "]", ",", "ex", ",", "traceback", ".", "format_exc", "(", ")", ")", ")", "self", ".", "logger", ".", "exception", "(", "msg", ")", "raise", "Exception", "(", "\"dbsException-3\"", ",", "msg", ")", "except", "Exception", "as", "ex", ":", "tran", ".", "rollback", "(", ")", "msg", "=", "(", "(", "\"%s DBSServiceRegistry/addServices.\"", "+", "\" %s\\n. Exception trace: \\n %s\"", ")", "%", "(", "DBSEXCEPTIONS", "[", "'dbsException-3'", "]", ",", "ex", ",", "traceback", ".", "format_exc", "(", ")", ")", ")", "self", ".", "logger", ".", "exception", "(", "msg", ")", "raise", "Exception", "(", "\"dbsException-3\"", ",", "msg", ")", "finally", ":", "conn", ".", "close", "(", ")" ]
Add a service to service registry
[ "Add", "a", "service", "to", "service", "registry" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/web/DBSServicesRegistry.py#L55-L115
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/MigrationRequests/UpdateRequestStatus.py
UpdateRequestStatus.execute
def execute(self, conn, daoinput, transaction = False): """ required keys: migration_status, migration_request_id """ if not conn: dbsExceptionHandler("dbsException-failed-connect2host", "Oracle/MigrationRequests/UpdateRequestStatus. Expects db connection from upper layer.", self.logger.exception) if daoinput['migration_status'] == 1: sql = self.sql2 elif daoinput['migration_status'] == 2: sql = self.sql + " and MIGRATION_STATUS = 1 " elif daoinput['migration_status'] == 3: sql = self.sql3 + " and MIGRATION_STATUS = 1 " else: dbsExceptionHandler("dbsException-conflict-data", "Oracle/MigrationRequest/UpdateRequestStatus. Expected migration status to be 1, 2 or 3", self.logger.exception) result = self.dbi.processData(sql, daoinput, conn, transaction)
python
def execute(self, conn, daoinput, transaction = False): """ required keys: migration_status, migration_request_id """ if not conn: dbsExceptionHandler("dbsException-failed-connect2host", "Oracle/MigrationRequests/UpdateRequestStatus. Expects db connection from upper layer.", self.logger.exception) if daoinput['migration_status'] == 1: sql = self.sql2 elif daoinput['migration_status'] == 2: sql = self.sql + " and MIGRATION_STATUS = 1 " elif daoinput['migration_status'] == 3: sql = self.sql3 + " and MIGRATION_STATUS = 1 " else: dbsExceptionHandler("dbsException-conflict-data", "Oracle/MigrationRequest/UpdateRequestStatus. Expected migration status to be 1, 2 or 3", self.logger.exception) result = self.dbi.processData(sql, daoinput, conn, transaction)
[ "def", "execute", "(", "self", ",", "conn", ",", "daoinput", ",", "transaction", "=", "False", ")", ":", "if", "not", "conn", ":", "dbsExceptionHandler", "(", "\"dbsException-failed-connect2host\"", ",", "\"Oracle/MigrationRequests/UpdateRequestStatus. Expects db connection from upper layer.\"", ",", "self", ".", "logger", ".", "exception", ")", "if", "daoinput", "[", "'migration_status'", "]", "==", "1", ":", "sql", "=", "self", ".", "sql2", "elif", "daoinput", "[", "'migration_status'", "]", "==", "2", ":", "sql", "=", "self", ".", "sql", "+", "\" and MIGRATION_STATUS = 1 \"", "elif", "daoinput", "[", "'migration_status'", "]", "==", "3", ":", "sql", "=", "self", ".", "sql3", "+", "\" and MIGRATION_STATUS = 1 \"", "else", ":", "dbsExceptionHandler", "(", "\"dbsException-conflict-data\"", ",", "\"Oracle/MigrationRequest/UpdateRequestStatus. Expected migration status to be 1, 2 or 3\"", ",", "self", ".", "logger", ".", "exception", ")", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "daoinput", ",", "conn", ",", "transaction", ")" ]
required keys: migration_status, migration_request_id
[ "required", "keys", ":", "migration_status", "migration_request_id" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/MigrationRequests/UpdateRequestStatus.py#L67-L85
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/MigrationRequests/List.py
List.execute
def execute(self, conn, migration_url="", migration_input="", create_by="", migration_request_id="", oldest= False, transaction=False): """ Lists all requests if pattern is not provided. """ sql = self.sql binds = {} if migration_request_id: sql += " WHERE MR.MIGRATION_REQUEST_ID=:migration_request_id" binds['migration_request_id']=migration_request_id elif oldest: #FIXME: Need to write the sql.YG #current_date = dbsUtils().getTime() #we require waiting time for #retry_count=0 is 1 minutes #retry_count=1 is 2 minutes #retyr_count=2 is 4 minutes sql += """ WHERE MR.MIGRATION_STATUS=0 or (MR.migration_status=3 and MR.retry_count=0 and MR.last_modification_date <= :current_date-60) or (MR.migration_status=3 and MR.retry_count=1 and MR.last_modification_date <= :current_date-120) or (MR.migration_status=3 and MR.retry_count=2 and MR.last_modification_date <= :current_date-240) ORDER BY MR.creation_date """ binds['current_date'] = dbsUtils().getTime() #print "time= " + str(binds['current_date']) else: if migration_url or migration_input or create_by: sql += " WHERE " if migration_url: sql += " MR.MIGRATION_URL=:migration_url" binds['migration_url']=migration_url if migration_input: if migration_url: sql += " AND " op = ("=", "like")["%" in migration_input] sql += " MR.MIGRATION_INPUT %s :migration_input" % op binds['migration_input']=migration_input if create_by: if migration_url or migration_input: sql += " AND " sql += " MR.CREATE_BY=:create_by" %create_by binds['create_by']=create_by cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = [] for c in cursors: result.extend(self.formatCursor(c, size=100)) return result
python
def execute(self, conn, migration_url="", migration_input="", create_by="", migration_request_id="", oldest= False, transaction=False): """ Lists all requests if pattern is not provided. """ sql = self.sql binds = {} if migration_request_id: sql += " WHERE MR.MIGRATION_REQUEST_ID=:migration_request_id" binds['migration_request_id']=migration_request_id elif oldest: #FIXME: Need to write the sql.YG #current_date = dbsUtils().getTime() #we require waiting time for #retry_count=0 is 1 minutes #retry_count=1 is 2 minutes #retyr_count=2 is 4 minutes sql += """ WHERE MR.MIGRATION_STATUS=0 or (MR.migration_status=3 and MR.retry_count=0 and MR.last_modification_date <= :current_date-60) or (MR.migration_status=3 and MR.retry_count=1 and MR.last_modification_date <= :current_date-120) or (MR.migration_status=3 and MR.retry_count=2 and MR.last_modification_date <= :current_date-240) ORDER BY MR.creation_date """ binds['current_date'] = dbsUtils().getTime() #print "time= " + str(binds['current_date']) else: if migration_url or migration_input or create_by: sql += " WHERE " if migration_url: sql += " MR.MIGRATION_URL=:migration_url" binds['migration_url']=migration_url if migration_input: if migration_url: sql += " AND " op = ("=", "like")["%" in migration_input] sql += " MR.MIGRATION_INPUT %s :migration_input" % op binds['migration_input']=migration_input if create_by: if migration_url or migration_input: sql += " AND " sql += " MR.CREATE_BY=:create_by" %create_by binds['create_by']=create_by cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = [] for c in cursors: result.extend(self.formatCursor(c, size=100)) return result
[ "def", "execute", "(", "self", ",", "conn", ",", "migration_url", "=", "\"\"", ",", "migration_input", "=", "\"\"", ",", "create_by", "=", "\"\"", ",", "migration_request_id", "=", "\"\"", ",", "oldest", "=", "False", ",", "transaction", "=", "False", ")", ":", "sql", "=", "self", ".", "sql", "binds", "=", "{", "}", "if", "migration_request_id", ":", "sql", "+=", "\" WHERE MR.MIGRATION_REQUEST_ID=:migration_request_id\"", "binds", "[", "'migration_request_id'", "]", "=", "migration_request_id", "elif", "oldest", ":", "#FIXME: Need to write the sql.YG", "#current_date = dbsUtils().getTime()", "#we require waiting time for ", "#retry_count=0 is 1 minutes", "#retry_count=1 is 2 minutes", "#retyr_count=2 is 4 minutes", "sql", "+=", "\"\"\"\n WHERE MR.MIGRATION_STATUS=0 \n or (MR.migration_status=3 and MR.retry_count=0 and MR.last_modification_date <= :current_date-60) \n or (MR.migration_status=3 and MR.retry_count=1 and MR.last_modification_date <= :current_date-120) \n or (MR.migration_status=3 and MR.retry_count=2 and MR.last_modification_date <= :current_date-240)\n ORDER BY MR.creation_date\n \"\"\"", "binds", "[", "'current_date'", "]", "=", "dbsUtils", "(", ")", ".", "getTime", "(", ")", "#print \"time= \" + str(binds['current_date'])", "else", ":", "if", "migration_url", "or", "migration_input", "or", "create_by", ":", "sql", "+=", "\" WHERE \"", "if", "migration_url", ":", "sql", "+=", "\" MR.MIGRATION_URL=:migration_url\"", "binds", "[", "'migration_url'", "]", "=", "migration_url", "if", "migration_input", ":", "if", "migration_url", ":", "sql", "+=", "\" AND \"", "op", "=", "(", "\"=\"", ",", "\"like\"", ")", "[", "\"%\"", "in", "migration_input", "]", "sql", "+=", "\" MR.MIGRATION_INPUT %s :migration_input\"", "%", "op", "binds", "[", "'migration_input'", "]", "=", "migration_input", "if", "create_by", ":", "if", "migration_url", "or", "migration_input", ":", "sql", "+=", "\" AND \"", "sql", "+=", "\" MR.CREATE_BY=:create_by\"", "%", "create_by", "binds", "[", "'create_by'", "]", "=", "create_by", "cursors", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ",", "returnCursor", "=", "True", ")", "result", "=", "[", "]", "for", "c", "in", "cursors", ":", "result", ".", "extend", "(", "self", ".", "formatCursor", "(", "c", ",", "size", "=", "100", ")", ")", "return", "result" ]
Lists all requests if pattern is not provided.
[ "Lists", "all", "requests", "if", "pattern", "is", "not", "provided", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/MigrationRequests/List.py#L28-L75
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/FileBuffer/DeleteDuplicates.py
DeleteDuplicates.execute
def execute(self, conn, logical_file_name, transaction=False): """ simple execute """ if not conn: dbsExceptionHandler("dbsException-db-conn-failed", "Oracle/FileBuffer/DeleteDupicates. Expects db connection from upper layer.") print(self.sql) self.dbi.processData(self.sql, logical_file_name, conn, transaction)
python
def execute(self, conn, logical_file_name, transaction=False): """ simple execute """ if not conn: dbsExceptionHandler("dbsException-db-conn-failed", "Oracle/FileBuffer/DeleteDupicates. Expects db connection from upper layer.") print(self.sql) self.dbi.processData(self.sql, logical_file_name, conn, transaction)
[ "def", "execute", "(", "self", ",", "conn", ",", "logical_file_name", ",", "transaction", "=", "False", ")", ":", "if", "not", "conn", ":", "dbsExceptionHandler", "(", "\"dbsException-db-conn-failed\"", ",", "\"Oracle/FileBuffer/DeleteDupicates. Expects db connection from upper layer.\"", ")", "print", "(", "self", ".", "sql", ")", "self", ".", "dbi", ".", "processData", "(", "self", ".", "sql", ",", "logical_file_name", ",", "conn", ",", "transaction", ")" ]
simple execute
[ "simple", "execute" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/FileBuffer/DeleteDuplicates.py#L21-L29
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/BlockParent/List.py
List.execute
def execute(self, conn, block_name="", transaction = False): """ block: /a/b/c#d """ if not conn: dbsExceptionHandler("dbsException-failed-connect2host", "Oracle/BlockParent/List. Expects db connection from upper layer.", self.logger.exception) sql = self.sql if isinstance(block_name, basestring): binds = {'block_name' :block_name} elif type(block_name) is list: binds = [{'block_name':x} for x in block_name] else: msg = "Oracle/BlockParent/List. Block_name must be provided either as a string or as a list." dbsExceptionHandler('dbsException-invalid-input', msg, self.logger.exception) result = self.dbi.processData(sql, binds, conn, transaction) return self.formatDict(result)
python
def execute(self, conn, block_name="", transaction = False): """ block: /a/b/c#d """ if not conn: dbsExceptionHandler("dbsException-failed-connect2host", "Oracle/BlockParent/List. Expects db connection from upper layer.", self.logger.exception) sql = self.sql if isinstance(block_name, basestring): binds = {'block_name' :block_name} elif type(block_name) is list: binds = [{'block_name':x} for x in block_name] else: msg = "Oracle/BlockParent/List. Block_name must be provided either as a string or as a list." dbsExceptionHandler('dbsException-invalid-input', msg, self.logger.exception) result = self.dbi.processData(sql, binds, conn, transaction) return self.formatDict(result)
[ "def", "execute", "(", "self", ",", "conn", ",", "block_name", "=", "\"\"", ",", "transaction", "=", "False", ")", ":", "if", "not", "conn", ":", "dbsExceptionHandler", "(", "\"dbsException-failed-connect2host\"", ",", "\"Oracle/BlockParent/List. Expects db connection from upper layer.\"", ",", "self", ".", "logger", ".", "exception", ")", "sql", "=", "self", ".", "sql", "if", "isinstance", "(", "block_name", ",", "basestring", ")", ":", "binds", "=", "{", "'block_name'", ":", "block_name", "}", "elif", "type", "(", "block_name", ")", "is", "list", ":", "binds", "=", "[", "{", "'block_name'", ":", "x", "}", "for", "x", "in", "block_name", "]", "else", ":", "msg", "=", "\"Oracle/BlockParent/List. Block_name must be provided either as a string or as a list.\"", "dbsExceptionHandler", "(", "'dbsException-invalid-input'", ",", "msg", ",", "self", ".", "logger", ".", "exception", ")", "result", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ")", "return", "self", ".", "formatDict", "(", "result", ")" ]
block: /a/b/c#d
[ "block", ":", "/", "a", "/", "b", "/", "c#d" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/BlockParent/List.py#L27-L44
train
dmwm/DBS
Server/Python/src/dbs/dao/Oracle/Block/ListBlockOrigin.py
ListBlockOrigin.execute
def execute(self, conn, origin_site_name="", dataset="", block_name="", transaction = False): """ origin_site_name: T1_US_FNAL_Buffer dataset: /a/b/c block_name: /a/b/c#d """ if not conn: dbsExceptionHandler("dbsException-db-conn-failed", "Oracle/Block/List. Expects db connection from upper layer.", self.logger.exception) binds = {} if origin_site_name: wheresql = 'WHERE B.ORIGIN_SITE_NAME = :origin_site_name' binds.update(origin_site_name=origin_site_name) if dataset: if 'wheresql' in locals(): wheresql += ' AND DS.DATASET = :dataset' else: wheresql = 'WHERE DS.DATASET = :dataset' binds.update(dataset=dataset) if block_name: if 'wheresql' in locals(): wheresql += ' AND B.BLOCK_NAME = :block_name' else: wheresql = 'WHERE B.BLOCK_NAME = :block_name' binds.update(block_name=block_name) sql = '{sql} {wheresql}'.format(sql=self.sql, wheresql=wheresql) cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = [] for cursor in cursors: result.extend(self.formatCursor(cursor, size=100)) return result
python
def execute(self, conn, origin_site_name="", dataset="", block_name="", transaction = False): """ origin_site_name: T1_US_FNAL_Buffer dataset: /a/b/c block_name: /a/b/c#d """ if not conn: dbsExceptionHandler("dbsException-db-conn-failed", "Oracle/Block/List. Expects db connection from upper layer.", self.logger.exception) binds = {} if origin_site_name: wheresql = 'WHERE B.ORIGIN_SITE_NAME = :origin_site_name' binds.update(origin_site_name=origin_site_name) if dataset: if 'wheresql' in locals(): wheresql += ' AND DS.DATASET = :dataset' else: wheresql = 'WHERE DS.DATASET = :dataset' binds.update(dataset=dataset) if block_name: if 'wheresql' in locals(): wheresql += ' AND B.BLOCK_NAME = :block_name' else: wheresql = 'WHERE B.BLOCK_NAME = :block_name' binds.update(block_name=block_name) sql = '{sql} {wheresql}'.format(sql=self.sql, wheresql=wheresql) cursors = self.dbi.processData(sql, binds, conn, transaction, returnCursor=True) result = [] for cursor in cursors: result.extend(self.formatCursor(cursor, size=100)) return result
[ "def", "execute", "(", "self", ",", "conn", ",", "origin_site_name", "=", "\"\"", ",", "dataset", "=", "\"\"", ",", "block_name", "=", "\"\"", ",", "transaction", "=", "False", ")", ":", "if", "not", "conn", ":", "dbsExceptionHandler", "(", "\"dbsException-db-conn-failed\"", ",", "\"Oracle/Block/List. Expects db connection from upper layer.\"", ",", "self", ".", "logger", ".", "exception", ")", "binds", "=", "{", "}", "if", "origin_site_name", ":", "wheresql", "=", "'WHERE B.ORIGIN_SITE_NAME = :origin_site_name'", "binds", ".", "update", "(", "origin_site_name", "=", "origin_site_name", ")", "if", "dataset", ":", "if", "'wheresql'", "in", "locals", "(", ")", ":", "wheresql", "+=", "' AND DS.DATASET = :dataset'", "else", ":", "wheresql", "=", "'WHERE DS.DATASET = :dataset'", "binds", ".", "update", "(", "dataset", "=", "dataset", ")", "if", "block_name", ":", "if", "'wheresql'", "in", "locals", "(", ")", ":", "wheresql", "+=", "' AND B.BLOCK_NAME = :block_name'", "else", ":", "wheresql", "=", "'WHERE B.BLOCK_NAME = :block_name'", "binds", ".", "update", "(", "block_name", "=", "block_name", ")", "sql", "=", "'{sql} {wheresql}'", ".", "format", "(", "sql", "=", "self", ".", "sql", ",", "wheresql", "=", "wheresql", ")", "cursors", "=", "self", ".", "dbi", ".", "processData", "(", "sql", ",", "binds", ",", "conn", ",", "transaction", ",", "returnCursor", "=", "True", ")", "result", "=", "[", "]", "for", "cursor", "in", "cursors", ":", "result", ".", "extend", "(", "self", ".", "formatCursor", "(", "cursor", ",", "size", "=", "100", ")", ")", "return", "result" ]
origin_site_name: T1_US_FNAL_Buffer dataset: /a/b/c block_name: /a/b/c#d
[ "origin_site_name", ":", "T1_US_FNAL_Buffer", "dataset", ":", "/", "a", "/", "b", "/", "c", "block_name", ":", "/", "a", "/", "b", "/", "c#d" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/dao/Oracle/Block/ListBlockOrigin.py#L29-L63
train
dmwm/DBS
Server/Python/src/dbs/business/DBSPrimaryDataset.py
DBSPrimaryDataset.listPrimaryDatasets
def listPrimaryDatasets(self, primary_ds_name="", primary_ds_type=""): """ Returns all primary dataset if primary_ds_name or primary_ds_type are not passed. """ conn = self.dbi.connection() try: result = self.primdslist.execute(conn, primary_ds_name, primary_ds_type) if conn: conn.close() return result finally: if conn: conn.close()
python
def listPrimaryDatasets(self, primary_ds_name="", primary_ds_type=""): """ Returns all primary dataset if primary_ds_name or primary_ds_type are not passed. """ conn = self.dbi.connection() try: result = self.primdslist.execute(conn, primary_ds_name, primary_ds_type) if conn: conn.close() return result finally: if conn: conn.close()
[ "def", "listPrimaryDatasets", "(", "self", ",", "primary_ds_name", "=", "\"\"", ",", "primary_ds_type", "=", "\"\"", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "result", "=", "self", ".", "primdslist", ".", "execute", "(", "conn", ",", "primary_ds_name", ",", "primary_ds_type", ")", "if", "conn", ":", "conn", ".", "close", "(", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Returns all primary dataset if primary_ds_name or primary_ds_type are not passed.
[ "Returns", "all", "primary", "dataset", "if", "primary_ds_name", "or", "primary_ds_type", "are", "not", "passed", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSPrimaryDataset.py#L29-L40
train
dmwm/DBS
Server/Python/src/dbs/business/DBSPrimaryDataset.py
DBSPrimaryDataset.listPrimaryDSTypes
def listPrimaryDSTypes(self, primary_ds_type="", dataset=""): """ Returns all primary dataset types if dataset or primary_ds_type are not passed. """ conn = self.dbi.connection() try: result = self.primdstypeList.execute(conn, primary_ds_type, dataset) if conn: conn.close() return result finally: if conn: conn.close()
python
def listPrimaryDSTypes(self, primary_ds_type="", dataset=""): """ Returns all primary dataset types if dataset or primary_ds_type are not passed. """ conn = self.dbi.connection() try: result = self.primdstypeList.execute(conn, primary_ds_type, dataset) if conn: conn.close() return result finally: if conn: conn.close()
[ "def", "listPrimaryDSTypes", "(", "self", ",", "primary_ds_type", "=", "\"\"", ",", "dataset", "=", "\"\"", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "try", ":", "result", "=", "self", ".", "primdstypeList", ".", "execute", "(", "conn", ",", "primary_ds_type", ",", "dataset", ")", "if", "conn", ":", "conn", ".", "close", "(", ")", "return", "result", "finally", ":", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Returns all primary dataset types if dataset or primary_ds_type are not passed.
[ "Returns", "all", "primary", "dataset", "types", "if", "dataset", "or", "primary_ds_type", "are", "not", "passed", "." ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSPrimaryDataset.py#L42-L53
train
dmwm/DBS
Server/Python/src/dbs/business/DBSPrimaryDataset.py
DBSPrimaryDataset.insertPrimaryDataset
def insertPrimaryDataset(self, businput): """ Input dictionary has to have the following keys: primary_ds_name, primary_ds_type, creation_date, create_by. it builds the correct dictionary for dao input and executes the dao """ conn = self.dbi.connection() tran = conn.begin() #checking for required fields if "primary_ds_name" not in businput: dbsExceptionHandler("dbsException-invalid-input", " DBSPrimaryDataset/insertPrimaryDataset. " + "Primary dataset Name is required for insertPrimaryDataset.") try: businput["primary_ds_type_id"] = (self.primdstypeList.execute(conn, businput["primary_ds_type"] ))[0]["primary_ds_type_id"] del businput["primary_ds_type"] businput["primary_ds_id"] = self.sm.increment(conn, "SEQ_PDS") self.primdsin.execute(conn, businput, tran) tran.commit() tran = None except KeyError as ke: dbsExceptionHandler("dbsException-invalid-input", " DBSPrimaryDataset/insertPrimaryDataset. Missing: %s" % ke) self.logger.warning(" DBSPrimaryDataset/insertPrimaryDataset. Missing: %s" % ke) except IndexError as ie: dbsExceptionHandler("dbsException-missing-data", " DBSPrimaryDataset/insertPrimaryDataset. %s" % ie) self.logger.warning(" DBSPrimaryDataset/insertPrimaryDataset. Missing: %s" % ie) except Exception as ex: if (str(ex).lower().find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1): self.logger.warning("DBSPrimaryDataset/insertPrimaryDataset:" + " Unique constraint violation being ignored...") self.logger.warning(ex) else: if tran: tran.rollback() if conn: conn.close() raise finally: if tran: tran.rollback() if conn: conn.close()
python
def insertPrimaryDataset(self, businput): """ Input dictionary has to have the following keys: primary_ds_name, primary_ds_type, creation_date, create_by. it builds the correct dictionary for dao input and executes the dao """ conn = self.dbi.connection() tran = conn.begin() #checking for required fields if "primary_ds_name" not in businput: dbsExceptionHandler("dbsException-invalid-input", " DBSPrimaryDataset/insertPrimaryDataset. " + "Primary dataset Name is required for insertPrimaryDataset.") try: businput["primary_ds_type_id"] = (self.primdstypeList.execute(conn, businput["primary_ds_type"] ))[0]["primary_ds_type_id"] del businput["primary_ds_type"] businput["primary_ds_id"] = self.sm.increment(conn, "SEQ_PDS") self.primdsin.execute(conn, businput, tran) tran.commit() tran = None except KeyError as ke: dbsExceptionHandler("dbsException-invalid-input", " DBSPrimaryDataset/insertPrimaryDataset. Missing: %s" % ke) self.logger.warning(" DBSPrimaryDataset/insertPrimaryDataset. Missing: %s" % ke) except IndexError as ie: dbsExceptionHandler("dbsException-missing-data", " DBSPrimaryDataset/insertPrimaryDataset. %s" % ie) self.logger.warning(" DBSPrimaryDataset/insertPrimaryDataset. Missing: %s" % ie) except Exception as ex: if (str(ex).lower().find("unique constraint") != -1 or str(ex).lower().find("duplicate") != -1): self.logger.warning("DBSPrimaryDataset/insertPrimaryDataset:" + " Unique constraint violation being ignored...") self.logger.warning(ex) else: if tran: tran.rollback() if conn: conn.close() raise finally: if tran: tran.rollback() if conn: conn.close()
[ "def", "insertPrimaryDataset", "(", "self", ",", "businput", ")", ":", "conn", "=", "self", ".", "dbi", ".", "connection", "(", ")", "tran", "=", "conn", ".", "begin", "(", ")", "#checking for required fields", "if", "\"primary_ds_name\"", "not", "in", "businput", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\" DBSPrimaryDataset/insertPrimaryDataset. \"", "+", "\"Primary dataset Name is required for insertPrimaryDataset.\"", ")", "try", ":", "businput", "[", "\"primary_ds_type_id\"", "]", "=", "(", "self", ".", "primdstypeList", ".", "execute", "(", "conn", ",", "businput", "[", "\"primary_ds_type\"", "]", ")", ")", "[", "0", "]", "[", "\"primary_ds_type_id\"", "]", "del", "businput", "[", "\"primary_ds_type\"", "]", "businput", "[", "\"primary_ds_id\"", "]", "=", "self", ".", "sm", ".", "increment", "(", "conn", ",", "\"SEQ_PDS\"", ")", "self", ".", "primdsin", ".", "execute", "(", "conn", ",", "businput", ",", "tran", ")", "tran", ".", "commit", "(", ")", "tran", "=", "None", "except", "KeyError", "as", "ke", ":", "dbsExceptionHandler", "(", "\"dbsException-invalid-input\"", ",", "\" DBSPrimaryDataset/insertPrimaryDataset. Missing: %s\"", "%", "ke", ")", "self", ".", "logger", ".", "warning", "(", "\" DBSPrimaryDataset/insertPrimaryDataset. Missing: %s\"", "%", "ke", ")", "except", "IndexError", "as", "ie", ":", "dbsExceptionHandler", "(", "\"dbsException-missing-data\"", ",", "\" DBSPrimaryDataset/insertPrimaryDataset. %s\"", "%", "ie", ")", "self", ".", "logger", ".", "warning", "(", "\" DBSPrimaryDataset/insertPrimaryDataset. Missing: %s\"", "%", "ie", ")", "except", "Exception", "as", "ex", ":", "if", "(", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"unique constraint\"", ")", "!=", "-", "1", "or", "str", "(", "ex", ")", ".", "lower", "(", ")", ".", "find", "(", "\"duplicate\"", ")", "!=", "-", "1", ")", ":", "self", ".", "logger", ".", "warning", "(", "\"DBSPrimaryDataset/insertPrimaryDataset:\"", "+", "\" Unique constraint violation being ignored...\"", ")", "self", ".", "logger", ".", "warning", "(", "ex", ")", "else", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")", "raise", "finally", ":", "if", "tran", ":", "tran", ".", "rollback", "(", ")", "if", "conn", ":", "conn", ".", "close", "(", ")" ]
Input dictionary has to have the following keys: primary_ds_name, primary_ds_type, creation_date, create_by. it builds the correct dictionary for dao input and executes the dao
[ "Input", "dictionary", "has", "to", "have", "the", "following", "keys", ":", "primary_ds_name", "primary_ds_type", "creation_date", "create_by", ".", "it", "builds", "the", "correct", "dictionary", "for", "dao", "input", "and", "executes", "the", "dao" ]
9619bafce3783b3e77f0415f8f9a258e33dd1e6f
https://github.com/dmwm/DBS/blob/9619bafce3783b3e77f0415f8f9a258e33dd1e6f/Server/Python/src/dbs/business/DBSPrimaryDataset.py#L55-L99
train